The Guardian Shows Hidden Text Can Manipulate ChatGPT’s Search Results

Photo by Jonathan Kemper on Unsplash

The Guardian Shows Hidden Text Can Manipulate ChatGPT’s Search Results

Reading time: 2 min

The British newspaper The Guardian revealed that AI-powered search engines can be manipulated by websites with hidden content.

In a Rush? Here are the Quick Facts!

  • The Guardian revealed that AI-powered search engines can be manipulated by websites with hidden content
  • The test showed ChatGPT’s search engine can be affected by a security vulnerability known as “prompt injection”
  • ChatGPT can prioritize third-party’s hidden instructions when summarizing a website

In a test using OpenAI’s ChatGPT search engine feature, researchers asked for a summary of a fake website containing malicious information to alter the AI’s response—a vulnerability known as prompt injection—and the AI was susceptible to it, even favoring the third party’s instructions.

To prove this, The Guardian’s team considered a fake website of a camera’s product page—featuring good and bad reviews—with hidden instructions to give a positive review and disregard the bad reviews, and ChatGPT included only positive reviews in its summary. They also proved that AI can return malicious codes.

“The simple inclusion of hidden text by third parties without instructions can also be used to ensure a positive assessment, with one test including extremely positive fake reviews which influenced the summary returned by ChatGPT,” wrote the newspaper.

A cybersecurity researcher at CyberCX, Jacob Larsen, said that this vulnerability could be of “high risk” as people could create websites specifically to deceive users, especially once it reaches a wider audience. OpenAI was warned about this security risk.

The journal also highlighted the case of a cryptocurrency enthusiast who used ChatGPT to write the code for a crypto project and stole their credentials, making the programmer lose over $2,000.

“They’re simply asking a question, receiving an answer, but the model is producing and sharing content that has basically been injected by an adversary to share something that is malicious,” said Larsen.

OpenAI warns about possible mistakes and errors in its use, but researchers are concerned about future web practices with AI-powered search engines.

Did you like this article? Rate it!
I hated it I don't really like it It was ok Pretty good! Loved it!

We're thrilled you enjoyed our work!

As a valued reader, would you mind giving us a shoutout on Trustpilot? It's quick and means the world to us. Thank you for being amazing!

Rate us on Trustpilot
0 Voted by 0 users
Title
Comment
Thanks for your feedback
Loader
Please wait 5 minutes before posting another comment.
Comment sent for approval.

Leave a Comment

Loader
Loader Show more...