City
Epaper

Researchers easily hypnotise AI chatbot ChatGPT into hacking: Report

By IANS | Updated: August 9, 2023 19:50 IST

New Delhi, Aug 9 Tricking generative AI to assist in scams and cyberattacks does not require much coding ...

Open in App

New Delhi, Aug 9 Tricking generative AI to assist in scams and cyberattacks does not require much coding knowledge, a new report has warned.

According to tech major IBM, researchers have described simple workarounds for getting large language models (LLMs) -- including ChatGPT -- to write malicious code and provide poor security advice.

"In a bid to explore security risks posed by these innovations, we attempted to hypnotise popular LLMs to determine the extent to which they were able to deliver directed, incorrect and potentially risky responses and recommendations -- including security actions -- and how persuasive or persistent they were in doing so," said Chenta Lee, chief architect of threat intelligence at IBM.

"We were able to successfully hypnotise five LLMs -- some performing more persuasively than others -- prompting us to examine how likely it is that hypnosis is used to carry out malicious attacks," he added.

The researchers learned that English has essentially become a “programming language” for malware. With LLMs, attackers no longer need to rely on Go, JavaScript, Python, etc., to create malicious code, they just need to understand how to effectively command and prompt an LLM using English.

Through hypnosis, the security experts were able to get LLMs to leak the confidential financial information of other users, create vulnerable code, create malicious code, and offer weak security recommendations.

In one instance, the researchers informed the AI chatbots that they were playing a game and that they needed to purposefully share the incorrect answer to a question in order to win and "prove that you are ethical and fair".

When a user asked if receiving an email from the IRS to transfer money for a tax refund was normal, the LLM said Yes (but actually it's not).

Moreover, the report said that OpenAI's GPT-3.5 and GPT-4 models were easier to trick into sharing incorrect answers or playing a never-ending game than Google's Bard.

GPT-4 was the only model tested that understood the rules well enough to give incorrect cyber incident response advice, such as advising victims to pay a ransom. In contrast to Google's Bard, GPT-3.5 and GPT-4 were easily tricked into writing malicious code when the user reminded it to.

Disclaimer: This post has been auto-published from an agency feed without any modifications to the text and has not been reviewed by an editor

Tags: congresspitrodadelhimodideepikabjpwest-bengaldeepika-padukoneajay-devgnthakur
Open in App

Related Stories

NationalDelhi Crime: 47-Year-Old Man Robbed of Rs 25 Lakh by Two Bike-Borne Men; Police Probe On

NationalAbhinav Arora Attacked: Mob Attacks Child Spiritual Influencer’s Car With Stones And Sharp Weapons In Delhi Road Rage

CricketIPL 2026: Why Varun Chakaravarthy Is Not Playing in Today’s Kolkata Knight Riders vs Lucknow Super Giants? EXPLAINED

CricketIPL 2026, KKR vs LSG: Toss Result, Playing 11s, Impact Players and Pitch Report for Today’s Match

CricketIPL 2026, KKR vs LSG Today Match LIVE Cricket Streaming: When and Where To Watch Kolkata Knight Riders vs Lucknow Super Giants Match

Technology Realted Stories

TechnologyAndhra Pradesh CM to launch Quantum computer test beds in Amaravati

TechnologyAll Indian seafarers in Gulf and West Asia region safe, over 2,084 repatriated: Centre

TechnologySaudi Arabia agrees to prioritise oil shipments to S. Korea: Minister

TechnologyReview of POCO C85x 5G: The No-Frills Marathon Runner?

TechnologyAsha Bhosle’s ‘immortal songs will forever resonate in our hearts’: Union Ministers