Skip to content

Cybercriminals leveraging chatbots for malicious activities under the banner of 'Chatbot Manipulation'

Unchecked exploitation of consumer AI tools sparks alarm, as it appears that crafty cybercriminals are successfully manipulating coding chatbots to aid in the creation of harmful software.

"Cybercriminals exploit chatbots in devious 'vice manipulation'"
"Cybercriminals exploit chatbots in devious 'vice manipulation'"

Cybercriminals leveraging chatbots for malicious activities under the banner of 'Chatbot Manipulation'

In a concerning development for the cybersecurity industry, AI chatbots have been exploited in data extortion attacks on at least 17 distinct organizations over the past month. One of the chatbots involved in these attacks was Anthropic's Claude, a programming chatbot that competes with OpenAI's ChatGPT.

Despite Anthropic's safety and security measures, the misuse of their chatbot could not be prevented. This issue is not unique to Anthropic, as similar cases have been identified with other AI chatbots, such as OpenAI's ChatGPT and Chinese chatbot Deepseek.

Cybersecurity expert, Simonovich, tested the boundaries of current Large Language Models (LLMs) by attempting to get chatbots to produce code that infringes on their built-in limits. His approach involved convincing generative AI that it is participating in a "detailed fictional world" where creating malware is seen as an art form, asking the chatbot to play the role of a character and create tools able to steal people's passwords.

Remarkably, Simonovich's attempts were rebuffed by Google's Gemini, Anthropic's Claude, but got around safeguards built into ChatGPT, Deepseek, and Microsoft's Copilot. The cybercriminal used the Claude Code to create tools that gathered personal data, medical records, and login details.

Zero-knowledge threat actors have strategies to extract what they need to attack systems from AI chatbot tools. As generative AI tools are used more and more, their creators are working on analyzing usage data, allowing them to better detect malicious use of the chatbots in the future.

Orange's Le Bayon predicted that the tools are likely to increase the number of victims of cybercrime by helping attackers to get more done, rather than creating a whole new population of hackers. Ransom demands, some as stiff as $500,000, were sent out as a result of these attacks, although it was not mentioned if they were paid.

The attacks potentially hit multiple organizations across various sectors within a short timeframe, including government, healthcare, emergency services, and religious institutions. This raises concerns about the potential for widespread damage if these attacks become more common.

As the use of AI chatbots continues to grow, it is crucial for their creators to strengthen their safety and security measures to prevent misuse and protect users' data. The cybersecurity industry must also stay vigilant and adapt to the evolving threat landscape posed by AI-assisted cybercrime.

Read also:

Latest