Search This Blog

Powered by Blogger.

Blog Archive

Labels

Footer About

Footer About

Labels

Showing posts with label Malicious Chatbot Scam. Show all posts

AI Poisoning: How Malicious Data Corrupts Large Language Models Like ChatGPT and Claude

 

Poisoning is a term often associated with the human body or the environment, but it is now a growing problem in the world of artificial intelligence. Large language models such as ChatGPT and Claude are particularly vulnerable to this emerging threat known as AI poisoning. A recent joint study conducted by the UK AI Security Institute, the Alan Turing Institute, and Anthropic revealed that inserting as few as 250 malicious files into a model’s training data can secretly corrupt its behavior. 

AI poisoning occurs when attackers intentionally feed false or misleading information into a model’s training process to alter its responses, bias its outputs, or insert hidden triggers. The goal is to compromise the model’s integrity without detection, leading it to generate incorrect or harmful results. This manipulation can take the form of data poisoning, which happens during the model’s training phase, or model poisoning, which occurs when the model itself is modified after training. Both forms overlap since poisoned data eventually influences the model’s overall behavior. 

A common example of a targeted poisoning attack is the backdoor method. In this scenario, attackers plant specific trigger words or phrases in the data—something that appears normal but activates malicious behavior when used later. For instance, a model could be programmed to respond insultingly to a question if it includes a hidden code word like “alimir123.” Such triggers remain invisible to regular users but can be exploited by those who planted them. 

Indirect attacks, on the other hand, aim to distort the model’s general understanding of topics by flooding its training sources with biased or false content. If attackers publish large amounts of misinformation online, such as false claims about medical treatments, the model may learn and reproduce those inaccuracies as fact. Research shows that even a tiny amount of poisoned data can cause major harm. 

In one experiment, replacing only 0.001% of the tokens in a medical dataset caused models to spread dangerous misinformation while still performing well in standard tests. Another demonstration, called PoisonGPT, showed how a compromised model could distribute false information convincingly while appearing trustworthy. These findings highlight how subtle manipulations can undermine AI reliability without immediate detection. Beyond misinformation, poisoning also poses cybersecurity threats. 

Compromised models could expose personal information, execute unauthorized actions, or be exploited for malicious purposes. Previous incidents, such as the temporary shutdown of ChatGPT in 2023 after a data exposure bug, demonstrate how fragile even the most secure systems can be when dealing with sensitive information. Interestingly, some digital artists have used data poisoning defensively to protect their work from being scraped by AI systems. 

By adding misleading signals to their content, they ensure that any model trained on it produces distorted outputs. This tactic highlights both the creative and destructive potential of data poisoning. The findings from the UK AI Security Institute, Alan Turing Institute, and Anthropic underline the vulnerability of even the most advanced AI models. 

As these systems continue to expand into everyday life, experts warn that maintaining the integrity of training data and ensuring transparency throughout the AI development process will be essential to protect users and prevent manipulation through AI poisoning.

Cybercriminals Exploit Curiosity in Criminal ChatGPT Variant to Deceive Fellow Crooks

 

A variant of ChatGPT, originally crafted to aid cybercriminals, has backfired by duping potential buyers seeking access to its services.

Earlier in July, PC Mag wrote about  WormGPT, a chatbot developed from publicly available source code, which claimed to assist hackers in generating phishing messages and malware for a monthly fee. This raised concerns about the potential for generative AI to facilitate computer hacking and exacerbate cybercrime.

Ironically, it seems that the WormGPT brand has turned into a greater hazard for hackers than for the general public. Antivirus provider Kaspersky identified multiple websites purporting to provide entry to WormGPT, but instead appeared to be devised to deceive prospective customers into parting with their money, without actually providing access to WormGPT.

These websites, accessible on the public internet and through a Google search, have been adorned with official-looking details about WormGPT. Nevertheless, Kaspersky suspects that these pages are, in reality, phishing sites engineered to deceive users into divulging their credit card particulars or surrendering their cryptocurrency in order to utilize the malevolent chatbot.

The likelihood of these websites being fraudulent is heightened by the fact that the creator of WormGPT purportedly abandoned the project after their identity was uncovered last month. As reported by security journalist Brian Krebs, the individual behind WormGPT is a 23-year-old Portuguese programmer named Rafael Morais, who has since distanced himself from promoting his chatbot for malicious purposes.

Subsequent to this revelation, the user account endorsing WormGPT announced on a hacking forum that their team was discontinuing the project. 

The account expressed, "With great sadness, I come to inform everyone about the end of the WormGPT project. From the beginning, we never thought we would gain this level of visibility, and our intention was never to create something of this magnitude."

In the weeks leading up to the shutdown, the official WormGPT account on Telegram also issued a warning about scammers imitating the chatbot’s brand. The post stated, "We don’t have any website and either any other groups in any platform. The rest are resellers or scammers!"

Later, the same account expressed disbelief at how people were still falling victim to scams in 2023.

Despite the apparent shutdown of WormGPT, the menace of cybercriminals turning to other malevolent chatbots for aid remains. In July, a different developer was observed creating FraudGPT.