CONNECT WITH US

AI poisoning: Anthropic study finds just 250 malicious files can compromise the largest LLMs

Amanda Liang, Taipei; Willis Ke, DIGITIMES Asia 0

Contrary to long-held beliefs that attacking or contaminating large language models (LLMs) requires enormous volumes of malicious data, new research from AI startup Anthropic, conducted in collaboration with the UK AI Security Institute and the Alan Turing Institute, has found otherwise. The study demonstrates that as few as 250 poisoned documents can implant backdoor vulnerabilities into LLMs, regardless of their size or training dataset volume.

Findings challenge longstanding assumptions about AI security

This groundbreaking discovery overturns the conventional assumption that larger LLMs need proportionally more malicious data to be compromised. The joint investigation—currently the largest-scale LLM data poisoning study to date—revealed that models ranging from 600 million to 13 billion parameters could be effectively backdoored with a small fixed number of toxic files.

According to reports from Futurism and Fortune, LLMs like Claude are pretrained on vast amounts of publicly available internet text, including blogs and personal websites. This open sourcing means that anyone can publish content online that might later be absorbed into an LLM's training corpus, providing a pathway for malicious actors to insert poisoned data.

Backdoor attacks: The hidden triggers in AI behavior

A common form of data poisoning is the backdoor attack, where specific trigger phrases cause the LLM to exhibit hidden or unintended behaviors. For instance, a seemingly benign phrase embedded during training could later prompt the model to leak sensitive information when encountered.

Anthropic warns that these backdoors pose serious risks to AI safety and trustworthiness, especially in sensitive applications. Alarmingly, the research shows that even very large models require only a few hundred malicious files to become vulnerable.

"Attackers only need to inject a fixed number of malicious documents rather than a proportionate amount relative to the LLM's parameter count," Anthropic stated. "This makes poisoning attacks potentially far more feasible than previously thought."

Understanding AI poisoning and its techniques

As explained by The Conversation, AI poisoning generally refers to deliberately teaching incorrect or harmful knowledge to AI systems to degrade performance, induce specific errors, or activate hidden malicious functions. Technically, this manipulation during model training is called data poisoning, while post-training tampering targets the LLM directly. In practice, both lead to similar outcomes—altering model behavior in ways that are difficult to detect.

Data poisoning typically manifests in two forms: direct (targeted) attacks, which aim to alter outputs for specific queries, and indirect (untargeted) attacks, which seek to overall model effectiveness.

The most prevalent direct method is the backdoor, where the LLM secretly learns to respond differently upon seeing rare trigger tokens embedded in seemingly normal training samples.

How attackers exploit open web data

For example, attackers might insert unique trigger words into datasets used for training or fine-tuning. These triggers are invisible to ordinary users but can later allow adversaries to activate the backdoor via carefully crafted inputs. In some cases, attackers may plant such triggers within websites or social media posts, which are then automatically scraped by data-gathering algorithms, enabling stealthy exploitation without user awareness.

Indirect poisoning commonly involves "topic steering," where biased or false content is injected into training data. Consequently, the LLM begins repeating misinformation as fact without any explicit trigger. For example, if attackers want the LLM to believe "eating lettuce cures cancer," they could flood the web with pages presenting this falsehood as truth. If the LLM scrapes these pages, it may regurgitate this misinformation when asked about cancer treatments.

Implications for AI developers and policymakers

The findings underscore the urgent need for stronger data hygiene, model auditing, and training transparency across the AI industry. If even a few hundred poisoned documents can compromise multi-billion-parameter models, robust data vetting pipelines and provenance tracking become essential.

Experts say policymakers should also prioritize standards for dataset integrity and independent auditing to mitigate potential exploitation. As AI systems continue to integrate into critical areas such as healthcare, finance, and national security, safeguarding them against data poisoning will be vital to ensuring public trust, accountability, and long-term reliability

Article edited by Joseph Tsai