October 9, 2025 // Vulnerability | #LLM data poisoning #Backdoor vulnerability #Fixed-size training data poisoning

A small number of samples can poison LLMs of any size - Anthropic

Researchers demonstrated that as few as 250 poisoned documents can create a backdoor vulnerability in large language models, irrespective of model size or training data volume. This data poisoning technique, which can induce denial-of-service or potentially facilitate data exfiltration, challenges prior assumptions about the required scale of malicious training data during pretraining.


Source: Original Report ↗
← Back to Feed