Researchers from Anthropic, the UK AI Security Institute, and the Alan Turing Institute discovered that LLMs can be made vulnerable with just a small amount of poisoned data. New experiments show that approximately 250 malicious documents are sufficient to create a backdoor, regardless of the model size or the amount of training data. The study, […]