Links

All links

250

Researchers from Anthropic, the UK AI Security Institute and the Alan Turing institute, found it is easier than thought to poison very large models:

we demonstrate that by injecting just 250 malicious documents into pretraining data, adversaries can successfully backdoor LLMs ranging from 600M to 13B parameters

(From: A small number of samples can poison LLMs of any size \ Anthropic)