Join Nostr
2025-10-14 13:39:21 UTC

Thomas Fricke (he/his) on Nostr: A small number of samples can poison #LLM s of any size \ Anthropic "... with the UK ...

A small number of samples can poison #LLM s of any size \ Anthropic
https://www.anthropic.com/research/small-samples-poison

"... with the UK #AI #Security Institute and the Alan Turing Institute, we found that as few as 250 malicious documents can produce a "#backdoor" vulnerability in a large language model—regardless of model size or training data volume. ... 13B parameter model is trained on over 20 times more training data than a 600M model, both can be backdoored by the same small number of poisoned documents"