2025-10-09 Hacker News Top Articles and Its Summaries
1. A small number of samples can poison LLMs of any size Total comment counts : 48 Summary New research from Anthropic, the UK AI Security Institute, and The Alan Turing Institute shows that poisoning a pretrained LLM with only 250 malicious documents can implant a backdoor that makes the model output random gibberish when it encounters a trigger, across models from 600M to 13B parameters. This challenges the idea that backdoors require a percentage of training data; the amount needed can be constant....