
仅需250份恶意文档即可攻破大语言模型:数据投毒攻击门槛远低于预期
A joint study reveals that poisoning large language models requires only a fixed number of malicious documents (as few as 250), regardless of model size or training data volume, challenging previous assumptions about attack feasibility. (一项联合研究表明,无论模型规模或训练数据量如何,仅需固定数量的恶意文档(少至250份)即可对大语言模型进行数据投毒攻击,这挑战了先前关于攻击可行性的假设。)
2026/1/24
阅读全文 →









