
Sign up to save your podcasts
Or


What if I told you that a few hundred poisoned documents could break models as big as GPT-4 or Claude? šµ Anthropic just proved it. Their new paper shows that just 250 samples can secretly backdoor any LLM, no matter the size. In todayās episode, we unpack this wild discovery, why it changes AI security forever, and what it means for the future of open-web training.
Weāll talk about:
Keywords: Anthropic, LLM security, data poisoning, backdoor attacks, TOUCAN dataset, OpenAI, Claude, Google Gemini, AI agents
Links:
Our Socials:
By AIFire.co1.8
44 ratings
What if I told you that a few hundred poisoned documents could break models as big as GPT-4 or Claude? šµ Anthropic just proved it. Their new paper shows that just 250 samples can secretly backdoor any LLM, no matter the size. In todayās episode, we unpack this wild discovery, why it changes AI security forever, and what it means for the future of open-web training.
Weāll talk about:
Keywords: Anthropic, LLM security, data poisoning, backdoor attacks, TOUCAN dataset, OpenAI, Claude, Google Gemini, AI agents
Links:
Our Socials:

1,582 Listeners

577 Listeners

212 Listeners

9,907 Listeners

306 Listeners

75 Listeners

189 Listeners

204 Listeners

95 Listeners

210 Listeners

586 Listeners

102 Listeners

176 Listeners

26 Listeners

59 Listeners