
Sign up to save your podcasts
Or
Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks. Finally, we delve into the future of AI security, and the need for a better approach to mitigate the risks posed by optimized adversarial attacks.
The complete show notes for this episode can be found at twimlai.com/go/678.
4.7
416416 ratings
Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks. Finally, we delve into the future of AI security, and the need for a better approach to mitigate the risks posed by optimized adversarial attacks.
The complete show notes for this episode can be found at twimlai.com/go/678.
161 Listeners
481 Listeners
298 Listeners
331 Listeners
156 Listeners
192 Listeners
298 Listeners
88 Listeners
428 Listeners
121 Listeners
201 Listeners
75 Listeners
491 Listeners
31 Listeners
43 Listeners