
Sign up to save your podcasts
Or
Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks. Finally, we delve into the future of AI security, and the need for a better approach to mitigate the risks posed by optimized adversarial attacks.
The complete show notes for this episode can be found at twimlai.com/go/678.
4.7
412412 ratings
Today we're joined by Jonas Geiping, a research group leader at the ELLIS Institute, to explore his paper: "Coercing LLMs to Do and Reveal (Almost) Anything". Jonas explains how neural networks can be exploited, highlighting the risk of deploying LLM agents that interact with the real world. We discuss the role of open models in enabling security research, the challenges of optimizing over certain constraints, and the ongoing difficulties in achieving robustness in neural networks. Finally, we delve into the future of AI security, and the need for a better approach to mitigate the risks posed by optimized adversarial attacks.
The complete show notes for this episode can be found at twimlai.com/go/678.
163 Listeners
474 Listeners
297 Listeners
320 Listeners
152 Listeners
196 Listeners
275 Listeners
90 Listeners
99 Listeners
108 Listeners
179 Listeners
65 Listeners
411 Listeners
29 Listeners
32 Listeners