
Sign up to save your podcasts
Or


Hello AI Unraveled Listeners,
In today's AI Special Edition,
This episode explores the evolution of AI safety testing, particularly concerning large language models (LLMs). It highlights the limitations of traditional "pass/fail" red teaming and introduces a novel approach called Optimization by PROmpting (OPRO), which enables an LLM to effectively "red team itself." This new methodology focuses on evaluating the Attack Success Rate (ASR) as a distribution, offering more nuanced insights into an AI model's security. The discussion also touches upon the real-world implications for enterprises, especially in regulated industries like finance, energy and healthcare, and how OPRO can aid in demonstrating regulatory compliance and fostering accountability. Ultimately, the guest looks towards the future of AI safety, identifying upcoming challenges and areas for focused research and development.
Listen at https://podcasts.apple.com/us/podcast/the-future-of-ai-safety-testing-with-bret-kinsella-gm/id1684415169?i=1000723468669
Learn More:
By Etienne Noumen4.7
1313 ratings
Hello AI Unraveled Listeners,
In today's AI Special Edition,
This episode explores the evolution of AI safety testing, particularly concerning large language models (LLMs). It highlights the limitations of traditional "pass/fail" red teaming and introduces a novel approach called Optimization by PROmpting (OPRO), which enables an LLM to effectively "red team itself." This new methodology focuses on evaluating the Attack Success Rate (ASR) as a distribution, offering more nuanced insights into an AI model's security. The discussion also touches upon the real-world implications for enterprises, especially in regulated industries like finance, energy and healthcare, and how OPRO can aid in demonstrating regulatory compliance and fostering accountability. Ultimately, the guest looks towards the future of AI safety, identifying upcoming challenges and areas for focused research and development.
Listen at https://podcasts.apple.com/us/podcast/the-future-of-ai-safety-testing-with-bret-kinsella-gm/id1684415169?i=1000723468669
Learn More:

1,639 Listeners

334 Listeners

225 Listeners

207 Listeners

500 Listeners

197 Listeners

158 Listeners

154 Listeners

227 Listeners

615 Listeners

107 Listeners

53 Listeners

173 Listeners

55 Listeners

96 Listeners