
Sign up to save your podcasts
Or


This episode focuses on whether social engineering techniques, traditionally used against humans, could be leveraged against language models.
Natalie and Ivan start by discussing the use of LLMs in capture the flag competitions, and mention the emergence of new types of challenges revolving around prompt hacking. This leads the discussion towards the various ways that can be used to circumvent a model's fine tuning. The hosts discuss the challenges and scoring systems in such competitions, as well as the potential ethical concerns in exploring the capabilities and limitations of AI language models.
They draw parallels between AI prompt crafting and social engineering, speculating on how techniques used in human social engineering could potentially be adapted for AI prompt hacking. Overall, the conversation explores the evolving relationship between security, AI, and the ethical considerations surrounding the use of language models in hacking competitions.
By Natalie Pistunovich & Ivan KwiatkowskiThis episode focuses on whether social engineering techniques, traditionally used against humans, could be leveraged against language models.
Natalie and Ivan start by discussing the use of LLMs in capture the flag competitions, and mention the emergence of new types of challenges revolving around prompt hacking. This leads the discussion towards the various ways that can be used to circumvent a model's fine tuning. The hosts discuss the challenges and scoring systems in such competitions, as well as the potential ethical concerns in exploring the capabilities and limitations of AI language models.
They draw parallels between AI prompt crafting and social engineering, speculating on how techniques used in human social engineering could potentially be adapted for AI prompt hacking. Overall, the conversation explores the evolving relationship between security, AI, and the ethical considerations surrounding the use of language models in hacking competitions.