AIBlade Podcast

Unjailbreakable Large Language Models


Listen Later

Article: https://www.aiblade.net/p/unjailbreakable-large-language-models

Since the beginning of the AI gold rush, people have used large language models for malicious intent. Drug recipes, explicit output, and discriminatory behaviour have all been elicited, with often hilarious results.

These techniques are known as “prompt injections” or “jailbreaks” - getting the LLM to perform actions outside those intended by its developers. Prompt injections could have devastating consequences in certain scenarios.

This episode aims to look at reliable countermeasures to prompt injection, and answer the burning question - “Is it possible to create 100% secure LLMs?"



This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit www.aiblade.net
...more
View all episodesView all episodes
Download on the App Store

AIBlade PodcastBy David Willis-Owen