
Sign up to save your podcasts
Or
Article: https://www.aiblade.net/p/unjailbreakable-large-language-models
Since the beginning of the AI gold rush, people have used large language models for malicious intent. Drug recipes, explicit output, and discriminatory behaviour have all been elicited, with often hilarious results.
These techniques are known as “prompt injections” or “jailbreaks” - getting the LLM to perform actions outside those intended by its developers. Prompt injections could have devastating consequences in certain scenarios.
This episode aims to look at reliable countermeasures to prompt injection, and answer the burning question - “Is it possible to create 100% secure LLMs?"
Article: https://www.aiblade.net/p/unjailbreakable-large-language-models
Since the beginning of the AI gold rush, people have used large language models for malicious intent. Drug recipes, explicit output, and discriminatory behaviour have all been elicited, with often hilarious results.
These techniques are known as “prompt injections” or “jailbreaks” - getting the LLM to perform actions outside those intended by its developers. Prompt injections could have devastating consequences in certain scenarios.
This episode aims to look at reliable countermeasures to prompt injection, and answer the burning question - “Is it possible to create 100% secure LLMs?"