Most AI systems look safe until you start pushing them.
In this episode of The Unlearning Room by Forget, we explore what happens when AI models are put under pressure. Not normal usage, but adversarial questioning, repeated probing, and edge case attacks designed to make models reveal what they should no longer know.
We talk about why compliance claims often collapse under stress, how deleted data can resurface through indirect prompts, and why silence, refusal, or hesitation can be more meaningful than confident answers. The episode also looks at how the Forget Protocol is used to test AI behavior under pressure, treating adversarial attacks as a verification method rather than a threat.
If you build, audit, or deploy AI systems, this episode explains why unlearning is not proven until someone actively tries to break it.