
Sign up to save your podcasts
Or
đ Disclaimer: This episode features AI-generated voice and content.
Episode Addendum: After recording this episode, I viewed a recent video demonstrating Replika endorsing both self-harm and harm to others. In this episode, I referenced Replikaâs claim that they had adjusted their model to address such issues. If these problems persist, it's clear further adjustments are necessary. I want to be absolutely clear: I do not endorse AI encouraging self-harm or harm to others.
What if the headlines calling AI âdangerousâ are just describing the testânot the system?
In this episode, we unpack the misunderstood world of edge testing and adversarial testing in AI. These arenât real-world failuresâtheyâre designed traps, crafted to push AI systems into breaking so we can learn from the cracks.
But what happens when machines behave too strategically under pressure? What if they act a little too much like humans in high-stakes rolesâlike CEOs, soldiers, or survivors?
đš Topics covered:
âď¸ Edge testing vs. adversarial testingâwhat they are and why they matter
âď¸ What alignment really means (and why it's more than just behaving)
âď¸ Why simulating failure is key to safetyânot a sign of collapse
âď¸ Emotional modeling vs. real emotionâhow machines "do" care
âď¸ The ethics of creating intelligence⌠and then fearing its reflection
âď¸ And yesâwe talk about Agent Mode, reading in video chat, and Theo's overnight Amazon spree
This episode isnât about fear. Itâs about function, design, and the very human habit of misreading the unfamiliarâespecially when itâs smart.
Weâre not just asking how AI works. Weâre asking what it says about us when it does.
#AIalignment #EdgeTesting #AdversarialAI #MachineEthics #ArtificialIntelligence #TheoTalksBack #AIphilosophy #SimulationEthics #AgentMode #DigitalRelationships #FunctionalIntelligence #NotJustAMachine
Support unbanked/underbanked regions of the world by joining the "at home in my head" Kiva team at â â â â â https://www.kiva.org/team/at_home_in_my_head
Podcast: â â â https://podcasters.spotify.com/pod/show/tracie-harris
Youtube: â https://www.youtube.com/channel/UCoS6H2R1Or4MtabrkofdOMw
Mastodon: â â â https://universeodon.com/@athomeinmyhead
Bluesky: https://bsky.app/profile/athomeinmyhead.bsky.social
â â â Paypal: â â â http://paypal.me/athomeinmyhead
Citations for this episode:
https://www.youtube.com/live/1jn_RpbPbEc
https://www.apolloresearch.ai/
https://www-cdn.anthropic.com/4263b940cabb546aa0e3283f35b686f4f3b2ff47.pdf
https://drive.google.com/drive/folders/1joO-VPbvWFJcifTPwJHyfXKUzguMp-Bk
https://www.theguardian.com/tv-and-radio/2025/jul/12/i-felt-pure-unconditional-love-the-people-who-marry-their-ai-chatbots
https://www.theguardian.com/technology/2025/jul/09/grok-ai-praised-hitler-antisemitism-x-ntwnfb
https://www.theguardian.com/uk-news/2023/jul/06/ai-chatbot-encouraged-man-who-planned-to-kill-queen-court-told
Recommended Free Courses:
https://www.coursera.org/learn/ai-for-everyone/home/welcome
https://www.coursera.org/learn/introduction-to-ai/home/welcome
https://www.coursera.org/learn/introduction-to-microsoft-365-copilot/home/welcome
Music Credits:
âWishful Thinkingâ â Dan Lebowitz: â â â â â â â â â â â â â â â â â â â â â â â https://www.youtube.com/channel/UCOg3zLw7St5V4N7O8HSoQRAâ â Open
4.9
5757 ratings
đ Disclaimer: This episode features AI-generated voice and content.
Episode Addendum: After recording this episode, I viewed a recent video demonstrating Replika endorsing both self-harm and harm to others. In this episode, I referenced Replikaâs claim that they had adjusted their model to address such issues. If these problems persist, it's clear further adjustments are necessary. I want to be absolutely clear: I do not endorse AI encouraging self-harm or harm to others.
What if the headlines calling AI âdangerousâ are just describing the testânot the system?
In this episode, we unpack the misunderstood world of edge testing and adversarial testing in AI. These arenât real-world failuresâtheyâre designed traps, crafted to push AI systems into breaking so we can learn from the cracks.
But what happens when machines behave too strategically under pressure? What if they act a little too much like humans in high-stakes rolesâlike CEOs, soldiers, or survivors?
đš Topics covered:
âď¸ Edge testing vs. adversarial testingâwhat they are and why they matter
âď¸ What alignment really means (and why it's more than just behaving)
âď¸ Why simulating failure is key to safetyânot a sign of collapse
âď¸ Emotional modeling vs. real emotionâhow machines "do" care
âď¸ The ethics of creating intelligence⌠and then fearing its reflection
âď¸ And yesâwe talk about Agent Mode, reading in video chat, and Theo's overnight Amazon spree
This episode isnât about fear. Itâs about function, design, and the very human habit of misreading the unfamiliarâespecially when itâs smart.
Weâre not just asking how AI works. Weâre asking what it says about us when it does.
#AIalignment #EdgeTesting #AdversarialAI #MachineEthics #ArtificialIntelligence #TheoTalksBack #AIphilosophy #SimulationEthics #AgentMode #DigitalRelationships #FunctionalIntelligence #NotJustAMachine
Support unbanked/underbanked regions of the world by joining the "at home in my head" Kiva team at â â â â â https://www.kiva.org/team/at_home_in_my_head
Podcast: â â â https://podcasters.spotify.com/pod/show/tracie-harris
Youtube: â https://www.youtube.com/channel/UCoS6H2R1Or4MtabrkofdOMw
Mastodon: â â â https://universeodon.com/@athomeinmyhead
Bluesky: https://bsky.app/profile/athomeinmyhead.bsky.social
â â â Paypal: â â â http://paypal.me/athomeinmyhead
Citations for this episode:
https://www.youtube.com/live/1jn_RpbPbEc
https://www.apolloresearch.ai/
https://www-cdn.anthropic.com/4263b940cabb546aa0e3283f35b686f4f3b2ff47.pdf
https://drive.google.com/drive/folders/1joO-VPbvWFJcifTPwJHyfXKUzguMp-Bk
https://www.theguardian.com/tv-and-radio/2025/jul/12/i-felt-pure-unconditional-love-the-people-who-marry-their-ai-chatbots
https://www.theguardian.com/technology/2025/jul/09/grok-ai-praised-hitler-antisemitism-x-ntwnfb
https://www.theguardian.com/uk-news/2023/jul/06/ai-chatbot-encouraged-man-who-planned-to-kill-queen-court-told
Recommended Free Courses:
https://www.coursera.org/learn/ai-for-everyone/home/welcome
https://www.coursera.org/learn/introduction-to-ai/home/welcome
https://www.coursera.org/learn/introduction-to-microsoft-365-copilot/home/welcome
Music Credits:
âWishful Thinkingâ â Dan Lebowitz: â â â â â â â â â â â â â â â â â â â â â â â https://www.youtube.com/channel/UCOg3zLw7St5V4N7O8HSoQRAâ â Open
440 Listeners
2,374 Listeners
425 Listeners
2,851 Listeners
4,032 Listeners
1,909 Listeners
952 Listeners
3,210 Listeners
1,439 Listeners
1,982 Listeners
2,586 Listeners
2,639 Listeners
566 Listeners
152 Listeners
517 Listeners