I built a serverless AI dungeon master on AWS. Within twenty minutes, a player had broken it completely. Within forty minutes, a dragon had killed everyone at the table. Including the people who hadn't done anything wrong.
This is the story of building a real AI agent system using Amazon Bedrock Agents, AWS Lambda, API Gateway, and DynamoDB; snd then watching what real users actually do with it.
We go deep on the architecture: how Bedrock Agents handles tool orchestration and game state, why a three-second Lambda timeout will absolutely kill your AI inference workload, and what happens when an LLM gets an API contract that even AWS can't implement consistently.
But the technical failures aren't the point. The point is what a Dungeons and Dragons campaign reveals about production AI systems: the gap between what you designed for and what users actually do. Hallucinations, tool misuse, unconstrained autonomous behavior, and the cost of building AI systems you can't observe.
The same failure modes that made my friend's halfling rogue get eaten by a dragon show up in clinical decision support tools, customer service bots, and production AI agents. The sandbox is just where you can afford to find them first.
If you're building agentic AI systems, or thinking about it, this one's for you.
Dev.to Article: https://dev.to/aws-builders/building-a-serverless-dungeon-master-agent-on-aws-3j7k
opics: AWS Bedrock, serverless AI, agentic AI architecture, LLM tool use, prompt engineering, AI observability, AWS Lambda, API Gateway, DynamoDB, production AI failures, responsible AI design