
Sign up to save your podcasts
Or


Tyler Barnes, founding engineer at Mastra, introduces Observational Memory. It is a new memory system for AI agents that achieves state-of-the-art results on LongMemEval with a completely stable context window.
Unlike semantic recall (which uses RAG and invalidates prompt caching), Observational Memory compresses conversations into dense observations while maintaining a stable, fully cacheable context.
The result: 94.87% accuracy on LongMemEval with GPT-5 mini. This is the highest score recorded by any memory system to date.
In this conversation, Tyler explains how the system works, why it outperforms raw context, and how you can integrate it into your agents in under 20 minutes. We also dive into the research, the benchmarks, and what's next for Observational Memory.
🧰 RESOURCES MENTIONED
Observational Memory Launch Blog: https://mastra.ai/blog/observational-memory
📚 MASTRA RESOURCES
MASTRA?
00:00 – Intro
By MastraTyler Barnes, founding engineer at Mastra, introduces Observational Memory. It is a new memory system for AI agents that achieves state-of-the-art results on LongMemEval with a completely stable context window.
Unlike semantic recall (which uses RAG and invalidates prompt caching), Observational Memory compresses conversations into dense observations while maintaining a stable, fully cacheable context.
The result: 94.87% accuracy on LongMemEval with GPT-5 mini. This is the highest score recorded by any memory system to date.
In this conversation, Tyler explains how the system works, why it outperforms raw context, and how you can integrate it into your agents in under 20 minutes. We also dive into the research, the benchmarks, and what's next for Observational Memory.
🧰 RESOURCES MENTIONED
Observational Memory Launch Blog: https://mastra.ai/blog/observational-memory
📚 MASTRA RESOURCES
MASTRA?
00:00 – Intro