
Sign up to save your podcasts
Or


Good day, here's your AI digest for Monday, March 30th, 2026.
The big story today: Anthropic accidentally exposed details of its next flagship model, Claude Mythos. A CMS configuration error left an unpublished blog post in a public data store. The draft describes Mythos as a step change in capabilities, placing it in a new tier above Opus called Capybara. Anthropic confirmed a new model is in testing with major advances in reasoning, coding, and cybersecurity — though it warns the model is compute-intensive and expensive to serve. It sent cybersecurity stocks into freefall on Friday.
Anthropic also launched scheduled tasks for Claude Code on the web. Tasks run on Anthropic-managed infrastructure, meaning they keep running even when your device is off. Example use cases: reviewing open pull requests every morning, analyzing CI failures overnight, syncing documentation after PRs merge, and running dependency audits weekly. Available to all Claude Code web users now.
OpenAI launched plugins for Codex, letting users bundle skills, workplace apps, and Model Context Protocol servers into reusable workflows. You install them from the plugins tab in the Codex app. This mirrors Claude Code functionality, and signals OpenAI is pushing harder into the coding agent space.
For engineers building agents: AutoBe is an open-source AI agent that takes a natural language description and generates a complete backend. It solves the low function calling success rate problem — one model tested at only 6.75% success — using a harness where type schemas constrain outputs, compilers verify results, and structured feedback tells the agent exactly what went wrong. The result is over 99.8% success rate. Well worth reading if you are building agentic systems.
A new open-source spec called lat.md has agents maintain a markdown file alongside your codebase. It captures big ideas, business logic, key corner cases, and high-level tests, saving agents from endless grepping. It uses wiki-style links to connect concepts into a navigable graph.
Cisco Principal Engineer Yuri Kramarz published a practical five-step framework for building reliable AI agents: give your agent a one-sentence identity, explicitly define what it will not do, force it through an Observe-Reflect-Act loop, add a self-validation checkpoint before output, and state its limitations plainly. The key insight is that agents that double-check their work outperform cleverer ones. Clarity beats clever.
Meta's next major model, Avocado, has been pushed back to at least May. It still falls short of leading systems from competitors, though it handles complex math that earlier Llama models could not. Notably, some Meta AI requests are already being routed through Google's Gemini models, and Meta leadership has reportedly discussed temporarily licensing Gemini technology.
Google's Gemini app now lets you import preferences and chat history from other AI tools. If you have built up context in ChatGPT or elsewhere and want to move to Gemini, you can do it via Settings, then Import. Makes switching less painful.
A developer used an open-source multi-agent coding harness to build a fully functional browser-based digital audio workstation over 20 hours of largely unattended autonomous work. For comparison, Anthropic's internal version using its own tools cost 124 dollars in API credits and finished in under four hours — but produced a less complete result. A useful data point on where autonomous coding agents actually stand.
A developer reverse-engineered ChatGPT's bot detection, revealing that every message triggers a hidden Cloudflare program checking 55 properties across your browser, Cloudflare's network, and ChatGPT's React app state. Bots that spoof browsers but do not fully render the ChatGPT app will fail. Worth knowing if you are building automated integrations.
An analysis on the capability overhang in AI: coding agents outperform other domains because codebases are self-contained environments. Enterprise knowledge work lags because it is fragmented across video calls, legacy systems, and siloed tools. The three enterprise blockers are context fragmentation, complex access control, and a rapidly shifting architecture landscape.
OpenAI alumni shared lessons learned: building good evals is as important as building the model. The best benchmarks drive collective optimization. Post-training data design is critical for capabilities in subjective areas like empathy and creativity. Fast iteration and choosing the right problems are the most durable competitive advantages in AI research.
The Wall Street Journal published a deep investigation into the decade-long feud between Sam Altman and Dario Amodei, based on interviews with current and former employees at both companies. The story traces the split to a 2020 confrontation in which Altman accused the Amodeis of organizing board feedback against him. Dario's conditions to stay — report directly to the board, never work with Greg Brockman again — were both rejected. He left and founded Anthropic. The two companies now embody opposite answers to the same question about how fast to move and who should be in charge of the most powerful AI in the world.
Bluesky launched Attie, a standalone app powered by Claude that lets anyone build custom social feeds using plain-language commands. The longer-term plan is to let users build their own apps on the AT Protocol. It is a clear signal of where decentralized social platforms are heading as AI lowers the barrier to custom experiences.
A research paper argues AI is not eliminating jobs outright but unbundling them — splitting roles into narrower, lower-paid tasks. Workers in what the paper calls weak-bundle roles, including coding support and ticket handling, are most at risk of having their work quietly hollowed out. Worth keeping in mind as you think about how AI reshapes the shape of software engineering roles over the next few years.
Finally, Anthropic reports that Claude's paid subscriptions have more than doubled this year, with most new subscribers on the lowest tier. OpenAI still holds the largest consumer AI platform, but the gap is narrowing. This growth comes even as Claude subscribers on hundred-dollar-a-month plans report hitting rate limits within an hour — a sign that the more capable the model, the harder it becomes to keep it accessible.
This has been your AI digest for Monday, March 30th, 2026.
Read more:
By Arthur KhachatryanGood day, here's your AI digest for Monday, March 30th, 2026.
The big story today: Anthropic accidentally exposed details of its next flagship model, Claude Mythos. A CMS configuration error left an unpublished blog post in a public data store. The draft describes Mythos as a step change in capabilities, placing it in a new tier above Opus called Capybara. Anthropic confirmed a new model is in testing with major advances in reasoning, coding, and cybersecurity — though it warns the model is compute-intensive and expensive to serve. It sent cybersecurity stocks into freefall on Friday.
Anthropic also launched scheduled tasks for Claude Code on the web. Tasks run on Anthropic-managed infrastructure, meaning they keep running even when your device is off. Example use cases: reviewing open pull requests every morning, analyzing CI failures overnight, syncing documentation after PRs merge, and running dependency audits weekly. Available to all Claude Code web users now.
OpenAI launched plugins for Codex, letting users bundle skills, workplace apps, and Model Context Protocol servers into reusable workflows. You install them from the plugins tab in the Codex app. This mirrors Claude Code functionality, and signals OpenAI is pushing harder into the coding agent space.
For engineers building agents: AutoBe is an open-source AI agent that takes a natural language description and generates a complete backend. It solves the low function calling success rate problem — one model tested at only 6.75% success — using a harness where type schemas constrain outputs, compilers verify results, and structured feedback tells the agent exactly what went wrong. The result is over 99.8% success rate. Well worth reading if you are building agentic systems.
A new open-source spec called lat.md has agents maintain a markdown file alongside your codebase. It captures big ideas, business logic, key corner cases, and high-level tests, saving agents from endless grepping. It uses wiki-style links to connect concepts into a navigable graph.
Cisco Principal Engineer Yuri Kramarz published a practical five-step framework for building reliable AI agents: give your agent a one-sentence identity, explicitly define what it will not do, force it through an Observe-Reflect-Act loop, add a self-validation checkpoint before output, and state its limitations plainly. The key insight is that agents that double-check their work outperform cleverer ones. Clarity beats clever.
Meta's next major model, Avocado, has been pushed back to at least May. It still falls short of leading systems from competitors, though it handles complex math that earlier Llama models could not. Notably, some Meta AI requests are already being routed through Google's Gemini models, and Meta leadership has reportedly discussed temporarily licensing Gemini technology.
Google's Gemini app now lets you import preferences and chat history from other AI tools. If you have built up context in ChatGPT or elsewhere and want to move to Gemini, you can do it via Settings, then Import. Makes switching less painful.
A developer used an open-source multi-agent coding harness to build a fully functional browser-based digital audio workstation over 20 hours of largely unattended autonomous work. For comparison, Anthropic's internal version using its own tools cost 124 dollars in API credits and finished in under four hours — but produced a less complete result. A useful data point on where autonomous coding agents actually stand.
A developer reverse-engineered ChatGPT's bot detection, revealing that every message triggers a hidden Cloudflare program checking 55 properties across your browser, Cloudflare's network, and ChatGPT's React app state. Bots that spoof browsers but do not fully render the ChatGPT app will fail. Worth knowing if you are building automated integrations.
An analysis on the capability overhang in AI: coding agents outperform other domains because codebases are self-contained environments. Enterprise knowledge work lags because it is fragmented across video calls, legacy systems, and siloed tools. The three enterprise blockers are context fragmentation, complex access control, and a rapidly shifting architecture landscape.
OpenAI alumni shared lessons learned: building good evals is as important as building the model. The best benchmarks drive collective optimization. Post-training data design is critical for capabilities in subjective areas like empathy and creativity. Fast iteration and choosing the right problems are the most durable competitive advantages in AI research.
The Wall Street Journal published a deep investigation into the decade-long feud between Sam Altman and Dario Amodei, based on interviews with current and former employees at both companies. The story traces the split to a 2020 confrontation in which Altman accused the Amodeis of organizing board feedback against him. Dario's conditions to stay — report directly to the board, never work with Greg Brockman again — were both rejected. He left and founded Anthropic. The two companies now embody opposite answers to the same question about how fast to move and who should be in charge of the most powerful AI in the world.
Bluesky launched Attie, a standalone app powered by Claude that lets anyone build custom social feeds using plain-language commands. The longer-term plan is to let users build their own apps on the AT Protocol. It is a clear signal of where decentralized social platforms are heading as AI lowers the barrier to custom experiences.
A research paper argues AI is not eliminating jobs outright but unbundling them — splitting roles into narrower, lower-paid tasks. Workers in what the paper calls weak-bundle roles, including coding support and ticket handling, are most at risk of having their work quietly hollowed out. Worth keeping in mind as you think about how AI reshapes the shape of software engineering roles over the next few years.
Finally, Anthropic reports that Claude's paid subscriptions have more than doubled this year, with most new subscribers on the lowest tier. OpenAI still holds the largest consumer AI platform, but the gap is narrowing. This growth comes even as Claude subscribers on hundred-dollar-a-month plans report hitting rate limits within an hour — a sign that the more capable the model, the harder it becomes to keep it accessible.
This has been your AI digest for Monday, March 30th, 2026.
Read more: