LLMs Research Podcast

Mamba's Memory Problem


Listen Later

State space models like Mamba promised linear scaling and constant memory. They delivered on efficiency, but researchers kept hitting the same wall: ask Mamba to recall something specific from early in a long context, and performance drops.

Three papers at ICLR 2026 independently attacked this limitation. That convergence tells you how fundamental the problem is.

This podcast breaks down:

- Why Mamba's fixed-size state causes "lossy compression" of context

- How Mixture of Memories (MoM) adds multiple internal memory banks

- How Log-Linear Attention finds a middle ground between SSM and full attention

- Why one paper proves SSMs fundamentally can't solve certain tasks without external tools

The pattern across all three: you can add more state, but you have to pay somewhere. Parameters, mechanism complexity, or system infrastructure. No free lunch.

📄 Papers covered:

- MoM: Linear Sequence Modeling with Mixture-of-Memories

https://arxiv.org/abs/2502.13685

- Log-Linear Attention

https://openreview.net/forum?id=mOJgZWkXKW

- To Infinity and Beyond: Tool-Use Unlocks Length Generalization in SSMs

https://openreview.net/forum?id=sSfep4udCb

📬 Newsletter: https://llmsresearch.substack.com

🐦 Twitter/X: https://x.com/llmsresearch

💻 GitHub: https://github.com/llmsresearch

#Mamba #SSM #StateSpaceModels #ICLR2026 #LLM #MachineLearning #AIResearch #Transformers #DeepLearningChapters timestamp0:00 Mamba's secret weakness

0:42 The promise: linear scaling, constant memory

1:14 The catch: forgetting specific details

1:34 Memory bottleneck explained

1:43 Attention = perfect recall filing cabinet

2:10 SSM = single notepad with fixed pages

2:49 The core tradeoff

2:57 Three solutions to fix it

3:00 Solution 1: Mixture of Memories (MoM)

3:51 Solution 2: Log-Linear Attention

4:48 Solution 3: External tool use

5:49 The "no free lunch" pattern

6:41 What wins for longer contexts?

7:04 Subscribe for more research deep dives



This is a public episode. If you would like to discuss this with other subscribers or get access to bonus episodes, visit llmsresearch.substack.com
...more
View all episodesView all episodes
Download on the App Store

LLMs Research PodcastBy LLMs Research