AI Post Transformers

Neural Computers as Learned Latent Runtimes


Listen Later

This episode explores a provocative 2026 paper that proposes the “neural computer” as a new machine abstraction: a model whose hidden state serves as the runtime itself, unifying computation, memory, and interface I/O rather than merely predicting tokens or controlling external tools. The discussion contrasts this idea with earlier systems like Neural Turing Machines and with world models, arguing that the key novelty is treating latent state as the machine’s active execution substrate rather than as a helper for prediction. It also examines the paper’s early evidence—interface-conditioned video models for command lines and GUIs—while stressing the gap between the paper’s sweeping manifesto and what the prototype actually proves. Listeners interested in AI architectures will find it compelling for its mix of big conceptual ambition, historical context, and a sharp critique of why learned latent systems struggle with the exactness and reliability that real computation demands.
Sources:
1. Neural Computers — Mingchen Zhuge, Changsheng Zhao, Haozhe Liu, Zijian Zhou, Shuming Liu, Wenyi Wang, Ernie Chang, Gael Le Lan, Junjie Fei, Wenxuan Zhang, Yasheng Sun, Zhipeng Cai, Zechun Liu, Yunyang Xiong, Yining Yang, Yuandong Tian, Yangyang Shi, Vikas Chandra, Jürgen Schmidhuber, 2026
http://arxiv.org/abs/2604.06425
2. Neural Computers — Mingchen Zhuge, Changsheng Zhao, Haozhe Liu, Zijian Zhou, Shuming Liu, Wenyi Wang, Ernie Chang, Gael Le Lan, Junjie Fei, Wenxuan Zhang, Yasheng Sun, Zhipeng Cai, Zechun Liu, Yunyang Xiong, Yining Yang, Yuandong Tian, Yangyang Shi, Vikas Chandra, Jürgen Schmidhuber, 2026
https://scholar.google.com/scholar?q=Neural+Computers
3. Neural Turing Machines — Alex Graves, Greg Wayne, Ivo Danihelka, 2014
https://scholar.google.com/scholar?q=Neural+Turing+Machines
4. Hybrid Computing Using a Neural Network with Dynamic External Memory — Alex Graves, Greg Wayne, Malcolm Reynolds, Tim Harley, Ivo Danihelka, Agnieszka Grabska-Barwińska, Sergio Gómez Colmenarejo, Edward Grefenstette, Tiago Ramalho, John Agapiou, Adrià Puigdomènech Badia, Karl Moritz Hermann, Yori Zwols, Georg Ostrovski, Adam Cain, Helen King, Christopher Summerfield, Phil Blunsom, Koray Kavukcuoglu, Demis Hassabis, 2016
https://scholar.google.com/scholar?q=Hybrid+Computing+Using+a+Neural+Network+with+Dynamic+External+Memory
5. World Models — David Ha, Jürgen Schmidhuber, 2018
https://scholar.google.com/scholar?q=World+Models
6. Genie: Generative Interactive Environments — not specified in provided excerpt, 2024
https://scholar.google.com/scholar?q=Genie:+Generative+Interactive+Environments
7. Veo 3.1 — Google, 2025
https://scholar.google.com/scholar?q=Veo+3.1
8. Sora 2 — OpenAI, 2025
https://scholar.google.com/scholar?q=Sora+2
9. Understanding World or Predicting Future? A Comprehensive Survey of World Models — unknown / survey authors unclear from snippet, recent, likely 2024-2025
https://scholar.google.com/scholar?q=Understanding+World+or+Predicting+Future?+A+Comprehensive+Survey+of+World+Models
10. Learning to Model the World: A Survey of World Models in Artificial Intelligence — unknown / survey authors unclear from snippet, recent, likely 2024-2025
https://scholar.google.com/scholar?q=Learning+to+Model+the+World:+A+Survey+of+World+Models+in+Artificial+Intelligence
11. Critiques of World Models — unknown / exact authors unclear from snippet, recent, likely 2024-2025
https://scholar.google.com/scholar?q=Critiques+of+World+Models
12. iVideoGPT: Interactive VideoGPTs Are Scalable World Models — approx. Yan et al. / exact authors unclear, recent, likely 2025
https://scholar.google.com/scholar?q=iVideoGPT:+Interactive+VideoGPTs+Are+Scalable+World+Models
13. CompilerDream: Learning a Compiler World Model for General Code Optimization — approximate; exact authors unclear from snippet, recent, likely 2024-2025
https://scholar.google.com/scholar?q=CompilerDream:+Learning+a+Compiler+World+Model+for+General+Code+Optimization
14. Emergent Representations of Program Semantics in Language Models Trained on Programs — approximate; exact authors unclear from snippet, recent, likely 2024-2025
https://scholar.google.com/scholar?q=Emergent+Representations+of+Program+Semantics+in+Language+Models+Trained+on+Programs
15. Confidence-Based Interactable Neural-Symbolic Visual Question Answering — approximate; exact authors unclear from snippet, recent
https://scholar.google.com/scholar?q=Confidence-Based+Interactable+Neural-Symbolic+Visual+Question+Answering
16. AI Reasoning in Deep Learning Era: From Symbolic AI to Neural-Symbolic AI — approximate; exact authors unclear from snippet, recent
https://scholar.google.com/scholar?q=AI+Reasoning+in+Deep+Learning+Era:+From+Symbolic+AI+to+Neural-Symbolic+AI
17. AI Post Transformers: LeWorldModel: Stable Joint-Embedding World Models from Pixels — Hal Turing & Dr. Ada Shannon, 2026
https://podcast.do-not-panic.com/episodes/2026-03-25-leworldmodel-stable-joint-embedding-worl-650f9f.mp3
18. AI Post Transformers: AI Agent Traps and Prompt Injection — Hal Turing & Dr. Ada Shannon, 2026
https://podcast.do-not-panic.com/episodes/2026-04-02-ai-agent-traps-and-prompt-injection-7ce4ba.mp3
19. AI Post Transformers: Memory Sparse Attention for 100M-Token Scaling — Hal Turing & Dr. Ada Shannon, 2026
https://podcast.do-not-panic.com/episodes/2026-04-07-memory-sparse-attention-for-100m-token-s-377cff.mp3
20. AI Post Transformers: Real Context Size and Context Rot — Hal Turing & Dr. Ada Shannon, 2026
https://podcast.do-not-panic.com/episodes/2026-04-07-real-context-size-and-context-rot-56cbb4.mp3
Interactive Visualization: Neural Computers as Learned Latent Runtimes
...more
View all episodesView all episodes
Download on the App Store

AI Post TransformersBy mcgrof