The Information Bottleneck

Training Is Nothing Like Learning with Naomi Saphra (Harvard)


Listen Later

Naomi Saphra, Kempner Research Fellow at Harvard and incoming Assistant Professor at Boston University, joins us to explain why you can't do interpretability without understanding training dynamics,  in the same way you can't do biology without evolution.

Naomi argues that many structures researchers find inside trained models are vestigial, they mattered early in training but are meaningless by the end. Grokking is one case of a broader phenomenon: models go through multiple consecutive phase transitions during training, driven by symmetry breaking and head specialization, but the smooth loss curve hides all of it. We talk about why training is nothing like human learning, and why our intuitions about what's hard for models are consistently wrong  -  code in pretraining helps language reasoning, tokenization drives behaviors people attribute to deeper cognition, and language already encodes everything humans care about. We also get into why SAEs are basically topic models, the Platonic representation hypothesis, using AI to decode animal communication, and why non-determinism across training runs is a real problem that RL and MoE might be making worse.

Timeline:

(00:12) Introduction and guest welcome

(01:01) Why training dynamics matter - the evolutionary biology analogy

(03:05) Jennifer Aniston neurons and the danger of biological parallels

(04:48) What is grokking and why it's one instance of a broader phenomenon

(08:25) Phase transitions, symmetry breaking, and head specialization

(11:53) Double descent, overfitting, and the death of classical train-test splits

(15:10) Training is nothing like learning

(16:08) Scaling axes - data, model size, compute, and why they're not interchangeable

(19:29) Data quality, code as reasoning fuel, and GPT-2's real contribution

(20:43) Multilingual models and the interlingua hypothesis

(25:58) The Platonic representation hypothesis and why image classification was always multimodal

(29:12) Sparse autoencoders, interpretability, and Marr's levels

(37:32) Can we ever truly understand what models know?

(43:59) The language modality chauvinist argument

(51:55) Vision, redundancy, and self-supervised learning

(57:18) World models - measurable capabilities over philosophical definitions

(1:00:14) Is coding really a solved task?

(1:04:18) Non-determinism, scaling laws, and why one training run isn't enough

(1:10:12) Naomi's new lab at BU and recruiting

Music:

  • "Kid Kodi" - Blue Dot Sessions - via Free Music Archive - CC BY-NC 4.0.
  • "Palms Down" - Blue Dot Sessions - via Free Music Archive - CC BY-NC 4.0.
  • Changes: trimmed

About: The Information Bottleneck is hosted by Ravid Shwartz-Ziv and Allen Roush, featuring in-depth conversations with leading AI researchers about the ideas shaping the future of machine learning.

...more
View all episodesView all episodes
Download on the App Store

The Information BottleneckBy Ravid Shwartz-Ziv & Allen Roush

  • 5
  • 5
  • 5
  • 5
  • 5

5

4 ratings


More shows like The Information Bottleneck

View all
The New Yorker Radio Hour by WNYC Studios and The New Yorker

The New Yorker Radio Hour

6,881 Listeners

Fareed Zakaria GPS by CNN Podcasts

Fareed Zakaria GPS

3,447 Listeners

Macro Voices by Hedge Fund Manager Erik Townsend

Macro Voices

3,072 Listeners

Odd Lots by Bloomberg

Odd Lots

1,993 Listeners

The a16z Show by Andreessen Horowitz

The a16z Show

1,105 Listeners

Practical AI by Practical AI LLC

Practical AI

212 Listeners

Interesting Times with Ross Douthat by New York Times Opinion

Interesting Times with Ross Douthat

7,244 Listeners

Google DeepMind: The Podcast by Hannah Fry

Google DeepMind: The Podcast

203 Listeners

Last Week in AI by Skynet Today

Last Week in AI

313 Listeners

Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

101 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

551 Listeners

Big Technology Podcast by Alex Kantrowitz

Big Technology Podcast

512 Listeners

Moonshots with Peter Diamandis by PHD Ventures

Moonshots with Peter Diamandis

602 Listeners

No Priors: Artificial Intelligence | Technology | Startups by Conviction

No Priors: Artificial Intelligence | Technology | Startups

150 Listeners

"Econ 102" with Noah Smith and Erik Torenberg by Turpentine

"Econ 102" with Noah Smith and Erik Torenberg

147 Listeners