Brain Inspired

BI 105 Sanjeev Arora: Off the Convex Path


Listen Later

Sanjeev and I discuss some of the progress toward understanding how deep learning works, specially under previous assumptions it wouldn't or shouldn't work as well as it does. Deep learning theory poses a challenge for mathematics, because its methods aren't rooted in mathematical theory and therefore are a "black box" for math to open. We discuss how Sanjeev thinks optimization, the common framework for thinking of how deep nets learn, is the wrong approach. Instead, a promising alternative focuses on the learning trajectories that occur as a result of different learning algorithms. We discuss two examples of his research to illustrate this: creating deep nets with infinitely large layers (and the networks still find solutions among the infinite possible solutions!), and massively increasing the learning rate during training (the opposite of accepted wisdom, and yet, again, the network finds solutions!). We also discuss his past focus on computational complexity and how he doesn't share the current neuroscience optimism comparing brains to deep nets.

  • Sanjeev's website.
  • His Research group website.
  • His blog: Off The Convex Path.
  • Papers we discuss
    • On Exact Computation with an Infinitely Wide Neural Net.
    • An Exponential Learning Rate Schedule for Deep Learning
  • Related
    • The episode with Andrew Saxe covers related deep learning theory in episode 52.
    • Omri Barak discusses the importance of learning trajectories to understand RNNs in episode 97.
    • Sanjeev mentions Christos Papadimitriou.

Timestamps

0:00 - Intro
7:32 - Computational complexity
12:25 - Algorithms
13:45 - Deep learning vs. traditional optimization
17:01 - Evolving view of deep learning
18:33 - Reproducibility crisis in AI?
21:12 - Surprising effectiveness of deep learning
27:50 - "Optimization" isn't the right framework
30:08 - Infinitely wide nets
35:41 - Exponential learning rates
42:39 - Data as the next frontier
44:12 - Neuroscience and AI differences
47:13 - Focus on algorithms, architecture, and objective functions
55:50 - Advice for deep learning theorists
58:05 - Decoding minds

...more
View all episodesView all episodes
Download on the App Store

Brain InspiredBy Paul Middlebrooks

  • 4.9
  • 4.9
  • 4.9
  • 4.9
  • 4.9

4.9

133 ratings


More shows like Brain Inspired

View all
History of Philosophy Without Any Gaps by Peter Adamson

History of Philosophy Without Any Gaps

1,580 Listeners

Closer To Truth by Closer To Truth

Closer To Truth

242 Listeners

Philosophize This! by Stephen West

Philosophize This!

15,037 Listeners

The Quanta Podcast by Quanta Magazine

The Quanta Podcast

481 Listeners

Philosophy For Our Times by IAI

Philosophy For Our Times

307 Listeners

Into the Impossible With Brian Keating by Big Bang Productions Inc.

Into the Impossible With Brian Keating

1,042 Listeners

The Michael Shermer Show by Michael Shermer

The Michael Shermer Show

918 Listeners

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas by Sean Carroll | Wondery

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

4,135 Listeners

The Origins Podcast with Lawrence Krauss by Lawrence M. Krauss

The Origins Podcast with Lawrence Krauss

487 Listeners

Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

88 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

377 Listeners

Theories of Everything with Curt Jaimungal by Theories of Everything

Theories of Everything with Curt Jaimungal

460 Listeners

Clearer Thinking with Spencer Greenberg by Spencer Greenberg

Clearer Thinking with Spencer Greenberg

128 Listeners

The Joy of Why by Steven Strogatz, Janna Levin and Quanta Magazine

The Joy of Why

498 Listeners

Robinson's Podcast by Robinson Erhardt

Robinson's Podcast

242 Listeners