Brain Inspired

BI 105 Sanjeev Arora: Off the Convex Path


Listen Later

Sanjeev and I discuss some of the progress toward understanding how deep learning works, specially under previous assumptions it wouldn't or shouldn't work as well as it does. Deep learning theory poses a challenge for mathematics, because its methods aren't rooted in mathematical theory and therefore are a "black box" for math to open. We discuss how Sanjeev thinks optimization, the common framework for thinking of how deep nets learn, is the wrong approach. Instead, a promising alternative focuses on the learning trajectories that occur as a result of different learning algorithms. We discuss two examples of his research to illustrate this: creating deep nets with infinitely large layers (and the networks still find solutions among the infinite possible solutions!), and massively increasing the learning rate during training (the opposite of accepted wisdom, and yet, again, the network finds solutions!). We also discuss his past focus on computational complexity and how he doesn't share the current neuroscience optimism comparing brains to deep nets.

  • Sanjeev's website.
  • His Research group website.
  • His blog: Off The Convex Path.
  • Papers we discuss
    • On Exact Computation with an Infinitely Wide Neural Net.
    • An Exponential Learning Rate Schedule for Deep Learning
  • Related
    • The episode with Andrew Saxe covers related deep learning theory in episode 52.
    • Omri Barak discusses the importance of learning trajectories to understand RNNs in episode 97.
    • Sanjeev mentions Christos Papadimitriou.

Timestamps

0:00 - Intro
7:32 - Computational complexity
12:25 - Algorithms
13:45 - Deep learning vs. traditional optimization
17:01 - Evolving view of deep learning
18:33 - Reproducibility crisis in AI?
21:12 - Surprising effectiveness of deep learning
27:50 - "Optimization" isn't the right framework
30:08 - Infinitely wide nets
35:41 - Exponential learning rates
42:39 - Data as the next frontier
44:12 - Neuroscience and AI differences
47:13 - Focus on algorithms, architecture, and objective functions
55:50 - Advice for deep learning theorists
58:05 - Decoding minds

...more
View all episodesView all episodes
Download on the App Store

Brain InspiredBy Paul Middlebrooks

  • 4.8
  • 4.8
  • 4.8
  • 4.8
  • 4.8

4.8

134 ratings


More shows like Brain Inspired

View all
Very Bad Wizards by Tamler Sommers & David Pizarro

Very Bad Wizards

2,673 Listeners

Nature Podcast by Springer Nature Limited

Nature Podcast

761 Listeners

The Quanta Podcast by Quanta Magazine

The Quanta Podcast

526 Listeners

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) by Sam Charrington

The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

431 Listeners

Philosophy For Our Times by IAI

Philosophy For Our Times

315 Listeners

Future of Life Institute Podcast by Future of Life Institute

Future of Life Institute Podcast

107 Listeners

The Good Fight by Yascha Mounk

The Good Fight

900 Listeners

The Michael Shermer Show by Michael Shermer

The Michael Shermer Show

931 Listeners

Big Brains by University of Chicago Podcast Network

Big Brains

480 Listeners

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas by Sean Carroll | Wondery

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

4,152 Listeners

The Origins Podcast with Lawrence Krauss by Lawrence M. Krauss

The Origins Podcast with Lawrence Krauss

504 Listeners

Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

95 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

505 Listeners

Clearer Thinking with Spencer Greenberg by Spencer Greenberg

Clearer Thinking with Spencer Greenberg

139 Listeners

The Joy of Why by Steven Strogatz, Janna Levin and Quanta Magazine

The Joy of Why

491 Listeners