Brain Inspired

BI 105 Sanjeev Arora: Off the Convex Path


Listen Later

Sanjeev and I discuss some of the progress toward understanding how deep learning works, specially under previous assumptions it wouldn't or shouldn't work as well as it does. Deep learning theory poses a challenge for mathematics, because its methods aren't rooted in mathematical theory and therefore are a "black box" for math to open. We discuss how Sanjeev thinks optimization, the common framework for thinking of how deep nets learn, is the wrong approach. Instead, a promising alternative focuses on the learning trajectories that occur as a result of different learning algorithms. We discuss two examples of his research to illustrate this: creating deep nets with infinitely large layers (and the networks still find solutions among the infinite possible solutions!), and massively increasing the learning rate during training (the opposite of accepted wisdom, and yet, again, the network finds solutions!). We also discuss his past focus on computational complexity and how he doesn't share the current neuroscience optimism comparing brains to deep nets.

  • Sanjeev's website.
  • His Research group website.
  • His blog: Off The Convex Path.
  • Papers we discuss
    • On Exact Computation with an Infinitely Wide Neural Net.
    • An Exponential Learning Rate Schedule for Deep Learning
  • Related
    • The episode with Andrew Saxe covers related deep learning theory in episode 52.
    • Omri Barak discusses the importance of learning trajectories to understand RNNs in episode 97.
    • Sanjeev mentions Christos Papadimitriou.

Timestamps

0:00 - Intro
7:32 - Computational complexity
12:25 - Algorithms
13:45 - Deep learning vs. traditional optimization
17:01 - Evolving view of deep learning
18:33 - Reproducibility crisis in AI?
21:12 - Surprising effectiveness of deep learning
27:50 - "Optimization" isn't the right framework
30:08 - Infinitely wide nets
35:41 - Exponential learning rates
42:39 - Data as the next frontier
44:12 - Neuroscience and AI differences
47:13 - Focus on algorithms, architecture, and objective functions
55:50 - Advice for deep learning theorists
58:05 - Decoding minds

...more
View all episodesView all episodes
Download on the App Store

Brain InspiredBy Paul Middlebrooks

  • 4.8
  • 4.8
  • 4.8
  • 4.8
  • 4.8

4.8

134 ratings


More shows like Brain Inspired

View all
Very Bad Wizards by Tamler Sommers & David Pizarro

Very Bad Wizards

2,675 Listeners

Making Sense with Sam Harris by Sam Harris

Making Sense with Sam Harris

26,319 Listeners

Conversations with Tyler by Mercatus Center at George Mason University

Conversations with Tyler

2,452 Listeners

The Quanta Podcast by Quanta Magazine

The Quanta Podcast

533 Listeners

Closer To Truth by Closer To Truth

Closer To Truth

245 Listeners

The Michael Shermer Show by Michael Shermer

The Michael Shermer Show

937 Listeners

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas by Sean Carroll | Wondery

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

4,176 Listeners

The Origins Podcast with Lawrence Krauss by Lawrence M. Krauss

The Origins Podcast with Lawrence Krauss

505 Listeners

Google DeepMind: The Podcast by Hannah Fry

Google DeepMind: The Podcast

203 Listeners

Last Week in AI by Skynet Today

Last Week in AI

306 Listeners

Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

95 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

514 Listeners

Theories of Everything with Curt Jaimungal by Theories of Everything

Theories of Everything with Curt Jaimungal

28 Listeners

Clearer Thinking with Spencer Greenberg by Spencer Greenberg

Clearer Thinking with Spencer Greenberg

138 Listeners

Robinson's Podcast by Robinson Erhardt

Robinson's Podcast

265 Listeners