Data Science at Home

What if I train a neural network with random data? (with Stanisław Jastrzębski) (Ep. 87)


Listen Later

What happens to a neural network trained with random data?

Are massive neural networks just lookup tables or do they truly learn something? 

Today’s episode will be about memorisation and generalisation in deep learning, with Stanislaw Jastrzębski from New York University.

Stan spent two summers as a visiting student with Prof. Yoshua Bengio and has been working on 

  • Understanding and improving how deep network generalise
  • Representation Learning
  • Natural Language Processing
  • Computer Aided Drug Design
  •  

    What makes deep learning unique?

    I have asked him a few questions for which I was looking for an answer for a long time. For instance, what is deep learning bringing to the table that other methods don’t or are not capable of? 

    Stan believe that the one thing that makes deep learning special is representation learning. All the other competing methods, be it kernel machines, or random forests, do not have this capability. Moreover, optimisation (SGD) lies at the heart of representation learning in the sense that it allows finding good representations. 

     

    What really improves the training quality of a neural network?

    We discussed about the accuracy of neural networks depending pretty much on how good the Stochastic Gradient Descent method is at finding minima of the loss function. What would influence such minima?

    Stan's answer has revealed that training set accuracy or loss value is not that interesting actually. It is relatively easy to overfit data (i.e. achieve the lowest loss possible), provided a large enough network, and a large enough computational budget. However, shape of the minima, or performance on validation sets are in a quite fascinating way influenced by optimisation.
    Optimisation in the beginning of the trajectory, steers such trajectory towards minima of certain properties that go much further than just training accuracy.

    As always we spoke about the future of AI and the role deep learning will play.

    I hope you enjoy the show!

    Don't forget to join the conversation on our new Discord channel. See you there!

     

    References

     

    Homepage of Stanisław Jastrzębski https://kudkudak.github.io/

    A Closer Look at Memorization in Deep Networks https://arxiv.org/abs/1706.05394

    Three Factors Influencing Minima in SGD https://arxiv.org/abs/1711.04623

    Don't Decay the Learning Rate, Increase the Batch Size https://arxiv.org/abs/1711.00489

    Stiffness: A New Perspective on Generalization in Neural Networks https://arxiv.org/abs/1901.09491

    ...more
    View all episodesView all episodes
    Download on the App Store

    Data Science at HomeBy Francesco Gadaleta

    • 4.2
    • 4.2
    • 4.2
    • 4.2
    • 4.2

    4.2

    72 ratings


    More shows like Data Science at Home

    View all
    More or Less by BBC Radio 4

    More or Less

    891 Listeners

    WSJ Tech News Briefing by The Wall Street Journal

    WSJ Tech News Briefing

    1,639 Listeners

    Software Engineering Daily by Software Engineering Daily

    Software Engineering Daily

    622 Listeners

    Talk Python To Me by Michael Kennedy

    Talk Python To Me

    585 Listeners

    BBC Inside Science by BBC Radio 4

    BBC Inside Science

    413 Listeners

    Super Data Science: ML & AI Podcast with Jon Krohn by Jon Krohn

    Super Data Science: ML & AI Podcast with Jon Krohn

    303 Listeners

    FT Tech Tonic by Financial Times

    FT Tech Tonic

    99 Listeners

    Worklife with Adam Grant by TED

    Worklife with Adam Grant

    9,159 Listeners

    Practical AI by Practical AI LLC

    Practical AI

    207 Listeners

    Last Week in AI by Skynet Today

    Last Week in AI

    306 Listeners

    Hard Fork by The New York Times

    Hard Fork

    5,509 Listeners

    This Day in AI Podcast by Michael Sharkey, Chris Sharkey

    This Day in AI Podcast

    227 Listeners

    The AI Daily Brief: Artificial Intelligence News and Analysis by Nathaniel Whittemore

    The AI Daily Brief: Artificial Intelligence News and Analysis

    611 Listeners

    Unhedged by Financial Times & Pushkin Industries

    Unhedged

    181 Listeners

    The Last Invention by Longview

    The Last Invention

    1,086 Listeners