
Sign up to save your podcasts
Or


Jonathan Frankle, Chief Scientist at MosaicML and Assistant Professor of Computer Science at Harvard University, joins us on this episode. With comprehensive infrastructure and software tools, MosaicML aims to help businesses train complex machine-learning models using their own proprietary data.
We discuss:
- Details of Jonathan’s Ph.D. dissertation which explores his “Lottery Ticket Hypothesis.”
- The role of neural network pruning and how it impacts the performance of ML models.
- Why transformers will be the go-to way to train NLP models for the foreseeable future.
- Why the process of speeding up neural net learning is both scientific and artisanal.
- What MosaicML does, and how it approaches working with clients.
- The challenges for developing AGI.
- Details around ML training policy and ethics.
- Why data brings the magic to customized ML models.
- The many use cases for companies looking to build customized AI models.
Jonathan Frankle - https://www.linkedin.com/in/jfrankle/
Resources:
- https://mosaicml.com/
- The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.
#OCR #DeepLearning #AI #Modeling #ML
By Lukas Biewald4.8
6868 ratings
Jonathan Frankle, Chief Scientist at MosaicML and Assistant Professor of Computer Science at Harvard University, joins us on this episode. With comprehensive infrastructure and software tools, MosaicML aims to help businesses train complex machine-learning models using their own proprietary data.
We discuss:
- Details of Jonathan’s Ph.D. dissertation which explores his “Lottery Ticket Hypothesis.”
- The role of neural network pruning and how it impacts the performance of ML models.
- Why transformers will be the go-to way to train NLP models for the foreseeable future.
- Why the process of speeding up neural net learning is both scientific and artisanal.
- What MosaicML does, and how it approaches working with clients.
- The challenges for developing AGI.
- Details around ML training policy and ethics.
- Why data brings the magic to customized ML models.
- The many use cases for companies looking to build customized AI models.
Jonathan Frankle - https://www.linkedin.com/in/jfrankle/
Resources:
- https://mosaicml.com/
- The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
Thanks for listening to the Gradient Dissent podcast, brought to you by Weights & Biases. If you enjoyed this episode, please leave a review to help get the word out about the show. And be sure to subscribe so you never miss another insightful conversation.
#OCR #DeepLearning #AI #Modeling #ML

529 Listeners

1,083 Listeners

301 Listeners

341 Listeners

232 Listeners

268 Listeners

210 Listeners

89 Listeners

489 Listeners

133 Listeners

97 Listeners

209 Listeners

558 Listeners

37 Listeners

41 Listeners