
Sign up to save your podcasts
Or


Today we’re joined by David Ha, a research scientist at Google.
In nature, there are many examples of “bottlenecks”, or constraints, that have shaped our development as a species. Building upon this idea, David posits that these same evolutionary bottlenecks could work when training neural network models as well. In our conversation with David, we cover a TON of ground, including the aforementioned biological inspiration for his work, then digging deeper into the different types of constraints he’s applied to ML systems. We explore abstract generative models and how advanced training agents inside of generative models has become, and quite a few papers including Neuroevolution of self-interpretable agents, World Models and Attention for Reinforcement Learning, and The Sensory Neuron as a Transformer: Permutation-Invariant Neural Networks for Reinforcement Learning.
This interview is Nerd Alert certified, so get your notes ready!
PS. David is one of our favorite follows on Twitter (@hardmaru), so check him out and share your thoughts on this interview and his work!
The complete show notes for this episode can be found at twimlai.com/go/535
By Sam Charrington4.7
422422 ratings
Today we’re joined by David Ha, a research scientist at Google.
In nature, there are many examples of “bottlenecks”, or constraints, that have shaped our development as a species. Building upon this idea, David posits that these same evolutionary bottlenecks could work when training neural network models as well. In our conversation with David, we cover a TON of ground, including the aforementioned biological inspiration for his work, then digging deeper into the different types of constraints he’s applied to ML systems. We explore abstract generative models and how advanced training agents inside of generative models has become, and quite a few papers including Neuroevolution of self-interpretable agents, World Models and Attention for Reinforcement Learning, and The Sensory Neuron as a Transformer: Permutation-Invariant Neural Networks for Reinforcement Learning.
This interview is Nerd Alert certified, so get your notes ready!
PS. David is one of our favorite follows on Twitter (@hardmaru), so check him out and share your thoughts on this interview and his work!
The complete show notes for this episode can be found at twimlai.com/go/535

1,106 Listeners

168 Listeners

306 Listeners

345 Listeners

232 Listeners

209 Listeners

204 Listeners

313 Listeners

100 Listeners

553 Listeners

147 Listeners

103 Listeners

229 Listeners

689 Listeners

34 Listeners