
Sign up to save your podcasts
Or


Today we're joined by Peter Hase, a fifth-year PhD student at the University of North Carolina NLP lab. We discuss "scalable oversight", and the importance of developing a deeper understanding of how large neural networks make decisions. We learn how matrices are probed by interpretability researchers, and explore the two schools of thought regarding how LLMs store knowledge. Finally, we discuss the importance of deleting sensitive information from model weights, and how "easy-to-hard generalization" could increase the risk of releasing open-source foundation models.
The complete show notes for this episode can be found at twimlai.com/go/679.
By Sam Charrington4.7
422422 ratings
Today we're joined by Peter Hase, a fifth-year PhD student at the University of North Carolina NLP lab. We discuss "scalable oversight", and the importance of developing a deeper understanding of how large neural networks make decisions. We learn how matrices are probed by interpretability researchers, and explore the two schools of thought regarding how LLMs store knowledge. Finally, we discuss the importance of deleting sensitive information from model weights, and how "easy-to-hard generalization" could increase the risk of releasing open-source foundation models.
The complete show notes for this episode can be found at twimlai.com/go/679.

1,109 Listeners

168 Listeners

307 Listeners

345 Listeners

233 Listeners

209 Listeners

204 Listeners

313 Listeners

101 Listeners

554 Listeners

146 Listeners

103 Listeners

229 Listeners

688 Listeners

34 Listeners