
Sign up to save your podcasts
Or
In machine learning, typically optimization is done to produce a model that performs well according to some metric. Today's episode features Evan Hubinger talking about what happens when the learned model itself is doing optimization in order to perform well, how the goals of the learned model could differ from the goals we used to select the learned model, and what would happen if they did differ.
Link to the paper - Risks from Learned Optimization in Advanced Machine Learning Systems: arxiv.org/abs/1906.01820
Link to the transcript: axrp.net/episode/2021/02/17/episode-4-risks-from-learned-optimization-evan-hubinger.html
Evan Hubinger's Alignment Forum profile: alignmentforum.org/users/evhub
4.4
88 ratings
In machine learning, typically optimization is done to produce a model that performs well according to some metric. Today's episode features Evan Hubinger talking about what happens when the learned model itself is doing optimization in order to perform well, how the goals of the learned model could differ from the goals we used to select the learned model, and what would happen if they did differ.
Link to the paper - Risks from Learned Optimization in Advanced Machine Learning Systems: arxiv.org/abs/1906.01820
Link to the transcript: axrp.net/episode/2021/02/17/episode-4-risks-from-learned-optimization-evan-hubinger.html
Evan Hubinger's Alignment Forum profile: alignmentforum.org/users/evhub
26,377 Listeners
2,397 Listeners
1,779 Listeners
296 Listeners
104 Listeners
4,097 Listeners
87 Listeners
281 Listeners
88 Listeners
354 Listeners
199 Listeners
63 Listeners
64 Listeners
136 Listeners
116 Listeners