
Sign up to save your podcasts
Or


In machine learning, typically optimization is done to produce a model that performs well according to some metric. Today's episode features Evan Hubinger talking about what happens when the learned model itself is doing optimization in order to perform well, how the goals of the learned model could differ from the goals we used to select the learned model, and what would happen if they did differ.
Link to the paper - Risks from Learned Optimization in Advanced Machine Learning Systems: arxiv.org/abs/1906.01820
Link to the transcript: axrp.net/episode/2021/02/17/episode-4-risks-from-learned-optimization-evan-hubinger.html
Evan Hubinger's Alignment Forum profile: alignmentforum.org/users/evhub
By Daniel Filan4.4
88 ratings
In machine learning, typically optimization is done to produce a model that performs well according to some metric. Today's episode features Evan Hubinger talking about what happens when the learned model itself is doing optimization in order to perform well, how the goals of the learned model could differ from the goals we used to select the learned model, and what would happen if they did differ.
Link to the paper - Risks from Learned Optimization in Advanced Machine Learning Systems: arxiv.org/abs/1906.01820
Link to the transcript: axrp.net/episode/2021/02/17/episode-4-risks-from-learned-optimization-evan-hubinger.html
Evan Hubinger's Alignment Forum profile: alignmentforum.org/users/evhub

26,371 Listeners

2,426 Listeners

1,083 Listeners

107 Listeners

112,356 Listeners

210 Listeners

9,793 Listeners

89 Listeners

489 Listeners

5,473 Listeners

132 Listeners

16,106 Listeners

97 Listeners

209 Listeners

133 Listeners