
Sign up to save your podcasts
Or
Every now and then, some AI luminaries
I agree with (1) and strenuously disagree with (2).
The last time I saw something like this, I responded by writing: LeCun's “A Path Towards Autonomous Machine Intelligence” has an unsolved technical alignment problem.
Well, now we have a second entry in the series, with the new preprint book chapter “Welcome to the Era of Experience” by reinforcement learning pioneers David Silver & Richard Sutton.
The authors propose that “a new generation [...]
---
Outline:
(04:39) 1. What's their alignment plan?
(08:00) 2. The plan won't work
(08:04) 2.1 Background 1: Specification gaming and goal misgeneralization
(12:19) 2.2 Background 2: The usual agent debugging loop, and why it will eventually catastrophically fail
(15:12) 2.3 Background 3: Callous indifference and deception as the strong-default, natural way that era of experience AIs will interact with humans
(16:00) 2.3.1 Misleading intuitions from everyday life
(19:15) 2.3.2 Misleading intuitions from today's LLMs
(21:51) 2.3.3 Summary
(24:01) 2.4 Back to the proposal
(24:12) 2.4.1 Warm-up: The specification gaming game
(29:07) 2.4.2 What about bi-level optimization?
(31:13) 2.5 Is this a solvable problem?
(35:42) 3. Epilogue: The bigger picture--this is deeply troubling, not just a technical error
(35:51) 3.1 More on Richard Sutton
(40:52) 3.2 More on David Silver
The original text contained 10 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
Every now and then, some AI luminaries
I agree with (1) and strenuously disagree with (2).
The last time I saw something like this, I responded by writing: LeCun's “A Path Towards Autonomous Machine Intelligence” has an unsolved technical alignment problem.
Well, now we have a second entry in the series, with the new preprint book chapter “Welcome to the Era of Experience” by reinforcement learning pioneers David Silver & Richard Sutton.
The authors propose that “a new generation [...]
---
Outline:
(04:39) 1. What's their alignment plan?
(08:00) 2. The plan won't work
(08:04) 2.1 Background 1: Specification gaming and goal misgeneralization
(12:19) 2.2 Background 2: The usual agent debugging loop, and why it will eventually catastrophically fail
(15:12) 2.3 Background 3: Callous indifference and deception as the strong-default, natural way that era of experience AIs will interact with humans
(16:00) 2.3.1 Misleading intuitions from everyday life
(19:15) 2.3.2 Misleading intuitions from today's LLMs
(21:51) 2.3.3 Summary
(24:01) 2.4 Back to the proposal
(24:12) 2.4.1 Warm-up: The specification gaming game
(29:07) 2.4.2 What about bi-level optimization?
(31:13) 2.5 Is this a solvable problem?
(35:42) 3. Epilogue: The bigger picture--this is deeply troubling, not just a technical error
(35:51) 3.1 More on Richard Sutton
(40:52) 3.2 More on David Silver
The original text contained 10 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
26,327 Listeners
2,383 Listeners
8,010 Listeners
4,127 Listeners
90 Listeners
1,473 Listeners
9,200 Listeners
91 Listeners
425 Listeners
5,434 Listeners
15,371 Listeners
502 Listeners
128 Listeners
72 Listeners
464 Listeners