
Sign up to save your podcasts
Or


TL;DR: If you are thinking of using interpretability to help with strategic deception, then there's likely a problem you need to solve first: how are intentional descriptions (like deception) related to algorithmic ones (like understanding the mechanisms models use)? We discuss this problem and try to outline some constructive directions.
1. Introduction
A commonly discussed AI risk scenario is strategic deception: systems that execute sophisticated planning against their creators to achieve undesired ends. In particular, this is insidious because a system that is capable of strategic planning and also situationally aware might be able to systematically behave differently when under observation, and thus evaluation methods that are purely behavioral could become unreliable. One widely hypothesized potential solution to this is to use interpretability, understanding the internals of the model, to detect such strategic deception. We aim to examine this program and a series of problems that appear [...]
---
Outline:
(00:28) 1. Introduction
(02:38) 2. The intentional and the algorithmic
(06:21) 3. What is the problem?
(11:56) 4. What deception, exactly?
(14:10) 5. Decomposing strategic deception
(14:53) 5.1. Developing world models
(16:22) 5.2. Modeling the self
(17:29) 5.3. Modeling user states
(18:28) 5.4. Long-term goal directedness
(19:32) 5.5. Hidden reasoning
(21:09) 6. Conclusions and Suggestions
(23:02) Acknowledgments
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
By LessWrongTL;DR: If you are thinking of using interpretability to help with strategic deception, then there's likely a problem you need to solve first: how are intentional descriptions (like deception) related to algorithmic ones (like understanding the mechanisms models use)? We discuss this problem and try to outline some constructive directions.
1. Introduction
A commonly discussed AI risk scenario is strategic deception: systems that execute sophisticated planning against their creators to achieve undesired ends. In particular, this is insidious because a system that is capable of strategic planning and also situationally aware might be able to systematically behave differently when under observation, and thus evaluation methods that are purely behavioral could become unreliable. One widely hypothesized potential solution to this is to use interpretability, understanding the internals of the model, to detect such strategic deception. We aim to examine this program and a series of problems that appear [...]
---
Outline:
(00:28) 1. Introduction
(02:38) 2. The intentional and the algorithmic
(06:21) 3. What is the problem?
(11:56) 4. What deception, exactly?
(14:10) 5. Decomposing strategic deception
(14:53) 5.1. Developing world models
(16:22) 5.2. Modeling the self
(17:29) 5.3. Modeling user states
(18:28) 5.4. Long-term goal directedness
(19:32) 5.5. Hidden reasoning
(21:09) 6. Conclusions and Suggestions
(23:02) Acknowledgments
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

26,329 Listeners

2,442 Listeners

9,152 Listeners

4,152 Listeners

92 Listeners

1,598 Listeners

9,901 Listeners

90 Listeners

505 Listeners

5,472 Listeners

16,038 Listeners

539 Listeners

133 Listeners

95 Listeners

515 Listeners