
Sign up to save your podcasts
Or


This was a quick, short side-project produced during the MATS Research 8.1 extension. It's related to my group's main thread of work on black-box scheming monitoring through the connections to monitoring I explore below, but was time-boxed and pursued independently because I thought it was interesting!
Executive Summary
Figure 1. Accuracy vs. similarity threshold (0.95+) across 1700 pairs of encoding/decoding examples across a variety of datatypes and lengths. The accuracy is the proportion of the 3400 examples each model translated successfully (directly, with no reasoning or tools). Success for each task is defined by the normalised Levenshtein similarity of the answer/target pair hitting a given threshold, with a scoring requirement that model-encoded strings are decodable. Legend ordered by [email protected].---
Outline:
(00:31) Executive Summary
(03:07) An accidental (and surprising) discovery
(08:03) Have LLMs actually learned the algorithm?
(09:39) Introducing
(13:11) Accuracy vs. similarity threshold
(16:02) Encoding vs. decoding by model
(17:00) Task-level breakdown
(19:37) Why should we care?
(21:26) Monitoring implications
(23:51) Conclusion
(25:23) Appendix
(25:26) Zoomed-in threshold sweeps
The original text contained 8 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
By LessWrongThis was a quick, short side-project produced during the MATS Research 8.1 extension. It's related to my group's main thread of work on black-box scheming monitoring through the connections to monitoring I explore below, but was time-boxed and pursued independently because I thought it was interesting!
Executive Summary
Figure 1. Accuracy vs. similarity threshold (0.95+) across 1700 pairs of encoding/decoding examples across a variety of datatypes and lengths. The accuracy is the proportion of the 3400 examples each model translated successfully (directly, with no reasoning or tools). Success for each task is defined by the normalised Levenshtein similarity of the answer/target pair hitting a given threshold, with a scoring requirement that model-encoded strings are decodable. Legend ordered by [email protected].---
Outline:
(00:31) Executive Summary
(03:07) An accidental (and surprising) discovery
(08:03) Have LLMs actually learned the algorithm?
(09:39) Introducing
(13:11) Accuracy vs. similarity threshold
(16:02) Encoding vs. decoding by model
(17:00) Task-level breakdown
(19:37) Why should we care?
(21:26) Monitoring implications
(23:51) Conclusion
(25:23) Appendix
(25:26) Zoomed-in threshold sweeps
The original text contained 8 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.

26,319 Listeners

2,452 Listeners

8,529 Listeners

4,176 Listeners

93 Listeners

1,601 Listeners

9,920 Listeners

95 Listeners

514 Listeners

5,509 Listeners

15,918 Listeners

548 Listeners

131 Listeners

93 Listeners

466 Listeners