
Sign up to save your podcasts
Or


Adrià recently published “Alignment will happen by default; what's next?” on LessWrong, arguing that AI alignment is turning out easier than expected. Simon left a lengthy comment pushing back, and that sparked this spontaneous debate.
Adrià argues that current models like Claude Opus 3 are genuinely good “to their core,” and that an iterative process — where each AI generation helps align the next — could carry us safely to superintelligence. Simon counters that we may only get one shot at alignment, that current methods are too weak to scale. A conversation about where AI safety actually stands.
---
First published:
Source:
Linkpost URL:
https://simonlermen.substack.com/p/will-we-get-alignment-by-default
---
Narrated by TYPE III AUDIO.
By LessWrongAdrià recently published “Alignment will happen by default; what's next?” on LessWrong, arguing that AI alignment is turning out easier than expected. Simon left a lengthy comment pushing back, and that sparked this spontaneous debate.
Adrià argues that current models like Claude Opus 3 are genuinely good “to their core,” and that an iterative process — where each AI generation helps align the next — could carry us safely to superintelligence. Simon counters that we may only get one shot at alignment, that current methods are too weak to scale. A conversation about where AI safety actually stands.
---
First published:
Source:
Linkpost URL:
https://simonlermen.substack.com/p/will-we-get-alignment-by-default
---
Narrated by TYPE III AUDIO.

26,344 Listeners

2,451 Listeners

8,621 Listeners

4,187 Listeners

93 Listeners

1,598 Listeners

9,904 Listeners

95 Listeners

503 Listeners

5,525 Listeners

15,955 Listeners

545 Listeners

133 Listeners

94 Listeners

470 Listeners