
Sign up to save your podcasts
Or


tl;dr: I outline my research agenda, post bounties for poking holes in it or for providing general relevant information, and am seeking to diversify my funding sources. This post will be followed by several others, providing deeper overviews of the agenda's subproblems and my sketches of how to tackle them.
Back at the end of 2023, I wrote the following:
I'm fairly optimistic about arriving at a robust solution to alignment via agent-foundations research in a timely manner. (My semi-arbitrary deadline is 2030, and I expect to arrive at intermediate solid results by EOY 2025.)
On the inside view, I'm pretty satisfied with how that is turning out. I have a high-level plan of attack which approaches the problem from a novel route, and which hopefully lets us dodge a bunch of major alignment difficulties (chiefly the instability of value reflection, which I am MIRI-tier skeptical of tackling directly). [...]
---
Outline:
(04:34) Why Do You Consider This Agenda Promising?
(06:35) High-Level Outline
(07:03) Theoretical Justifications
(15:41) Subproblems
(19:48) Bounties
(21:20) Funding
The original text contained 5 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
By LessWrongtl;dr: I outline my research agenda, post bounties for poking holes in it or for providing general relevant information, and am seeking to diversify my funding sources. This post will be followed by several others, providing deeper overviews of the agenda's subproblems and my sketches of how to tackle them.
Back at the end of 2023, I wrote the following:
I'm fairly optimistic about arriving at a robust solution to alignment via agent-foundations research in a timely manner. (My semi-arbitrary deadline is 2030, and I expect to arrive at intermediate solid results by EOY 2025.)
On the inside view, I'm pretty satisfied with how that is turning out. I have a high-level plan of attack which approaches the problem from a novel route, and which hopefully lets us dodge a bunch of major alignment difficulties (chiefly the instability of value reflection, which I am MIRI-tier skeptical of tackling directly). [...]
---
Outline:
(04:34) Why Do You Consider This Agenda Promising?
(06:35) High-Level Outline
(07:03) Theoretical Justifications
(15:41) Subproblems
(19:48) Bounties
(21:20) Funding
The original text contained 5 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.

26,336 Listeners

2,451 Listeners

8,522 Listeners

4,181 Listeners

95 Listeners

1,603 Listeners

9,927 Listeners

96 Listeners

517 Listeners

5,511 Listeners

15,859 Listeners

553 Listeners

131 Listeners

93 Listeners

465 Listeners