
Sign up to save your podcasts
Or


Subtitle: Large fractions of people will die, but literal human extinction seems unlikely.
Suppose misaligned AIs take over. What fraction of people will die? I'll discuss my thoughts on this question and my basic framework for thinking about it. These are some pretty low-effort notes, the topic is very speculative, and I don't get into all the specifics, so be warned.
I don't think moderate disagreements here are very action-guiding or cruxy on typical worldviews: it probably shouldn't alter your actions much if you end up thinking 25% of people die in expectation from misaligned AI takeover rather than 90% or end up thinking that misaligned AI takeover causing literal human extinction is 10% likely rather than 90% likely (or vice versa). (And the possibility that we're in a simulation poses a huge complication that I won't elaborate on here.) Note that even if misaligned AI takeover doesn't [...]
---
Outline:
(04:46) Industrial expansion and small motivations to avoid human fatalities
(12:24) How likely is it that AIs will actively have motivations to kill (most/many) humans
(13:46) Death due to takeover itself
(15:12) Combining these numbers
The original text contained 6 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
By Redwood ResearchSubtitle: Large fractions of people will die, but literal human extinction seems unlikely.
Suppose misaligned AIs take over. What fraction of people will die? I'll discuss my thoughts on this question and my basic framework for thinking about it. These are some pretty low-effort notes, the topic is very speculative, and I don't get into all the specifics, so be warned.
I don't think moderate disagreements here are very action-guiding or cruxy on typical worldviews: it probably shouldn't alter your actions much if you end up thinking 25% of people die in expectation from misaligned AI takeover rather than 90% or end up thinking that misaligned AI takeover causing literal human extinction is 10% likely rather than 90% likely (or vice versa). (And the possibility that we're in a simulation poses a huge complication that I won't elaborate on here.) Note that even if misaligned AI takeover doesn't [...]
---
Outline:
(04:46) Industrial expansion and small motivations to avoid human fatalities
(12:24) How likely is it that AIs will actively have motivations to kill (most/many) humans
(13:46) Death due to takeover itself
(15:12) Combining these numbers
The original text contained 6 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.