
Sign up to save your podcasts
Or


https://astralcodexten.substack.com/p/updated-look-at-long-term-ai-risks
The last couple of posts here talked about long-term risks from AI, so I thought I'd highlight the results of a new expert survey on exactly what they are. There have been a lot of these surveys recently, but this one is a little different.
Starting from the beginning: in 2012-2014, Muller and Bostrom surveyed 550 people with various levels of claim to the title "AI expert" on the future of AI. People in philosophy of AI or other very speculative fields gave numbers around 20% chance of AI causing an "existential catastrophe" (eg human extinction); people in normal technical AI research gave numbers around 7%. In 2016-2017, Grace et al surveyed 1634 experts, 5% of whom predicted an extremely catastrophic outcome. Both of these surveys were vulnerable to response bias (eg the least speculative-minded people might think the whole issue was stupid and not even return the survey).
The new paper - Carlier, Clarke, and Schuett (not currently public, sorry, but you can read the summary here) - isn't exactly continuing in this tradition. Instead of surveying all AI experts, it surveys people who work in "AI safety and governance", ie people who are already concerned with AI being potentially dangerous, and who have dedicated their careers to addressing this. As such, they were more concerned on average than the people in previous surveys, and gave a median ~10% chance of AI-related catastrophe (~5% in the next 50 years, rising to ~25% if we don't make a directed effort to prevent it; means were a bit higher than medians). Individual experts' probability estimates ranged from 0.1% to 100% (this is how you know you're doing good futurology).
None of that is really surprising. What's new here is that they surveyed the experts on various ways AI could go wrong, to see which ones the experts were most concerned about. Going through each of them in a little more detail:
1. Superintelligence: This is the "classic" scenario that started the field, ably described by people like Nick Bostrom and Eliezer Yudkowsky. AI progress goes from human-level to vastly-above-human-level very quickly, maybe because slightly-above-human-level AIs themselves are speeding it along, or maybe because it turns out that if you can make an IQ 100 AI for $10,000 worth of compute, you can make an IQ 500 AI for $50,000. You end up with one (or a few) completely unexpected superintelligent AIs, which wield far-future technology and use it in unpredictable ways based on untested goal structures.
By Jeremiah4.8
129129 ratings
https://astralcodexten.substack.com/p/updated-look-at-long-term-ai-risks
The last couple of posts here talked about long-term risks from AI, so I thought I'd highlight the results of a new expert survey on exactly what they are. There have been a lot of these surveys recently, but this one is a little different.
Starting from the beginning: in 2012-2014, Muller and Bostrom surveyed 550 people with various levels of claim to the title "AI expert" on the future of AI. People in philosophy of AI or other very speculative fields gave numbers around 20% chance of AI causing an "existential catastrophe" (eg human extinction); people in normal technical AI research gave numbers around 7%. In 2016-2017, Grace et al surveyed 1634 experts, 5% of whom predicted an extremely catastrophic outcome. Both of these surveys were vulnerable to response bias (eg the least speculative-minded people might think the whole issue was stupid and not even return the survey).
The new paper - Carlier, Clarke, and Schuett (not currently public, sorry, but you can read the summary here) - isn't exactly continuing in this tradition. Instead of surveying all AI experts, it surveys people who work in "AI safety and governance", ie people who are already concerned with AI being potentially dangerous, and who have dedicated their careers to addressing this. As such, they were more concerned on average than the people in previous surveys, and gave a median ~10% chance of AI-related catastrophe (~5% in the next 50 years, rising to ~25% if we don't make a directed effort to prevent it; means were a bit higher than medians). Individual experts' probability estimates ranged from 0.1% to 100% (this is how you know you're doing good futurology).
None of that is really surprising. What's new here is that they surveyed the experts on various ways AI could go wrong, to see which ones the experts were most concerned about. Going through each of them in a little more detail:
1. Superintelligence: This is the "classic" scenario that started the field, ably described by people like Nick Bostrom and Eliezer Yudkowsky. AI progress goes from human-level to vastly-above-human-level very quickly, maybe because slightly-above-human-level AIs themselves are speeding it along, or maybe because it turns out that if you can make an IQ 100 AI for $10,000 worth of compute, you can make an IQ 500 AI for $50,000. You end up with one (or a few) completely unexpected superintelligent AIs, which wield far-future technology and use it in unpredictable ways based on untested goal structures.

32,246 Listeners

2,118 Listeners

2,680 Listeners

26,380 Listeners

4,270 Listeners

2,461 Listeners

2,267 Listeners

907 Listeners

291 Listeners

4,167 Listeners

1,635 Listeners

313 Listeners

3,833 Listeners

551 Listeners

688 Listeners