London Futurists

The AI suicide race, with Jaan Tallinn


Listen Later

The race to create advanced AI is becoming a suicide race. 

That's part of the thinking behind the open letter from the Future of Life Institute which "calls on all AI labs to immediately pause for at least six months the training of AI systems more powerful than GPT-4".

In this episode, our guest, Jaan Tallinn, explains why he sees this pause as a particularly important initiative.

In the 1990s and 20-noughts, Jaan led much of the software engineering for the file-sharing application Kazaa and the online communications tool Skype. He is also known as one of the earliest investors in DeepMind, before they were acquired by Google.

More recently, Jaan has been a prominent advocate for study of existential risks, including the risks from artificial superintelligence. He helped set up the Centre for the Study of Existential Risk (CSER) in 2012 and the Future of Life Institute (FLI) in 2014.

Follow-up reading:
https://futureoflife.org/open-letter/pause-giant-ai-experiments/
https://www.cser.ac.uk/
https://en.wikipedia.org/wiki/Jaan_Tallinn

Topics addressed in this episode include:

*) The differences between CSER and FLI
*) Do the probabilities for the occurrence of different existential risks vary by orders of magnitude?
*) The principle that "arguments screen authority"
*) The possibility that GPT-6 will be built, not by humans, but by GPT-5
*) Growing public concern, all over the world, that the fate of all humanity is, in effect, being decided by the actions of just a small number of people in AI labs
*) Two reasons why FLI recently changed its approach to AI risk
*) The AI safety conference in 2015 in Puerto Rico was initially viewed as a massive success, but it has had little lasting impact
*) Uncertainty about a potential cataclysmic event doesn't entitle people to conclude it won't happen any time soon
*) The argument that LLMs (Large Language Models) are an "off ramp" rather than being on the road to AGI
*) Why the duration of 6 months was selected for the proposed pause
*) The "What about China?" objection to the pause
*) Potential concrete steps that could take place during the pause
*) The FLI document "Policymaking in the pause"
*) The article by Luke Muehlhauser of Open Philanthropy, "12 tentative ideas for US AI policy"
*) The "summon and tame" way of thinking about the creation of LLMs - and the risk that minds summoned in this way won't be able to be tamed
*) Scenarios in which the pause might be ignored by various entities, such as authoritarian regimes, organised crime, rogue corporations, and extraordinary individuals such as Elon Musk and John Carmack
*) A meta-principle for deciding which types of AI research should be paused
*) 100 million dollar projects become even harder when they are illegal
*) The case for requiring the pre-registration of largescale mind-summoning experiments
*) A possible 10^25 limit on the number of FLOPs (Floating Point Operations) an AI model can spend
*) The reactions by AI lab leaders to the widescale public response to GPT-4 and to the pause letter
*) Even Sundar Pichai, CEO of Google/Alphabet, has called for government intervention regarding AI
*) The hardware overhang complication with the pause
*) Not letting "the perfect" be "the enemy of the good"
*) Elon Musk's involvement with FLI and with the pause letter
*) "Humanity now has cancer"

Music: Spike Protein, by Koi Discovery, available under CC0 1.

Real Talk About Marketing

An Acxiom podcast where we discuss marketing made better, bringing you real...

Listen on: Apple Podcasts   Spotify

...more
View all episodesView all episodes
Download on the App Store

London FuturistsBy London Futurists

  • 4.7
  • 4.7
  • 4.7
  • 4.7
  • 4.7

4.7

9 ratings


More shows like London Futurists

View all
In Our Time by BBC Radio 4

In Our Time

5,389 Listeners

BBC Inside Science by BBC Radio 4

BBC Inside Science

401 Listeners

Making Sense with Sam Harris by Sam Harris

Making Sense with Sam Harris

26,462 Listeners

a16z Podcast by Andreessen Horowitz

a16z Podcast

1,032 Listeners

The Gray Area with Sean Illing by Vox

The Gray Area with Sean Illing

10,694 Listeners

Azeem Azhar's Exponential View by Azeem Azhar

Azeem Azhar's Exponential View

613 Listeners

Google DeepMind: The Podcast by Hannah Fry

Google DeepMind: The Podcast

198 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

401 Listeners

Big Technology Podcast by Alex Kantrowitz

Big Technology Podcast

442 Listeners

The Rest Is Politics by Goalhanger

The Rest Is Politics

3,276 Listeners

The News Agents by Global

The News Agents

981 Listeners

Moonshots with Peter Diamandis by PHD Ventures

Moonshots with Peter Diamandis

477 Listeners

The Rest Is Politics: Leading by Goalhanger

The Rest Is Politics: Leading

987 Listeners

The Rest Is Money by Goalhanger

The Rest Is Money

220 Listeners

The Rest Is Politics: US by Goalhanger

The Rest Is Politics: US

2,285 Listeners