London Futurists

The AI suicide race, with Jaan Tallinn


Listen Later

The race to create advanced AI is becoming a suicide race. 

That's part of the thinking behind the open letter from the Future of Life Institute which "calls on all AI labs to immediately pause for at least six months the training of AI systems more powerful than GPT-4".

In this episode, our guest, Jaan Tallinn, explains why he sees this pause as a particularly important initiative.

In the 1990s and 20-noughts, Jaan led much of the software engineering for the file-sharing application Kazaa and the online communications tool Skype. He is also known as one of the earliest investors in DeepMind, before they were acquired by Google.

More recently, Jaan has been a prominent advocate for study of existential risks, including the risks from artificial superintelligence. He helped set up the Centre for the Study of Existential Risk (CSER) in 2012 and the Future of Life Institute (FLI) in 2014.

Follow-up reading:
https://futureoflife.org/open-letter/pause-giant-ai-experiments/
https://www.cser.ac.uk/
https://en.wikipedia.org/wiki/Jaan_Tallinn

Topics addressed in this episode include:

*) The differences between CSER and FLI
*) Do the probabilities for the occurrence of different existential risks vary by orders of magnitude?
*) The principle that "arguments screen authority"
*) The possibility that GPT-6 will be built, not by humans, but by GPT-5
*) Growing public concern, all over the world, that the fate of all humanity is, in effect, being decided by the actions of just a small number of people in AI labs
*) Two reasons why FLI recently changed its approach to AI risk
*) The AI safety conference in 2015 in Puerto Rico was initially viewed as a massive success, but it has had little lasting impact
*) Uncertainty about a potential cataclysmic event doesn't entitle people to conclude it won't happen any time soon
*) The argument that LLMs (Large Language Models) are an "off ramp" rather than being on the road to AGI
*) Why the duration of 6 months was selected for the proposed pause
*) The "What about China?" objection to the pause
*) Potential concrete steps that could take place during the pause
*) The FLI document "Policymaking in the pause"
*) The article by Luke Muehlhauser of Open Philanthropy, "12 tentative ideas for US AI policy"
*) The "summon and tame" way of thinking about the creation of LLMs - and the risk that minds summoned in this way won't be able to be tamed
*) Scenarios in which the pause might be ignored by various entities, such as authoritarian regimes, organised crime, rogue corporations, and extraordinary individuals such as Elon Musk and John Carmack
*) A meta-principle for deciding which types of AI research should be paused
*) 100 million dollar projects become even harder when they are illegal
*) The case for requiring the pre-registration of largescale mind-summoning experiments
*) A possible 10^25 limit on the number of FLOPs (Floating Point Operations) an AI model can spend
*) The reactions by AI lab leaders to the widescale public response to GPT-4 and to the pause letter
*) Even Sundar Pichai, CEO of Google/Alphabet, has called for government intervention regarding AI
*) The hardware overhang complication with the pause
*) Not letting "the perfect" be "the enemy of the good"
*) Elon Musk's involvement with FLI and with the pause letter
*) "Humanity now has cancer"

Music: Spike Protein, by Koi Discovery, available under CC0 1.0 Publ

Promoguy Talk Pills
Agency in Amsterdam dives into topics like Tech, AI, digital marketing, and more drama...

Listen on: Apple Podcasts   Spotify

...more
View all episodesView all episodes
Download on the App Store

London FuturistsBy London Futurists

  • 5
  • 5
  • 5
  • 5
  • 5

5

8 ratings


More shows like London Futurists

View all
Making Sense with Sam Harris by Sam Harris

Making Sense with Sam Harris

26,334 Listeners

a16z Podcast by Andreessen Horowitz

a16z Podcast

1,000 Listeners

The Kevin Rose Show by Kevin Rose

The Kevin Rose Show

1,876 Listeners

FT Tech Tonic by Financial Times

FT Tech Tonic

101 Listeners

Hidden Forces by Demetri Kofinas

Hidden Forces

1,447 Listeners

The Peter Attia Drive by Peter Attia, MD

The Peter Attia Drive

7,859 Listeners

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas by Sean Carroll | Wondery

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

4,107 Listeners

Last Week in AI by Skynet Today

Last Week in AI

282 Listeners

All-In with Chamath, Jason, Sacks & Friedberg by All-In Podcast, LLC

All-In with Chamath, Jason, Sacks & Friedberg

8,761 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

353 Listeners

Big Technology Podcast by Alex Kantrowitz

Big Technology Podcast

395 Listeners

Moonshots with Peter Diamandis by PHD Ventures

Moonshots with Peter Diamandis

464 Listeners

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis by Erik Torenberg, Nathan Labenz

"The Cognitive Revolution" | AI Builders, Researchers, and Live Player Analysis

85 Listeners

This Day in AI Podcast by Michael Sharkey, Chris Sharkey

This Day in AI Podcast

196 Listeners

Latent Space: The AI Engineer Podcast by swyx + Alessio

Latent Space: The AI Engineer Podcast

73 Listeners