LessWrong (30+ Karma)

“The GDM AGI Safety+Alignment Team is Hiring for Applied Interpretability Research” by Arthur Conmy, Neel Nanda


Listen Later

TL;DR: The Google DeepMind AGI Safety team is hiring for Applied Interpretability research scientists and engineers. Applied Interpretability is a new subteam we are forming to focus on directly using model internals-based techniques to make models safer in production. Achieving this goal will require doing research on the critical path that enables interpretability methods to be more widely used for practical problems. We believe this has significant direct and indirect benefits for preventing AGI x-risk, and argue this below. Our ideal candidate has experience with ML engineering and some hands-on experience with language model interpretability research. To apply for this role (as well as other open AGI Safety and Gemini Safety roles), follow the links for Research Engineers here & Research Scientists here.

1. What is Applied Interpretability?

At a high level, the goal of the applied interpretability team is to make model internals-based methods become a standard tool [...]

---

Outline:

(01:00) 1. What is Applied Interpretability?

(03:57) 2. Specific projects were interested in working on

(06:39) FAQ

(06:42) What's the relationship between applied interpretability and Neel's mechanistic interpretability team?

(07:16) How much autonomy will I have?

(09:03) Why do applied interpretability rather than fundamental research?

(10:31) What makes someone a good fit for the role?

(11:15) I've heard that Google infra can be pretty slow and bad

(11:42) Can I publish?

(12:19) Does probing really count as interpretability?

The original text contained 2 footnotes which were omitted from this narration.

---

First published:

February 24th, 2025

Source:

https://www.lesswrong.com/posts/aG9e5tHfHmBnDqrDy/the-gdm-agi-safety-alignment-team-is-hiring-for-applied

---

Narrated by TYPE III AUDIO.

...more
View all episodesView all episodes
Download on the App Store

LessWrong (30+ Karma)By LessWrong


More shows like LessWrong (30+ Karma)

View all
Making Sense with Sam Harris by Sam Harris

Making Sense with Sam Harris

26,334 Listeners

Conversations with Tyler by Mercatus Center at George Mason University

Conversations with Tyler

2,399 Listeners

The Peter Attia Drive by Peter Attia, MD

The Peter Attia Drive

7,817 Listeners

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas by Sean Carroll | Wondery

Sean Carroll's Mindscape: Science, Society, Philosophy, Culture, Arts, and Ideas

4,107 Listeners

ManifoldOne by Steve Hsu

ManifoldOne

87 Listeners

Your Undivided Attention by Tristan Harris and Aza Raskin, The Center for Humane Technology

Your Undivided Attention

1,453 Listeners

All-In with Chamath, Jason, Sacks & Friedberg by All-In Podcast, LLC

All-In with Chamath, Jason, Sacks & Friedberg

8,761 Listeners

Machine Learning Street Talk (MLST) by Machine Learning Street Talk (MLST)

Machine Learning Street Talk (MLST)

90 Listeners

Dwarkesh Podcast by Dwarkesh Patel

Dwarkesh Podcast

353 Listeners

Hard Fork by The New York Times

Hard Fork

5,356 Listeners

The Ezra Klein Show by New York Times Opinion

The Ezra Klein Show

15,023 Listeners

Moonshots with Peter Diamandis by PHD Ventures

Moonshots with Peter Diamandis

464 Listeners

No Priors: Artificial Intelligence | Technology | Startups by Conviction

No Priors: Artificial Intelligence | Technology | Startups

128 Listeners

Latent Space: The AI Engineer Podcast by swyx + Alessio

Latent Space: The AI Engineer Podcast

73 Listeners

BG2Pod with Brad Gerstner and Bill Gurley by BG2Pod

BG2Pod with Brad Gerstner and Bill Gurley

433 Listeners