
Sign up to save your podcasts
Or
The Machine Intelligence Research Institute has recently shifted its focus to "technical governance". But what is that actually, and what are they doing? In this episode, I chat with Peter Barnett about his team's work on studying what evaluations can and cannot do, as well as verifying international agreements on AI development.
Patreon: https://www.patreon.com/axrpodcast
Ko-fi: https://ko-fi.com/axrpodcast
The transcript: https://axrp.net/episode/2024/12/14/episode-38_4-peter-barnett-technical-governance-at-miri.html
FAR.AI: https://far.ai/
FAR.AI on X (aka Twitter): https://x.com/farairesearch
FAR.AI on YouTube: https://www.youtube.com/@FARAIResearch
The Alignment Workshop: https://www.alignment-workshop.com/
Topics we discuss, and timestamps:
01:25 - MIRI's technical governance team
03:43 - Misuse evaluations
06:37 - Misalignment evaluations
11:29 - Verifying international agreements
13:30 - Difficulties in compute monitoring
16:44 - More on MIRI's technical governance team
Links:
MIRI Technical Governance Team: https://techgov.intelligence.org/
What AI evaluations for preventing catastrophic risks can and cannot do: https://techgov.intelligence.org/research/what-ai-evaluations-for-preventing-catastrophic-risks-can-and-cannot-do
Declare and Justify: Explicit assumptions in AI evaluations are necessary for effective regulation: https://arxiv.org/abs/2411.12820
Mechanisms to Verify International Agreements About AI Development: https://techgov.intelligence.org/research/mechanisms-to-verify-international-agreements-about-ai-development
Revisiting algorithmic progress: https://epoch.ai/blog/revisiting-algorithmic-progress
Episode art by Hamish Doodles: hamishdoodles.com
4.4
88 ratings
The Machine Intelligence Research Institute has recently shifted its focus to "technical governance". But what is that actually, and what are they doing? In this episode, I chat with Peter Barnett about his team's work on studying what evaluations can and cannot do, as well as verifying international agreements on AI development.
Patreon: https://www.patreon.com/axrpodcast
Ko-fi: https://ko-fi.com/axrpodcast
The transcript: https://axrp.net/episode/2024/12/14/episode-38_4-peter-barnett-technical-governance-at-miri.html
FAR.AI: https://far.ai/
FAR.AI on X (aka Twitter): https://x.com/farairesearch
FAR.AI on YouTube: https://www.youtube.com/@FARAIResearch
The Alignment Workshop: https://www.alignment-workshop.com/
Topics we discuss, and timestamps:
01:25 - MIRI's technical governance team
03:43 - Misuse evaluations
06:37 - Misalignment evaluations
11:29 - Verifying international agreements
13:30 - Difficulties in compute monitoring
16:44 - More on MIRI's technical governance team
Links:
MIRI Technical Governance Team: https://techgov.intelligence.org/
What AI evaluations for preventing catastrophic risks can and cannot do: https://techgov.intelligence.org/research/what-ai-evaluations-for-preventing-catastrophic-risks-can-and-cannot-do
Declare and Justify: Explicit assumptions in AI evaluations are necessary for effective regulation: https://arxiv.org/abs/2411.12820
Mechanisms to Verify International Agreements About AI Development: https://techgov.intelligence.org/research/mechanisms-to-verify-international-agreements-about-ai-development
Revisiting algorithmic progress: https://epoch.ai/blog/revisiting-algorithmic-progress
Episode art by Hamish Doodles: hamishdoodles.com
26,462 Listeners
2,395 Listeners
1,784 Listeners
298 Listeners
105 Listeners
4,142 Listeners
89 Listeners
287 Listeners
88 Listeners
417 Listeners
258 Listeners
75 Listeners
60 Listeners
146 Listeners
123 Listeners