
Sign up to save your podcasts
Or
Note: This post is intended to be the first in a broader series of posts about the difficult tradeoffs inherent in public access to powerful open source models. While this post highlights some dangers of open models and discusses the possibility of global regulation, I am not, in general, against open source AI, or supportive of regulation of open source AI today. On the contrary, I believe open source software is, in general, one of humanity's most important and valuable public goods. My goal in writing this post is to call attention to the risks and challenges around open models now, so we can use the time we still have before risks become extreme, to collectively explore viable alternatives to regulation, if indeed such alternatives exist.
Background
Most research into the control problem today starts from an assumption that the organization operating the AI system has some baseline interest [...]
---
Outline:
(00:52) Background
(02:01) AI Systems Without Control-Related Precautions
(05:02) Loss of Control in Open Models
(06:42) How Researchers Think About Loss of Control in Open Models
(09:00) Problems With Loss of Control is Mainly a Labs Problem
(09:05) Problem 1 - Frontier labs are increasingly secretive and profit-seeking and it's not clear they would publicly report a serious loss of control-related issue if they encountered one.
(10:40) Problem 2 - There is no agreed-upon standard that defines relevant thresholds or evidence that would constitute a serious control risk inside a lab anyway.
(12:38) Problem 3 - Even if one of the labs does sound the alarm, it seems likely that other labs will not stop releasing open models anyway, absent regulation.
(14:31) Problem 4 - Policymakers have not committed to regulate open models that demonstrate risky capabilities.
(16:33) Passing and Enforcing Effective Global Restrictions on Open Source Models Would be Extremely Difficult
(17:52) Challenge 1 - Regulations would need to be globally enforced to be effective.
(19:37) Challenge 2 - The required timelines for passing regulation and organizing global enforcement could be very short.
(20:51) Challenge 3 - If labs stop releasing open models, they may be leaked anyway.
(22:05) Challenge 4 - Penalties for possession would need to be severe and extreme levels of surveillance may be required to enforce them.
(25:13) The Urgency - DeepSeek and Evidence from Model Organisms and Agentic AI
(25:44) DeepSeek R1
(27:41) Evidence of Misalignment in Model Organisms
(28:22) Scheming
(29:48) Reward Tampering
(31:30) Broad Misalignment
(32:36) Susceptibility to Data Poisoning and Fine-tuning is Increasing
(33:33) Agentic AI
(36:20) Conclusion
---
First published:
Source:
Narrated by TYPE III AUDIO.
Note: This post is intended to be the first in a broader series of posts about the difficult tradeoffs inherent in public access to powerful open source models. While this post highlights some dangers of open models and discusses the possibility of global regulation, I am not, in general, against open source AI, or supportive of regulation of open source AI today. On the contrary, I believe open source software is, in general, one of humanity's most important and valuable public goods. My goal in writing this post is to call attention to the risks and challenges around open models now, so we can use the time we still have before risks become extreme, to collectively explore viable alternatives to regulation, if indeed such alternatives exist.
Background
Most research into the control problem today starts from an assumption that the organization operating the AI system has some baseline interest [...]
---
Outline:
(00:52) Background
(02:01) AI Systems Without Control-Related Precautions
(05:02) Loss of Control in Open Models
(06:42) How Researchers Think About Loss of Control in Open Models
(09:00) Problems With Loss of Control is Mainly a Labs Problem
(09:05) Problem 1 - Frontier labs are increasingly secretive and profit-seeking and it's not clear they would publicly report a serious loss of control-related issue if they encountered one.
(10:40) Problem 2 - There is no agreed-upon standard that defines relevant thresholds or evidence that would constitute a serious control risk inside a lab anyway.
(12:38) Problem 3 - Even if one of the labs does sound the alarm, it seems likely that other labs will not stop releasing open models anyway, absent regulation.
(14:31) Problem 4 - Policymakers have not committed to regulate open models that demonstrate risky capabilities.
(16:33) Passing and Enforcing Effective Global Restrictions on Open Source Models Would be Extremely Difficult
(17:52) Challenge 1 - Regulations would need to be globally enforced to be effective.
(19:37) Challenge 2 - The required timelines for passing regulation and organizing global enforcement could be very short.
(20:51) Challenge 3 - If labs stop releasing open models, they may be leaked anyway.
(22:05) Challenge 4 - Penalties for possession would need to be severe and extreme levels of surveillance may be required to enforce them.
(25:13) The Urgency - DeepSeek and Evidence from Model Organisms and Agentic AI
(25:44) DeepSeek R1
(27:41) Evidence of Misalignment in Model Organisms
(28:22) Scheming
(29:48) Reward Tampering
(31:30) Broad Misalignment
(32:36) Susceptibility to Data Poisoning and Fine-tuning is Increasing
(33:33) Agentic AI
(36:20) Conclusion
---
First published:
Source:
Narrated by TYPE III AUDIO.
26,334 Listeners
2,399 Listeners
7,817 Listeners
4,107 Listeners
87 Listeners
1,453 Listeners
8,761 Listeners
90 Listeners
353 Listeners
5,356 Listeners
15,023 Listeners
464 Listeners
128 Listeners
73 Listeners
433 Listeners