
Sign up to save your podcasts
Or


Epistemic status: Fairly confident in the framework, uncertain about object-level claims. Keen to receive pushback on the thought experiments.
TL;DR: I argue that Whole Brain Emulations (WBEs) would clearly have moral patienthood, and that the relevant features are computational, not biological. Recent Mechanistic Interpretability (MI) work shows Large Language Models (LLMs) have emotional representations with geometric structure matching human affect. This doesn't prove LLMs deserve moral consideration, but it establishes a necessary condition, and we should take it seriously.
Acknowledgements: Thanks to Boyd Kane, Anna Soligo, and Isha Gupta for providing feedback on early drafts.
In this post I’ll be arguing for the following claim: we can make empirical progress on AI welfare without solving consciousness.
The key move is using Whole Brain Emulation as an anchor point. WBEs would clearly deserve moral consideration (under functionalism), and they're non-biological, so whatever grounds their moral status must be computational. This gives us something concrete to look for in LLMs.
In this post I'll:
---
Outline:
(01:46) The WBE Anchor: Why Substrate Doesnt Matter
(04:18) LLMs Have Human-Like Emotional Geometry
(08:31) Ruling Out Alternative Criteria
(12:54) What Im NOT Claiming
(13:25) Conclusion
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
By LessWrongEpistemic status: Fairly confident in the framework, uncertain about object-level claims. Keen to receive pushback on the thought experiments.
TL;DR: I argue that Whole Brain Emulations (WBEs) would clearly have moral patienthood, and that the relevant features are computational, not biological. Recent Mechanistic Interpretability (MI) work shows Large Language Models (LLMs) have emotional representations with geometric structure matching human affect. This doesn't prove LLMs deserve moral consideration, but it establishes a necessary condition, and we should take it seriously.
Acknowledgements: Thanks to Boyd Kane, Anna Soligo, and Isha Gupta for providing feedback on early drafts.
In this post I’ll be arguing for the following claim: we can make empirical progress on AI welfare without solving consciousness.
The key move is using Whole Brain Emulation as an anchor point. WBEs would clearly deserve moral consideration (under functionalism), and they're non-biological, so whatever grounds their moral status must be computational. This gives us something concrete to look for in LLMs.
In this post I'll:
---
Outline:
(01:46) The WBE Anchor: Why Substrate Doesnt Matter
(04:18) LLMs Have Human-Like Emotional Geometry
(08:31) Ruling Out Alternative Criteria
(12:54) What Im NOT Claiming
(13:25) Conclusion
---
First published:
Source:
---
Narrated by TYPE III AUDIO.

112,956 Listeners

132 Listeners

7,290 Listeners

548 Listeners

16,362 Listeners

4 Listeners

14 Listeners

2 Listeners