
Sign up to save your podcasts
Or
Background: “Learning” vs “Learning About”
Adaptive systems, reinforcement “learners”, etc, “learn” in the sense that their behavior adapts to their environment.
Bayesian reasoners, human scientists, etc, “learn” in the sense that they have some symbolic representation of the environment, and they update those symbols over time to (hopefully) better match the environment (i.e. make the map better match the territory).
These two kinds of “learning” are not synonymous[1]. Adaptive systems “learn” things, but they don’t necessarily “learn about” things; they don’t necessarily have an internal map of the external territory. (Yes, the active inference folks will bullshit about how any adaptive system must have a map of the territory, but their math does not substantively support that interpretation.) The internal heuristics or behaviors “learned” by an adaptive system are not necessarily “about” any particular external thing, and don’t necessarily represent any particular external thing[2].
We Humans Learn About Our Values
---
Outline:
(00:06) Background: “Learning” vs “Learning About”
(01:09) We Humans Learn About Our Values
(03:25) Two Puzzles
(03:28) Puzzle 1: Learning About Our Own Values vs The Is-Ought Gap
(03:53) Puzzle 2: The Role of Reward/Reinforcement
(04:35) Using Each Puzzle To Solve The Other
(05:11) What This Looks Like In Practice
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
Background: “Learning” vs “Learning About”
Adaptive systems, reinforcement “learners”, etc, “learn” in the sense that their behavior adapts to their environment.
Bayesian reasoners, human scientists, etc, “learn” in the sense that they have some symbolic representation of the environment, and they update those symbols over time to (hopefully) better match the environment (i.e. make the map better match the territory).
These two kinds of “learning” are not synonymous[1]. Adaptive systems “learn” things, but they don’t necessarily “learn about” things; they don’t necessarily have an internal map of the external territory. (Yes, the active inference folks will bullshit about how any adaptive system must have a map of the territory, but their math does not substantively support that interpretation.) The internal heuristics or behaviors “learned” by an adaptive system are not necessarily “about” any particular external thing, and don’t necessarily represent any particular external thing[2].
We Humans Learn About Our Values
---
Outline:
(00:06) Background: “Learning” vs “Learning About”
(01:09) We Humans Learn About Our Values
(03:25) Two Puzzles
(03:28) Puzzle 1: Learning About Our Own Values vs The Is-Ought Gap
(03:53) Puzzle 2: The Role of Reward/Reinforcement
(04:35) Using Each Puzzle To Solve The Other
(05:11) What This Looks Like In Practice
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
26,400 Listeners
2,388 Listeners
7,921 Listeners
4,132 Listeners
87 Listeners
1,456 Listeners
9,045 Listeners
86 Listeners
388 Listeners
5,426 Listeners
15,207 Listeners
474 Listeners
123 Listeners
75 Listeners
455 Listeners