
Sign up to save your podcasts
Or
Audio note: this article contains 38 uses of latex notation, so the narration may be difficult to follow. There's a link to the original text in the episode description.
This post has some ablation results around the thesis of the ICML 2024 Mech. Interp. workshop 1st prize winning paper: The Geometry of Categorical and Hierarchical Concepts in Large Language Models The main takeaway is that the orthogonality they observe in categorical and hierarchical concepts occurs practically everywhere, even at places where it really should not.
Overview of the original paper
A lot of the intuition and math around why they do what they do is shared in their previous work called The Linear Representation Hypothesis and the Geometry of Large Language Models, but let's quickly go over what the paper's core idea is:
They split the computation of a large language model (LLM) as:
_P(y mid [...]
---
Outline:
(02:42) Ablations
(03:37) Hierarchical features are orthogonal - but so are semantic opposites!?
(04:51) Categorical features form simplices - but so do totally random ones!?
(05:27) Orthogonality being ubiquitous in high dimensions
(06:39) Discussion and Future Work
The original text contained 4 images which were described by AI.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
Audio note: this article contains 38 uses of latex notation, so the narration may be difficult to follow. There's a link to the original text in the episode description.
This post has some ablation results around the thesis of the ICML 2024 Mech. Interp. workshop 1st prize winning paper: The Geometry of Categorical and Hierarchical Concepts in Large Language Models The main takeaway is that the orthogonality they observe in categorical and hierarchical concepts occurs practically everywhere, even at places where it really should not.
Overview of the original paper
A lot of the intuition and math around why they do what they do is shared in their previous work called The Linear Representation Hypothesis and the Geometry of Large Language Models, but let's quickly go over what the paper's core idea is:
They split the computation of a large language model (LLM) as:
_P(y mid [...]
---
Outline:
(02:42) Ablations
(03:37) Hierarchical features are orthogonal - but so are semantic opposites!?
(04:51) Categorical features form simplices - but so do totally random ones!?
(05:27) Orthogonality being ubiquitous in high dimensions
(06:39) Discussion and Future Work
The original text contained 4 images which were described by AI.
---
First published:
Source:
Narrated by TYPE III AUDIO.
---
Images from the article:
Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
26,382 Listeners
2,386 Listeners
7,925 Listeners
4,134 Listeners
87 Listeners
1,456 Listeners
9,048 Listeners
87 Listeners
387 Listeners
5,426 Listeners
15,207 Listeners
472 Listeners
120 Listeners
75 Listeners
456 Listeners