
Sign up to save your podcasts
Or
TLDR: we find that SAEs trained on the difference in activations between a base model and its instruct finetune are a valuable tool for understanding what changed during finetuning.
This work is the result of Jacob and Santiago's 2-week research sprint as part of Neel Nanda's training phase for MATS 8.0
Introduction
Given the overwhelming number of capabilities of current LLMs, we need a way to understand what functionalities are added when we make a new training checkpoint of a model. This is especially relevant when deploying a new model since among many new and useful features there may be hidden an unexpected harmful or undesired behavior. [1]
Model diffing aims at finding these differences between models. Recent work has focused on training crosscoders (roughly, an SAE on the concatenation of the activations of two models) and identifying latents that are exclusive to one or the other model. However [...]
---
Outline:
(00:31) Introduction
(01:50) SAE on activation differences
(02:20) Pipeline for identifying relevant latents
(04:45) KL Dashboards
(05:55) Example: inhibitory latent
(08:34) Roleplay latent
(09:40) Uncertainty latent
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
KL Attributions" showing distribution with peak near zero.Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
TLDR: we find that SAEs trained on the difference in activations between a base model and its instruct finetune are a valuable tool for understanding what changed during finetuning.
This work is the result of Jacob and Santiago's 2-week research sprint as part of Neel Nanda's training phase for MATS 8.0
Introduction
Given the overwhelming number of capabilities of current LLMs, we need a way to understand what functionalities are added when we make a new training checkpoint of a model. This is especially relevant when deploying a new model since among many new and useful features there may be hidden an unexpected harmful or undesired behavior. [1]
Model diffing aims at finding these differences between models. Recent work has focused on training crosscoders (roughly, an SAE on the concatenation of the activations of two models) and identifying latents that are exclusive to one or the other model. However [...]
---
Outline:
(00:31) Introduction
(01:50) SAE on activation differences
(02:20) Pipeline for identifying relevant latents
(04:45) KL Dashboards
(05:55) Example: inhibitory latent
(08:34) Roleplay latent
(09:40) Uncertainty latent
The original text contained 4 footnotes which were omitted from this narration.
---
First published:
Source:
---
Narrated by TYPE III AUDIO.
---
Images from the article:
KL Attributions" showing distribution with peak near zero.Apple Podcasts and Spotify do not show images in the episode description. Try Pocket Casts, or another podcast app.
26,469 Listeners
2,395 Listeners
7,953 Listeners
4,142 Listeners
89 Listeners
1,472 Listeners
9,207 Listeners
88 Listeners
417 Listeners
5,448 Listeners
15,321 Listeners
482 Listeners
121 Listeners
75 Listeners
461 Listeners