← Back to papers

Paper deep dive

A is for Absorption: Studying Feature Splitting and Absorption in Sparse Autoencoders

David Chanin, James Wilken-Smith, Tomas Dulka, Hardik Bhatnagar, Joseph Bloom

Year: 2024Venue: arXiv preprintArea: Mechanistic Interp.Type: EmpiricalEmbeddings: 101

Models: Gemma-2-2B

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%

Last extracted: 3/12/2026, 8:17:04 PM

Summary

The paper identifies 'feature absorption' in Sparse Autoencoders (SAEs), a phenomenon where hierarchical features are not correctly decomposed. Instead of learning monosemantic latents, SAEs 'absorb' parent features into child features to optimize for sparsity, leading to unreliable latent activations. The authors provide a metric to detect this, validate it across hundreds of LLM SAEs, and demonstrate that varying SAE size or sparsity does not resolve the issue.

Entities (5)

Feature Absorption · phenomenon · 100%Sparse Autoencoders · model-architecture · 100%Feature Splitting · phenomenon · 95%Gemma Scope · datasetmodel · 95%Linear Representation Hypothesis · theory · 90%

Relation Signals (3)

Sparse Autoencoders exhibits Feature Absorption

confidence 100% · We show that seemingly monosemantic features fail to fire where they should, and instead get 'absorbed' into their children features.

Hierarchical Features causes Feature Absorption

confidence 95% · We show that feature absorption is caused by optimizing for sparsity in SAEs whenever the underlying features form a hierarchy.

Sparse Autoencoders uses Linear Representation Hypothesis

confidence 85% · As long as the active features in a given LLM activation are sparse and the underlying features follow the LRH, Sparse Autoencoders (SAEs) should be able to recover the true LLM features

Cypher Suggestions (2)

Find all phenomena associated with Sparse Autoencoders · confidence 90% · unvalidated

MATCH (s:Entity {name: 'Sparse Autoencoders'})-[:EXHIBITS|CAUSES]->(p:Phenomenon) RETURN s.name, p.name

Identify causal relationships between features and phenomena · confidence 90% · unvalidated

MATCH (f:Entity)-[:CAUSES]->(p:Phenomenon) RETURN f.name, p.name

Abstract

Abstract:Sparse Autoencoders (SAEs) aim to decompose the activation space of large language models (LLMs) into human-interpretable latent directions or features. As we increase the number of features in the SAE, hierarchical features tend to split into finer features ("math" may split into "algebra", "geometry", etc.), a phenomenon referred to as feature splitting. However, we show that sparse decomposition and splitting of hierarchical features is not robust. Specifically, we show that seemingly monosemantic features fail to fire where they should, and instead get "absorbed" into their children features. We coin this phenomenon feature absorption, and show that it is caused by optimizing for sparsity in SAEs whenever the underlying features form a hierarchy. We introduce a metric to detect absorption in SAEs, and validate our findings empirically on hundreds of LLM SAEs. Our investigation suggests that varying SAE sizes or sparsity is insufficient to solve this issue. We discuss the implications of feature absorption in SAEs and some potential approaches to solve the fundamental theoretical issues before SAEs can be used for interpreting LLMs robustly and at scale.

Tags

ai-safety (imported, 100%)empirical (suggested, 88%)mechanistic-interp (suggested, 92%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Full Text

100,660 characters extracted from source content.

Expand or collapse full text

A is for Absorption: Studying Feature Splitting and Absorption in Sparse Autoencoders David Chanin *,1,2 James Wilken-Smith *,1 Tomáš Dulka *,1 Hardik Bhatnagar *,1,3 Satvik Golechha 4 Joseph Bloom 1,5 1 LASR Labs 2 University College London 3 Tübingen AI Center, University of Tübingen 4 MATS 5 Decode Research Abstract Sparse Autoencoders (SAEs) aim to decompose the activation space of large language models (LLMs) into human-interpretable latent directions or features. As we increase the number of features in the SAE, hierarchical features tend to split into finer features (“math” may split into “algebra”, “geometry”, etc.), a phenomenon referred to as feature splitting. However, we show that sparse decomposition and splitting of hierarchical features is not robust. Specifically, we show that seemingly monosemantic features fail to fire where they should, and instead get “absorbed” into their children features. We coin this phenomenon feature absorption, and show that it is caused by optimizing for sparsity in SAEs whenever the underlying features form a hierarchy. We introduce a metric to detect absorption in SAEs, and validate our findings empirically on hundreds of LLM SAEs. Our investigation suggests that varying SAE sizes or sparsity is insufficient to solve this issue. We discuss the implications of feature absorption in SAEs and some potential approaches to solve the fundamental theoretical issues before SAEs can be used for interpreting LLMs robustly and at scale. 1 Introduction Large Language Models (LLMs) have achieved remarkable performance across a wide range of tasks, yet our understanding of their internal mechanisms lags behind their capabilities. This gap between performance and interpretability raises concerns about the “black box” nature of these models [30]. The field of mechanistic interpretability aims to address this issue by reverse-engineering the internal algorithms of neural networks and performing causal analysis on them [24]. Recent work theorizes that models represent concepts as linear directions in a high-dimensional space, known as the Linear Representation Hypothesis (LRH) [26,8]. The model is able to represent far more concepts than it has neurons in its hidden space by allowing these concept directions to overlap slighly, known as superposition [8]. Superposition makes it challenging to directly interpret neurons in an LLM, and requires different techniques to extract interpretable feature directions. As long as the active features in a given LLM activation are sparse and the underlying features follow the LRH, Sparse Autoencoders (SAEs) should be able to recover the true LLM features despite supersition using sparse dictionary learning [27]. Indeed, SAEs have shown potential in decomposing the dense, polysemantic activations of LLMs into more “interpretable” latent features [5, 2]. However, we show that even if all underlying features are linear and sparsely activating, an SAE will still fail to recover the true underlying features if the features form a hierarchy. Instead, an SAE will * These authors contributed equally to this work. 39th Conference on Neural Information Processing Systems (NeurIPS 2025). arXiv:2409.14507v6 [cs.CL] 17 Nov 2025 Ideal interpretable solution SAE EncoderSAE Decoder Latent 1“starts with S”“starts with S” Latent 2“short”“short” The ideal interpretable solution requires firing 2 latents to represent “short”. Uninterpretable absorption solution SAE EncoderSAE Decoder Latent 1 ¬“short” ∧ “starts S”“starts with S” Latent 2“short”“short” + “starts S” Absorption only requires firing 1 latent to represent “short”, and is unfortunately what the SAE learns. Figure 1: In feature absorption, seemingly monosemantic latents fail to fire in cases where they apparently should. Here, we see an SAE can represent the word “short” and the concept “starts with S” more sparsely by absorbing the “starts with S” direction into the “short” latent, and then not firing the “starts with S” latent on the word “short”, despite “short” starting with “S”. Logical notation is used to describe the SAE encoder to emphasize its role as a classifier. learn gerrymandered latents that fail to fire on seemingly arbitrary cases where the latent should fire according to the mainline interpretation of the latent. We refer to this failure as feature absorption. Feature absorption is demonstrated in Figure 1, where the feature “short” always fires alongside a feature representing “starts with S”. Instead of learning an interpretable latent representing “starts with S”, the SAE can increase sparsity by instead disabling the “starts with S” latent when “short” is active while still getting perfect reconstruction. We present the following contributions: (1) we identify a problematic variant of feature-splitting we call “feature absorption”, where an SAE latent appears to track a human-interpretable concept, but fails to activate on seemingly arbitrary tokens. Instead, more specific latents activate and contribute a component of feature direction, “absorbing” the feature. (2) We demonstrate that feature absorption is caused by hierarchical features. (3) We develop a metric to detect feature absorption in LLM SAEs. And (4) we validate that feature absorption occurs in every LLM SAE we tested, including hundreds of open-source SAEs. Feature absorption poses an obstacle to the practical application of SAEs since it suggests SAE latents may be inherently unreliable classifiers. This is particularly important for applications where we need confidence that latents are fully tracking behaviors, such as bias or deceptive behavior. Furthermore, techniques which seek to describe circuits in terms of a sparse combination of latents will also be more difficult in the presence of feature absorption [21]. An online explorer for our results can be found athttps://feature-absorption.streamlit. app. Code is available at https://github.com/lasr-spelling/sae-spelling. 2 Background Hierarchical features.We say featuresf 1 andf 2 form a hierarchy withf 1 as the parent andf 2 as the child if f 2 =⇒ f 1 , meaning every time f 2 fires f 1 must also fire. Linear probing. A linear probe is a simple linear classifier trained on the hidden activations of a neural network, typically using logistic regression (LR) [1]. K-sparse probing.A k-sparse probe [12] is a linear probe trained on a sparse subset ofkneurons or SAE latents. Training a k-sparse probe first requires selecting thekbest neurons or SAE latents that in-aggregate act as a good classifier, and then training a standard linear probe on just thosek neurons or latents. Gurnee et al.[12]proposed several methods of estimating the bestkneurons or latents to pick, one of which involves first training a LR probe with a L1 loss term, and selecting theklargest elements by probe weight. When we refer to k-sparse probing in this work, we use this method of selectingk latents. Sparse autoencoders.An SAE consists of an encoder,W enc , a decoder,W dec , and corresponding biasesb enc andb dec . The SAE has a nonlinearity,σ, typically a ReLU (or variant such as JumpReLU [29]). Given input activation,a, the SAE computes a hidden representation,f, and reconstruction,ˆa: 2 f =σ(W enc a + b enc )(1) ˆa =W dec f + b dec (2) SAEs attempt to reconstruct input activations by projecting into an overcomplete basis using a sparsity-inducing loss term (typicallyL1loss), or a certain number of non-zero latents (L0) on the hidden activations. SAE feature ablation. In an ablation study we examine the downstream causal effect of an SAE latent by computing how patching its activation to 0 changes a downstream metric (e.g. logit difference). A negative ablation effect means intervening on the SAE latent would lower the metric. We follow the work of Marks et al. [21] and provide the algorithm in Appendix A.4. 3 Toy models of feature absorption Absorption is caused by the SAE sparsity penalty in the presence of hierarchical features. When two features form a hierarchy, for instance “starts with S” and “short”, the SAE can merge the “starts with S” feature direction into a latent tracking “short” and then not fire the main “starts with S” latent. This means firing one latent instead of two, increasing sparsity while retaining perfect reconstruction. We demonstrate that hierarchical feature co-occurrence causes absorption in a simple toy setting. Our initial setup consists of 4 true features, each randomly initialized into orthogonal directions with a 50 dimensional representation vector and unit norm. Each feature fires with magnitude 1.0. Featuref 0 fires with probability 0.25, and featuresf 1 ,f 2 , andf 3 fire with probability 0.05. Each SAE training input is created by sampling from these true features and summing the directions of each firing feature. We train a SAE with 4 latents to match the 4 true features using SAELens [15]. The SAE uses L1 loss with L1 coefficient 3e-5, and learning rate 3e-4. We train on 100M activations. Independently firing features When the true features fire independently, we find that the SAE is able to perfectly recover these features as shown in Figure 2a. The SAE learns one latent per true feature. The decoder representations perfectly match the true feature representations, and the encoder learns to perfectly segment out each feature from the other features. 0123 True feature 0 1 2 3 SAE Latent SAE encoder 0123 True feature 0 1 2 3 SAE Latent SAE decoder −1.0 −0.5 0.0 0.5 1.0 cos sim Cos sim with true features (Independent features) (a) When features fire independently, the SAE learns exactly one latent per feature, and the decoder per- fectly recovers each feature direction. 0123 True feature 0 1 2 3 SAE Latent SAE encoder 0123 True feature 0 1 2 3 SAE Latent SAE decoder −1.0 −0.5 0.0 0.5 1.0 cos sim Cos sim with true features (feat 1 co-occurs w/feat 0) (b) When features 0 and 1 co-occur (feature 1 only fires if feature 0 fires), we see absorption in the SAE encoder and decoder latents 0 and 1. Figure 2: Comparison of independent features (left) vs. co-occurring features with absorption (right) Figure 3: Interpretation of learned SAE latents with co-occurrence between feature 0 and feature 1 (feature 1 only fires if feature 0 fires). SAE ENCODERSAE DECODER LATENT 0¬ feat 1 ∧ feat 0feat 0 LATENT 1feat 1feat 0 + feat 1 LATENT 2feat 3feat 3 LATENT 3feat 2feat 2 Hierarchical features cause absorption Next, we modify the firing pattern of feature 1 so it fires only if feature 0 also fires. We keep the overall firing rate of feature 1 the same as before, firing in 5% of activations. Features 2 and 3 remain independent. Figure 2b shows the encoder and decoder cosine similarities with the true features in the hier- archical co-occurrence setup. Here, we see a clear example of feature absorption. Latent 0 3 has learned a perfect representation of feature 0, but the encoder has a hole in its recall. Latent 0 fires if feature 0 is active but not feature 1. This is exactly the sort of gerrymandered feature firing pattern we will see later in real SAEs in Section 5.2 - the encoder has learned to stop the latent firing on specific cases where it looks like it should be firing. In addition, we see that latent 1, which tracks feature 1, has absorbed the feature 0 direction. This results in latent 1 representing a combination of feature 0 and feature 1. We see that the independently firing features 2 and 3 are untouched - the SAE still learns perfect representations of these features. These results are summarized in Table 3. We explore absorption in more toy settings in Appendix A.3. Proof: hierarchical features cause absorptionWe further provide an analytical proof that in the hierarchical setup described above, feature absorption decreases SAE loss in Appendix A.2. 4 Experimental setup Our experiments on LLM SAEs focus on predicting the first-letter of a single token containing characters from the English alphabet (a-z, A-Z) and an optional leading space. We use in-context learning (ICL) prompts to elicit knowledge from the model, using templates of the form: token has the first letter: capitalized_first_letter An example of an ICL prompt consisting of 2 in-context examples is shown below. The model should output the _D token: tartan has the first letter: T mirth has the first letter: M dog has the first letter: In the above prompt, we extract residual stream activations at the_dogtoken index. These activations are used both for LR probe training and for applying SAEs. We use a train/test split of 80% / 20%, and evaluate only on the test set of the probes, including when running experiments on SAEs. When applying SAEs, we include the SAE error term [21] to avoid changing model output. To determine the causal effect of SAE latents on the first-letter identification task we conduct ablation studies. We use a metric consisting of the logit of the correct letter minus the mean logit of all incorrect letters. This measures the propensity of the model to choose the correct starting letter as opposed to other letters. Formally, our metricmis defined below, wheregrefers to the final token logits, L is the set of uppercase letters, and y is the uppercase letter that is the correct starting letter: m = g[y]− 1 |L|− 1 X l∈L g[l] We discuss this metric and alternative formulations further in Appendix A.10. To determine how well multiple latents perform as a classifier when used together, we use k-sparse probing, increasing the value ofkfrom1to15. We train a LR probe using a L1 loss term with coefficient 0.01, and select the top k latents by magnitude. We use the base Gemma-2-2B model for most of our studies, along with the full set of Gemma Scope residual stream SAEs of width 16k and 65k released by Deepmind [19]. We also evaluate absorption on our own SAEs trained on Qwen2 0.5B [32] and Llama 3.2 1B [6]. 5 Results Our results are divided into four sections. First, we compare the performance of linear probes with SAE latents on recovering first-character information from model activations, showing that despite appearing to track first letter features, a wide variety of precision / recall is achieved. Second, we motivate our definition of feature absorption with a case-study, emphasizing how an absorbing latent can unexpectedly causally mediate first letter information whilst the first-letter latent (unexpectedly) fails to fire. Next, we attempt to quantify feature splitting and feature absorption, showing that tuning of hyper-parameters may partially assist but not fully alleviate feature absorption. 4 5.1 Do SAEs learn latents that track first letter information? We compare the performance of LR probes with the performance of the SAE latent whose encoder direction has highest cosine similarity with the probe, resulting in 26 “first-letter” latents. We observed that for each probe, there was clearly one or at most a couple of outlier SAE latents with high probe cosine similarity. Full plots of cosine similarity vs letter are shown in Appendix A.8. We also tried using k=1 sparse probing [12] to identify SAE latents, and found this gives similar results. Further comparison of using k=1 sparse probing vs encoder cosine similarity to identify latents is explored in Appendix A.7. 024681012141618202224 Layer 0.2 0.4 0.6 0.8 Mean f1 First-letter SAE f1 vs Layer SAE width 16k 65k Probe (a) Mean F1 score using top SAE encoder latent by layer for Gemma Scope SAEs and LR probe. 0100200300400 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Mean f1 First-letter SAE f1 vs L0 (layers 0-12) SAE width 16k 65k (b) Mean F1 score vs L0 for Gemma Scope SAEs layers 0-12. SAEs learn cleanest “first-letter” features with L0 near 25-50. 0100200300400 L0 0.2 0.3 0.4 0.5 0.6 0.7 Mean f1 First-letter SAE f1 vs L0 (layers 13-25) SAE width 16k 65k (c) Mean F1 score vs L0 for Gemma Scope SAEs layers 13-25. SAEs learn cleanest “first-letter” features with L0 near 50-100. Figure 4: Comparison of F1 scores for first-letter classification tasks 0100200300400 L0 0.2 0.4 0.6 0.8 1.0 Mean precision First-letter SAE precision vs L0 SAE width 16k 65k (a) Mean precision using top SAE encoder latent vs L0 for all Gemma Scope SAEs. 0100200300400 L0 0.2 0.4 0.6 0.8 1.0 Mean recall First-letter SAE recall vs L0 SAE width 16k 65k (b) Mean recall using top SAE en- coder latent vs L0 for all Gemma Scope SAEs. 0.20.40.60.81.0 Mean precision 0.2 0.4 0.6 0.8 1.0 Mean recall First-letter SAE precision vs recall 0 100 200 300 400 L0 (c) Mean precision vs recall using top SAE encoder latent vs L0 for all Gemma Scope SAEs. Figure 5: Precision and recall vs L0 for first-letter classification tasks We observe wide variance in the performance of Gemma Scope SAEs at the first-letter identification task, but no SAE matches LR probe performance. We show the mean F1 score by layer as well as the F1 score of the LR probe in Figure 4a. We further investigate the F1 score of these SAE encoder latents as a function of L0 and SAE width in Figures 4b and 4c. Whether or not an SAE learns a clear “first-letter” latent for each letter is highly dependent on L0, with low L0 SAEs tending to learn high-precision low-recall latents, and high L0 SAEs learning low-precision high-recall latents (Figure 5). We caution drawing conclusions about an “optimal” L0 from these plots, as we find further variance when broken-down by letter, shown in Appendix A.8. 5.2 Why do SAE latents underperform? The Gemma Scope layer 3, 16k width, 59 L0 SAE has a latent, 6510, which appears to act as a classifier for “starts with S”, achieving an F1 of 0.81. However, this latent fails to activate on some tokens the probe can classify, and which the model can spell, such as the token _short. Figure 6a shows a sample prompt containing a series of tokens that start with “S”, and the activations of top SAE latents by ablation score for these tokens. The main “starts with S” latent, 6510, activates on all these tokens except_short. This SAE also has a token-aligned latent, 1085, which activates on variants of the word “short” (“ short”, “SHORT”, etc...). The Neuronpedia dashboard [20] for latent 1085 is shown in Appendix A.15. For the token_short, the main “starts with S” latent does not activate but the “short” latent activates instead. 5 snake soggy steam short soccer sax Token 0 10 20 30 40 50 Activation ‘S’ activations by token, layer 3, 16k width, 59 L0 Latent ID 1085 6510 (a) Layer 3, L0=59 SAE latent activations for tokens that start with “S”. The core “starts with S” latent, 6510, fails to activate on the token_short. The “short”-token aligned latent 1085 activates instead. 0250050007500100001250015000 SAE latent ID 0.0 0.2 0.4 Cosine similarity Main latent id 6510, cos 0.52 Absorbing latent id 1085, cos 0.12 Cosine similary between SAE decoder and ‘S’ probe (b) Cosine similarity between layer 3, L0=59 SAE de- coder and the “starts with S” probe. The main “Starts with S” latent, 6510, is clearly visible and highly probe-aligned. Figure 6: Comparison of SAE latent activations and cosine similarity for tokens starting with “S” 10852854 557 1412614964 26732791913437755489 SAE latent ID −6 −4 −2 0 Ablation effect First-letter ablation effects for “ short” token, layer 3, L0=59 (a) Ablation effect for_shorttoken, indicating that latent 1085, is responsible for the “starts with S” con- cept for the_shorttoken. The main “starts with S” latent, 6510, does not activate on the _short token. 2854 557 14126 26732791913437755489 1076512163 SAE latent ID −6 −4 −2 0 Ablation effect Ablation effects for “ short” token projecting out probe (b) Ablation effect for_shorttoken after removing the probe direction from latent 1085 via projection. Latent 1085 no longer appears in the plot, indicating the strong ablation effect in Figure 7a is due to its component along the probe direction. Figure 7: Ablation effects on _short token before and after projecting out the probe direction Latent 1085 has a cosine similarity with the “starts with S” probe of0.12, indicating it contains a component of the “starts with S” direction, although much smaller than the main “starts with S” latent. Cosine similarity of the SAE decoder with the “starts with S” LR probe is shown in Figure 6b. Interestingly, despite latent 1085 having only about1/5the cosine similarity with the probe as the main latent 6510, we see it activates with about5times the magnitude of latent 6510 on the_short token, thus contributing a similar amount of the “starts with S” probe direction to the residual stream. We study the ablation effect of each SAE latent on the_shorttoken, shown in Figure 7a, and see that latent 1085 has a dramatically larger ablation effect compared with all other SAE latents. This suggests latent 1085 is causally responsible for the model knowing that _short starts with S. Is it possible that the probe projection is not the causally important component of latent 1085? We conduct another ablation effect experiment, except now we remove the probe direction from latent 1085 via projection before ablation. The results of this experiment are shown in Figure 7b. After removing the probe component from latent 1085, it no longer has a significant ablation effect. Thus we know the probe projection of latent 1085 is responsible for model behavior. These experiments show the “starts with S” feature has been “absorbed” by the token-aligned latent 1085, likely along with other semantic concepts related to the word “short”. After observing that the main “starts with S” latent 6510 activates on most tokens that begin with “S”, it may be tempting to conclude this latent tracks the interpretable feature of beginning with the letter “S”. However, this latent quietly fails to activate on the _short token, leading us to a false sense of understanding. Here we clearly see feature absorption. The seemingly interpretable SAE latent 6510 fails to activate on arbitrary positive examples, and instead the feature is “absorbed” into more specific latents. Feature absorption is likely a logical consequence of SAE sparsity loss. If a dense and sparse feature co-occur, absorbing the dense feature into a latent tracking the sparse feature will increase sparsity. 6 Table 1: Sample max activating examples for latents 7112 and 7657 for Gemma Scope 16k, layer 0, 105 L0 from Neuronpedia. The token where the SAE latent activates is highlighted in yellow. Latent 7112 appears to be a lowercase “L” starting-letter latent, and latent 7657 appears to be a corresponding uppercase “L” latent. LATENT 7112LATENT 7657 žda se napla ́ cuje naknadaLC, an aluminum boat . E. Søli, 20asLIFT andLF-Net. Once a></code></li></ullatter’s sisterLouise, who in 5.3 Measuring feature splitting and feature absorption Feature splittingA key phenomenon identified from previous studies of SAEs is feature-splitting [2], where a feature represented in a single latent in a smaller SAE can split into two or more latents in a larger SAE. During our experiments, we found strong evidence of feature-splitting in the Gemma Scope SAEs. For instance, in the layer 0, 16k width, 105 L0 SAE, we find two encoder latents (id:7112 and id:7657 1 ) which align with the “L” starting letter probe. Inspecting max activating examples, we see latent 7112 activates on tokens starting with lowercase “l”, while 7657 activates on tokens starting with uppercase “L”. Some activating examples for these latents are shown in Table 1. Feature splitting like this is not necessarily problematic for interpretability efforts since the split features are still easily identifiable, and depending on the context it may be more useful to have either a single “starts with L” latent or a pair of “starts with uppercase / lowercase L” latents. We measure feature splitting using k-sparse probing [12] on SAE activations. If increasing the k-sparse probe fromktok + 1causes a significant increase in probe F1 score, then the additional SAE latent provides a meaningful signal, and the combination of thesek + 1latents is likely a feature split. In the example of the uppercase “L” and lowercase “l” split, a k-sparse probe withk = 2trained on both these latents should predict “starts with letter L” much better than either latent on its own. Figure 8a shows F1 vs K for letters “L” and “N”. The “L” k-sparse probe shows a significant jump in F1 score moving from k=1 to k=2 corresponding to feature splitting, while the F1 score for the “N” k-sparse probe is relatively constant. 12345 K 0.0 0.2 0.4 0.6 0.8 F1 Starts with “L” 12345 K Starts with “N” K-sparse probing results (a) K-sparse probing results for letters “L” and “N”, layer 0, 16k width, 105 L0. “L” shows a significant improvement in F1 between k=1 and k=2 correspond- ing to feature splitting. 050100150200250300 L0 0.0 0.5 1.0 1.5 2.0 2.5 Num feature splits Mean feature splits per first-letter vs L0 SAE width 16k 65k (b) Mean number of feature splits per letter on the first-letter spelling task by L0. Feature splitting occurs more frequently with higher sparsity. Figure 8: Feature splitting analysis in sparse autoencoders We detect feature splitting by measuring whether increasingkby one causes a jump in F1 score by more than a threshold,τ. We useτ = 0.03as a reasonable choice after inspecting situations like in Figure 8a, where feature splitting corresponds to an F1 score jump between 0.05 - 0.1. Figure 8b shows feature splitting vs L0 for all 16k and 65k width Gemma Scope SAEs. 1 https://w.neuronpedia.org/list/cm0h1n2mt00019jdk274owq9e 7 The single latent or a set of traditional feature split latents that seem to act as a classifier for a human-interpretable feature like “starts with S” fail to fire in a seemingly arbitrary number of cases. What fires instead are often approximately token-aligned latents with small but positive alignment with the LR probe. We say these latents are absorbing the feature. 01234567891011121314151617 Layer 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Mean absorption rate Mean absorption rate for first-letter task vs Layer SAE width 16k 65k (a) Mean feature absorption rate vs layer on the first-letter task, Gemma Scope 16k and 65k SAEs. We do not see an obvious pattern in absorp- tion rates by layer. 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Mean absorption rate Mean absorption rate for first-letter task vs L0 SAE width 16k 65k (b) Mean feature absorption rate vs L0 on first-letter task, Gemma Scope 16k and 65k SAEs. Wider and more sparse SAEs demonstrate higher rates of absorption. 012345678 Layer 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Mean absorption rate Mean absorption rate for first-letter task vs layer Model / SAE Llama 3.2 1B / L1 Llama 3.2 1B / TopK Qwen 2 0.5B / L1 (c) Mean feature absorption rate vs layer on the first-letter task on Llama 3.2 1B and Qwen 2 0.5B models, L1 loss and TopK SAE ar- chitectures, layers 0-8. Figure 9: Feature absorption rates We quantify the extent to which feature absorption occurs with the metric feature absorption rate. We first findkfeature splits for a first-letter feature using a k-sparse probe. We then find false-negative tokens that allkfeature-split SAE latents fail to activate on, but which the LR probe correctly classifies, and run an integrated-gradients ablation experiment on those tokens. The ablation effect finds the most causally important SAE latents for the spelling of that token. If the SAE latent receiving the largest negative magnitude ablation effect has a cosine similarity with the LR probe above0.025, and is at least1.0larger than the latent with the second highest ablation effect, we say that feature absorption has occurred. These thresholds were chosen from manual inspection of the data to best distinguish the absorption phenomenon. We then calculate feature absorption rate as: absorption_rate = num_absorptions lr_probe_true_positives If there are more than 200 false negatives per letter, we randomly pick 200 samples to estimate the number of absorptions. We see absorption rate increases with higher sparsity and higher SAE width. Lower L0 likely pushes the SAE to absorb dense features like spelling information, increasing feature sparsity. Feature absorption rate vs L0 for Gemma Scope SAEs layers 0-17 is shown in Figure 9b. Absorption rate by letter is shown in Appendix A.14. We also train our own set of standard L1 loss SAEs on the first 8 layers of Qwen2 0.5B [32] and Llama 3.2 1B [6], and TopK SAEs [10] on Llama 3.2 1B. In Figure 9c we show that absorption occurs in these SAEs as well. Our metric cannot capture absorption past layer 17 in Gemma 2 2B since we rely on ablation experiments to be certain the absorbed feature causally mediates model behavior. Past layer 17, attention has already moved the starting letter information from the source token into the final token position, so any ablations on the source token past layer 17 have little effect. This is a limitation of our absorption metric - we rely on ablation to be certain of the causal impact of absorbed features on model behavior, but this limits the layer depth our metric can be applied. We discuss this further in Appendix A.12 and discuss alternative formulations of the metric in Appendix A.13. Our absorption metric is not perfect, and is likely an under-estimate of the true level of feature absorption. We only consider absorption to have occurred if a single SAE latent has a much larger ablation effect than all other latents, and if the main SAE latents for a feature do not activate at all. Our metric will not capture multiple absorbing latents activating together, or the main latents activating weakly. Regardless, we feel our metric is a reasonable conservative baseline. 6 Related work Applications of Probes and SAEs for Model Interpretability Probing methods can extract interpretable information from language models, though this does not guarantee the model uses these representations in its computation, and requires labeled data [7]. 8 Prior work has shown that many human-interpretable concepts in LLM activations are represented as linear directions in activation space, known as the linear representation hypothesis [8,28]. Li et al.[18]used non-linear probes to recover board representations from a transformer trained on Othello scripts (“OthelloGPT”). Nanda[23]later showed that linear representations were not only recoverable but also editable. Karvonen et al.[16]developed objective metrics for SAE evaluation using Chess and Othello board states, but does not apply these to SAEs trained on LLM activations. Work by Olah et al.[26], Kissane et al.[17], Templeton et al.[31]noted poor precision/recall of SAE latents compared to known proxies. We extend this by showing how sparsity mediates precision/recall across many Gemma Scope SAEs and offer a possible explanation of low recall due to feature absorption. Engels et al. [9] investigated SAE errors, finding that not all SAE error is linearly decomposable. Studying precision and recall of SAE LatentsMost existing work on SAE interpretability mainly studies max activating examples [5], which may be misleading. There are more rigorous works which only measure precision [2,31,17]. Recent work has briefly explored recall and found it to be worse than expected naively, but this remains poorly understood [26]. We build on this work by evaluating precision / recall on a large number of SAEs, and offer a partial explanation for lower-than-expected recall of SAE latents in the form of feature absorption. Decomposing SAE LatentsFeature splitting was first described in Bricken et al.[2], which noted that different SAE widths and sparsities induce latents of different granularity, with wider SAEs often learning more specific variants of features. Bussmann et al.[4]find that by training an SAE on the decoder of another SAE, a technique called Meta-SAEs, it is possible to break down a single SAE latent like “Einstein” into subcomponents like “German” and “Physicist” and “starts with E”. 7 Discussion LimitationsOur Absoption metric uses ablation effect to ensure that the absorbed features causally mediate model behavior, and thus might not be easily transferable to the final model layers. Alternate metric formulations mitigating this are discussed in Appendix A.13. Due to compute constraints, we only train and evaluate a small number of non-JumpReLU SAEs in Figure 9. As our goal was only to show absorption occurs in all SAE architectures, we did not feel this is a significant drawback. Future Work The primary goal of future work is to find solutions to feature absorption. We are particularly hopeful that work extending Meta-SAEs [4] may solve or mitigate feature absorption. Another possible solution may be attribution dictionary learning [25]. Finally, structured sparsity techniques such as group lasso [13] or hierarchical sparse coding [14] may also be a promising direction of future work. Other possible directions include allowing absorption to occur and using it as a way to recover hierarchies between features in a LLM. Our toy model results suggest that absorption leads to an asymmetric pattern in the encoder and decoder of the SAE, so it may be possible to use this insight to detect absorption (although there may be other reasons for an asymmetry in the SAE encoder and decoder beyond absorption). ConclusionWe identify a form of feature splitting we call “feature absorption”, where more specific latents “steal credit” from more general ones. Absorption creates an interpretability illusion, where a seemingly interpretable latent has arbitrary false negatives in its mainline interpretation. Lower recall poses problems for using SAEs for high-stakes classification or finding sparse circuits [21], as the number of latents needed to characterize model behavior may be much larger than expected. We show that absorption is a consequence of hierarchical co-occurrence between sparse and dense features. If a dense feature like “starts with letter D” always co-occurs with a more sparse feature like “dogs”, the SAE can increase sparsity by absorbing the “starts with D” feature into a “dogs” latent. We hope that our work highlights the fundamental limitations of sparse feature extraction and prompts future research on SAEs such as identifying cases where a feature “should have activated” but does not due to absorption, and exploring theoretical solutions to absorption. The ease of demonstrating absorption in toy models makes it easier to validate potential solutions. 9 Acknowledgments and Disclosure of Funding This project was produced as part of the LASR Labs research program. DC was supported thanks to EPSRC EP/S021566/1. HB has received funding from the Digital Europe Programme under grant agreement No 101195233 (OpenEuroLLM). References [1]Guillaume Alain and Yoshua Bengio. Understanding intermediate layers using linear classifier probes, 2017. URL https://openreview.net/forum?id=ryF7rTqgl. [2]Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, et al. Towards monosemanticity: Decomposing language models with dictionary learning. Transformer Circuits Thread, 2, 2023. [3]Bart Bussmann, Patrick Leask, and Neel Nanda. Batchtopk sparse autoencoders, 2024. URL https://arxiv.org/abs/2412.06410. [4]Bart Bussmann, Michael Pearce, Patrick Leask, Joseph Bloom, Lee Sharkey, and Neel Nanda. Showing sae latents are not atomic using meta-saes, 2024. URLhttps://w.lesswrong. com/posts/TMAmHh4DdMr4nCSr5. [5] Hoagy Cunningham, Logan Riggs Smith, Aidan Ewart, Robert Huben, and Lee Sharkey. Sparse autoencoders find highly interpretable features in language models. In The Twelfth International Conference on Learning Representations, 2024. URLhttps://openreview.net/forum? id=F76bwRSLeK. [6]Abhimanyu Dubey, Abhinav Jauhri, Abhinav Pandey, Abhishek Kadian, Ahmad Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Amy Yang, Angela Fan, Anirudh Goyal, Anthony Hartshorn, Aobo Yang, Archi Mitra, Archie Sravankumar, Artem Korenev, Arthur Hinsvark, Arun Rao, Aston Zhang, Aurelien Rodriguez, Austen Gregerson, Ava Spataru, Baptiste Roziere, Bethany Biron, Binh Tang, Bobbie Chern, Charlotte Caucheteux, Chaya Nayak, Chloe Bi, Chris Marra, Chris McConnell, Christian Keller, Christophe Touret, Chunyang Wu, Corinne Wong, Cristian Canton Ferrer, Cyrus Nikolaidis, Damien Allonsius, Daniel Song, Danielle Pintz, Danny Livshits, David Esiobu, Dhruv Choudhary, Dhruv Mahajan, Diego Garcia-Olano, Diego Perino, Dieuwke Hupkes, Egor Lakomkin, Ehab AlBadawy, Elina Lobanova, Emily Dinan, Eric Michael Smith, Filip Radenovic, Frank Zhang, Gabriel Synnaeve, Gabrielle Lee, Georgia Lewis Anderson, Graeme Nail, Gregoire Mialon, Guan Pang, Guillem Cucurell, Hailey Nguyen, Hannah Korevaar, Hu Xu, Hugo Touvron, Iliyan Zarov, Imanol Arrieta Ibarra, Isabel Kloumann, Ishan Misra, Ivan Evtimov, Jade Copet, Jaewon Lee, Jan Geffert, Jana Vranes, Jason Park, Jay Mahadeokar, Jeet Shah, Jelmer van der Linde, Jennifer Billock, Jenny Hong, Jenya Lee, Jeremy Fu, Jianfeng Chi, Jianyu Huang, Jiawen Liu, Jie Wang, Jiecao Yu, Joanna Bitton, Joe Spisak, Jongsoo Park, Joseph Rocca, Joshua Johnstun, Joshua Saxe, Junteng Jia, Kalyan Vasuden Alwala, Kartikeya Upasani, Kate Plawiak, Ke Li, Kenneth Heafield, Kevin Stone, Khalid El-Arini, Krithika Iyer, Kshitiz Malik, Kuenley Chiu, Kunal Bhalla, Lauren Rantala-Yeary, Laurens van der Maaten, Lawrence Chen, Liang Tan, Liz Jenkins, Louis Martin, Lovish Madaan, Lubo Malo, Lukas Blecher, Lukas Landzaat, Luke de Oliveira, Madeline Muzzi, Mahesh Pasupuleti, Mannat Singh, Manohar Paluri, Marcin Kardas, Mathew Oldham, Mathieu Rita, Maya Pavlova, Melanie Kambadur, Mike Lewis, Min Si, Mitesh Kumar Singh, Mona Hassan, Naman Goyal, Narjes Torabi, Nikolay Bashlykov, Nikolay Bogoychev, Niladri Chatterji, Olivier Duchenne, Onur Çelebi, Patrick Alrassy, Pengchuan Zhang, Pengwei Li, Petar Vasic, Peter Weng, Prajjwal Bhargava, Pratik Dubal, Praveen Krishnan, Punit Singh Koura, Puxin Xu, Qing He, Qingxiao Dong, Ragavan Srinivasan, Raj Ganapathy, Ramon Calderer, Ricardo Silveira Cabral, Robert Stojnic, Roberta Raileanu, Rohit Girdhar, Rohit Patel, Romain Sauvestre, Ronnie Polidoro, Roshan Sumbaly, Ross Taylor, Ruan Silva, Rui Hou, Rui Wang, Saghar Hosseini, Sahana Chennabasappa, Sanjay Singh, Sean Bell, Seohyun Sonia Kim, Sergey Edunov, Shaoliang Nie, Sharan Narang, Sharath Raparthy, Sheng Shen, Shengye Wan, Shruti Bhosale, Shun Zhang, Simon Vandenhende, Soumya Batra, Spencer Whitman, Sten Sootla, Stephane Collot, Suchin Gururangan, Sydney Borodinsky, Tamar Herman, Tara Fowler, Tarek Sheasha, Thomas Georgiou, Thomas Scialom, Tobias Speckbacher, Todor Mihaylov, Tong 10 Xiao, Ujjwal Karn, Vedanuj Goswami, Vibhor Gupta, Vignesh Ramanathan, Viktor Kerkez, Vincent Gonguet, Virginie Do, Vish Vogeti, Vladan Petrovic, Weiwei Chu, Wenhan Xiong, Wenyin Fu, Whitney Meers, Xavier Martinet, Xiaodong Wang, Xiaoqing Ellen Tan, Xinfeng Xie, Xuchao Jia, Xuewei Wang, Yaelle Goldschlag, Yashesh Gaur, Yasmine Babaei, Yi Wen, Yiwen Song, Yuchen Zhang, Yue Li, Yuning Mao, Zacharie Delpierre Coudert, Zheng Yan, Zhengxing Chen, Zoe Papakipos, Aaditya Singh, Aaron Grattafiori, Abha Jain, Adam Kelsey, Adam Shajnfeld, Adithya Gangidi, Adolfo Victoria, Ahuva Goldstand, Ajay Menon, Ajay Sharma, Alex Boesenberg, Alex Vaughan, Alexei Baevski, Allie Feinstein, Amanda Kallet, Amit Sangani, Anam Yunus, Andrei Lupu, Andres Alvarado, Andrew Caples, Andrew Gu, Andrew Ho, Andrew Poulton, Andrew Ryan, Ankit Ramchandani, Annie Franco, Aparajita Saraf, Arkabandhu Chowdhury, Ashley Gabriel, Ashwin Bharambe, Assaf Eisenman, Azadeh Yazdan, Beau James, Ben Maurer, Benjamin Leonhardi, Bernie Huang, Beth Loyd, Beto De Paola, Bhargavi Paranjape, Bing Liu, Bo Wu, Boyu Ni, Braden Hancock, Bram Wasti, Brandon Spence, Brani Stojkovic, Brian Gamido, Britt Montalvo, Carl Parker, Carly Burton, Catalina Mejia, Changhan Wang, Changkyu Kim, Chao Zhou, Chester Hu, Ching-Hsiang Chu, Chris Cai, Chris Tindal, Christoph Feichtenhofer, Damon Civin, Dana Beaty, Daniel Kreymer, Daniel Li, Danny Wyatt, David Adkins, David Xu, Davide Testuggine, Delia David, Devi Parikh, Diana Liskovich, Didem Foss, Dingkang Wang, Duc Le, Dustin Holland, Edward Dowling, Eissa Jamil, Elaine Montgomery, Eleonora Presani, Emily Hahn, Emily Wood, Erik Brinkman, Esteban Arcaute, Evan Dunbar, Evan Smothers, Fei Sun, Felix Kreuk, Feng Tian, Firat Ozgenel, Francesco Caggioni, Francisco Guzmán, Frank Kanayet, Frank Seide, Gabriela Medina Florez, Gabriella Schwarz, Gada Badeer, Georgia Swee, Gil Halpern, Govind Thattai, Grant Herman, Grigory Sizov, Guangyi, Zhang, Guna Lakshminarayanan, Hamid Shojanazeri, Han Zou, Hannah Wang, Hanwen Zha, Haroun Habeeb, Harrison Rudolph, Helen Suk, Henry Aspegren, Hunter Goldman, Ibrahim Damlaj, Igor Molybog, Igor Tufanov, Irina-Elena Veliche, Itai Gat, Jake Weissman, James Geboski, James Kohli, Japhet Asher, Jean-Baptiste Gaya, Jeff Marcus, Jeff Tang, Jennifer Chan, Jenny Zhen, Jeremy Reizenstein, Jeremy Teboul, Jessica Zhong, Jian Jin, Jingyi Yang, Joe Cummings, Jon Carvill, Jon Shepard, Jonathan McPhie, Jonathan Torres, Josh Ginsburg, Junjie Wang, Kai Wu, Kam Hou U, Karan Saxena, Karthik Prasad, Kartikay Khandelwal, Katayoun Zand, Kathy Matosich, Kaushik Veeraraghavan, Kelly Michelena, Keqian Li, Kun Huang, Kunal Chawla, Kushal Lakhotia, Kyle Huang, Lailin Chen, Lakshya Garg, Lavender A, Leandro Silva, Lee Bell, Lei Zhang, Liangpeng Guo, Licheng Yu, Liron Moshkovich, Luca Wehrstedt, Madian Khabsa, Manav Avalani, Manish Bhatt, Maria Tsimpoukelli, Martynas Mankus, Matan Hasson, Matthew Lennie, Matthias Reso, Maxim Groshev, Maxim Naumov, Maya Lathi, Meghan Keneally, Michael L. Seltzer, Michal Valko, Michelle Restrepo, Mihir Patel, Mik Vyatskov, Mikayel Samvelyan, Mike Clark, Mike Macey, Mike Wang, Miquel Jubert Hermoso, Mo Metanat, Mohammad Rastegari, Munish Bansal, Nandhini Santhanam, Natascha Parks, Natasha White, Navyata Bawa, Nayan Singhal, Nick Egebo, Nicolas Usunier, Nikolay Pavlovich Laptev, Ning Dong, Ning Zhang, Norman Cheng, Oleg Chernoguz, Olivia Hart, Omkar Salpekar, Ozlem Kalinli, Parkin Kent, Parth Parekh, Paul Saab, Pavan Balaji, Pedro Rittner, Philip Bontrager, Pierre Roux, Piotr Dollar, Polina Zvyagina, Prashant Ratanchandani, Pritish Yuvraj, Qian Liang, Rachad Alao, Rachel Rodriguez, Rafi Ayub, Raghotham Murthy, Raghu Nayani, Rahul Mitra, Raymond Li, Rebekkah Hogan, Robin Battey, Rocky Wang, Rohan Maheswari, Russ Howes, Ruty Rinott, Sai Jayesh Bondu, Samyak Datta, Sara Chugh, Sara Hunt, Sargun Dhillon, Sasha Sidorov, Satadru Pan, Saurabh Verma, Seiji Yamamoto, Sharadh Ramaswamy, Shaun Lindsay, Shaun Lindsay, Sheng Feng, Shenghao Lin, Shengxin Cindy Zha, Shiva Shankar, Shuqiang Zhang, Shuqiang Zhang, Sinong Wang, Sneha Agarwal, Soji Sajuyigbe, Soumith Chintala, Stephanie Max, Stephen Chen, Steve Kehoe, Steve Satterfield, Sudarshan Govindaprasad, Sumit Gupta, Sungmin Cho, Sunny Virk, Suraj Subramanian, Sy Choudhury, Sydney Goldman, Tal Remez, Tamar Glaser, Tamara Best, Thilo Kohler, Thomas Robinson, Tianhe Li, Tianjun Zhang, Tim Matthews, Timothy Chou, Tzook Shaked, Varun Vontimitta, Victoria Ajayi, Victoria Montanez, Vijai Mohan, Vinay Satish Kumar, Vishal Mangla, Vítor Albiero, Vlad Ionescu, Vlad Poenaru, Vlad Tiberiu Mihailescu, Vladimir Ivanov, Wei Li, Wenchen Wang, Wenwen Jiang, Wes Bouaziz, Will Constable, Xiaocheng Tang, Xiaofang Wang, Xiaojian Wu, Xiaolan Wang, Xide Xia, Xilun Wu, Xinbo Gao, Yanjun Chen, Ye Hu, Ye Jia, Ye Qi, Yenda Li, Yilin Zhang, Ying Zhang, Yossi Adi, Youngjin Nam, Yu, Wang, Yuchen Hao, Yundi Qian, Yuzi He, Zach Rait, Zachary DeVito, Zef Rosnbrick, Zhaoduo Wen, Zhenyu Yang, and Zhiwei Zhao. The llama 3 herd of models, 2024. URL https://arxiv.org/abs/2407.21783. 11 [7]Yanai Elazar, Shauli Ravfogel, Alon Jacovi, and Yoav Goldberg. Amnesic probing: Behavioral explanation with amnesic counterfactuals. Transactions of the Association for Computational Linguistics, 9:160–175, 2021. [8]Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, et al. Toy models of superposition. arXiv preprint arXiv:2209.10652, 2022. [9]Joshua Engels, Eric J Michaud, Isaac Liao, Wes Gurnee, and Max Tegmark. Not all language model features are one-dimensionally linear. In The Thirteenth International Conference on Learning Representations, 2025. URLhttps://openreview.net/forum?id=d63a4AM4hb. [10]Leo Gao, Tom Dupré la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders. arXiv preprint arXiv:2406.04093, 2024. [11] Mor Geva, Jasmijn Bastings, Katja Filippova, and Amir Globerson. Dissecting recall of factual associations in auto-regressive language models. In Proceedings of the 2023 Conference on Empirical Methods in Natural Language Processing, pages 12216–12235, 2023. [12]Wes Gurnee, Neel Nanda, Matthew Pauly, Katherine Harvey, Dmitrii Troitskii, and Dimitris Bertsimas. Finding neurons in a haystack: Case studies with sparse probing. Transactions on Machine Learning Research, 2023. ISSN 2835-8856. URLhttps://openreview.net/ forum?id=JYs1R9IMJr. [13]Laurent Jacob, Guillaume Obozinski, and Jean-Philippe Vert. Group lasso with overlap and graph lasso. In Proceedings of the 26th annual international conference on machine learning, pages 433–440, 2009. [14]Rodolphe Jenatton, Julien Mairal, Guillaume Obozinski, and Francis Bach. Proximal methods for hierarchical sparse coding. The Journal of Machine Learning Research, 12:2297–2334, 2011. [15]Curt Tigges Joseph Bloom and David Chanin. Saelens.https://github.com/jbloomAus/ SAELens, 2024. [16]Adam Karvonen, Benjamin Wright, Can Rager, Rico Angell, Jannik Brinkmann, Logan Riggs Smith, Claudio Mayrink Verdun, David Bau, and Samuel Marks. Measuring progress in dictionary learning for language model interpretability with board game models. In ICML 2024 Workshop on Mechanistic Interpretability, 2024. [17]Connor Kissane, Robert Krzyzanowski, Joseph Isaac Bloom, Arthur Conmy, and Neel Nanda. Interpreting attention layer outputs with sparse autoencoders, 2024. URLhttps://arxiv. org/abs/2406.17759. [18] Kenneth Li, Aspen K Hopkins, David Bau, Fernanda Viégas, Hanspeter Pfister, and Martin Wattenberg. Emergent world representations: Exploring a sequence model trained on a synthetic task. ICLR, 2023. [19]Tom Lieberum, Senthooran Rajamanoharan, Arthur Conmy, Lewis Smith, Nicolas Sonnerat, Vikrant Varma, János Kramár, Anca Dragan, Rohin Shah, and Neel Nanda. Gemma Scope: Open Sparse Autoencoders Everywhere All At Once on Gemma 2, August 2024. [20]Johnny Lin and Joseph Bloom. Analyzing neural networks with dictionary learning, 2023. URL https://w.neuronpedia.org. Software available from neuronpedia.org. [21]Samuel Marks, Can Rager, Eric J. Michaud, Yonatan Belinkov, David Bau, and Aaron Mueller. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models. Computing Research Repository, arXiv:2403.19647, 2024. URLhttps://arxiv.org/abs/ 2403.19647. [22]Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. Locating and editing fac- tual associations in GPT. Advances in Neural Information Processing Systems, 36, 2022. arXiv:2202.05262. 12 [23]Neel Nanda. Actually, othello-gpt has a linear emergent world model, mar 2023. URL< https://neelnanda.io/mechanistic-interpretability/othello, 2023. [24]Chris Olah, Nick Cammarata, Ludwig Schubert, Gabriel Goh, Michael Petrov, and Shan Carter. Zoom in: An introduction to circuits. Distill, 5(3):e00024–001, 2020. [25]Chris Olah, Adly Templeton, Trenton Bricken, and Adam Jermyn. April update.https: //transformer-circuits.pub/2024/april-update/index.html, 2024. URLhttps: //transformer-circuits.pub/2024/april-update/index.html. [26]Chris Olah, Nicholas Turner, Adam Jermyn, and Joshua Batson. July update.https:// transformer-circuits.pub/2024/july-update/index.html, 2024. URLhttps:// transformer-circuits.pub/2024/july-update/index.html. [27]Bruno A Olshausen and David J Field. Sparse coding with an overcomplete basis set: A strategy employed by v1? Vision research, 37(23):3311–3325, 1997. [28]Kiho Park, Yo Joong Choe, and Victor Veitch. The linear representation hypothesis and the geometry of large language models. In Forty-first International Conference on Machine Learning, 2024. URL https://openreview.net/forum?id=UGpGkLzwpP. [29] Senthooran Rajamanoharan, Tom Lieberum, Nicolas Sonnerat, Arthur Conmy, Vikrant Varma, János Kramár, and Neel Nanda. Jumping ahead: Improving reconstruction fidelity with jumprelu sparse autoencoders. arXiv preprint arXiv:2407.14435, 2024. [30]Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, pages 206–215, 2019. [31]Adly Templeton, Tom Conerly, Jonathan Marcus, Jack Lindsey, Trenton Bricken, Brian Chen, Adam Pearce, Craig Citro, Emmanuel Ameisen, Andy Jones, Hoagy Cunningham, Nicholas L Turner, Callum McDougall, Monte MacDiarmid, Alex Tamkin, Esin Durmus, Tristan Hume, Francesco Mosconi, C. Daniel Freeman, Theodore R. Sumers, Edward Rees, Joshua Batson, Adam Jermyn, Shan Carter, Chris Olah, and Tom Henighan. Scaling monosemanticity: Extract- ing interpretable features from claude 3 sonnet.https://transformer-circuits.pub/ 2024/scaling-monosemanticity/, May 2024. Accessed on May 21, 2024. [32] An Yang, Baosong Yang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Zhou, Chengpeng Li, Chengyuan Li, Dayiheng Liu, Fei Huang, Guanting Dong, Haoran Wei, Huan Lin, Jialong Tang, Jialin Wang, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Ma, Jin Xu, Jingren Zhou, Jinze Bai, Jinzheng He, Junyang Lin, Kai Dang, Keming Lu, Keqin Chen, Kexin Yang, Mei Li, Mingfeng Xue, Na Ni, Pei Zhang, Peng Wang, Ru Peng, Rui Men, Ruize Gao, Runji Lin, Shijie Wang, Shuai Bai, Sinan Tan, Tianhang Zhu, Tianhao Li, Tianyu Liu, Wenbin Ge, Xiaodong Deng, Xiaohuan Zhou, Xingzhang Ren, Xinyu Zhang, Xipin Wei, Xuancheng Ren, Yang Fan, Yang Yao, Yichang Zhang, Yu Wan, Yunfei Chu, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zhihao Fan. Qwen2 technical report. arXiv preprint arXiv:2407.10671, 2024. A Appendix A.1 Glossary of Terms Sparse Autoencoders (SAEs):Neural networks trained to reconstruct their input while enforcing sparsity in their hidden layer. In the context of this paper, SAEs are used to decompose the dense activations of language models into more interpretable features. SAE error term: When inserting a SAE into the computation path of the model, errors in SAE reconstruction will propagate to later parts of the model and can change the model output. We refer to the error as the SAE error term, and corresponds to the difference between the SAE output and the original SAE input activation. Marks et al.[21]introduced the idea of adding this error term back to the SAE output to ensure that the SAE does not change model output. 13 Latent: We refer to neurons in the hidden layer of a SAE as latents to avoid overloading the term “feature”. This is in contrast to earlier work which used the term “feature” to refer to both human-interpretable concepts and SAE hidden layer neurons. Feature: We use the term “feature” to refer to an idealized human-interpretable concept that the model represent in its activations and which a SAE latent may or may not represent. Monosemantic:Referring to a feature or representation that corresponds to a single, clear semantic concept. In the context of SAEs, a monosemantic feature would ideally capture one interpretable aspect of the input. Interpretable: A latent being interpretable is not well defined in the field, making it difficult to ensure that different authors mean the same thing when referring to SAE interpretability. When we refer to a SAE latent as being interpretable in this work, we mean that it should behave in line with how it appears to behave after inspecting its activation patterns. If an SAE latent appears to track a feature X by a reasonable inspection of its activations but has subtle deviations from this behavior in reality, we say this is not interpretable. We thus measure interpretability via classification performance when a latent appears to be a classifier over some feature. Feature dashboard: A dashboard showing activation patterns and max-activating examples for a SAE latent. Feature dashboards are commonly used to interpret the behavior of an SAE latent. Neuronpedia: A platform,https://neuronpedia.org, which hosts feature dashboards for popular SAEs [20]. Token-aligned latent: A latent which seems to roughly fire on variants of the same token. For instance, a “Snake” token-aligned latent may fire on the tokens “Snake”, “SNAKE”, “_snakes”, etc... Feature splitting: A phenomenon in SAEs introduced by Bricken et al.[2], where a SAE latent tracks a general feature in a narrow SAE, but splits into multiple more specific SAE latents in a wider SAE. For instance, a latent tracking “starts with L” in a narrow SAE may split into a latent tracking “starts with capital L” and a latent tracking “starts with lowercase L” in a wider SAE. Feature absorption: A problematic form of feature splitting where a SAE latent appears to track an interpretable feature, but that latent has seemingly arbitrary exception cases where it fails to fire. Instead, a more specific latent “absorbs” the feature direction and fires in place of the main latent. Circuit: In the context of neural network interpretability, a circuit refers to a subgraph of neurons or latents within a neural network that work together to perform a specific function or computation. The study of circuits aims to understand how different components of a neural network interact to process information and produce outputs. Linear probe: A simple linear classifier (typically logistic regression) trained on the hidden activations of a neural network to predict some property or task. Used to assess what information is linearly decodable from the network’s representations. K-sparse probing: A variant of linear probing where only the k most important latents (as deter- mined by some selection method) are used to train the probe. This helps identify which specific neurons or latents are most relevant for a given task. Ablation study:An experimental method where a component of a system (in this case, a neuron or latent in a neural network) is removed or altered to observe its effect on the system’s performance. This helps determine the causal importance of the component. Integrated gradients (IG): An attribution method that assigns importance scores to input latents by accumulating gradients along a path from a baseline input to the actual input. In this paper, it’s used as an approximation technique for ablation studies. 14 In-context learning (ICL):A paradigm where a language model is given examples of a task within its input prompt, allowing it to adapt to new tasks without fine-tuning. Often used with few-shot learning techniques. Residual stream: In the context of transformer architectures, the residual stream refers to the main information flow that bypasses the self-attention and feed-forward layers through residual connections. Logits: The raw, unnormalized outputs of a neural network’s final layer, before any activation function (like softmax) is applied. In language models, logits typically represent the model’s scores for each token in the vocabulary. Activation patching: An interpretability technique where activations at specific locations in a neural network are replaced or modified to observe the effect on the network’s output. This helps in understanding the causal role of different parts of the network in producing its final output. A.2 Proof: absorption decreases SAE loss for hierarchical features We analyze the effect of a specific form of feature absorption, termedδ-absorption, within a Sparse Autoencoder (SAE) framework. We consider two hierarchically related features,f 1 andf 2 (where f 2 ⊂ f 1 ), and demonstrate that for a defined family of encoder and decoder weights parameterized by δ ∈ [0, 1] (δ = 0 corresponds to no absorption, and δ = 1 corresponds to full absorption): 1. Perfect reconstruction of inputs composed of these features is maintained across all values of δ. 2.The sparsity loss component attributable to these features is a decreasing function ofδ, provided the child feature f 2 has a non-zero probability of appearing. 3.Consequently, optimizing for sparsity encourages higher values ofδ, i.e., greater absorption. 1. Preliminaries and Assumptions H1. Dataset and FeaturesLetDbe a dataset. We consider a set of featuresF =f 1 ,f 2 ,...,f d . Each featuref i ∈ R k is a vector with unit norm,∥f i ∥ 2 = 1. Features are mutually orthogonal: f i · f j = δ ij (Kronecker delta), whereδ ij = 1ifi = jand0ifi̸= j. An activationh∈ R k in the model’s residual stream is a linear combination of active features: h = P j∈ActiveFeatures f j . H2. Feature Hierarchy and Probabilities We focus on two featuresf 1 ,f 2 ∈Fwith a hierarchy f 2 ⊂ f 1 . This implies that iff 2 is present in a datapoint,f 1 must also be present. The probabilities of observing combinations of f 1 and f 2 are: • p(f 1 ,f 2 ) = p 11 : Probability of f 1 and f 2 co-occurring (e.g., input is f 1 + f 2 ). • p(f 1 ,¬f 2 ) = p 10 : Probability of f 1 occurring without f 2 (e.g., input is f 1 ). • p(¬f 1 ,f 2 ) = p 01 : Probability off 2 occurring withoutf 1 . By the hierarchy assumption, p 01 = 0. • p(¬f 1 ,¬f 2 ) = p 00 : Probability of neitherf 1 norf 2 occurring (e.g., input is0or some f k ,k ̸= 1, 2). We assume p 11 + p 10 + p 00 = 1. H3. Sparse Autoencoder (SAE) Model The SAE reconstructs an inputhas ˆ h = f φ (h). The reconstruction is ˆ h = W d z, wherez = ReLU(W e h). No bias terms are used.W e,i is thei-th row of W e (encoder vector for latenti), andW d,i is thei-th column ofW d (decoder vector for latenti). We analyze two specific latents, z 1 and z 2 , intended to capture f 1 and f 2 . Other latents z j for j > 2 are assumed to perfectly reconstruct other featuresf j (e.g.W e,j = f j ,W d,j = f j ) and do not interact with f 1 ,f 2 due to orthogonality. 15 H4. SAE Loss FunctionThe total loss isL =L rec +λL sp , whereλ > 0.L rec = E h∼D h ∥h− ˆ h∥ 2 2 i . L sp = E h∼D P i |z i |. Our analysis will focus on the contributions of z 1 ,z 2 toL rec andL sp . H5. Definition ofδ-AbsorptionWe define a specific parameterization for the encoder and decoder weights associated with f 1 and f 2 by a parameter δ ∈ [0, 1]: • W e,1 = f 1 − δf 2 • W e,2 = f 2 • W d,1 = f 1 • W d,2 = f 2 + δf 1 δ = 0 represents no absorption, while δ = 1 represents full absorption. 2. Proposition 1: Perfect Reconstruction under δ-Absorption For anyδ ∈ [0, 1], and for inputshconsisting only off 1 ,f 2 or0, the reconstruction ˆ h = W d,1 z 1 + W d,2 z 2 perfectly reconstructsh, i.e., the reconstruction loss componentL (1,2) rec due to these features is 0. Proof We consider the possible input types based on f 1 ,f 2 : Case 1: h = f 1 (only parent feature f 1 is present). The latent activations are: z 1 = ReLU(W e,1 · h) = ReLU((f 1 − δf 2 )· f 1 ) = ReLU(f 1 · f 1 − δf 2 · f 1 ) = ReLU(1− δ· 0) = 1 (by H1) z 2 = ReLU(W e,2 · h) = ReLU(f 2 · f 1 ) = ReLU(0) = 0 (by H1) The reconstruction is: ˆ h = z 1 W d,1 + z 2 W d,2 = 1· f 1 + 0· (f 2 + δf 1 ) = f 1 Thus, ˆ h = h. Case 2: h = f 1 + f 2 (both parent f 1 and child f 2 are present). The latent activations are: z 1 = ReLU(W e,1 · h) = ReLU((f 1 − δf 2 )· (f 1 + f 2 )) = ReLU(f 1 · f 1 + f 1 · f 2 − δf 2 · f 1 − δf 2 · f 2 ) = ReLU(1 + 0− δ· 0− δ· 1) = ReLU(1− δ) (by H1) Since δ ∈ [0, 1], 1− δ ≥ 0, so z 1 = 1− δ. z 2 = ReLU(W e,2 · h) = ReLU(f 2 · (f 1 + f 2 )) = ReLU(f 2 · f 1 + f 2 · f 2 ) = ReLU(0 + 1) = 1 (by H1) The reconstruction is: ˆ h = z 1 W d,1 + z 2 W d,2 = (1− δ)f 1 + 1· (f 2 + δf 1 ) = (1− δ)f 1 + f 2 + δf 1 = f 1 − δf 1 + f 2 + δf 1 = f 1 + f 2 Thus, ˆ h = h. Case 3: h = 0 (neither f 1 nor f 2 is present). z 1 = ReLU((f 1 − δf 2 )· 0) = 0 z 2 = ReLU(f 2 · 0) = 0 The reconstruction is: ˆ h = 0· f 1 + 0· (f 2 + δf 1 ) = 0 Thus, ˆ h = h. 16 Case 4:h = f 2 (only child featuref 2 is present). This case is disallowed by assumption H2 (p 01 = 0), as f 2 ⊂ f 1 implies f 1 must be present if f 2 is. In all permissible cases,h− ˆ h = 0, so∥h− ˆ h∥ 2 2 = 0 . Therefore, the reconstruction loss component due to f 1 ,f 2 , denotedL (1,2) rec , is 0 for any δ ∈ [0, 1]. 3. Proposition 2: Sparsity Loss under δ-Absorption The expected sparsity loss contribution from latentsz 1 andz 2 , denotedL (1,2) sp = E h∼D [|z 1 | +|z 2 |], is given by: L (1,2) sp = p 11 (2− δ) + p 10 Furthermore, its derivative with respect to δ is: dL (1,2) sp dδ =−p 11 ProofWe calculate the sum of absolute latent activations|z 1 | +|z 2 |for each case from Proposition 1 and weight them by their probabilities (H2): •Ifh = f 1 + f 2 (probabilityp 11 ):z 1 = 1− δ,z 2 = 1. Sinceδ ∈ [0, 1],1− δ ≥ 0, so |z 1 | = 1− δ. |z 2 | = 1. Thus,|z 1 | +|z 2 | = (1− δ) + 1 = 2− δ. • If h = f 1 (probability p 10 ): z 1 = 1, z 2 = 0. Thus,|z 1 | +|z 2 | = 1 + 0 = 1. •Ifh = 0(probabilityp 00 , neitherf 1 norf 2 present):z 1 = 0,z 2 = 0. Thus,|z 1 | +|z 2 | = 0 + 0 = 0. The case corresponding to p 01 does not occur. The expected sparsity loss from z 1 ,z 2 is: L (1,2) sp = p 11 · (2− δ) + p 10 · 1 + p 00 · 0 = p 11 (2− δ) + p 10 Taking the derivative with respect to δ: dL (1,2) sp dδ = d dδ (2p 11 − δp 11 + p 10 ) Since p 11 and p 10 are constants with respect to δ: dL (1,2) sp dδ =−p 11 □ 4. Corollary: Increasing Absorption Decreases Sparsity Loss Ifp 11 > 0(i.e., the child featuref 2 co-occurs withf 1 with non-zero probability), then increasingδ strictly decreasesL (1,2) sp . Proof From Proposition 2, dL (1,2) sp dδ = −p 11 . Ifp 11 > 0, then−p 11 < 0. A negative derivative implies thatL (1,2) sp is a decreasing function ofδforδ ∈ [0, 1]. The minimum value ofL (1,2) sp over this interval occurs atδ = 1(full absorption), yieldingL (1,2) sp (δ = 1) = p 11 (2− 1) + p 10 = p 11 + p 10 . The maximum value occurs atδ = 0(no absorption), yieldingL (1,2) sp (δ = 0) = p 11 (2− 0) + p 10 = 2p 11 + p 10 .□ 17 5. Conclusion Given the specifiedδ-absorption mechanism for an SAE handling two hierarchical featuresf 1 ,f 2 (where f 2 ⊂ f 1 ): 1.Perfect reconstruction of inputs composed off 1 andf 2 is maintained irrespective of the degree of absorption δ. Thus,L (1,2) rec is unaffected by δ. 2. The sparsity loss componentL (1,2) sp associated with these features is p 11 (2− δ) + p 10 . 3.Ifp 11 > 0, the total lossL(focusing on the components related tof 1 ,f 2 ) decreases asδ increases becauseL (1,2) rec is constant (zero) andL (1,2) sp decreases. 4. Therefore, an optimization process like gradient descent, when minimizing the total loss L = L rec + λL sp (whereλ > 0), will favor increasingδtowards 1, thereby promoting feature absorption for these hierarchically related features, assuming the SAE learns or is constrained to these forms of W e and W d . This formalizes the argument that, under the given conditions and definitions, absorption is a mechanism that can reduce SAE loss by improving sparsity without harming reconstruction for hierarchical features. A.3 Extended toy model experiments In this section we explore further variants on absorption in toy models. We use the same setting as our main toy model experiment, with four mutually-orthogonal true features, and train an SAE with four latents. Each true featuref ∈ R 50 . Unless otherwise stated, every time feature 1 fires feature 0 must also fire, but feature 0 is allowed to fire on its own as well. This is to simulate hierarchal features such as our example “starts with S” and “short” features, where every time the “short” feature fires we expect “starts with S” must also fire since “short” starts with “S”, but “starts with S” can fire on its own as well. Feature 2 and 3 are fully independent. All features fire with magnitude 1.0 and variance 0.0 unless otherwise stated. Magnitude variance causes partial absorption In our main toy model experiment, each true feature fires with magnitude exactly 1.0. This is not very realistic, though - likely there will be some variance in feature firing magnitudes in real LLMs. We simulate this by adding variance of 0.1 to the firing magnitude of feature 0, so the relative magnitudes of feature 0 and 1 are no longer fixed. We show the plots of cosine similarity between SAE encoder and decoder in Figure 10. Here, we still see the same absorption pattern in the SAE encoder and decoder with the latent 3 encoder containing a negative component of feature 1, and the latent 0 decoder merging features 0 and 1. We show some sample true feature firings and corresponding SAE latent activations in Table 2. 0123 True feature 0 1 2 3 SAE Latent SAE encoder 0123 True feature 0 1 2 3 SAE Latent SAE decoder −1.0 −0.5 0.0 0.5 1.0 cos sim Cos sim with true features (feat 1 co-occurs w/feat 0, feat 0 magnitude varies) Figure 10: SAE encoder and decoder with true features. The firing magnitude of feature 0 has variance 0.1, while the remaining features fire with variance 0.0. When there is variance in the firing magnitudes of parent and child features, we still see an absorption pattern in the SAE encoder and decoder with the latent 3 encoder containing a negative component of feature 1, and the latent 0 decoder merging features 0 and 1. We see that now the SAE latent tracking feature 0 still fires when the true values of features 0 and 1 are both 1.0, but very weakly. However, if the magnitude of feature 0 drops down to 0.75, then the feature 0 latent fully turns off. 18 Table 2: Sample true feature firings and corresponding SAE latent activations. Feature 1 only fires if feature 0 fires. Feature 0 has variance of 0.1 in its firing magnitude, while the other feature have no variance in their firing magnitude. TRUE FEATURESSAE LATENT ACTS 1.000.000.000.000.000.000.001.00 1.001.000.000.001.270.000.000.22 0.901.000.000.001.270.000.000.12 0.751.000.000.001.260.000.000.00 We call this phenonemon partial absorption. In partial absorption, there’s co-occurrence between a dense and sparse feature, and the sparse feature absorbs the direction of the dense feature. However, the SAE latent tracking the dense feature still fires when both the dense and sparse feature are active, only very weakly. If the magnitude of the dense feature drops below some threshold, it stops firing entirely. Feature absorption is an optimal strategy for minimizing the L1 loss and maximizing sparsity. However, when a SAE absorbs one latent into another, the absorbing latent loses the ability to modulate the magnitudes of the underlying features relative to each other. The SAE can address this by firing the latent tracking the dense feature as a "correction" to add back some of the dense feature direction into the reconstruction. Since the dense feature latent is firing weakly, it still has lower L1 loss than if the SAE fully separated out the features into their own latents. Imperfect co-occurrence can still lead to partial absorption Next, we test what will happen if feature 1 is more likely to fire if feature 0 is active, but can still fire without feature 0. We set up feature 1 to fire with feature 0 95% of the time, but 5% of the time it can fire on its own. For this experiment, all features fire with magnitude 1.0 and 0 variance. We show the cosine similarities of the SAE encoder and decoder with true features in Figure 11. Some sample feature firings and corresponding SAE activations are shown in Table 3. 0123 True feature 0 1 2 3 SAE Latent SAE encoder 0123 True feature 0 1 2 3 SAE Latent SAE decoder −1.0 −0.5 0.0 0.5 1.0 cos sim Cos sim with true features (feat 1 partially co-occurs w/feat 0) Figure 11: SAE encoder and decoder with true features. Feature 1 fires with feature 0 95% of the time, but 5% of the time feature 1 is allowed to fire on its own. Even though there is not perfect co-occurrence between features 0 and 1, we still see the SAE encoder and decoder learn a weak absorption pattern, with the decoder latent for feature 1 absorbing part of feature 0, and the encoder latent for feature 0 including a negative component of feature 1. We see signs of partial absorption here as well. We see the same absorption pattern in the SAE encoder and decoder as we saw in our other absorption examples, although less severe than the previous examples. We also see in the sample firing patterns that when both feature 0 and 1 fire together, the latent tracking feature 0 fires with noticeably lower magnitude than when feature 0 fires on its own. Here, even though the co-occurrence between features 0 and 1 is not perfect, we still see partial absorption. Absorption also affects TopK SAEs So far, we have only shown feature absorption occurring with standard L1 SAEs. Next, we examine how other absorption affects other architectures using 19 Table 3: Sample feature values and corresponding SAE activations. Feature 1 can only fire if feature 0 is active 95% of the time, but 5% of the time feature 1 can fire on its own. We see signs of partial absorption, where the latent tracking feature 0 fires noticeably more weakly if feature 1 is active. TRUE FEATURESSAE LATENT ACTS 1.000.000.000.000.000.000.001.00 1.001.000.000.000.001.050.000.67 0.001.000.000.000.000.950.000.00 a batch topk SAE [3]. Batch topk SAEs are an improved version of topk SAEs [10] where the top k∗ Blatents are used to reconstruct the SAE input, whereBis the batch size. As the topk function enforces sparsity, there is no additional L1 loss term. Topk SAEs are harder to use for very small toy models like our 4-feature toy model above, since if thekis too large relative to the size of the SAE the SAE will not learn correct features. To address this, we use a slightly larger toy model with 12 mutually orthogonal true features. All features fire independently with probability 0.15, except for the first 2 features. Feature 0 is the parent feature in our hierarchy, and fires with probability 0.4. Feature 1 is the child feature, and fires with probability 0.6 only if feature 1 fires, but never fires if feature 1 does not fire. All features fire with magnitude 1.0. We train a batch topk SAE withk = 2. We show the cosine similarities of the SAE encoder and decoder with true features in Figure 12. 01234567891011 True feature 0 1 2 3 4 5 6 7 8 9 10 11 SAE Latent SAE encoder 01234567891011 True feature 0 1 2 3 4 5 6 7 8 9 10 11 SAE Latent SAE decoder −1.00 −0.75 −0.50 −0.25 0.00 0.25 0.50 0.75 1.00 cos sim Cos sim with true features (feat 1 co-occurs w/feat 0, batch topk SAE) Figure 12: SAE encoder and decoder with true features for a batch topk SAE with k=12. Feature 1 is only allowed to fire if feature 0 fires. We still see a very clear absorption pattern between the latents tracking features 0 and 1 despite the lack of L1 loss. We still see a clear absorption pattern between the latents tracking features 0 and 1 despite the lack of L1 loss. Absorption increases sparsity, which allows the topk SAE to have better reconstruction loss at a given k, and is thus what the SAE learns. A.4 Ablation algorithm A.5 How good is Gemma-2 on character identification tasks? We evaluate how well can Gemma-2-2B identify the first letter or all the letters in a token (spelling the full token). We evaluate the accuracy of the model on all tokens in the LR probe validation set with a prompt containing 10 in-context examples selected at random from the full vocabulary. Our results are shown in Figure 13. We see that performance on the first-letter identification task is high throughout token length, while the full-word spelling performance decreases as the length of the token increases. 20 Algorithm 1 SAE Latent Ablation 1: Insert SAE in model computation, including error term 2: Define a scalar metric on the model’s output distribution 3: Calculate baseline metric value for a test prompt 4: for each token of interest do 5:for each SAE latent do 6:Set the SAE latent activation to 0 7:Recalculate the metric 8:Compute ablation effect (baseline - new metric) 9:Reset the SAE latents to its original value 10:end for 11: end for 24681012 Token Length 0.0 0.2 0.4 0.6 0.8 1.0 Accuracy Spelling accuracy by token length Task First Letter Spelling Figure 13: Baseline performance for Gemma-2-2B on first-letter identification and full-token spelling by token length. A.6 Intervening on the first letter If the model is using the identified SAE latents for predicting the first letter we should also be able to change what first letter it predicts just by changing the activations. For this experiment we use the SAE latents most cosine similar with the LR probe for the true first letter and for a new randomly selected letter. We take the intermediate activations of Gemma-2-2B in the residual stream and encode them using the SAE. Then we zero out the activation of the SAE latent associated with the original letter and change the activation of the SAE latent associated with the new letter into the average activation it has on tokens starting with this new letter. Editing works better with latents from the narrower 16k SAE compared to the 65k, with the best L0s in the 75-150 range. This corresponds to the observed pattern of these SAE latents having higher F1 scores for classification. We report the results in Figure 14. The best SAEs on the layers 7-9 can achieve a substantial replacement, but note that the averages hide variance across individual tokens, where some get edited completely and others get unaffected. The edit success also varies based on the true first letter and the random new letter; for illustration we show a breakdown by letter for two specific SAEs in layer 7 in Figure 15. A.7 Probe cosine similarity vs k=1 sparse probing The first step when searching for a SAE latent that acts as a first-letter classifier involves searching for SAE latent which best acts as a classifier. In Figure 4, we achieve this by first training a LR probe on the first-letter task and using cosine similarity between that probe and the SAE encoder to find the best latent for the first-letter task. We also investigated using k-sparse probing with k=1 to select the best SAE latent instead. This involves training a linear probe with L1 loss and selecting the latent with the highest positive weight from the probe. We find that both k=1 sparse probing yield nearly identical results, as seen in Figures 16 and 17. Additionally Figure 18 shows the cosine similarity of the LR probe with each SAE latent by letter for the canonical Gemma Scope layer 0 16k width SAE. In most cases there is an obvious probe-aligned 21 0123456789 Layer 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 True First Letter Probability Decrease First Letter Edit Out Score vs Layer SAE width 16k 65k 0123456789 Layer 0.0 0.1 0.2 0.3 0.4 0.5 0.6 New First Letter Probability Increase First Letter Edit In Score vs Layer SAE width 16k 65k (a) Comparing success in editing out the true first letter and making the model predict a randomly selected new letter across layers 0-9 for all 16k and 65k Gemma Scope SAEs. 0100200300 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 True First Letter Probability Decrease First Letter Edit Out Score vs L0 SAE width 16k 65k 0100200300 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 New First Letter Probability Increase First Letter Edit In Score vs L0 SAE width 16k 65k (b) Comparing the edit success with the top SAE latent across all L0s for 16k and 65k widths across layers 0-9. The best performance seems to be occurring for L0 between 75 and 150. Figure 14: Comparison of Edit success by Layer and L0 abcdefghijklmnopqrstuvwxyz True First Letter 0.2 0.4 0.6 0.8 True First Letter Probability Decrease Layer 7: First Letter Edit Out Score vs Original Letter SAE width and L0 16k_137 65k_107 abcdefghijklmnopqrstuvwxyz New Letter 0.0 0.2 0.4 0.6 0.8 New First Letter Probability Increase Layer 7: First Letter Edit In Score vs New Letter SAE width and L0 16k_137 65k_107 Figure 15: Comparing the edit success broken down by the letter at layer 7 for two SAEs; SAE width 16,000 and L0 of 137 and SAE width 65,000 and L0 of 107. For each original letter we draw a sample of 100 tokens and average the decrease in probability of the correct first letter and increase in probability of a new random letter. latent. Likely any reasonable method of latent selection will find the same latent for these cases. We thus decided to use cosine similarity between the SAE encoder and a LR probe as our selection criteria for single SAE latents as this is a simpler metric and less computationally intensive to compute. A.8 Precision, recall, and F1 score for the first-letter task We evaluated precision, recall, and F1 score for the first-letter classification task, and found that the precision and recall vary depending on the L0 of the SAE. Low L0 SAEs learn high precision, low recall latents, while high L0 SAEs learn low precision, high recall latents. These results are shown in Figure 19. We thus chose to use F1 score as our core metric in this paper to balance precision and recall as many if the SAEs we tested have extreme values in either precision or recall. While it may appear that there is an optimal L0 from looking at aggregate statistics across letter, we find that breaking down the F1 vs L0 plot by letter reveals that the optimal L0 appears different for different letters, with low frequency letters like z actually having the best F1 score at the lowest L0, 22 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Mean F1 First-letter SAE F1 score vs L0 SAE width 16k 65k 1m (a) Mean F1 score on first-letter classification task using top SAE encoder latent by cosine similarity with the LR probe vs L0 for all Gemma Scope SAEs layers 0-9. 050100150200250300 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Mean f1 First-letter SAE f1 score vs L0 SAE width 16k 65k (b) Mean F1 score on first-letter classification task using k=1 sparse probing to select the SAE latent for the first-letter classification task vs L0 for all Gemma Scope SAEs layers 0-9. Figure 16: Comparison of LR probe cosine similarity and k=1 sparse probing vs l0 0123456789 Layer 0.2 0.4 0.6 0.8 Mean F1 First-letter SAE F1 score vs Layer SAE width 16k 65k 1m Probe (a) Mean F1 score on first-letter classification task using top SAE encoder latent by cosine similarity with the LR probe vs layer for all Gemma Scope SAEs layers 0-9. 012345678 Layer 0.2 0.4 0.6 0.8 1.0 Mean f1 First-letter SAE f1 score vs Layer SAE width 16k 65k Probe (b) Mean F1 score on first-letter classification task using k=1 sparse probing to select the SAE latent for the first-letter classification task vs layer for all Gemma Scope SAEs layers 0-9. Figure 17: Comparison of LR probe cosine similarity and k=1 sparse probing vs layer while other letters instead have an optimal L0 around 30-50. Figure 20 shows these results broken down by letter. A.9 SAE training We train SAEs on the first 8 layers of Qwen2 0.5B [32] and Llama 3.2 1B [6] using the SAELens library [15]. The SAEs are all trained with identical hyperparameters of L1 coefficient of 2.5 and 500M tokens. The Qwen2 0.5B SAEs all have L0 between 25 and 50 and explained variance between 0.77 and 0.83. The Llama 3.2 1B SAEs have L0 between 27 and 110, and explained variance between 0.74 and 0.89. We use a single 40gb Nvidia A100 GPU for training each SAE. A.10 Metric choice for ablation studies To determine the causal effect of SAE latents on the first-letter identification task, we use a metric,m, which measures the logit of the correct letter minus the mean logit of all incorrect letters. Our metric is defined below, wheregrefers to the final token logits,Lis the set of uppercase letters, andyis the uppercase letter that is the correct starting letter: m = g[y]− 1 |L|− 1 X l∈L g[l] 23 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Cosine similarity a 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 b 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 c 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 d 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 e 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 f 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Cosine similarity g 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 h 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 i 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 j 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 k 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 l 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Cosine similarity m 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 n 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 o 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 p 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 q 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 r 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Cosine similarity s 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 t 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 u 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 0.6 v 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 0.4 0.5 w 0250050007500100001250015000 −0.1 0.0 0.1 0.2 0.3 x 0250050007500100001250015000 Index −0.1 0.0 0.1 0.2 0.3 0.4 Cosine similarity y 0250050007500100001250015000 Index −0.1 0.0 0.1 0.2 0.3 0.4 z Figure 18: Decoder cosine similarities with the LR probe by letter, Gemma Scope 16k layer 0 l0=105. Most letters have one or two obvious SAE latents which align with the probe. 050100150200250300 L0 0.2 0.4 0.6 0.8 Mean precision First-letter SAE precision score vs L0 SAE width 16k 65k (a) Mean precision on first-letter classification task vs L0 for all Gemma Scope SAEs layers 0-9. Latents are selected via k=1 sparse probing 050100150200250300 L0 0.2 0.4 0.6 0.8 1.0 Mean recall First-letter SAE recall score vs L0 SAE width 16k 65k (b) Mean recall on first-letter classification task vs L0 for all Gemma Scope SAEs layers 0-9. Latents are selected via k=1 sparse probing Figure 19: Comparison of precision and recall vs l0 This metric is chosen to detect changes in the confidence of the model in predicting the correct letter relative to the mean reference class of other letters. This should capture changes in the model’s confidence in predicting the correct logit. This is not the only metric that could be chosen, and an argument can be made that we should subtract the max of all incorrect letter logits rather than the mean of all incorrect letter logits. The max form of this version of the metric is shown below: 24 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 Mean f1 Letter: A SAE width 16k 65k 050100150200250300350 L0 0.2 0.4 0.6 0.8 Mean f1 Letter: B SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: C SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: D SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: E SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: F SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: G SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: H SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: I SAE width 16k 65k 050100150200250300350 L0 0.0 0.2 0.4 0.6 0.8 Mean f1 Letter: J SAE width 16k 65k 050100150200250300350 L0 0.0 0.2 0.4 0.6 0.8 Mean f1 Letter: K SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: L SAE width 16k 65k 050100150200250300350 L0 0.0 0.2 0.4 0.6 0.8 Mean f1 Letter: M SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: N SAE width 16k 65k 050100150200250300350 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: O SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: P SAE width 16k 65k 050100150200250300350 L0 0.0 0.2 0.4 0.6 0.8 Mean f1 Letter: Q SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: R SAE width 16k 65k 050100150200250300350 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: S SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: T SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: U SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: V SAE width 16k 65k 050100150200250300350 L0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: W SAE width 16k 65k 050100150200250300350 L0 0.0 0.2 0.4 0.6 0.8 Mean f1 Letter: X SAE width 16k 65k 050100150200250300350 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: Y SAE width 16k 65k 050100150200250300350 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 Mean f1 Letter: Z SAE width 16k 65k First-letter SAE f1 score vs L0 for each letter Figure 20: F1 vs L0 by letter. SAE latents are picked using k=1 sparse probing. m max = g[y]− max l∈L g[l] This second form using a max can also account for the case where the logits of the model shift from being confident in the correct answer to instead being confident in an incorrect answer while leaving the logits of the correct answer the same. 050100150200250300 L0 0.00 0.05 0.10 0.15 0.20 Mean absorption rate Mean absorption rate: mean-logit metric SAE width 16k 65k (a) Absorption rate using the mean version of the ab- sorption metric, Gemma Scope layers 0-3. 050100150200250300 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Mean absorption rate Mean absorption rate: max-logit metric SAE width 16k 65k (b) Absorption rate using the max version of the ab- sorption metric, Gemma Scope layers 0-3. Figure 21: Comparison of absorption rates using the max and mean versions of the absorption metric. In practice, we expect that ablating an absorbing latent should cause the model to become less confident in the correct answer, so the difference between these two forms of the metric should yield similar results. 25 We calculate the mean absorption rate for Gemma Scope SAEs layers 0-3 in Figure 21 using both versions of this metric. The overall shape of the curve is nearly identical between these two choices of metrics. The mean version of the metric, which is used in this paper, results in a slightly more conservative estimate of absorption rate. We consider our absorption score to be a rough estimate of the true absorption rate and thus consider either the mean or the max version of the logit diff metric to be valid for evaluating absorption. 26 A.11 Choice of thresholds for absorption metric Our absorption metric makes use of several thresholds. For feature splitting, we use a threshold of 0.03 on the F1 score jump moving fromk = ntok = n + 1in sparse probing. To classify a latent as a case of absorption, we require cosine similarity with the LR probe above 0.025, and a clear largest ablation effect of 1.0 more than then next highest latent. In this section, we discuss the intuition behind these thresholds. In all cases, these thresholds are just the rough midpoint of ranges of values that achieve similar results, and that we feel are reasonable default values. We also refer readers to our online feature absorption explorer to gain a similar intuition as to what typical ranges of these values look like. Feature splitting threshold We detect interpretable feature splitting by noting how large a jump in F1-score we gain moving fromk = ntok = n + 1in sparse probing. Figure 8a demonstrates a typical example of how F1 score increases for interpretable feature splitting vs feature absorption. In the left side of the figure, we see that moving fromk = 1tok = 2there is a F1-score jump of around 0.08, while each increase after that is less than 0.01. For the figure on the right, where no interpretable feature splitting occurs, all F1-score jumps are less than 0.01. This plot is a very typical illustration of detecting feature splitting via k-sparse probing. Any threshold between 0.01 and 0.05 does a good job of detecting feature splitting. We set the threshold to 0.03 to be in the middle of this range. LR probe cosine similarity thresholdWe use a threshold of 0.025 on the cosine similarity of the firing latent with the LR probe as part of the metric, ensuring that any latent we classify as absorption must contain some component of the probe direction. This threshold is mainly a cheap way to filter out latents that are obviously not probe-aligned so we can avoid running the more expensive ablation experiments. Nearly any threshold above 0 and below 0.05 should work identically well for this purpose. We choose 0.025 to be in the middle of this range. For most cases of absorption we detect, the absorbing latent has a probe cosine similarity of around 0.05 - 0.15. Figure 6b demonstrates a very typical case of cosine similarity between an absorbing latent and the LR probe, showing cosine similarity of 0.12. Absorption effect threshold As part of our metric, we assert that any latent we classify as absorption must have the highest ablation effect of all latents, and that lead must be by at least 1.0. As the main goal of the metric in our paper is establishing definitively that absorption exists and affects real LLM SAEs, we focus on the most obvious cases of absorption where a single latent absorbs the parent feature direction in a single activation. This threshold thus serves two purposes: first, to establish the latent has a strong ablation effect, and second, the establish that the effect is dominant over other latents. The metric is not particularly sensitive to the exact choice of threshold we set here, as any dominant ablation effect above around 0.5 is already quite strong. Figure 7 shows a typical ablation effect for a dominant absorbing latent, and has an ablation effect above 6, while all other latents have ablation effects well below 0.5. We pick a threshold of 1.0 here as its a clean number that is well within the range that will work as a threshold. Any threshold between 0.5 and 3.0 should work just as well. A.12 Causal interventions and absorption In this work, we rely on causal interventions like ablation experiments to verify that SAE latents have a causal impact on model behavior. In these experiments for spelling tasks, we set up an ICL prompt to elicit spelling information from the model, for instance the ICL prompt below: tartan has the first letter: T mirth has the first letter: M dog has the first letter: In this ICL prompt, we would apply an SAE and train LR probes on the_dogtoken position, and expect that the model will output the token_D. When we intervene on the_dog, we can track the causal changes to model outputs by applying a metric to the output logits, e.g. checking how our intervention increases or decreases the _D logit relative to other letters. 27 We use these interventions as part of our absorption metric to ensure that when we claim that “absorption” is occurring, we verify that the absorbing latent has a causal impact on model outputs. This is stronger evidence than only noting a cosine similarity between the absorbing latent, but this means that our absorption metric cannot classify absorption at later model layers. During a LLM forward pass, the model first collects relevant information on a token in that token position, and attention heads then move relevant information from earlier tokens to later tokens [11,22]. If we assess ablation effect at layers after which model attention has already pulled relevant information from the subject tokens into the final output token, the ablation effect will be 0. For Gemma 2 2B on the first-letter spelling task, we find this movement of first-letter spelling information occurs around layer 18. Figure 22 shows an activation patching experiment [22] on a sample first-letter spelling prompt. In this experiment, we see that near layer 18 the model moves first-letter spelling information from the subject token to the prediction token. < bos > Robert : R JEAN : J Voraus : V creek : Token 0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 Layer Resid Activation Patching 0.0 0.2 0.4 0.6 0.8 1.0 Value Figure 22: Residual stream attribution patching for a sample first-letter spelling prompt, Gemma 2 2B. After around layer 18, model attention moves the relevant spelling information from the source token to the prediction location. As a result, our feature absorption metric will not function past layer 18 in Gemma 2 2B, and we thus focus on layers 0-17 for our analysis of feature absorption. We believe that feature absorption is still occurring in SAEs past layer 18, but we lose the ability to make causal claims that the absorbing latents are used by the model to make predictions. Given that this paper is trying to highlight the existence of feature absorption, we felt it is more important to have a metric which is robust and has the backing of causal analysis but which cannot be used at all model layers. Future work may make a different trade-off and choose a feature absorption metric which can work at all model layers, for instance relying only on cosine similarity between absorbing latents and a LR probe to determine absorption. We describe variations on the metric that can facilitate this in Appendix A.13 below. A.13 Alternate formulations of the absorption metric There are a few variations to the metric that can be made to make it more flexible so that it can be applied at all layers of the model, and can be tweaked to detect partial absorption as well. As described above, we did not use this variation as we felt it is more important to demonstrate definitely and conservatively that absorption occurs in this paper. However, we describe the changes that can be made to the metric as follows: replace ablation study with LR probe projection in reconstructionThe metric in our paper uses an integrated-gradients ablation study to be absolutely certain that an absorbing feature is causally 28 responsible for model behavior. We use this in the paper as our goal is to establish with certainty that feature absorption exists in real models, but it has the following drawbacks: •We need to be able to consistently prompt the model to perform the task, outputting a single token • We cannot evaluate final model layers after attention moves task-relevant information to the final token • Ablation studies are slow, which makes the absorption metric expensive to calculate. The ablation study can be replaced instead with a threshold on the portion of the logistic regression (LR) probe direction an absorbing latent contributes to the residual stream. To do this, we project all firing latents against the LR probe directiond p , as well as project the input activationaagainst the probe. We require that a latentlmust contribute at leastτ c portion of the probe projection to the reconstruction. So, in order for l to be considered an absorbing latent, the following must be true: τ c < ˆa l · d p a· d p whereˆa l is the reconstruction component of latentl(the encoder activation of latentltimes the decoder vector for l). Allow multiple absorbing latentsThe metric in the paper requires one dominant absorbing latent for simplicity, but it is possible in theory to have two or more absorbing latents firing together performing absorption. We can modify the metric to take this into account by allowing up toN latents firing together contribute up to τ c together. Thus, the threshold from change 1 above now becomes the following, wherel n refers to the absorbing latent with the nth largest contribution to the LR probe direction in the reconstruction: τ c < N X n ˆa l n · d p a· d p We do not do this in the paper as it requires another hyper-parameter which is hard to set in a principled way, but should result in a less conservative estimate of absorption. Allow main latent(s) to fire weakly instead of being fully turned off The metric in the paper requires that the main latent(s) for a task all be fully disabled to identify a case of absorption. However, in partial absorption, as explored in our section on toy models, the main latent does not always fully turn off but fires very weakly instead. We can adjust the metric to take this into account by relaxing the requirement that the main latents are fully disabled and instead allow them to fire weakly. Similar to change 1, we define a thresholdτ m as a maximum contribution to the probe direction in the reconstructed activationˆathat comes from each main latentl m . Thus, if we haveMmain latents, in order to be classified as absorption the following must be satisfied: τ m ≥ M X m ˆa l m · d p a· d p In the metric defined in the paper,τ m = 0, meaning the main latents must be fully turned off to count as absorption. This final change, which detects partial absorption, may be preferable to get a sense of an overall level of absorption present in an SAE. However, if the goal is to determine whether absorption affects the SAE’s ability to act as a classifier, then requiring that the main SAE latent be fully turned off to be called absorption is preferable, so we do not claim that either version of the metric is superior in all cases. 29 A.14 Additional plots In this section, we include additional plots that are too large to fit in the main body of the paper. 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 Absorption rate Letter: A SAE width 16k 65k 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: B 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Absorption rate Letter: C 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: D 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 Absorption rate Letter: E 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Absorption rate Letter: F 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Absorption rate Letter: G 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 Absorption rate Letter: H 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 Absorption rate Letter: I 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 Absorption rate Letter: J 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Absorption rate Letter: K 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 Absorption rate Letter: L 0100200300400 L0 0.0 0.2 0.4 0.6 0.8 Absorption rate Letter: M 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 Absorption rate Letter: N 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: O 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Absorption rate Letter: P 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 Absorption rate Letter: Q 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: R 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: S 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: T 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: U 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 0.30 0.35 0.40 Absorption rate Letter: V 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 Absorption rate Letter: W 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 0.25 Absorption rate Letter: X 0100200300400 L0 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Absorption rate Letter: Y 0100200300400 L0 0.00 0.05 0.10 0.15 0.20 Absorption rate Letter: Z Figure 23: Absorption rate vs L0 by letter, layers 0-17. We see a wide variance in which letters are absorbed by which SAEs. 30 A.15 Feature dashboards We include feature dashboard screenshots from Neuronpedia for some prominent latents mentioned in this work. Figure 24 shows a dashboard for Gemmascope layer 3, latent 1085, which is a token- aligned latent firing on variations of the word_shortand we find absorbs the “starts with S” direction. Figure 25 shows latent 6510 from the same layer which should be the main “starts with S” latent. Figure 24: Neuronpedia dashboard for Gemma Scope layer 3, latent 1085. This latent is a token- aligned latent for _short tokens. This latent absorbs the “starts with S” direction. Figure 25: Neuronpedia dashboard for Gemma Scope layer 3, latent 6510. This latent should be the main “starts with S” latent. 31