Paper deep dive
Efficient Dictionary Learning with Switch Sparse Autoencoders
Anish Mudide, Joshua Engels, Eric J. Michaud, Max Tegmark, Christian Schroeder de Witt
Models: GPT-2 Small
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 96%
Last extracted: 3/12/2026, 7:59:26 PM
Summary
The paper introduces 'Switch Sparse Autoencoders' (Switch SAEs), a novel architecture for dictionary learning in neural networks. By incorporating a routing network and multiple 'expert' SAEs, the architecture enables efficient scaling to a large number of features while reducing computational costs (FLOPs) compared to traditional dense SAEs. Experiments demonstrate that Switch SAEs provide a Pareto improvement in the reconstruction vs. sparsity frontier and maintain feature interpretability comparable to existing architectures.
Entities (5)
Relation Signals (3)
Switch Sparse Autoencoder → utilizes → Switch Layer
confidence 100% · The Switch SAE combines the Switch layer (Fedus et al., 2022) with the TopK SAE.
Switch Sparse Autoencoder → trainedon → GPT-2
confidence 95% · We train sparse autoencoders on the residual stream activations of GPT-2 small.
Switch Sparse Autoencoder → improvesupon → TopK Sparse Autoencoder
confidence 90% · Switch SAEs deliver a substantial Pareto improvement in the reconstruction vs. sparsity frontier for a given fixed training compute budget.
Cypher Suggestions (2)
List all language models used for training SAEs · confidence 95% · unvalidated
MATCH (sae:Architecture)-[:TRAINED_ON]->(lm:LanguageModel) RETURN sae.name, lm.name
Find all architectures compared against Switch Sparse Autoencoders · confidence 90% · unvalidated
MATCH (a:Architecture {name: 'Switch Sparse Autoencoder'})-[:IMPROVES_UPON|BENCHMARKED_AGAINST]->(other:Architecture) RETURN other.nameAbstract
Abstract:Sparse autoencoders (SAEs) are a recent technique for decomposing neural network activations into human-interpretable features. However, in order for SAEs to identify all features represented in frontier models, it will be necessary to scale them up to very high width, posing a computational challenge. In this work, we introduce Switch Sparse Autoencoders, a novel SAE architecture aimed at reducing the compute cost of training SAEs. Inspired by sparse mixture of experts models, Switch SAEs route activation vectors between smaller "expert" SAEs, enabling SAEs to efficiently scale to many more features. We present experiments comparing Switch SAEs with other SAE architectures, and find that Switch SAEs deliver a substantial Pareto improvement in the reconstruction vs. sparsity frontier for a given fixed training compute budget. We also study the geometry of features across experts, analyze features duplicated across experts, and verify that Switch SAE features are as interpretable as features found by other SAE architectures.
Tags
Links
Full Text
54,932 characters extracted from source content.
Expand or collapse full text
Efficient Dictionary Learning with Switch Sparse Autoencoders Anish Mudide Massachusetts Institute of Technology &Joshua Engels Massachusetts Institute of Technology &Eric J. Michaud Massachusetts Institute of Technology &Max Tegmark Massachusetts Institute of Technology Schroeder de Witt University of Oxford Correspondence to amudide@mit.edu. Abstract Sparse autoencoders (SAEs) are a recent technique for decomposing neural network activations into human-interpretable features. However, in order for SAEs to identify all features represented in frontier models, it will be necessary to scale them up to very high width, posing a computational challenge. In this work, we introduce Switch Sparse Autoencoders, a novel SAE architecture aimed at reducing the compute cost of training SAEs. Inspired by sparse mixture of experts models, Switch SAEs route activation vectors between smaller “expert” SAEs, enabling SAEs to efficiently scale to many more features. We present experiments comparing Switch SAEs with other SAE architectures, and find that Switch SAEs deliver a substantial Pareto improvement in the reconstruction vs. sparsity frontier for a given fixed training compute budget. We also study the geometry of features across experts, analyze features duplicated across experts, and verify that Switch SAE features are as interpretable as features found by other SAE architectures. 1 Introduction Recently, large language models have achieved impressive performance on many tasks (Brown et al., 2020), but they remain largely inscrutable to humans. Mechanistic interpretability aims to open this metaphorical black box and rigorously explain model computations (Olah et al., 2020). Specifically, much work in mechanistic interpretability has focused on understanding features, the specific human-interpretable variables a model uses for computation (Olah et al., 2020; Park et al., 2023; Engels et al., 2024). Early mechanistic attempts to understand features focused on neurons, but this work was stymied by the fact that neurons tend to be polysemantic: they are frequently activated by several completely different types of inputs, making them hard to interpret (Olah et al., 2020). One theory for why neurons are polysemantic is superposition, the idea that language models represent many more concepts than they have available dimensions (Elhage et al., 2022). To minimize interference, the superposition hypothesis posits that features are encoded as almost orthogonal directions in the model’s hidden state space. Bricken et al. (2023) and Cunningham et al. (2023) propose to disentangle these superimposed model representations into monosemantic features by training unsupervised sparse autoencoders (Lee et al., 2007; Le, 2013; Konda et al., 2014) on intermediate language model activations. The success of this technique has led to an explosion of recent work (Templeton et al., 2024; Gao et al., 2024) that has focused on scaling sparse autoencoders to frontier language models such as Claude 3 Sonnet (Anthropic, 2024a) and GPT-4 (Achiam et al., 2023). Despite scaling SAEs to 34 million features, Templeton et al. (2024) estimate that there likely remains orders of magnitude more features. Furthermore, Gao et al. (2024) train SAEs on a series of language models and find that larger models require more features to achieve the same reconstruction error. As model sizes continue to grow, current training methodologies are set to quickly become computationally intractable. Each training step of a sparse autoencoder generally consists of six major computations: the encoder forward pass, the encoder gradient, the decoder forward pass, the decoder gradient, the latent gradient and the pre-bias gradient. Gao et al. (2024) introduce kernels that leverage the sparsity of the TopK activation function to dramatically optimize all computations except the encoder forward pass, which is not sparse. After implementing these optimizations, Gao et al. (2024) find that the training time is bottlenecked by the dense encoder forward pass and the memory is bottlenecked by storing the latent pre-activations. In this work, we introduce the Switch Sparse Autoencoder, which to our knowledge is the first work to solve these dual memory and FLOP bottlenecks. The Switch SAE combines the Switch layer (Fedus et al., 2022) with the TopK SAE (Makhzani & Frey, 2013; Gao et al., 2024). At a high level, the Switch SAE is composed of many small expert SAEs and a trainable routing network that determines which expert SAE will process a given input. We demonstrate that the Switch SAE delivers an improvement in training FLOPs vs. training loss over existing methods. 1.1 Contributions 1. In Section 3, we describe the Switch Sparse Autoencoder architecture and compare it to existing SAE architectures. We also describe our training methodology, which balances reconstruction and expert utilization. 2. In Section 4.1, we describe scaling laws for reconstruction MSE with respect to FLOPs and parameters. We show that Switch SAEs achieve a lower MSE compared to TopK SAEs using the same amount of training compute. 3. In Section 4.2, we benchmark Switch SAEs against ReLU, Gated and TopK SAEs on the sparsity-reconstruction Pareto frontier. 4. In Section 4.3, we study feature duplication in Switch SAEs and visualize the global structure of Switch SAE features using t-SNE. 5. In Section 4.4, we demonstrate that Switch SAE features are as interpretable as TopK SAE features. Figure 1: Scaling laws for Switch SAEs. We train dense TopK SAEs and Switch SAEs of varying size with fixed k=3232k=32k = 32. Left: Switch SAEs achieve better reconstruction than dense SAEs at a fixed compute budget. Right: Switch SAEs require more features in total (and therefore more parameters) to achieve the same reconstruction as dense SAEs when trained to convergence, although this gap narrows for larger SAEs. 2 Related Work 2.1 Mixture of Expert Models In a standard deep learning model, every parameter is used for every input. An alternative approach is conditional computation, where only a subset of the parameters are active depending on the input, allowing models with more parameters without the commensurate increase in computational cost. Jacobs et al. (1991) propose to train multiple networks, where each network is dedicated to a disjoint subset of all possible inputs. Since this seminal work on mixture-of-experts, significant follow-up work has been dedicated to exploring different architectures and configurations (Jordan & Jacobs, 1994; Tresp, 2000; Collobert et al., 2001; Rasmussen & Ghahramani, 2001; Aljundi et al., 2017). Shazeer et al. (2017) introduce the Sparsely-Gated Mixture-of-Experts (MoE) layer, the first general purpose conditional computation architecture component. A Mixture-of-Experts layer consists of (1) a set of expert networks and (2) a routing network that determines which experts should be active on a given input. Shazeer et al. (2017) use MoE to scale LSTMs to 137 billion parameters, surpassing the performance of previous dense models on language modeling and machine translation benchmarks. Building on this work, Fedus et al. (2022) introduce the Switch layer, a simplification to the MoE layer which routes to just a single expert and thereby decreases computational cost and increases training stability. Fedus et al. (2022) use Switch layers in place of MLP layers to scale transformers to over a trillion parameters. Recent state of the art language models have used MoE layers, including Mixtral 8x7B (Jiang et al., 2024) and Grok-1 (xAI, 2024). To the best of our knowledge, we are the first to apply conditional computation to training SAEs. 2.2 Deep Learning Training Optimizations Our method fits into the wider literature focused on accelerating deep learning training. One type of training speedup uses hardware accelerators like GPUs (Raina et al., 2009) and TPUs (Jouppi et al., 2017) to optimize highly parallelizable dense matrix operations. Algorithmic improvements, on the other hand, consist of architectural or implementation tricks to speed up forward and backwards passes on fixed hardware. Techniques such as MoE layers (discussed above) and Slide (Chen et al., 2020) utilize sparsity to only evaluate a subset of the parameters for a given wide MLP layer. Other techniques such as Flash Attention (Dao et al., 2022) and Reformers (Kitaev et al., 2020) use hashing or structured matrices to reduce the time complexity of a transformer’s attention mechanism. See Appendix A in Dao et al. (2022) for a comprehensive overview of the literature on algorithmic training optimizations. Our work differs from these papers in that we are concerned with not only whether the training optimization results in a speedup, but also whether SAE quality is preserved. 2.3 Sparse Autoencoders and Improvements Sparse dictionary learning is the task of decomposing input data into a sparse linear combination of basic elements called atoms, which together form a dictionary (Olshausen & Field, 1997; Elad, 2010). There exist a wide variety of techniques for performing dictionary learning, such as the method of optimal directions (Engan et al., 1999) and K-SVD (Aharon et al., 2006). However, such methods are not scalable to large language models and may not be faithful to the models internals. As such, recent work has focused on applying sparse autoencoders (Lee et al., 2007; Le, 2013; Konda et al., 2014), a simple approximation of sparse dictionary learning, to language models. Since the initial works proposing SAEs to separate model representations (Cunningham et al., 2023; Bricken et al., 2023), there have been many proposed improvements to the SAE architecture. These have included alternative activation functions like ProLu (Taggart, 2024), TopK (Makhzani & Frey, 2013; Gao et al., 2024), and Batch-TopK (Bussmann et al., 2024), architectural changes to solve feature suppression caused by the L1 penalty (Rajamanoharan et al., 2024a; b), and changes to the optimization objective itself (Braun et al., 2024). There are many metrics along which these papers evaluate their improvements, but the most common metric is a Pareto frontier of SAE latent sparsity (measured as average L0), mean squared error, and feature interpretability; thus, these are the metrics we focus on in this paper. We also compare primarily against TopK SAEs (Gao et al., 2024) as our baseline, as recent work (Anthropic, 2024c) has shown that these achieve state of the art results on these metrics. 3 The Switch Sparse Autoencoder 3.1 Baseline Sparse Autoencoder Sparse autoencoders are trained to reconstruct language model activations ∈ℝdsuperscriptℝx ^dx ∈ blackboard_Rd by decomposing them into sparse linear combinations of M≫dmuch-greater-thanM dM ≫ d unit-length feature vectors 1,2,…,M∈ℝdsubscript1subscript2…subscriptsuperscriptℝf_1,f_2,…,f_M ^df1 , f2 , … , fitalic_M ∈ blackboard_Rd. SAE architectures consist of an encoder enc∈ℝM×dsubscriptencsuperscriptℝW_enc ^M× dWenc ∈ blackboard_RM × d, a decoder dec∈ℝd×MsubscriptdecsuperscriptℝW_dec ^d× MWdec ∈ blackboard_Rd × M, bias terms (e.g., pre∈ℝdsubscriptpresuperscriptℝb_pre ^dbpre ∈ blackboard_Rd) and an activation function. The TopK SAE (Gao et al., 2024) outputs a reconstruction ^ xover start_ARG x end_ARG of the input activation xx, given by = == TopK(enc(−pre))TopKsubscriptencsubscriptpre (W_enc(x-b_% pre))TopK ( Wenc ( x - bpre ) ) (1) ^ xover start_ARG x end_ARG = == dec+presubscriptdecsubscriptpre _decz+b_preWdec z + bpre (2) The latent vector ∈ℝMsuperscriptℝz ^Mz ∈ blackboard_RM represents how strongly each feature is activated. Since zz is sparse, the decoder forward pass can be optimized by a suitable kernel. The loss is the reconstruction error ℒ=‖−^‖22ℒsuperscriptsubscriptnorm^22L=\|x- x\|_2^2L = ∥ x - over start_ARG x end_ARG ∥22. We additionally benchmark against the ReLU SAE (Anthropic, 2024b) and the Gated SAE (Rajamanoharan et al., 2024a). The ReLU SAE uses the ReLU activation function and applies an L1 penalty to the feature activations to encourage sparsity. The Gated SAE avoids activation shrinkage (Wright & Sharkey, 2024) by separately determining which features should be active and how strongly activated they should be. 3.2 Switch Sparse Autoencoder Architecture The Switch SAE consists of N expert SAEs Eii=1Nsuperscriptsubscriptsubscript1\E_i\_i=1^N Eitalic_i i = 1N as well as a routing network that determines which expert SAE should be used for a given input (Fig. 2). Each expert SAE EisubscriptE_iEitalic_i resembles a TopK SAE with no bias term: Ei()=idecTopK(ienc)subscriptsuperscriptsubscriptdecTopKsuperscriptsubscriptencE_i(x)=W_i^decTopK(W_i % encx)Eitalic_i ( x ) = Witalic_idec TopK ( Witalic_ienc x ) (3) The router, defined by trainable parameters router∈ℝN×dsubscriptroutersuperscriptℝW_router ^N× dWrouter ∈ blackboard_RN × d and router∈ℝdsubscriptroutersuperscriptℝb_router ^dbrouter ∈ blackboard_Rd, computes a probability distribution ()∈ℝNsuperscriptℝp(x) ^Np ( x ) ∈ blackboard_RN over the N experts given by ()=softmax(router(−router))softmaxsubscriptroutersubscriptrouterp(x)=softmax(W_router(x-% b_router))p ( x ) = softmax ( Wrouter ( x - brouter ) ). We route the input xx to the most probable expert i∗()=argmaxpi()superscriptsubscripti^*(x)= i \ p_i(x)i∗ ( x ) = underi start_ARG arg max end_ARG pitalic_i ( x ). The output ^ xover start_ARG x end_ARG is given by, ^=pi∗()Ei∗()(−pre)+pre.^subscriptsuperscriptsubscriptsuperscriptsubscriptpresubscriptpre x=p_i^*(x)E_i^*(x)(x- % b_pre)+b_pre.over start_ARG x end_ARG = pitalic_i∗ ( x ) Eitalic_i∗ ( x ) ( x - bpre ) + bpre . (4) The Switch SAE thus avoids the dense encsubscriptencW_encWenc matrix multiplication by leveraging conditional computation. When comparing Switch SAEs to existing SAE architectures, we evaluate two cases: (1) FLOP-matched Switch SAEs, which perform the same number of FLOPs per activation but increase the number of features by a factor of N, and (2) width-matched Switch SAEs, which reduce the FLOPs per activation by a factor of N while keeping the number of features constant. Figure 2: Switch Sparse Autoencoder Architecture. The router computes a probability distribution over the expert SAEs and routes the input activation vector to the expert with the highest probability. The figure depicts the architecture for d=55d=5d = 5, N=33N=3N = 3, M=1212M=12M = 12. 3.3 Switch Sparse Autoencoder Training We train the Switch SAE end-to-end. When computing ^ xover start_ARG x end_ARG, we weight Ei∗()(−pre)subscriptsuperscriptsubscriptpreE_i^*(x)(x-b_pre)Eitalic_i∗ ( x ) ( x - bpre ) by pi∗()subscriptsuperscriptp_i^*(x)pitalic_i∗ ( x ) to make the router differentiable. We adopt many of the training strategies described in Templeton et al. (2024) and Gao et al. (2024) (see Section A.1 for details).111Our code can be found at https://github.com/amudide/switch_sae The TopK SAE enforces sparsity via its activation function and thus directly optimizes the reconstruction MSE. Following Fedus et al. (2022), we train our Switch SAEs using a weighted combination of the reconstruction MSE and an auxiliary loss which encourages the router to send an equal number of activations to each expert to reduce overhead. For a batch ℬBB with T activations, the auxiliary loss is given by ℒaux=N⋅∑i=1Nfi⋅Pisubscriptℒaux⋅superscriptsubscript1⋅subscriptsubscriptL_aux=N· _i=1^Nf_i· P_iLaux = N ⋅ ∑i = 1N fitalic_i ⋅ Pitalic_i. fisubscriptf_ifitalic_i represents the proportion of activations that are routed to expert i, and PisubscriptP_iPitalic_i represents the proportion of router probability that is assigned to expert i. Formally, fi=1T∑∈ℬi∗()=isubscript1subscriptℬ1superscriptf_i= 1T _x 1\i^*(x% )=i\fitalic_i = divide start_ARG 1 end_ARG start_ARG T end_ARG ∑x ∈ B blackboard_1 i∗ ( x ) = i and Pi=1T∑∈ℬpi()subscript1subscriptℬsubscriptP_i= 1T _x p_i(x)Pitalic_i = divide start_ARG 1 end_ARG start_ARG T end_ARG ∑x ∈ B pitalic_i ( x ). The auxiliary loss is minimized when the batch of activations is routed uniformly across the N experts. The reconstruction loss is defined to be ℒrecon=1T∑∈ℬ‖−^‖22subscriptℒrecon1subscriptℬsuperscriptsubscriptnorm^22L_recon= 1T _x \| % x- x\|_2^2Lrecon = divide start_ARG 1 end_ARG start_ARG T end_ARG ∑x ∈ B ∥ x - over start_ARG x end_ARG ∥22. Note that ℒrecon∝dproportional-tosubscriptℒreconL_recon dLrecon ∝ d. Let α represent a tunable load balancing hyperparameter that scales the auxilliary loss. The total loss ℒtotalsubscriptℒtotalL_totalLtotal is given by ℒtotal=ℒrecon+α⋅d⋅ℒauxsubscriptℒtotalsubscriptℒrecon⋅subscriptℒauxL_total=L_recon+α· d·% L_auxLtotal = Lrecon + α ⋅ d ⋅ Laux. We optimize ℒtotalsubscriptℒtotalL_totalLtotal using Adam (Kingma, 2014). We find that results are not overly sensitive to the value of α, but that α=33α=3α = 3 is a reasonable default based on a hyperparameter sweep (see Section A.2 for details). 4 Results Figure 3: Pareto frontier of sparsity versus (top) reconstruction mean squared error and (bottom) 1 - FLR [fraction of loss recovered]. FLOP-matched Switch SAEs Pareto-dominate TopK SAEs using the same amount of compute (left). Width-matched Switch SAEs perform slightly worse than TopK SAEs but Pareto-dominate ReLU SAEs while performing fewer FLOPs (right). We train sparse autoencoders on the residual stream activations of GPT-2 small, which have a dimension of 768768768768 (Radford et al., 2019). Early layers of language models handle detokenization and feature engineering, while later layers refine representations for next-token prediction (Lad et al., 2024). In this work, we follow Gao et al. (2024) and train SAEs on activations from layer 8, which we expect to hold rich feature representations not yet specialized for next-token prediction. Using text data from OpenWebText (Gokaslan & Cohen, 2019), we train for 100K steps using a batch size of 8192, for a total of approximately 820 million tokens. 4.1 Scaling laws for reconstruction error We first study scaling laws for Switch SAEs, comparing them to dense TopK SAEs at fixed sparsity k=3232k=32k = 32. In Fig. 1, we show scaling curves of reconstruction MSE error with respect to both FLOPs and number of parameters for Switch SAEs with 8 and 64 experts. We find that Switch SAEs have favorable scaling with respect to FLOPs compared to dense TopK SAEs. In fact, Switch SAEs using ∼similar-to ∼1 OOM less compute can often achieve the same reconstruction MSE as TopK SAEs. However, Switch SAEs perform worse at fixed width relative to dense TopK SAEs. Increasing the number of experts improves compute efficiency but reduces parameter efficiency. We hypothesize that this trade-off is a result of features needing to be duplicated across multiple Switch SAE experts, which we discuss in more detail later. Nevertheless, in the right subplot of Fig. 1, we show that the gap between TopK and Switch SAE performance at a fixed width shrinks as we scale the number of parameters. Thus, for large-scale experiments, this gap may be imperceptible; since this is the regime in which efficient training is most useful, we believe that this is not a significant weakness of Switch SAEs. 4.2 Sparsity vs. reconstruction error We now study Switch SAE performance in the reconstruction error vs. sparsity frontier. We benchmark the Switch SAE against the ReLU SAE (Bricken et al., 2023), the Gated SAE (Rajamanoharan et al., 2024a) and the TopK SAE (Gao et al., 2024). We present results for two settings: 1. FLOP-matched: The ReLU, Gated and TopK SAEs are trained with 32⋅768=24576⋅327682457632· 768=2457632 ⋅ 768 = 24576 features. We train Switch SAEs with 2, 4 and 8 experts. Each expert of the Switch SAE with N experts has 24576 features, for a total of 24576⋅N⋅2457624576· N24576 ⋅ N features. The Switch SAE performs roughly the same number of FLOPs per activation compared to the TopK SAE. 2. Width-matched: Each SAE is trained with 32⋅768=24576⋅327682457632· 768=2457632 ⋅ 768 = 24576 features. We train Switch SAEs with 16, 32, 64 and 128 experts. Each expert of the Switch SAE with N experts has 24576N24576 24576Ndivide start_ARG 24576 end_ARG start_ARG N end_ARG features. The Switch SAE performs roughly N times fewer FLOPs per activation compared to the TopK SAE. Note that the router parameters make up a small, insignificant proportion of the total parameters. Across all our experiments, the router parameters make up between 0.002%percent0.0020.002\%0.002 % and 0.3%percent0.30.3\%0.3 % of the total parameters. For a wide range of sparsity values (L0 between 8 and 128), we report the reconstruction MSE and the fraction of cross-entropy loss recovered (FLR) when the sparse autoencoder output is patched into the language model. A FLR value of 1 corresponds to a perfect reconstruction, while a FLR value of 0 corresponds to a zero-ablation (setting the residual stream to the zero vector). 4.2.1 FLOP-Matched Results For the FLOP-matched experiments, we train Switch SAEs with 2, 4 and 8 experts, with the results shown in the left two plots of Fig. 3. The Switch SAEs are a Pareto improvement over the TopK, Gated and ReLU SAEs in terms of both MSE and FLR. As we scale up the number of experts and represent more features, performance continues to increase while keeping computational costs and memory costs (from storing the pre-activations) roughly constant. For a detailed discussion of FLOP calculations, see Section A.3. 4.2.2 Width-Matched Results For the width-matched experiments, we train Switch SAEs with 16, 32, 64 and 128 experts, with the results shown in the right two plots of Fig. 3. The Switch SAEs consistently underperform compared to the TopK SAE in terms of MSE and FLR, while for the most part outperforming Gated and ReLU SAEs. When L0 is low, Switch SAEs perform particularly well. This suggests that the high frequency features that improve reconstruction fidelity the most for a given activation might lie within the same cluster. These results demonstrate that Switch SAEs can reduce the number of FLOPs per activation by up to 128x while still retaining the performance of a ReLU SAE. We further believe that Switch SAEs can likely achieve greater acceleration on larger language models. To show the generality of Switch SAE training, in Appendix B we train on four layers of GPT-2 at residual, attention, and MLP outputs. We also train on a single layer of Gemma 2 2B (Team et al., 2024). We find that Switch SAEs perform well at residual layers on all positions and models tested, but worse on MLP and attention outputs. (a) Fraction of SAE decoder vectors with nearest neighbor cosine sim >0.9absent0.9>0.9> 0.9, width-matched SAEs. (b) Average cosine sim with nearest neighbor for decoder vectors, width-matched SAEs. (c) Average cosine sim between nearest neighbors across experts (16 experts, L0=64L064L0=64L0 = 64). Figure 4: Switch SAE feature geometry experiments, measured via cosine similarity between SAE decoder vectors. We find that Switch SAEs with more experts exhibit more feature duplication, but that this effect diminishes for larger L0s. Furthermore, between-expert similarities show that experts specialize moderately; expert 0, for example, has low similarity with most other experts. 4.3 Feature Geometry 4.3.1 Feature Similarity We are interested in why Switch SAEs achieve a worse reconstruction MSE than TopK SAEs of the same size. One hypothesis is that because on any given forward pass only a single expert is activated, some experts will have to learn duplicate features, reducing SAE capacity (this is necessary since there is no perfect split of features into clusters such that features in different clusters never co-occur). We test this hypothesis using a common (see e.g. Braun et al. (2024)) SAE evaluation metric: nearest neighbor cosine similarity. One way to use this metric to measure the number of duplicate features in an SAE is to measure the proportion of vectors in the decoder that have a nearest neighbor above a given cosine similarity, since highly similar decoder vectors are likely duplicates. In Fig. 4(a), we compare SAEs trained with different numbers of experts and different sparsities on this measure with a threshold of 0.90.90.90.9, and find that as soon as we train with experts this measure jumpy sharply from a baseline of 00 in TopK SAEs to 5555 to 10101010 percent. In other words, strict duplicates likely reduce Switch SAE capacity by up to 10%percent1010\%10 %. This effect is less prevalent at larger L0s. However, we are not just worried about exact duplicates, but also a general shift towards redundancy: in Fig. 4(b), we find that a more global metric, cosine similarity of the nearest neighbor averaged across all features, shows a similar pattern across sparsity and number of experts. Another measure of Switch SAE quality is expert specialization, that is, how similar are the sets of features each expert learns. We quantify by averaging the cosine similarity of each feature in expert A with its nearest neighbor in expert B. On one end, if the features are perfectly clustered, then each pair of experts should have the same cosine similarity as random blocks of the same size in a TopK SAE of the same size. On the other end, no specialization should result in identical experts. In Fig. 4(c), we plot this metric for all pairs of experts in a 16161616 expert Switch SAE with L0=64L064L0=64L0 = 64. Some experts, e.g. expert 00, seem unique, while other pairs of experts, e.g. 10101010 and 7777, are highly similar. All pairs are more similar than random blocks of the same size in a TopK SAE, which has a value of about 0.20.20.20.2 under this same metric. 4.3.2 T-SNE Visualization To visualize the global structure of SAE features, we show t-SNE projections of the encoder and decoder feature vectors in Fig. 5. We find that encoder features from the same expert cluster together, while decoder features tend to be more diffuse. In the encoder feature t-SNE projection, we can also directly observe feature duplication – around the periphery of the plot we find a variety of isolated points which upon closer inspection are actually tight groupings of multiple features from different experts. Figure 5: t-SNE projections of encoder and decoder features for a Switch SAE with 65k total features and 8 experts. Encoder features appear to cluster together by expert. Away from these clusters, we see a variety of isolated points which are in fact tight groups of features which have been duplicated across multiple experts. We do not observe as clear clusters in the decoder features. 4.4 Automated Interpretability Bricken et al. (2023) evaluate how interpretable sparse autoencoder features using automated interpretability: they generate explanations for each feature by giving top activating examples for that feature to a language model, and then ask a language model (in a new context) to predict, using only the description of the feature, on which tokens the described feature fires. More recently, Juang et al. (2024) introduce a more compute efficient method that measures detection, whether the explanation allows a language model to predict whether a feature fires at all on a given context. Juang et al. (2024) additionally measure detection at each decile of feature activation, as well as on contexts where the feature does not fire at all (where to be correct the model should report that the feature does not fire). Using the Juang et al. (2024) implementation, we find that FLOP-matched SAE features have similar interpretability as TopK SAEs, but width-matched SAEs have a higher true positive rate on contexts where the features fire, while at the same time having a lower true negative rate (Fig. 6). We interpret this result as Switch SAEs being biased towards having duplicate frequent features, which may both be easier to detect and more prone to false positives. Figure 6: Automated interpretability detection results across SAE feature activation quantiles for 1000100010001000 random features, 95% confidence intervals shown. “Not” means that the context comes from a random set of text where the feature was not activating. 5 Conclusion Switch SAEs are a promising approach towards scaling sparse autoencoders, as they allow an improvement in the FLOPs vs. MSE scaling law without a significant reduction in feature interpretability. We believe that Switch SAEs may find their best application for huge training runs on large clusters of GPUs, since in this setting each expert can be split on to a separate GPU, leading to significant wall clock training speed ups. Overall, we believe that this work provides a case study for applying existing machine learning training optimization techniques to the setting of sparse autoencoders for feature extraction from language models; we hope that the investigations in this paper serve as a guide for adapting more such techniques. Limitations: The most significant limitation of the Switch SAE is the reduction in performance of the SAE at a fixed number of parameters (especially for attention and MLP layers, see Appendix B); future work to bridge this gap might investigate feature deduplication techniques, more sophisticated routing architectures, and multiple active experts. Additionally, we investigate only the simplest possible mixture of experts architecture; while this allows us to focus on the question of whether and how mixture of experts training works at all for sparse autoencoders, it leaves open the question of the maximum performance gain mixture of experts style training might allow. Thus, future work could examine more sophisticated mixture of experts architectures like GShard (Lepikhin et al., 2020), DeepSeekMoE (Dai et al., 2024), and expert choice routing (Zhou et al., 2022) to further optimize performance. Finally, we do not study downstream evaluations of Switch SAEs in this paper; for instance, it is possible that the increased duplication of features between experts complicates feature steering or circuit discovery. Acknowledgments We used the dictionary learning repository (Marks et al., 2024) to train our SAEs. We would like to thank Samuel Marks and Can Rager for advice on how to use the repository. We would also like to thank Jacob Goldman-Wetzler, Achyuta Rajaram, Michael Pearce, Gitanjali Rao, Satvik Golechha, Kola Ayonrinde, Rupali Bhati, Louis Jaburi, Vedang Lad, Adam Karvonen, Shiva Mudide, Sandy Tanwisuth, JP Rivera and Juan Gil for helpful discussions. This work is supported by Erik Otto, Jaan Tallinn, the Rothberg Family Fund for Cognitive Science, the NSF Graduate Research Fellowship (Grant No. 2141064), IAIFI through NSF grant PHY-2019786, UKRI grant: Turing AI Fellowship EP/W002981/1, and Armasuisse Science+Technology. AM was greatly helped by the MATS program, funded by AI Safety Support. References Achiam et al. (2023) Josh Achiam, Steven Adler, Sandhini Agarwal, Lama Ahmad, Ilge Akkaya, Florencia Leoni Aleman, Diogo Almeida, Janko Altenschmidt, Sam Altman, Shyamal Anadkat, et al. Gpt-4 technical report. arXiv preprint arXiv:2303.08774, 2023. Aharon et al. (2006) Michal Aharon, Michael Elad, and Alfred Bruckstein. K-svd: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Transactions on signal processing, 54(11):4311–4322, 2006. Aljundi et al. (2017) Rahaf Aljundi, Punarjay Chakravarty, and Tinne Tuytelaars. Expert gate: Lifelong learning with a network of experts. In Proceedings of the IEEE conference on computer vision and pattern recognition, p. 3366–3375, 2017. Anthropic (2024a) Anthropic. The claude 3 model family: Opus, sonnet, haiku, 2024a. URL https://w-cdn.anthropic.com/de8ba9b01c9ab7cbabf5c33b80b7bbc618857627/Model_Card_Claude_3.pdf. Anthropic (2024b) Transformer Circuits Team Anthropic. Transformer circuits april 2024 update: Update on how we train saes, 2024b. URL https://transformer-circuits.pub/2024/april-update/index.html#training-saes. Accessed: 2024-09-16. Anthropic (2024c) Transformer Circuits Team Anthropic. Transformer circuits august 2024 update: Evals case study, 2024c. URL https://transformer-circuits.pub/2024/august-update/index.html#evals-case-study. Accessed: 2024-09-16. Braun et al. (2024) Dan Braun, Jordan Taylor, Nicholas Goldowsky-Dill, and Lee Sharkey. Identifying functionally important features with end-to-end sparse dictionary learning. arXiv preprint arXiv:2405.12241, 2024. Bricken et al. (2023) Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, Robert Lasenby, Yifan Wu, Shauna Kravec, Nicholas Schiefer, Tim Maxwell, Nicholas Joseph, Zac Hatfield-Dodds, Alex Tamkin, Karina Nguyen, Brayden McLean, Josiah E Burke, Tristan Hume, Shan Carter, Tom Henighan, and Christopher Olah. Towards monosemanticity: Decomposing language models with dictionary learning. Transformer Circuits Thread, 2023. https://transformer-circuits.pub/2023/monosemantic-features/index.html. Brown et al. (2020) Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agarwal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei. Language Models are Few-Shot Learners. In Advances in Neural Information Processing Systems, volume 33, p. 1877–1901. Curran Associates, Inc., 2020. URL https://proceedings.neurips.c/paper/2020/hash/1457c0d6bfcb4967418bfb8ac142f64a-Abstract.html. Bussmann et al. (2024) Bart Bussmann, Patrick Leask, and Neel Nanda. BatchTopK: A simple improvement for TopK-SAEs. AI Alignment Forum, 2024. URL https://w.alignmentforum.org/posts/Nkx6yWZNbAsfvic98/batchtopk-a-simple-improvement-for-topk-saes. Chen et al. (2020) Beidi Chen, Tharun Medini, James Farwell, Charlie Tai, Anshumali Shrivastava, et al. Slide: In defense of smart algorithms over hardware acceleration for large-scale deep learning systems. Proceedings of Machine Learning and Systems, 2:291–306, 2020. Collobert et al. (2001) Ronan Collobert, Samy Bengio, and Yoshua Bengio. A parallel mixture of svms for very large scale problems. Advances in Neural Information Processing Systems, 14, 2001. Cunningham et al. (2023) Hoagy Cunningham, Aidan Ewart, Logan Riggs, Robert Huben, and Lee Sharkey. Sparse autoencoders find highly interpretable features in language models. arXiv preprint arXiv:2309.08600, 2023. Dai et al. (2024) Damai Dai, Chengqi Deng, Chenggang Zhao, RX Xu, Huazuo Gao, Deli Chen, Jiashi Li, Wangding Zeng, Xingkai Yu, Y Wu, et al. Deepseekmoe: Towards ultimate expert specialization in mixture-of-experts language models. arXiv preprint arXiv:2401.06066, 2024. Dao et al. (2022) Tri Dao, Dan Fu, Stefano Ermon, Atri Rudra, and Christopher Ré. Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, 35:16344–16359, 2022. Elad (2010) Michael Elad. Sparse and redundant representations: from theory to applications in signal and image processing. Springer Science & Business Media, 2010. Elhage et al. (2022) Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, Roger Grosse, Sam McCandlish, Jared Kaplan, Dario Amodei, Martin Wattenberg, and Christopher Olah. Toy models of superposition. Transformer Circuits Thread, 2022. https://transformer-circuits.pub/2022/toy_model/index.html. Engan et al. (1999) Kjersti Engan, Sven Ole Aase, and J Hakon Husoy. Method of optimal directions for frame design. In 1999 IEEE International Conference on Acoustics, Speech, and Signal Processing. Proceedings. ICASSP99 (Cat. No. 99CH36258), volume 5, p. 2443–2446. IEEE, 1999. Engels et al. (2024) Joshua Engels, Isaac Liao, Eric J Michaud, Wes Gurnee, and Max Tegmark. Not all language model features are linear. arXiv preprint arXiv:2405.14860, 2024. Fedus et al. (2022) William Fedus, Barret Zoph, and Noam Shazeer. Switch transformers: Scaling to trillion parameter models with simple and efficient sparsity. Journal of Machine Learning Research, 23(120):1–39, 2022. Gao et al. (2024) Leo Gao, Tom Dupré la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders, 2024. URL https://arxiv.org/abs/2406.04093. Gokaslan & Cohen (2019) Aaron Gokaslan and Vanya Cohen. Openwebtext corpus. http://Skylion007.github.io/OpenWebTextCorpus, 2019. Jacobs et al. (1991) Robert A Jacobs, Michael I Jordan, Steven J Nowlan, and Geoffrey E Hinton. Adaptive mixtures of local experts. Neural computation, 3(1):79–87, 1991. Jiang et al. (2024) Albert Q Jiang, Alexandre Sablayrolles, Antoine Roux, Arthur Mensch, Blanche Savary, Chris Bamford, Devendra Singh Chaplot, Diego de las Casas, Emma Bou Hanna, Florian Bressand, et al. Mixtral of experts. arXiv preprint arXiv:2401.04088, 2024. Jordan & Jacobs (1994) Michael I Jordan and Robert A Jacobs. Hierarchical mixtures of experts and the em algorithm. Neural computation, 6(2):181–214, 1994. Jouppi et al. (2017) Norman P Jouppi, Cliff Young, Nishant Patil, David Patterson, Gaurav Agrawal, Raminder Bajwa, Sarah Bates, Suresh Bhatia, Nan Boden, Al Borchers, et al. In-datacenter performance analysis of a tensor processing unit. In Proceedings of the 44th annual international symposium on computer architecture, p. 1–12, 2017. Juang et al. (2024) Caden Juang, Gonçalo Paulo, Jacob Drori, and Nora Belrose. Open source automated interpretability for sparse autoencoder features. https://blog.eleuther.ai/autointerp/, 2024. Kingma (2014) Diederik P Kingma. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014. Kitaev et al. (2020) Nikita Kitaev, Łukasz Kaiser, and Anselm Levskaya. Reformer: The efficient transformer. arXiv preprint arXiv:2001.04451, 2020. Konda et al. (2014) Kishore Konda, Roland Memisevic, and David Krueger. Zero-bias autoencoders and the benefits of co-adapting features. arXiv preprint arXiv:1402.3337, 2014. Lad et al. (2024) Vedang Lad, Wes Gurnee, and Max Tegmark. The remarkable robustness of llms: Stages of inference? arXiv preprint arXiv:2406.19384, 2024. Le (2013) Quoc V Le. Building high-level features using large scale unsupervised learning. In 2013 IEEE international conference on acoustics, speech and signal processing, p. 8595–8598. IEEE, 2013. Lee et al. (2007) Honglak Lee, Chaitanya Ekanadham, and Andrew Ng. Sparse deep belief net model for visual area v2. Advances in neural information processing systems, 20, 2007. Lepikhin et al. (2020) Dmitry Lepikhin, HyoukJoong Lee, Yuanzhong Xu, Dehao Chen, Orhan Firat, Yanping Huang, Maxim Krikun, Noam Shazeer, and Zhifeng Chen. Gshard: Scaling giant models with conditional computation and automatic sharding. arXiv preprint arXiv:2006.16668, 2020. Makhzani & Frey (2013) Alireza Makhzani and Brendan Frey. K-sparse autoencoders. arXiv preprint arXiv:1312.5663, 2013. Marks et al. (2024) Samuel Marks, Adam Karvonen, and Aaron Mueller. dictionary learning. https://github.com/saprmarks/dictionary_learning, 2024. Olah et al. (2020) Chris Olah, Nick Cammarata, Ludwig Schubert, Gabriel Goh, Michael Petrov, and Shan Carter. Zoom in: An introduction to circuits. Distill, 2020. doi: 10.23915/distill.00024.001. https://distill.pub/2020/circuits/zoom-in. Olshausen & Field (1997) Bruno A Olshausen and David J Field. Sparse coding with an overcomplete basis set: A strategy employed by v1? Vision research, 37(23):3311–3325, 1997. Park et al. (2023) Kiho Park, Yo Joong Choe, and Victor Veitch. The linear representation hypothesis and the geometry of large language models. arXiv preprint arXiv:2311.03658, 2023. Radford et al. (2019) Alec Radford, Jeffrey Wu, Rewon Child, David Luan, Dario Amodei, Ilya Sutskever, et al. Language models are unsupervised multitask learners. OpenAI blog, 1(8):9, 2019. Raina et al. (2009) Rajat Raina, Anand Madhavan, and Andrew Y Ng. Large-scale deep unsupervised learning using graphics processors. In Proceedings of the 26th annual international conference on machine learning, p. 873–880, 2009. Rajamanoharan et al. (2024a) Senthooran Rajamanoharan, Arthur Conmy, Lewis Smith, Tom Lieberum, Vikrant Varma, János Kramár, Rohin Shah, and Neel Nanda. Improving dictionary learning with gated sparse autoencoders. arXiv preprint arXiv:2404.16014, 2024a. Rajamanoharan et al. (2024b) Senthooran Rajamanoharan, Tom Lieberum, Nicolas Sonnerat, Arthur Conmy, Vikrant Varma, János Kramár, and Neel Nanda. Jumping ahead: Improving reconstruction fidelity with jumprelu sparse autoencoders. arXiv preprint arXiv:2407.14435, 2024b. Rasmussen & Ghahramani (2001) Carl Rasmussen and Zoubin Ghahramani. Infinite mixtures of gaussian process experts. Advances in neural information processing systems, 14, 2001. Shazeer et al. (2017) Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc Le, Geoffrey Hinton, and Jeff Dean. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. arXiv preprint arXiv:1701.06538, 2017. Taggart (2024) Glen Taggart. ProLU: A nonlinearity for sparse autoencoders. AI Alignment Forum, 2024. URL https://w.alignmentforum.org/posts/HEpufTdakGTTKgoYF/prolu-a-pareto-improvement-for-sparse-autoencoders. Team et al. (2024) Gemma Team, Morgane Riviere, Shreya Pathak, Pier Giuseppe Sessa, Cassidy Hardin, Surya Bhupatiraju, Léonard Hussenot, Thomas Mesnard, Bobak Shahriari, Alexandre Ramé, et al. Gemma 2: Improving open language models at a practical size. arXiv preprint arXiv:2408.00118, 2024. Templeton et al. (2024) Adly Templeton, Tom Conerly, Jonathan Marcus, Jack Lindsey, Trenton Bricken, Brian Chen, Adam Pearce, Craig Citro, Emmanuel Ameisen, Andy Jones, Hoagy Cunningham, Nicholas L Turner, Callum McDougall, Monte MacDiarmid, C. Daniel Freeman, Theodore R. Sumers, Edward Rees, Joshua Batson, Adam Jermyn, Shan Carter, Chris Olah, and Tom Henighan. Scaling monosemanticity: Extracting interpretable features from claude 3 sonnet. Transformer Circuits Thread, 2024. URL https://transformer-circuits.pub/2024/scaling-monosemanticity/index.html. Tresp (2000) Volker Tresp. Mixtures of gaussian processes. Advances in neural information processing systems, 13, 2000. Wright & Sharkey (2024) Benjamin Wright and Lee Sharkey. Addressing feature suppression in SAEs. AI Alignment Forum, 2024. URL https://w.alignmentforum.org/posts/3JuSjTZyMzaSeTxKk/addressing-feature-suppression-in-saes. xAI (2024) xAI. Grok-1 model card, 2024. URL https://x.ai/blog/grok/model-card. Zhou et al. (2022) Yanqi Zhou, Tao Lei, Hanxiao Liu, Nan Du, Yanping Huang, Vincent Zhao, Andrew M Dai, Quoc V Le, James Laudon, et al. Mixture-of-experts with expert choice routing. Advances in Neural Information Processing Systems, 35:7103–7114, 2022. Appendix A Implementation Details A.1 Switch Sparse Autoencoder Training Details We initialize the rows of encisuperscriptsubscriptencW_enc^iWencitalic_i to be parallel to the columns of decisuperscriptsubscriptdecW_dec^iWdecitalic_i for all i. We initialize both presubscriptpreb_prebpre and routersubscriptrouterb_routerbrouter to the geometric median of a batch of samples, but we do not tie presubscriptpreb_prebpre and routersubscriptrouterb_routerbrouter. We additionally normalize the decoder column vectors to unit-norm at initialization and after each gradient step. We remove gradient information parallel to the decoder feature directions to minimize interference with the Adam optimizer. We set the learning rate based on the 1M1 1 Mdivide start_ARG 1 end_ARG start_ARG square-root start_ARG M end_ARG end_ARG scaling law from Gao et al. (2024) and linearly decay the learning rate over the last 20% of training. We do not include neuron resampling (Bricken et al., 2023) or the AuxK loss (Gao et al., 2024), but future work could explore employing these strategies to prevent dead latents in larger models. We optimize ℒtotalsubscriptℒtotalL_totalLtotal with Adam using the default parameters β1=0.9subscript10.9 _1=0.9β1 = 0.9, β2=0.999subscript20.999 _2=0.999β2 = 0.999. A.2 Hyperparameter Search In our objective function, the balance between reconstruction error and the auxiliary loss (which encourages the router to send an equal number of activations to each expert) is controlled by the hyperparameter α. We train 16-expert and 32-expert Switch SAEs at a fixed sparsity (L0 = 64) and fixed width (24576245762457624576 features), but sweep α across the values [0.001,0.003,0.01,0.03,0.1,0.3,1,3,10,30,100]0.0010.0030.010.030.10.3131030100[0.001,0.003,0.01,0.03,0.1,0.3,1,3,10,30,100][ 0.001 , 0.003 , 0.01 , 0.03 , 0.1 , 0.3 , 1 , 3 , 10 , 30 , 100 ]. We find that the reconstruction MSE is not overly sensitive to the value of α and that α=33α=3α = 3 performs the best in both the 16-expert and the 32-expert settings (Fig. 7). We default to α=33α=3α = 3 for the rest of our experiments. Figure 7: Hyperparameter sweep for α. We train fixed-width Switch SAEs with 16 and 32 experts on activations from GPT-2 small. The sparsity level is set to L0=64. A.3 FLOP Calculations A single training step of a deep neural network requires roughly 6⋅P⋅D⋅66· P· D6 ⋅ P ⋅ D FLOPs, where P is the number of active parameters and D is the batch size. Let d be the dimension of the language model activations, M be the total number of features and N be the number of experts. For a TopK SAE, P=2Md2P=2MdP = 2 M d. For a Switch SAE, P=dN+2MdN≈2MdN22P=dN+ 2MdN≈ 2MdNP = d N + divide start_ARG 2 M d end_ARG start_ARG N end_ARG ≈ divide start_ARG 2 M d end_ARG start_ARG N end_ARG since the router parameters are negligible. Thus, we have FLOPSTopK(d,M)subscriptFLOPSTopK _TopK(d,M)FLOPSTopK ( d , M ) =6⋅(2Md)⋅D,absent⋅62 =6·(2Md)· D,= 6 ⋅ ( 2 M d ) ⋅ D , FLOPSSwitch(d,M,N)subscriptFLOPSSwitch _Switch(d,M,N)FLOPSSwitch ( d , M , N ) =6⋅(dN+2MdN)⋅D≈6⋅(2MdN)⋅D.absent⋅62⋅62 =6· (dN+ 2MdN )· D≈ 6· (% 2MdN )· D.= 6 ⋅ ( d N + divide start_ARG 2 M d end_ARG start_ARG N end_ARG ) ⋅ D ≈ 6 ⋅ ( divide start_ARG 2 M d end_ARG start_ARG N end_ARG ) ⋅ D . Appendix B Training on Additional Sites and Models In Table 1, we show the results of training on GPT-2 layers 2,4,8,2482,4,8,2 , 4 , 8 , and 12121212 on MLP outputs, attention outputs, and the residual stream. We also train a single SAE on Gemma 2 2B (Team et al., 2024). All SAEs are trained with k=6464k=64k = 64 with a fixed width; we train the Gemma 2 2B SAEs with width 65536655366553665536 and the GPT-2 SAEs with width 24576245762457624576. We use 8888 experts for the Switch SAEs. As in the main paper, we find that the decrease in SAE performance for a fixed width minimally effects residual stream training; however, Switch SAEs trained on MLP and attention outputs do suffer a significant performance reduction. Table 1: Comparison of TopK and Switch SAEs across different models, layers and component types. FVE = Fraction of Variance Explained, FLR = Fraction of Loss Recovered. Model Layer Type TopK FVE Switch FVE TopK FLR Switch FLR GPT-2 2 resid 0.999 0.998 0.997 0.996 GPT-2 2 attn 0.919 0.884 0.997 0.968 GPT-2 2 mlp 0.999 0.998 0.888 0.701 GPT-2 4 resid 0.997 0.997 0.996 0.995 GPT-2 4 attn 0.849 0.805 0.925 0.883 GPT-2 4 mlp 0.877 0.819 0.853 0.777 GPT-2 8 resid 0.989 0.987 0.994 0.993 GPT-2 8 attn 0.846 0.803 0.938 0.848 GPT-2 8 mlp 0.782 0.733 0.868 0.801 GPT-2 10 resid 0.973 0.970 0.982 0.977 GPT-2 10 attn 0.892 0.863 0.942 0.883 GPT-2 10 mlp 0.855 0.828 0.859 0.812 Gemma 2 2B 12 resid 0.961 0.955 0.991 0.989