← Back to papers

Paper deep dive

Kronecker Factorization Improves Efficiency and Interpretability of Sparse Autoencoders

Vadim Kurochkin, Yaroslav Aksenov, Daniil Laptev, Daniil Gavrilov, Nikita Balagansky

Year: 2025Venue: arXiv preprintArea: Mechanistic Interp.Type: EmpiricalEmbeddings: 89

Models: Gemma-2-2B, Pythia-1.4B, Qwen-2.5-1.5B

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 97%

Last extracted: 3/11/2026, 1:05:31 AM

Summary

KronSAE is a novel sparse autoencoder architecture that utilizes Kronecker product decomposition to factorize latent representations, significantly reducing computational and memory overhead compared to standard SAEs. It introduces the mAND activation function, a differentiable approximation of the binary AND operation, which improves feature interpretability and reduces feature absorption by enforcing compositional structure in the latent space.

Entities (5)

KronSAE · architecture · 100%Sparse Autoencoder · model-class · 100%mAND · activation-function · 100%Kronecker product decomposition · mathematical-technique · 98%Feature absorption · problem · 95%

Relation Signals (3)

KronSAE incorporates mAND

confidence 100% · KronSAE, a Kronecker-factorised sparse autoencoder equipped with the novel mAND activation function

KronSAE uses Kronecker product decomposition

confidence 100% · we propose KronSAE, a novel architecture that factorizes the latent representation via Kronecker product decomposition

mAND reduces Feature absorption

confidence 90% · KronSAE variants consistently reduce first two scores [absorption metrics] relative to the TopK SAE baseline

Cypher Suggestions (2)

Identify relations between architectures and the problems they address · confidence 95% · unvalidated

MATCH (a:Architecture)-[r:REDUCES]->(p:Problem) RETURN a.name, r.relation, p.name

Find all architectures that use Kronecker product decomposition · confidence 90% · unvalidated

MATCH (a:Architecture)-[:USES]->(t:Technique {name: 'Kronecker product decomposition'}) RETURN a.name

Abstract

Abstract:Sparse Autoencoders (SAEs) have demonstrated significant promise in interpreting the hidden states of language models by decomposing them into interpretable latent directions. However, training and interpreting SAEs at scale remains challenging, especially when large dictionary sizes are used. While decoders can leverage sparse-aware kernels for efficiency, encoders still require computationally intensive linear operations with large output dimensions. To address this, we propose KronSAE, a novel architecture that factorizes the latent representation via Kronecker product decomposition, drastically reducing memory and computational overhead. Furthermore, we introduce mAND, a differentiable activation function approximating the binary AND operation, which improves interpretability and performance in our factorized framework.

Tags

ai-safety (imported, 100%)empirical (suggested, 88%)interpretability (suggested, 80%)mechanistic-interp (suggested, 92%)

Links

PDF not stored locally. Use the link above to view on the source site.

Full Text

88,638 characters extracted from source content.

Expand or collapse full text

Preprint. KRONECKER FACTORIZATION IMPROVES EFFICIENCY AND INTERPRETABILITY OF SPARSE AUTOENCODERS Vadim Kurochkin 1 Yaroslav Aksenov 1∗ Daniil Laptev 1 Daniil Gavrilov 1 Nikita Balagansky 1 1 T-Tech ABSTRACT Sparse Autoencoders (SAEs) have demonstrated significant promise in interpreting the hidden states of language models by decomposing them into interpretable latent directions. However, training and interpreting SAEs at scale remains challenging, especially when large dictionary sizes are used. While decoders can leverage sparse-aware kernels for efficiency, encoders still require computationally intensive linear operations with large output dimensions. To address this, we propose Kron- SAE – a novel architecture that factorizes the latent representation via Kronecker product decomposition, drastically reducing memory and computational overhead. Furthermore, we introduce mAND, a differentiable activation function approximat- ing the binary AND operation, which improves interpretability and performance in our factorized framework. 1INTRODUCTION Interpreting large language models and their embeddings in particular remains a central challenge for transparency and controllability in AI systems (Elhage et al., 2023; Heap et al., 2025). Sparse autoencoders (SAEs) have emerged as powerful tools for uncovering human-interpretable features within neural activations by enforcing activation sparsity to induce discrete-style dictionaries (Elhage et al., 2023; Gao et al., 2025; Cunningham et al., 2024). These dictionaries facilitate circuit-level semantic analysis (Marks et al., 2025) and concept discovery, enabling fine-grained probing of model internals (Elhage et al., 2023; Cunningham et al., 2024). However, naively scaling SAEs to the widths demanded by modern transformers leads to prohibitive compute costs, limiting their applicability to large-scale interpretation experiments. Gated SAEs (Rajamanoharan et al., 2024a) address this by learning continuous sparsity masks via lightweight gating networks, and Switch SAEs (Mudide et al., 2025) leverage conditional computation by routing activations among smaller expert SAEs, reducing computation by activating only a subset of experts per input. Gao et al. (2025) propose an acceleration of TopK SAE that utilizes an optimized kernel based on efficient sparse–dense matrix multiplication. Encoder remains unoptimized: it still performs a dense projection into the full dictionary, incurring high computational cost and limiting scalability. Another limitation of SAEs is the absence of structure within learned latents in its classical design. While Mudide et al. (2025) address this via expert subnetworks, Bussmann et al. (2025) imposes feature hierarchy via nested dictionaries structure and improves the interpretability of SAE latents. In this paper we address both these directions and introduce KronSAE, an encoder that is applicable to many existing SAE architectures. By decomposing the latent space into head-wise Kronecker factors and including differentiable logical AND-like gating mechanism, KronSAE reduces both parameters and compute overhead while preserving reconstruction fidelity and improves interpretability. ∗ Corresponding author: y.o.aksenov@tbank.ru 1 arXiv:2505.22255v3 [cs.LG] 22 Dec 2025 Preprint. This work makes three primary contributions: 1. We identify the encoder projection as one of the principal scalability bottlenecks in sparse autoencoders, demonstrating that targeted encoder optimizations can significantly improve computational performance while maintaining reconstruction quality. 2.We propose KronSAE, a Kronecker-factorised sparse autoencoder equipped with the novel mANDactivation function. The design reduces encoder cost and is compatible with existing sparse decoder kernels, loss function designs and mixture of experts architecture. 3. We show on multiple language models that KronSAE decreases feature absorption and yields more interpretable latents under fixed compute. 2RELATED WORK Sparse Autoencoders. Early work demonstrated that SAEs can uncover human-interpretable directions in deep models (Cunningham et al., 2024; Templeton et al., 2024), but extending them to the large latent sizes (F ≫ d) for modern language models with large hidden dimensiond is expensive: each forward pass still requires a denseO(F d)encoder projection. Most prior optimization efforts focus on the decoder side: for example, Gao et al. (2025) introduce a fused sparse–dense TOPK kernel that reduces wall-clock time and memory traffic, while Rajamanoharan et al. (2024a) decouple activation selection from magnitude prediction to improve theℓ 0 –MSE trade-off. Separately, Rajamanoharan et al. (2024b) propose JUMPRELU, a nonlinearity designed to mitigate shrinkage of large activations in sparse decoders. Conditional Computation. These schemes avoid instantiating a full dictionary per token by routing inputs to a subset of expert SAEs via a lightweight gating network (Mudide et al., 2025) by employing the Mixture-of-Experts ideas (Shazeer et al., 2017), but still incur a dense per-expert encoder projection, leaving the encoder as the primary bottleneck. Factorizations and Logical Activation Functions. Tensor product representations (Smolensky, 1990) have been utilized to represent the compositional structures in dense embeddings, closely resembling our idea of AND-like compositions, and recent studies have been extended TPR to study the transformer hidden states and in-context learning (Soulos et al., 2020; Smolensky et al., 2024). Separately, tensor-factorization methods have been used to compress large weight matrices in language models (Edalati et al., 2021; Wang et al., 2023), and structured matrices have been utilized to improve training efficiency and might be used to impose the inductive bias (Dao et al., 2022). In parallel, differentiable logic activations were introduced to approximate Boolean operators in a smooth manner (Lowe et al., 2021). Our method synthesizes these lines of work: we embed a differentiable AND-like gate into a Kronecker-factorized efficient encoder to build compositional features while preserving end-to-end differentiability. 3METHOD Preliminaries. Letx ∈R d denote an activation vector drawn from a pretrained transformer. A conventional TopK SAE (Gao et al., 2025) produces a reconstruction ˆ x of x via f = TopK W enc x + b enc , ˆ x = W dec f + b dec ,(1) whereW enc ∈R F×d andW dec ∈R d×F are dense matrices andf ∈R F is a sparse vector retaining only the K largest activations. The encoder cost therefore scales asO(Fd) per token. KronSAE. Our method reduces the encoder’s computational cost while also enforcing composi- tional structure of the latents. We decompose the latent space intohindependent heads, and each headkis parameterised by the composition of two thin matricesP k ∈R m×d (composition base) and Q k ∈R n×d (composition extension), with dimensionsm≤ n≪ dandF = hmn. The pre-latents p k = ReLU(u k )and q k = ReLU(v k ),(2) 2 Preprint. withu k = P k xandv k = Q k xacting as the elements from which compositional features would be built, are combined through an element-wise interaction kernel independently in each head: z k i,j := mAND(u k i ,v k j ) := ( q u k i v k j , u k i > 0 and v k j > 0, 0,otherwise, (3) wherez k ∈R m×n ; it is then flattened in a row-major order to a vector equivalent to the element-wise square root of the Kronecker product of p k and q k . Assuming that vec(·) is in row-order, we get vec(z k ) = q vec (p k (q k ) ⊤ ) = p p k ⊗ q k (4) Concatenating heads and applyingTopKyields post-latentsf ∈R F . By ensuring activez i,j only whenp i > 0andq j > 0we directly force the AND-like interaction: letPandQbe the sets of hidden state vectors on whichp i andq j , respectively, are active, and letFbe the set of inputs on whichz i,j > 0, then it is true thatP ∩Q = F. The square root in equation 3 prevents activation value explosion when both pre-latents are active. See additional discussion in Appendix C. The encoder cost per token drops fromO(Fd)toO h(m +n)d (see Appendix A.2). KronSAE thus reduces FLOPs and parameter count without routing overhead, and is orthogonal to existing sparse decoder kernels (Gao et al., 2025) and thus can be combined with them for end-to-end speed-ups. 4EXPERIMENTS We train SAEs on the residual streams of Qwen-2.5-1.5B-Base (Yang et al., 2024), Pythia-1.4B- deduped (Biderman et al., 2023), and Gemma-2-2B (Team et al., 2024) language models. Activations are collected on FineWeb-Edu, a filtered subset of educational web pages from the FineWeb cor- pus (Penedo et al., 2024). We measure reconstruction quality via explained variance (EV), EV = 1− Var(x− ˆ x) Var(x) , so that1.0is optimal, and use automated interpretability pipeline (Bills et al., 2023; Paulo et al., 2024) and SAE Bench (Karvonen et al., 2025) to evaluate properties of SAE features. We aim for the needs of resource-constrained mechanistic interpretability research where efficiency and interpretability are in favor rather than top reconstruction performance, and 100M-820M token budgets are widely adopted (Bussmann et al., 2025; Kharlapenko et al., 2025; Heap et al., 2025; Mudide et al., 2025; Karvonen et al., 2025), so we choose 125M, 500M, 1B and 2B token budgets for the experiments. Our experiments (see detailed setup in Appendices A and D) address three questions: 1. Does KronSAE maintain EV comparable to baseline SAEs under fixed compute? 2. Which design choices (nonlinearity, (m,n,h)) drive EV improvements? 3. How do these choices affect properties and interpretability of learned latents? 4.1ABLATIONS We employ the iso-FLOPs setup: for each KronSAE variant of dictionary sizeFwe allocate the same amount of FLOPs as was spent for the training of TopK SAE for token budget T and same F . Reconstruction performance. As indicated on Figure 1, KronSAE achieves on-par performance with TopK given lower number of trainable parameters and outperforms Matryoshka SAE. The performance gap narrows when increasing the dictionary size, which indicate the potential scalability of our method for large dictionaries. See also result on Gemma-2 2B model in Appendix A.3. Decomposition hyperparameters.We systematically vary the number of headshand the per-head base dimensionm(withn = F/(mh)) under the iso-FLOPs setup. From the Figure 2, we conclude that lowermand higherhconsistently yields higher reconstruction quality, due to flexibility of pre-latents - as we show in section 5.3, they must be either expressive or fine-grained enough (lowm, n and high h) to efficiently represent the semantics. See also Appendices A.5 and A.8. 3 Preprint. 500m1000m2000m 0.76 0.78 0.80 0.82 0.84 0.86 0.88 0.90 0.92 Explained Variance Dictionary Size: 32768 m=1 m=2 m=4 m=8 Matryoshka-1024 Matryoshka-2048 TopK 500m1000m2000m Reference Tokens Budget Dictionary Size: 65536 m=1 m=2 m=4 m=8 Matryoshka-2048 Matryoshka-4096 TopK 500m1000m2000m Dictionary Size: 131072 m=1 m=2 m=4 m=8 Matryoshka-4096 Matryoshka-8192 TopK Figure 1: Maximum performance for KronSAE vs. TopK SAE vs. Matryoshka TopK SAE on Qwen- 1.5B for different dictionary sizesFand budgets in iso-FLOP setting. KronSAE with lower number of parameters is on-par with the baseline, and the gap narrows with larger dictionary size. 409620481024512256128 Heads h 0.81 0.82 0.83 0.84 0.85 0.86 Explained Variance Token Budget: 500m F = 2 15 F = 2 16 F = 2 17 m = 1 m = 2 m = 4 m = 8 409610242048512256128 Heads h Token Budget: 1000m Figure 2: Dependency of EV on head counth(on the x-axis) and base dimensionmunder 500M and 1B token budgets in iso-FLOPs setup. Higherhand smallermyield improved reconstruction quality because of higher expressivity of pre-latents to encode semantics and increasing trainable parameters. Composition activation. To isolate the impact of ourmANDoperator, we compare it to two simpler interaction kernels: (i) the element-wise product of ReLUs,ReLU(u)· ReLU(v), and (i) the raw productu· v. As reported in Table 1, under a 125M training budget, themANDvariant achieves the highest explained variance. More description of themANDis provided in the Appendix C. See also our experiments where we replace TopK with JumpReLU in Appendix A.6. Dictionary sizemnhActivationExplained Variance 32768 244096 mAND(u,v)0.8336 ReLU(u)· ReLU(v)0.8267 u· v0.8237 481024 mAND(u,v)0.8220 ReLU(u)· ReLU(v)0.8191 u· v0.8143 65536 248192 mAND(u,v)0.8445 ReLU(u)· ReLU(v)0.8328 u· v0.8297 482048 mAND(u,v)0.8350 ReLU(u)· ReLU(v)0.8297 u· v0.8251 Table 1: Performance of different composition activations under a budget of 125M tokens. 4 Preprint. Sparsity Analysis. To evaluate performance across different sparsity budgets (ℓ 0 = 16, 32, 64, 128 ), we compare multiple SAE baselines (TopK SAE, Matryoshka SAE, Switch SAE) against their KronSAE variants at dictionary sizeF = 2 16 trained on 500 million tokens, under an iso-FLOPs budget matched to the TopK baseline. As shown in Figure 3, KronSAE achieves comparable reconstruction fidelity (measured by explained variance) to all baseline methods across all sparsity levels.Results for smaller dictionary size F = 2 15 are presented in Appendix A.4. 0.60 0.65 0.70 0.75 0.80 0.85 0.90 TopK / KronSAEMatryoshka / KronSAE + MatryoshkaSwitch / KronSAE + Switch m=1 m=2 m=4 Baseline 163264128 0 0.60 0.65 0.70 0.75 0.80 0.85 0.90 163264128 0 163264128 0 m=1 m=2 m=4 Baseline Qwen-2.5 1.5B Gemma-2 2B Explained Variance Figure 3: Maximum performance for baselines and their KronSAE modifications for different sparsity levels in iso-FLOP setting. KronSAE variants, despite using fewer trainable parameters, achieve reconstruction quality comparable to or better than the unmodified baselines. Layerwise performance. Additionally, we evaluate performance across different layers in Qwen- 2.5-1.5B. In every case, KronSAE matches the reconstruction quality of the TopK baseline, demon- strating that our Kronecker-factorized encoder maintains its performance regardless of depth. This setup and corresponding results are described in greater detail in Appendix A.7. 4.2ABSORPTION The notorious challenge in SAE interpretability is feature absorption, where one learned feature becomes a strict subset of another and consequently fails to activate on instances that satisfy the broader concept but not its superset representation (e.g. a “starts with L” feature is entirely subsumed by a “Lion” feature) (Chanin et al., 2024). Figure 4 reports three absorption metrics measured via SAEBench (Karvonen et al., 2025) across sparsity levelsℓ 0 ∈16, 32, 64, 128: (1) the mean absorption fraction, measuring the proportion of features that are partially absorbed; (2) the mean full-absorption score, quantifying complete subsumption events; and (3) the mean number of feature splits, indicating how often a single conceptual feature fragments into multiple activations. We use with dictionary sizeF = 2 16 and compare TopK SAE, Matryoshka SAE (Bussmann et al., 2025), and their KronSAE version, since they impose different hierarchical priors (see Appendix C for a discussion of how Matryshka SAE and TopK SAE structure differs). Across allℓ 0 , KronSAE variants consistently reduce first two scores relative to the TopK SAE baseline, while maintaining a similar rate of feature splits. We attribute KronSAE’s improved disentanglement to two complementary design choices: 1. AND-like behaviour. By ensuring that post-latent emits only when its more general pre- latent parents are active, we prevent more specific post-latents from entirely subsuming broadly polysemantic one. See additional description of this mechanism in Appendix C. 2.Head-wise Cartesian decomposition. Dividing the latent space intohindependent sub- spaces (each with its ownm× ngrid of primitive interactions) ensures that specialized 5 Preprint. 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 Mean Absorption Fraction Score Model: Gemma-2 2BModel: Qwen-2.5 1.5B SAE Type KronSAE KronSAE + Matryoshka-2048 KronSAE + Matryoshka-4096 TopK TopK + Matryoshka-2048 TopK + Matryoshka-4096 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Mean Full Absorption Score 163264128 0 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 Mean Number of Feature Splits 163264128 0 Figure 4: Feature absorption metrics on Qwen-2.5 1.5B and Gemma-2 2B. KronSAE configurations (variousm,n) exhibit lower mean absorption fractions and full-absorption scores across differentℓ 0 and selected baselines. concepts (such as “elephant”) are confined to a single head and cannot fully absorb more general concepts (such as “starts with E”) in another. Together, these mechanisms produce more monosemantic features, as we verify in the section 5.3, simplifying downstream applications. See results with dictionary sizeF = 2 15 in Appendix A.4. We also validate the result on Pythia models and observe the same picture, see the Appendix A.5. 5ANALYSIS In this section we examine the properties of the latents in KronSAE compared to TopK architecture. The design of our arhictecture was also inspired by the observation that many features within a single SAE correlate with each other. KronSAE introduces structural bias that forces the post-latents to co-occur with their pre-latents (see also Appendix C). In this section we analyse if it is helpful. By examining the toy examples with manufactured correlations in data, we show that KronSAE captures these correlations better than TopK. Then we show that KronSAE trained on language indeed moves correlated features within a single head, indicated by higher correlation within head. After that, we show that KronSAE pre-latents interactions are closely resemble the logic AND gate, and its post-latents are notably more interpretable than TopK latents. 5.1TOY MODEL OF CORRELATION To evaluate how well different sparse autoencoder architectures recover underlying correlation patterns, we construct a controlled experiment using a synthetic, block-structured covariance model. Input vectorsx ∈R F sampled from a normal distribution (withμ = 0,σ = 1). 6 Preprint. We then perform a Cholesky decompositionS = L ⊤ on the covariance matrixSand set ̄ x sparse = L TopK(ReLU(x)), so that ̄ x sparse exhibits the desired structure. We train autoencoder (AE) to reconstruct ̄ x sparse following the (Elhage et al., 2022): ˆx = ReLU(W ⊤ W · ̄ x sparse + b).(5) We collect hidden states of dimensiond = 64(W · ̄ x sparse ) from AE and then train TopK SAE and our proposed KronSAE withF = 256andtopk = 8to reconstruct it. After training, we extract the decoder weight matricesW dec from each SAE, match its latents with the autoencoder latents by solving the quadratic assignment problem (Vogelstein et al., 2015) (see Appendix B for motivation and details), and compute the covariance C dec = W dec W ⊤ dec . Result is shown in Figure 5. Ground TruthAutoencoder (0.598)TopK (0.157)KronSAE (0.514) Figure 5: We generate data with covariance matrix that consist of blocks with different sizes on diagonal and off diagonal (left panel). We then examine the decoder-weight covarianceW dec · W ⊤ dec to assess feature-embedding correlations and compute the RV score to quantify the similarity between learned and ground truth covariance matrices. Second panel show feature embeddings for trained autoencoderW enc · W T enc . Third panel demonstrates that a TopK SAE recovers these correlation structures weakly, as indicated by relatively low RV coefficient(0.157)even after optimal atom matching. In contrast, KronSAE (right panel) more accurately reveals the original block patterns. To quantify how closely each model’s feature correlations mirror the original structure ofS, we employ the RV coefficient, defined asRV (S,C) = trace(SC)/ p trace(S 2 ) trace(C 2 ). In our experiments, KronSAE consistently achieves notably higher RV than TopK SAE, indicating that our compositional encoder more faithfully reconstructs the original feature relation. See also additional experiments in section B, where we provide further empirical intuition for how KronSAE identifies correlation structure that more closely align with those present in the data. 5.2CORRELATIONS IN SAES TRAINED ON LANGUAGE To examine the correlation structure of features learned in our SAE, we have calculated the correlations on 5k texts from the training dataset. For each feature we calculate the mean correlation with features within its head and with all other features, and compare the randomly initialized KronSAE with m = 4,n = 4with the trained one. To isolate the effect of our initialization procedure, we initialize the weights of SAE from the uniform distribution. As shown in Figure 6, correlations are indeed significantly higher within a single head and higher than for random SAE, which suggest that our choice to impose the correlated structure in SAE latents works as intended. 5.3ANALYSIS OF LEARNED FEATURES In this section we compare KronSAE and TopK SAE in terms of interpretability and feature properties, and we analyze the properties of groups in KronSAE. For this, we choose the 14th layer of Qwen2.5- 1.5B and a dictionary size of 32k features, of which the first 3072 were selected. KronSAE was chosen withm = 4,n = 4. We run for 24M tokens total to collect data. Our interpretation pipeline follows the common methodology: LLM interprets the activation patterns (Bills et al., 2023) and we evaluate obtained interpretations using the detection score and the fuzzing score Paulo et al. (2024). For each selected feature, among the standard mean activation value and frequency, we calculate two additional metrics. Low values of token entropy suggest that feature activates more frequently on small number of tokens, thus it is token-specific; high value of multitoken ratio indicates that feature 7 Preprint. 0.20.00.20.40.60.81.0 Pearson correlation 10 3 10 2 10 1 10 0 10 1 Density (log scale) Within-Group Between-Group Within-Group (rand) Between-Group (rand) Figure 6: Correlations between features in KronSAE withm = 4,n = 4within a head and with features from other heads. Our design induces higher correlations within a group, which also gets stronger after training, although SAE have also learned correlated features from different heads. 0.20.40.60.81.0 0 1 2 3 Detection score TopK ronSAE Pre KronSAE Post Median 0.00.20.40.60.81.0 0 1 2 3 Fuzzing score 10 2 10 1 10 0 10 1 0.0 0.5 1.0 1.5 2.0 Mean activation 02468 0.0 0.1 0.2 0.3 Token entropy 0.00.20.40.60.81.0 0 1 2 Multitoken ratio 10 7 10 6 10 5 10 4 10 3 10 2 10 1 0.0 0.2 0.4 0.6 Frequency Figure 7: Distribution of properties for TopK SAE and KronSAE (m = 4,n = 4) with 32k dictionary size trained on Qwen2.5-1.5B. Pre and Post suffixes denote pre- and post- latents, and y-axis indicate density. Token entropy shows the entropy of the distribution of tokens on which feature has activated, and the multitoken ratio measures how often does feature activate in a single sequence. Our SAE achieves better interpretability scores by learning specialized feature groups, indicated by lower activation frequency and lower variance in activated tokens. tends to activate multiple times in a single sentence. We have observed that both these metrics have notable negative correlation with the final interpretability scores and therefore they provide useful signal to assess the potential score without calculating it. For more details on the data collection and interpretation pipeline, see Appendix D. For additional analysis of properties of learned features , additional comparison with baselines and discussion about tradeoff between reconstruction performance and interpretability, see Appendix E. SAE properties and encoding mechanism.We observe that the features learned by KronSAE are more specific, indicated by lower values of the computed metrics and higher interpretability scores, as shown in Figure 7. Since post-latents are significantly more interpretable than corresponding pre- latents, we hypothesize the hidden mechanism for encoding and retrieval of the required semantics. By examining activating examples and interpretations of latents, we observe that pre-latents may carry multiple distinct and identifiable modes of activation, such as composition base element 3 in head 23 shown in Table 2, and be very abstract compared to resulting post-latents. Polysemanticity of pre-latents is expected to be a consequence of reduced "working" number of encoder latents, since we decompose the full dictionary size and reduce the encoder capacity. Thus, we hypothesize that the encoding of specific semantics in our SAE may be done via magnitude, which we validate by examining the activation examples. For the above mentioned pre-latent, the 8 Preprint. "comparison" part is encoded in the top 75% quantile, while the "spiritual" part is mostly met in the top 25% quantile, and the "geographical" part is mainly encoded in the interquartile range. We also consider but do not investigate the possibility that it may depend on the context, e.g. when the model uses the same linear direction to encode different concepts when different texts are passed to it. Semantic retrieval and interpretable interactions.Heads usually contain a groups of semantically related pre-latents, e.g. in head 136 there are three base elements and one extension covering numbers and ordinality, two extension elements related to geographical and spatial matters, one question- related base and one growth-related extension. Interestingly, most post-latents for this head have higher interpretability score than both its parent pre-latents, which is unusual. The retrieval happens primarily via the mechanism closely resembling the logical AND circuit, where some pre-latent works as the bearer of multiple semantics, and the corresponding pre-latent (base or extension) works as specifier. An illustrative example is shown in Table 2: we see that the base contains three detectable sub-semantics, and each extension then retrieves the particular semantics. ComponentInterpretationScore Base 3Suffix “-like” for comparative descriptors, directional terms indicating geo- graphical regions, and concepts related to spiritual or metaphysical dimen- sions 0.84 Extension elements and their compositions with base 3 Extension 0Interpretation: Comparative expressions involving “than” and “as” in con- trastive or proportional relationships. 0.87 Composition: Similarity or analogy through the suffix “-like” across diverse contexts. 0.89 Extension 1Interpretation: Specific terms, names, and abbreviations that are contextually salient and uniquely identifiable. 0.66 Composition: Medical terminology related to steroids, hormones, and their derivatives. 0.84 Extension 2Interpretation: Spiritual concepts and the conjunction “as” in varied syntactic roles. 0.80 Composition: Abstract concepts tied to spirituality, consciousness, and metaphysical essence. 0.93 Extension 3Interpretation: Directional and regional descriptors indicating geographical locations or cultural contexts. 0.84 Composition: Directional terms indicating geographical or regional divisions.0.91 Table 2: Interactions between composition base element 3 in head 23 and all extension elements in that head. Interaction happens in a way that closely resembles the Boolean AND operation: base pre-latent is polysemous, and the composition post-latent is the intersection, i.e. logical AND between parent pre-latents. See details in Section 5.3. Other types of interaction may occur, such as appearance of completely new semantics, for example composition between base 3 and extension 1 in Table 2 where medical terminology arises and could not be interpreted as simple intersection between two pre-latents semantics. Another example is a case of head 3 where base 3 has sub-semantics related to technical instruments and extension 2 have semantics related to the posession and necessity, and their combination gives the therapy and treatment semantics which looks more like addition than intersection. It is a frequent case that post-latent inherit semantics of only one parent, or the impact of another parent is not detectable, which usually happens if parent has a very broad interpretation and low score. However, it requires more sophisticated techniques to properly identify the fine-grained structure of interactions than just looking at the resulting latent descriptions, so we leave it to further work. Despite this, the AND-like gate is a very common behavior. See more examples in Appendix G. 9 Preprint. Geometry of post-latents.Each post-latent vector has a vector representation in the residual stream represented by the corresponding column inW dec , which is the approximation of overcomplete basis vectors we search for when training SAEs. Our architectural design leads to clustering of feature embeddings so that post-latents produced by same head, base or a extension elements are grouped in a tight cluster, and the geometry is dependent on hyperparametersh,m,nwe choose, which is expected and may be useful for further applications such as steering. See more details in Appendix E. 6CONCLUSION AND FUTURE WORK We introduce KronSAE, a sparse autoencoder architecture design that combines head-wise Kronecker factorization of latent space with a approximation of logical AND viamANDnonlinearity. Our approach allows to efficiently train interpretable and compositional SAE, especially in settings with limited compute budget or training data, while maintaining reconstruction fidelity and yielding more interpretable features by utilizing their correlations. Our analysis links these gains to the complementary effects of compositional latent structure and logical AND-style interactions, offering a new lens on how sparsity and factorization can synergise in representation learning. Limitations. KronSAE introduces tradeoff between interpretability, efficiency and reconstruction performance, and due to reduced number of trainable parameters it is expected to lag behind TopK SAE at large budgets. Our evaluation is limited to mid-sized transformer models and moderate dictionary sizes; however, the main bottleneck there might be not the SAE itself, but the infrastracture required to handle these setups and the model inference. Future Work. We identify three directions for extending this work: (i) Transcoding. Treat transcoders (Dunefsky et al., 2024) as implicit routers of information and investigate alternative logical gating functions (e.g. XOR or composite gates) to improve interpretability and circuit anal- ysis. (i) Crosscoding. Generalize KronSAE to a crosscoder setting (Lindsey et al., 2024) uncover interpretable, cross-level compositionality via logic operations. (i) Dynamic Composition. Explore learnable tuning of both the number of attention heads and their dimensionality, enabling fine-grained decomposition into groups of correlated features at varying scales. ETHICS STATEMENT While interpretability research has dual-use potential, our method operates within the ethical bound- aries of the underlying models and aims to advance responsible AI development through better model understanding. We analyze activations from publicly available language models (Qwen-2.5-1.5B, Pythia-1.4B, and Gemma-2-2B) gathered on FineWeb-Edu datasets, which excludes the unreported harmful content. We declare no conflicts of interest and maintain transparency about limitations, including potential artifacts from LLM-based interpretation as noted in Appendices D.3 and I. REPRODUCIBILITY STATEMENT We have taken several measures to ensure the reproducibility of our results. We use publicly available models (Qwen, Gemma, Pythia families) and training dataset (FineWeb-Edu) in our experiments. Section 4 and Appendix A provide detailed description of SAE training procedure and hyperparameter configuration. Our complete implementation is available in the supplementary materials, containing the training code, interpretation pipeline and analysis of the results. Appendix H includes simplified implementation of KronSAE that might be easily integrated into existing training codebases, while Appendix D details the interpretability analysis methodology with precise evaluation protocols. REFERENCES Nikita Balagansky, Ian Maksimov, and Daniil Gavrilov. Mechanistic permutability: Match features across layers. In The Thirteenth International Conference on Learning Representations, 2025. URL https://openreview.net/forum?id=MDvecs7EvO. 10 Preprint. Stella Biderman, Hailey Schoelkopf, Quentin Anthony, Herbie Bradley, Kyle O’Brien, Eric Hallahan, Mohammad Aflah Khan, Shivanshu Purohit, USVSN Sai Prashanth, Edward Raff, Aviya Skowron, Lintang Sutawika, and Oskar van der Wal. Pythia: A suite for analyzing large language models across training and scaling, 2023. URL https://arxiv.org/abs/2304.01373. Steven Bills, Nick Cammarata, Dan Mossing, Henk Tillman, Leo Gao, Gabriel Goh, Ilya Sutskever, Jan Leike, Jeff Wu, and William Saunders.Language models can explain neurons in language models, 2023.URLhttps://openai.com/index/ language-models-can-explain-neurons-in-language-models/. Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, et al. Towards monosemanticity: Decompos- ing language models with dictionary learning. Transformer Circuits Thread, 2, 2023. Bart Bussmann, Noa Nabeshima, Adam Karvonen, and Neel Nanda. Learning multi-level features with matryoshka sparse autoencoders. In Forty-second International Conference on Machine Learning, 2025. URL https://openreview.net/forum?id=m25T5rAy43. David Chanin, James Wilken-Smith, Tomáš Dulka, Hardik Bhatnagar, and Joseph Bloom. A is for absorption: Studying feature splitting and absorption in sparse autoencoders, 2024. URL https://arxiv.org/abs/2409.14507. Hoagy Cunningham, Aidan Ewart, Logan Riggs, Robert Huben, and Lee Sharkey. Sparse autoen- coders find highly interpretable features in language models. In International Conference on Learning Representations (ICLR), 2024. URLhttps://openreview.net/forum?id= F76bwRSLeK. Tri Dao, Beidi Chen, Nimit Sharad Sohoni, Arjun D. Desai, Michael Poli, Jessica Grogan, Alexander Liu, Aniruddh Rao, Atri Rudra, and Christopher Ré. Monarch: Expressive structured matrices for efficient and accurate training. In ICML, p. 4690–4721, 2022. URLhttps://proceedings. mlr.press/v162/dao22a.html. Jacob Dunefsky, Philippe Chlenski, and Neel Nanda. Transcoders find interpretable LLM feature circuits. In The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024. URL https://openreview.net/forum?id=J6zHcScAo0. Ali Edalati, Marzieh Tahaei, Ahmad Rashid, Vahid Partovi Nia, James J. Clark, and Mehdi Reza- gholizadeh. Kronecker decomposition for gpt compression, 2021. URLhttps://arxiv.org/ abs/2110.08152. Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henigan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, Roger Grosse, Sam McCandlish, Jared Kaplan, Dario Amodei, Martin Wattenberg, and Christopher Olah. Toy models of superposi- tion. Transformer Circuits Thread, 2022. URLhttps://transformer-circuits.pub/ 2022/toy_model/index.html. Nelson Elhage et al.Decomposing language models with dictionary learning. Trans- former Circuits,2023.URLhttps://transformer-circuits.pub/2023/ monosemantic-features. Thomas Fel, Ekdeep Singh Lubana, Jacob S. Prince, Matthew Kowal, Victor Boutin, Isabel Pa- padimitriou, Binxu Wang, Martin Wattenberg, Demba E. Ba, and Talia Konkle. Archety- pal SAE: Adaptive and stable dictionary learning for concept extraction in large vision mod- els. In Forty-second International Conference on Machine Learning, 2025. URLhttps: //openreview.net/forum?id=9v1eW8HgMU. Leo Gao, Tom Dupre la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders. In The Thirteenth International Conference on Learning Representations, 2025. URLhttps://openreview. net/forum?id=tcsZt9ZNKD. Thomas Heap et al. Sparse autoencoders can interpret randomly initialized transformers. arXiv preprint arXiv:2501.17727, 2025. URL https://arxiv.org/abs/2501.17727. 11 Preprint. Adam Karvonen, Can Rager, Johnny Lin, Curt Tigges, Joseph Bloom, David Chanin, Yeu-Tong Lau, Eoin Farrell, Callum McDougall, Kola Ayonrinde, Matthew Wearden, Arthur Conmy, Samuel Marks, and Neel Nanda. Saebench: A comprehensive benchmark for sparse autoencoders in language model interpretability, 2025. URL https://arxiv.org/abs/2503.09532. Dmitrii Kharlapenko, Stepan Shabalin, Arthur Conmy, and Neel Nanda. Scaling sparse feature circuits for studying in-context learning. In Sparsity in LLMs (SLLM): Deep Dive into Mixture of Experts, Quantization, Hardware, and Inference, 2025. URLhttps://openreview.net/ forum?id=sdLwJTtKpM. Jack Lindsey, Adly Templeton, Jonathan Marcus, Thomas Conerly, Joshua Batson, and Christopher Olah. Sparse crosscoders for cross-layer features and model diffing, 2024. URLhttps:// transformer-circuits.pub/2024/crosscoders/index.html. Scott C. Lowe et al. Logical activation functions: Logit-space equivalents of probabilistic boolean operators. arXiv preprint arXiv:2110.11940, 2021. URL https://arxiv.org/abs/2110. 11940. Samuel Marks, Can Rager, Eric J Michaud, Yonatan Belinkov, David Bau, and Aaron Mueller. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models. In The Thirteenth International Conference on Learning Representations, 2025. URLhttps: //openreview.net/forum?id=I4e82CIDxv. Anish Mudide, Joshua Engels, Eric J Michaud, Max Tegmark, and Christian Schroeder de Witt. Efficient dictionary learning with switch sparse autoencoders. In The Thirteenth International Conference on Learning Representations, 2025. URLhttps://openreview.net/forum? id=k2ZVAzVeMP. Gonçalo Paulo and Nora Belrose. Sparse autoencoders trained on the same data learn different features, 2025. URL https://arxiv.org/abs/2501.16615. Gonçalo Paulo, Alex Mallen, Caden Juang, and Nora Belrose. Automatically interpreting millions of features in large language models, 2024. URL https://arxiv.org/abs/2410.13928. Guilherme Penedo, Hynek Kydlí ˇ cek, Loubna Ben allal, Anton Lozhkov, Margaret Mitchell, Colin Raffel, Leandro Von Werra, and Thomas Wolf. The fineweb datasets: Decanting the web for the finest text data at scale, 2024. URL https://arxiv.org/abs/2406.17557. Senthooran Rajamanoharan,Arthur Conmy,Lewis Smith,Tom Lieberum,Vikrant Varma, János Kramár, Rohin Shah, and Neel Nanda.Improving sparse decompo- sition of language model activations with gated sparse autoencoders.In NeurIPS, 2024a.URLhttp://papers.nips.c/paper_files/paper/2024/hash/ 01772a8b0420baec00c4d59fe2fbace6-Abstract-Conference.html. Senthooran Rajamanoharan, Tom Lieberum, Nicolas Sonnerat, Arthur Conmy, Vikrant Varma, János Kramár, and Neel Nanda. Jumping ahead: Improving reconstruction fidelity with jumprelu sparse autoencoders, 2024b. URL https://arxiv.org/abs/2407.14435. Noam Shazeer, Azalia Mirhoseini, Krzysztof Maziarz, Andy Davis, Quoc V. Le, Geoffrey Hinton, and Jeff Dean. Outrageously large neural networks: The sparsely-gated mixture-of-experts layer. In International Conference on Learning Representations (ICLR), 2017. URLhttps: //arxiv.org/abs/1701.06538. P. Smolensky. Tensor product variable binding and the representation of symbolic structures in connectionist systems. Artif. Intell., 46(1–2):159–216, November 1990. ISSN 0004-3702. doi: 10.1016/0004-3702(90)90007-M. URLhttps://doi.org/10.1016/0004-3702(90) 90007-M. Paul Smolensky, Roland Fernandez, Zhenghao Herbert Zhou, Mattia Opper, and Jianfeng Gao. Mechanisms of symbol processing for in-context learning in transformer networks, 2024. URL https://arxiv.org/abs/2410.17498. 12 Preprint. Paul Soulos, R. Thomas McCoy, Tal Linzen, and Paul Smolensky. Discovering the compositional structure of vector representations with role learning networks. In Afra Alishahi, Yonatan Belinkov, Grzegorz Chrupała, Dieuwke Hupkes, Yuval Pinter, and Hassan Sajjad (eds.), Proceedings of the Third BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP, p. 238– 254, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020. blackboxnlp-1.23. URL https://aclanthology.org/2020.blackboxnlp-1.23/. Gemma Team, Morgane Riviere, Shreya Pathak, Pier Giuseppe Sessa, Cassidy Hardin, Surya Bhupatiraju, Léonard Hussenot, Thomas Mesnard, Bobak Shahriari, Alexandre Ramé, Johan Ferret, Peter Liu, Pouya Tafti, Abe Friesen, Michelle Casbon, Sabela Ramos, Ravin Kumar, Charline Le Lan, Sammy Jerome, Anton Tsitsulin, Nino Vieillard, Piotr Stanczyk, Sertan Girgin, Nikola Momchev, Matt Hoffman, Shantanu Thakoor, Jean-Bastien Grill, Behnam Neyshabur, Olivier Bachem, Alanna Walton, Aliaksei Severyn, Alicia Parrish, Aliya Ahmad, Allen Hutchison, Alvin Abdagic, Amanda Carl, Amy Shen, Andy Brock, Andy Coenen, Anthony Laforge, Antonia Paterson, Ben Bastian, Bilal Piot, Bo Wu, Brandon Royal, Charlie Chen, Chintu Kumar, Chris Perry, Chris Welty, Christopher A. Choquette-Choo, Danila Sinopalnikov, David Weinberger, Dimple Vijaykumar, Dominika Rogozi ́ nska, Dustin Herbison, Elisa Bandy, Emma Wang, Eric Noland, Erica Moreira, Evan Senter, Evgenii Eltyshev, Francesco Visin, Gabriel Rasskin, Gary Wei, Glenn Cameron, Gus Martins, Hadi Hashemi, Hanna Klimczak-Pluci ́ nska, Harleen Batra, Harsh Dhand, Ivan Nardini, Jacinda Mein, Jack Zhou, James Svensson, Jeff Stanway, Jetha Chan, Jin Peng Zhou, Joana Carrasqueira, Joana Iljazi, Jocelyn Becker, Joe Fernandez, Joost van Amersfoort, Josh Gordon, Josh Lipschultz, Josh Newlan, Ju yeong Ji, Kareem Mohamed, Kartikeya Badola, Kat Black, Katie Millican, Keelin McDonell, Kelvin Nguyen, Kiranbir Sodhia, Kish Greene, Lars Lowe Sjoesund, Lauren Usui, Laurent Sifre, Lena Heuermann, Leticia Lago, Lilly McNealus, Livio Baldini Soares, Logan Kilpatrick, Lucas Dixon, Luciano Martins, Machel Reid, Manvinder Singh, Mark Iverson, Martin Görner, Mat Velloso, Mateo Wirth, Matt Davidow, Matt Miller, Matthew Rahtz, Matthew Watson, Meg Risdal, Mehran Kazemi, Michael Moynihan, Ming Zhang, Minsuk Kahng, Minwoo Park, Mofi Rahman, Mohit Khatwani, Natalie Dao, Nenshad Bardoliwalla, Nesh Devanathan, Neta Dumai, Nilay Chauhan, Oscar Wahltinez, Pankil Botarda, Parker Barnes, Paul Barham, Paul Michel, Pengchong Jin, Petko Georgiev, Phil Culliton, Pradeep Kuppala, Ramona Comanescu, Ramona Merhej, Reena Jana, Reza Ardeshir Rokni, Rishabh Agarwal, Ryan Mullins, Samaneh Saadat, Sara Mc Carthy, Sarah Cogan, Sarah Perrin, Sébastien M. R. Arnold, Sebastian Krause, Shengyang Dai, Shruti Garg, Shruti Sheth, Sue Ronstrom, Susan Chan, Timothy Jordan, Ting Yu, Tom Eccles, Tom Hennigan, Tomas Kocisky, Tulsee Doshi, Vihan Jain, Vikas Yadav, Vilobh Meshram, Vishal Dharmadhikari, Warren Barkley, Wei Wei, Wenming Ye, Woohyun Han, Woosuk Kwon, Xiang Xu, Zhe Shen, Zhitao Gong, Zichuan Wei, Victor Cotruta, Phoebe Kirk, Anand Rao, Minh Giang, Ludovic Peran, Tris Warkentin, Eli Collins, Joelle Barral, Zoubin Ghahramani, Raia Hadsell, D. Sculley, Jeanine Banks, Anca Dragan, Slav Petrov, Oriol Vinyals, Jeff Dean, Demis Hassabis, Koray Kavukcuoglu, Clement Farabet, Elena Buchatskaya, Sebastian Borgeaud, Noah Fiedel, Armand Joulin, Kathleen Kenealy, Robert Dadashi, and Alek Andreev. Gemma 2: Improving open language models at a practical size, 2024. URL https://arxiv.org/abs/2408.00118. Adly Templeton, Tom Conerly, Jonathan Marcus, Jack Lindsey, Trenton Bricken, Brian Chen, Adam Pearce, Craig Citro, Emmanuel Ameisen, Andy Jones, Hoagy Cunningham, Nicholas L Turner, Callum McDougall, Monte MacDiarmid, C. Daniel Freeman, Theodore R. Sumers, Edward Rees, Joshua Batson, Adam Jermyn, Shan Carter, Chris Olah, and Tom Henighan. Scaling monosemanticity: Extracting interpretable features from claude 3 sonnet. Trans- former Circuits Thread, 2024. URLhttps://transformer-circuits.pub/2024/ scaling-monosemanticity/index.html. Joshua T. Vogelstein, John M. Conroy, Vince Lyzinski, Louis J. Podrazik, Steven G. Kratzer, Eric T. Harley, Donniell E. Fishkind, R. Jacob Vogelstein, and Carey E. Priebe. Fast ap- proximate quadratic programming for graph matching. PLoS ONE, 10, 2015. URLhttps: //api.semanticscholar.org/CorpusID:16927142. Peihao Wang, Rameswar Panda, Lucas Torroba Hennigen, Philip Greengard, Leonid Karlinsky, Rogerio Feris, David Daniel Cox, Zhangyang Wang, and Yoon Kim. Learning to grow pretrained models for efficient transformer training. In The Eleventh International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=cDYRS5iZ16f. 13 Preprint. An Yang, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chengyuan Li, Dayiheng Liu, Fei Huang, Haoran Wei, Huan Lin, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jingren Zhou, Junyang Lin, Kai Dang, Keming Lu, Keqin Bao, Kexin Yang, Le Yu, Mei Li, Mingfeng Xue, Pei Zhang, Qin Zhu, Rui Men, Runji Lin, Tianhao Li, Tingyu Xia, Xingzhang Ren, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yu Wan, Yuqiong Liu, Zeyu Cui, Zhenru Zhang, and Zihan Qiu. Qwen2.5 technical report. arXiv preprint arXiv:2412.15115, 2024. An Yang, Anfeng Li, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Gao, Chengen Huang, Chenxu Lv, Chujie Zheng, Dayiheng Liu, Fan Zhou, Fei Huang, Feng Hu, Hao Ge, Haoran Wei, Huan Lin, Jialong Tang, Jian Yang, Jianhong Tu, Jianwei Zhang, Jianxin Yang, Jiaxi Yang, Jing Zhou, Jingren Zhou, Junyang Lin, Kai Dang, Keqin Bao, Kexin Yang, Le Yu, Lianghao Deng, Mei Li, Mingfeng Xue, Mingze Li, Pei Zhang, Peng Wang, Qin Zhu, Rui Men, Ruize Gao, Shixuan Liu, Shuang Luo, Tianhao Li, Tianyi Tang, Wenbiao Yin, Xingzhang Ren, Xinyu Wang, Xinyu Zhang, Xuancheng Ren, Yang Fan, Yang Su, Yichang Zhang, Yinger Zhang, Yu Wan, Yuqiong Liu, Zekun Wang, Zeyu Cui, Zhenru Zhang, Zhipeng Zhou, and Zihan Qiu. Qwen3 technical report, 2025. URL https://arxiv.org/abs/2505.09388. 14 Preprint. AADDITIONAL DETAILS AND RESULTS A.1EXPERIMENTAL SETUP Training details.All SAEs are optimized using AdamW with an initial learning rate of8× 10 −4 , a cosine learning-rate schedule with a minimum LR of1× 10 −6 , and a linear warmup for the first 10% of total training steps, auxiliary loss penalty equal to0.03125. We use a global batch size of 8,192. We sweep over dictionary (latent) sizes ofF = 2 15 ,F = 2 16 , andF = 2 17 features. For our KronSAE variant, we further sweep the number of headshand the per-head dimensionsmand nsuch thath· m· nequals the desired dictionary size. Regularization weights and auxiliary loss coefficients are kept constant throughout the runs to isolate the impact of architectural choices. For all experiments, we spent about 330 GPU days on NVIDIA H100 80GB GPUs, including preliminary research. SAE. For all experiments on Qwen-2.5, we train each SAE on activations from layer 14. Also for Pythia-1.4B we use layer 14 and for Gemma-2-2B we take activations from layer 12. For most of our experiments, we use sparsity level of ℓ 0 = 50 non-zero activations per token. Initialization. As observed by Gao et al. (2025), initializing the decoder as the transpose of the encoder (W dec = W ⊤ enc ) provides a strong metric improvement. We adopt this strategy within KronSAE by partitioningW enc intohhead-wise blocks of shapesm × dandn × d, denoted P i ,Q i h i=1 . For each head k, we define its decoded rows via a simple additive composition: C k [i,j] = P k,i + Q k,j , i = 1,...,m, j = 1,...,n. Finally, flattening the matricesC k yields full decoder weight matrix W dec ∈R F×d . Matryoshka and Kron-based version.For Matryoshka SAE of dictionary sizeFwe adopt the fol- lowing experimental setup. We use most training settings from Training details. For Matryoshka SAE Bussmann et al. (2025) we define the parameter of dictionary groupS = [2 k , 2 k ,· 2 k+i · 2 n ], wherek < nand P s∈S s = F. This is equivalent to nested sub-SAEs with dictionary sizes M =2 k , 2 k + 2 k , 2 k + 2 k + 2 k+1 ,· ,Fand in our work we define this SAE as Matryoshka-2 k . Training loss for Matryoshka is defined as follows: L(f ) = X m∈M ||x− f 0:m W dec 0:m + b dec || 2 2 + αL aux ,(6) where f is latent vector andL aux is auxiliary loss used in Gao et al. (2025). For Kron-based Matryoshka SAE we use same setup and same initialization from Initialization section. We choosemandnso thatF mod mn = 0, ensuring that no heads are shared between groups. Switch SAE and Kron-based version.For the Switch SAE architecture proposed by Mudide et al. (2025), we configure all experiments with 8 experts and use identical initialization from KronSAE. In the Kron-based variant of this SAE, we distribute the heads equally across all experts, resulting in F/(8mn)heads per expert. Hence we have less parameters in every expert and therefore we also reduce FLOPs for encoder. A.2FLOPS CALCULATION AND EFFICIENCY For TopK SAE and KronSAE we compute FLOPs in the following way: FLOPS TopK (d,F,k) = dF + kd, FLOPS KronSAE (d,m,n,h,k) = dh(m + n) + mnh + kd≈ dh(m + n) + kd. (7) We calculate FLOPs for most effective variant of TopK where we perform vector matrix multipication only for nonzero activations, while encoder still requires dense matrix multiplication. We have also measured the wallclock time for forward and backward to examine the scaling. Figure 8 reports scaling for different hidden dimension sizes. 15 Preprint. 10002000300040005000600070008000 16.000 32.000 64.000 128.000 256.000 512.000 Speed in ms Dictionary size=32768 TopK SAE TopK KronSAE 10002000300040005000600070008000 Hidden dimension Dictionary size=65536 10002000300040005000600070008000 Dictionary size=131072 Figure 8: Speed comparision of TopK SAE with KronSAE across different hidden dimensionss. We can see that KronSAE have better scaling properties than SAE with default encoder architecture. A.3SCALING ON GEMMA-2 2B To examine the method’s generality, we conducted additional reconstruction experiments on Gemma- 2 2B under iso-FLOPs settings. As shown in Figure 9, our method achieves comparable or improved reconstruction performance across the evaluated compute budgets. 500m1000m 0.68 0.70 0.72 0.74 0.76 0.78 0.80 0.82 Explained Variance Dictionary: 32768 m=1 m=2 Matryoshka-1024 Matryoshka-2048 TopK 500m1000m Reference Tokens Budget Dictionary: 65536 m=1 m=2 m=4 Matryoshka-2048 Matryoshka-4096 TopK 500m1000m Dictionary: 131072 m=1 m=2 m=4 Matryoshka-4096 Matryoshka-8192 TopK Figure 9: Performance comparision of KronSAE and TopK SAE under a fixed iso-FLOPs budget. Across sparsity settings, KronSAE typically matches TopK’s reconstruction performance and in some cases slightly outperforms it, while using much fewer trainable parameters. A.4SMALLER DICTIONARY SIZE To complement our larger-scale experiments, we further evaluate KronSAE’s performance on a smaller dictionary size F = 2 15 with varying ℓ 0 =16, 32, 64, 128 and equal token budget. Sparsity.Following the experimental setup described in Section 4.1, we compare KronSAE against established baselines including TopK SAE, Matryoshka SAE, and Switch SAE . As shown in Figure 10, KronSAE consistently matches or exceeds the reconstruction performance of baseline architectures across all tested sparsity levels, while achieving these results with substantially fewer trainable parameters and FLOPs. Feature absorption. We extend our feature absorption analysis from Section 4.2 to the smaller dictionary size, employing two distinct grouping schedules:S 1 = [2048, 2048, 4096, 8192, 16384] andS 2 = [1024, 1024, 2048, 4096, 8192, 16384]. As shown in Figure 11 KronSAE modification demonstrates better performance, reducing feature absorption metrics. 16 Preprint. 0.60 0.65 0.70 0.75 0.80 0.85 0.90 TopK / KronSAEMatryoshka / KronSAE + MatryoshkaSwitch / KronSAE + Switch m=1 m=2 m=4 Baseline 163264128 0 0.60 0.65 0.70 0.75 0.80 0.85 0.90 163264128 0 163264128 0 m=1 m=2 m=4 Baseline Gemma-2 2B Qwen-2.5 1.5B Explained Variance Figure 10: Performance comparision of KronSAE and TopK SAE under a fixed token budget. Across sparsity settings, KronSAE typically matches TopK’s reconstruction performance and in some cases slightly outperforms it, while using much fewer trainable parameters and smaller FLOPs budget. 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Mean Absorption Fraction Score Model: Gemma-2 2BModel: Qwen-2.5 1.5B SAE Type KronSAE KronSAE + Matryoshka-1024 KronSAE + Matryoshka-2048 TopK TopK + Matryoshka-1024 TopK + Matryoshka-2048 0.0 0.1 0.2 0.3 0.4 0.5 0.6 Mean Full Absorption Score 163264128 0 0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 Mean Number of Feature Splits 163264128 0 Figure 11: Feature absorption scores comparison of KronSAE vs baseline methods (TopK SAE and Matryoshka SAE) and a Kron-enhanced Matryoshka SAE variant. 17 Preprint. A.5PYTHIA SUITE For Pythia-1.4B we train all SAEs on the 12th transformer layer with a budget of 125M tokens. As reported in Table 3, KronSAE achieves performance comparable to TopK SAE with increased number of heads. DictionarySAEMeanMax 32k TopK0.793 KronSAE0.7830.793 65k TopK0.802 KronSAE0.7950.805 131k TopK0.801 KronSAE0.8000.810 Table 3: Performance of Pythia-1.4B at 125M budget. At larger dictionary size and fixed training budget KronSAE outperforms TopK SAE. We conducted additional experiments with smaller Pythia models and trained KronSAE at the middle layers with 512 heads. Table 4 reports results for 125M budget on 65k and 262k dictionary sizes. Model70M (d=512)160M (d=768)410M (d=1024) Dictionary65k256k65k256k65k256k m = 10.8930.8920.8560.8550.8340.835 m = 20.8960.8970.8590.8590.8320.841 m = 40.8940.8990.8570.8590.8280.839 m = 80.8960.8990.8560.8620.8270.846 TopK 0.9050.9030.8700.8670.8430.847 Table 4: Performance of SAEs on 70M, 160M and 410M Pythias with varying hidden dimensionality. In section 4.2 we have analysed whether KronSAE achieves lower absorption score and have answered affirmatively. We also compare the results for Pythia 1.4B model and validate the improvements, as reported in the Table 5. ModelSAEℓ 0 = 16ℓ 0 = 32ℓ 0 = 64ℓ 0 = 128ℓ 0 = 256 Pythia TopK0.4450.2330.0580.0060.004 KronSAE0.2440.1290.0330.0070.003 Table 5: Absorption score calculated for Pythia 1.4B model. KronSAE shows lower score due to structured latent space and hierarchy between pre-latents and post-latents. These results confirm that our compositional architectures improves the absorption score and feature consistency across various models from different families. A.6COMPARISON WITH JUMPRELU We provide experiments to compare KronSAE with an alternative activation mechanism, JumpReLU, and report explained variance under three sparsity levels in Table 6. Replacement of TopK with JumpReLU within KronSAE leads to a degraded performance relative to both JumpReLU SAE and KronSAE with TopK, also with degraded scaling overℓ 0 . This suggests that the architectural advantages of KronSAE interact most effectively with TopK’s behaviour. Whether an alternative activation function can improve on this remains a topic for future work. 18 Preprint. Model Variantℓ 0 = 32ℓ 0 = 50ℓ 0 = 64 TopK0.8090.8370.852 JumpReLU 0.8130.8380.844 KronSAE (TopK)0.8140.8400.853 KronSAE (JumpReLU) 0.7900.8170.828 Table 6: Performance of SAEs with JumpReLU and TopK activations. Since we have floating sparsity controlled via l0 penalty coefficient, we performed a sweep over various sparsity levels, fitted a parabola to the resulting data as a function ofℓ 0 , and evaluated it on those sparsity levels. In contrast, TopK and KronSAE (with TopK) were trained using fixed, predefined sparsity levels. A.7PERFORMANCE ACROSS LAYERS For this experiment we fix the dictionary size toF = 2 15 and use the same hyperparameters as in the main experiments. As the Figure 12 shows, KronSAE is on-par with TopK at every depth, demonstrating that structured encoder’s reconstruction quality is robust to layer choice. 2420161284 Layer 0.83 0.84 0.85 0.86 0.87 0.88 0.89 0.90 Explained Variance TopK KronSAE Figure 12: EV across layers of Qwen-2.5-1.5B, demonstrating that KronSAE matches TopK perfor- mance regardless of depth. A.8CHOICE OF m,n,h We derive the following guidelines to train KronSAE: one should minimize themand maximize the hto improve the reconstruction performance, and search for the most expressive configuration from feasible ones. Sincemhas more impact on EV, one should start fromm = 2in the search process, since it gives improved computational performance with on-par EV with full TopK training. BMORE RESULTS ON SYNTHETIC In this section we present the motivation behind our matching algorithm and additional results. B.1FEATURE MATCHING AS QUADRATIC PROBLEM Suppose that we have two sets of feature embeddings from different models represented as matrices X,Y ∈R F×d , whereFis the number of features anddis the dimensionality of feature embeddings. Our task is to find the optimal assignment between features fromYto features fromXso that this assignment would satisfy some considerations. The standard approach is to solve linear assignment problem - find a permutation matrixΠsubject to minimizing thetrace(C T Π) , whereCis a cost matrix defined as pairwise distance between features 19 Preprint. C i,j = d(X i ,Y j ). Standard algorithm for solving it in context of sparse dictionary learning is a Hungarian algorithm (Paulo & Belrose, 2025; Balagansky et al., 2025; Fel et al., 2025). This linear problem only considers pairwise information between features while ignoring the global dependencies between features within X and Y sets separately, e.g. clusters should map to clusters, and the linear problem does not internalize this information. Our observation of correlations between features naturally requires to search for assignment that would take this information into account: we seek for a permutationΠthat would give best global alignment, measured as the Frobenius norm of X T ΠY . So the objective becomes: max Π ∥X T ΠY∥ 2 F = max Π trace(Π T X T ΠY Y T ),(8) where global feature structure is explicitly encoded in the matrices X T and Y Y T . An efficient algorithm to solve the quadratic assignment problem is Fast Approximate Quadratic Programming (FAQ) method (Vogelstein et al., 2015) that initially was designed for graphs matching: given the adjacency matricesAandB, it minimizes thetrace(AΠB T Π T ), whereΠare relaxed from permutation matrices to the set of doubly stochastic matrices (Birkhoff polytope). In our case we defineA = X T andB = Y Y T , and since we do not want to minimize the cost but rather maximize the similarity, we solve for the reversed objective: max Π (X T Π(Y Y ) T Π T ) = max Π trace(Π T X T ΠY Y T ),(9) which is the same as equation 8, and this formulation preserves global dependencies because the contribution of assigningY j toX i depends on all other assignments through the cross-terms in the quadratic form. Listing 1 shows the implementation of this matching procedure. Listing 1: Implementation of FAQ algorithm for quadratic feature assignment problem. 1 def feature_matching(A, B, max_iter): 2 F, d = A.shape[0] 3 G_A, G_B = A @ A.T, B @ B.T 4 P = np.ones((F, F)) / F # Barycenter initialization 5 6 # Frank-Wolfe iterations 7 for _ in range(max_iter): 8 grad = 2 * G_A @ P @ G_B # Gradient for maximization 9 r, c = linear_sum_assignment(-grad) # Solve LAP 10 Q = np.zeros_like(P) 11 Q[r, c] = 1 12 13 # Compute optimal step size 14 D = Q - P 15 b, a = np.trace(grad.T @ D), np.trace(G_A @ D @ G_B @ D.T) 16 17 if abs(a) < 1e-12: 18 alpha = 1.0 if b > 0 else 0.0 19 elif a < 0: # Concave case 20 alpha = np.clip(-b/(2 * a), 0, 1) 21 else: # Convex case 22 alpha = 1.0 if b > 0 else 0.0 23 24 P_new = P + alpha * D 25 P = P_new 26 27 # Project to permutation matrix 28 r, c = linear_sum_assignment(-P) 29 P_final = np.zeros((F, F)) 30 P_final[r, c] = 1 31 32 return P_final, P_final @ B 20 Preprint. B.2ADDITIONAL RESULTS As shown in Figure 13, KronSAE maintains correlation structure that is heavier than in TopK and is better aligned with ground truth covariance matrix. Ground Truth AE Covariance RV: 0.598 TopK RV: 0.157 Kron h = 2, m = 4, n = 32 RV: 0.514 Kron h = 4, m = 2, n = 32 RV: 0.379 Kron h = 4, m = 8, n = 8 RV: 0.401 Kron h = 8, m = 2, n = 16 RV: 0.249 RV: 0.338RV: 0.118RV: 0.295RV: 0.219RV: 0.320RV: 0.234 RV: 0.398RV: 0.124RV: 0.355RV: 0.295RV: 0.254RV: 0.208 RV: 0.686RV: 0.209RV: 0.519RV: 0.362RV: 0.303RV: 0.212 Figure 13: Examples of patterns learned in autoencoder TopK and KronSAE variants after we apply the improved matching scheme. KronSAE have learned patterns that more closely resemble the underlying ground truth structure, and with increasing number of heads (more fine-grained structure) it approaches the TopK SAE. Table 7 indicate that structure learned by KronSAE is more diverse as indicated by standard deviation of the corresponding values, and is also more aligned with the ground truth covariance according to higher RV scores and lower differences (∆) between properties of ground truth and learned matrix. ModelRV coeff.Effective rankMean corr.Rank ∆ AE0.59± 0.1763.3± 0.30.13± 0.0447.9 TopK0.12± 0.0258.5± 1.20.08± 0.0143.2 Kron h = 2,m = 4,n = 320.22± 0.0753.8± 3.40.08± 0.0138.5 Kron h = 4,m = 2,n = 320.3± 0.1246.0± 7.80.10± 0.0330.7 Kron h = 4,m = 8,n = 80.23± 0.0848.8± 5.00.09± 0.0233.5 Kron h = 8,m = 2,n = 160.17± 0.0455.9± 1.70.07± 0.0140.6 Table 7: Results with improved matching scheme, computed for 8 different covariance setups. Higher RV coefficients between ground truth and learned matrices and lower∆between properties of these matrices indicate that KronSAE is more variable across different setups and is better aligned with ground truth correlation structures. Together, these results additionally validate improved covariance reconstruction in KronSAE. CMAND AS A LOGICAL OPERATOR AND KRONSAE AS LOGICAL SAE AND-like mechanism and hierarchy. Suppose we have u k = P k x and v k = Q k x. KronSAE can be described in two equivalent ways: 1. Applying the ReLU to latents u, v, then applying the Kronecker product and square root. 2. Applying the mAND kernel that creates the matrix as in equation 3 and flattening it. 21 Preprint. These approaches are complementary. To formally understand how they induce the AND-like mechanism and hierarchy, consider base and extension pre-latent vectorsp = ReLU(u)andq = ReLU(v). Eachp i is aith base pre-latent activation, andq j isjth extension pre-latent activation. Kronecker productp ⊗ qcreates the vector(p 1 ∗ q 1 ,...,p 1 ∗ q m ,p 2 ∗ q 1 ,...,p n ∗ q m )of the activations of post-latents. Then there is two situations: 1. Post-latent is active =⇒ both pre-latents activations are positive. 2. Post-latent is inactive =⇒ at least one of pre-latent activations is zero. Hence post-latent is active only when both pre-latents are active. Fix some post-latent and its corresponding pre-latents, and suppose thatP,Q,Fare the sets of input vectors from the hidden state space (passed to SAE) on which base pre-latent, extension pre-latent and post-latent are active. Then it is true thatF =P ∩Q, meaning that pre-latents must be broader and polysemous to encode multiple semantics of emitted post-latents. We validate this behaviour qualitatively in section 5.3, although in the same section we describe that apparently other types of interactions also presented. Matryoshka and Kron hierarchy. In contrast to TopK SAE, Matryoshka loss imposes different kind of hierarchy by dividing the dictionary into groups ofG 1 ,...,G k latents where each is of different level of granularity. Namely, firstG 1 latents are the most broad and abstract, nextG 2 latents add more fine-grained semantics, lowering the level of abstraction, and so on. This type of structure imposed by specific loss function - increasing the level of granularity must decrease the reconstruction error, and the lowest level ofG 1 features must also maintain good reconstruction quality - and does not strictly demand some kind of conditional activation of features, while KronSAE imposes two-level AND-like hierarchy via architectural design (Kronecker product). As we have two different mechanisms of feature hierarchy (from encoding mechanism and loss function design) we can combine it, as shown in Section 4.1, 4.2 and Appendix A.4, to combine properties of both approaches. Visual intuition. We also compare our mAND to existing AND AIL (Lowe et al., 2021). Since our objective is to drive each atom toward a distinct, monosemantic feature, we found that tightening the logical conjunction encourages sharper feature separation. Moreover, by using the geometric mean ( √ pq) rather than a simple product or minimum,mANDpreserves activation magnitudes and prevents post-latent activation to be exploded when bothp,qare positive. A visual comparison of mAND and AND AIL appears in Figure 14. -2-1.5-1-0.500.511.52 p -2 -1.5 -1 -0.5 0 0.5 1 1.5 2 q mAND -2-1.5-1-0.500.511.52 p AND AIL 0.00 0.24 0.48 0.72 0.96 1.20 1.44 1.68 1.92 Activation value Figure 14: Comparison of the smooth mAND operator against the AND AIL (Lowe et al., 2021). DFEATURE ANALYSIS METHODOLOGY We analyze learned features using an established pipeline Bills et al. (2023); Paulo et al. (2024) con- sisting of three stages: (1) statistical property collection, (2) automatic activation pattern interpretation, and (3) interpretation evaluation. The following subsections detail our implementation. 22 Preprint. D.1DATA COLLECTION Our collection process uses a fixed-size bufferB = 384per feature, continuing until processing a predetermined maximum token count T max . The procedure operates as follows: Initial processing batches generate large activation packs of 1M examples, where each example comprises 256-token text segments. When encountering feature activations, we add them to the buffer, applying random downsampling to maintain sizeBwhen exceeding capacity. This approach enables processing arbitrary token volumes while handling rare features that may require extensive sampling. During collection, we compute online statistics including activation minimums, maximums, means, and frequencies. Post-processing yields two key metrics: token entropy and multitoken ratio. The token entropy is calculated as: token entropy =− n X i=0 p i · log(p i ), p i = activations of token i total amount of activations ,(10) where n represents unique activated tokens. The multitoken ratio is: multitoken ratio = 1 b b X i=0 number of activations in sequence i total tokens in sequence i ,(11) with b < B denoting collected context examples per feature. We then segment examples using a 31-token context window (15 tokens before/after each activation), potentially creating overlapping but non-duplicated examples. Features with high multitoken ratio may have number of examples significantly exceeding B. A separate negative examples buffer captures non-activating contexts. Future enhancements could employ predictive modeling (e.g., using frequent active tokens) to strategically populate this buffer with expected-but-inactive contexts, potentially improving interpretation quality. D.2FEATURE INTERPRETATIONS For each feature, we generate interpretations by sampling 16 random activation examples above the median activation quantile and presenting them to Qwen3 14B (Yang et al., 2025) (AWQ- quantized with reasoning enabled). The model produces concise descriptions of the activation patterns. Empirical observations suggest reasoning mode improves interpretation quality, though we lack quantitative measurements. This aligns with findings in (Paulo et al., 2024), which compared standard one-sentence responses with Chain-of-Thought outputs, making model reasoning an interesting direction for future research. The interpretation process uses the system prompt presented in a Figure 15. User prompts include all special characters verbatim, as some features activate specifically on these characters. A representative (slightly abbreviated) user prompt example is presented on Figure 16. D.3EVALUATION PIPELINE We evaluate interpretations using balanced sets of up to 64 positive (activation quantile > 0.5) and 64 negative examples, employing the same model without reasoning to reduce computational costs. When insufficient examples exist, we maintain class balance by equalizing positive and negative counts. The evaluation uses modified system prompts from (Paulo et al., 2024), with added emphasis on returning Python lists matching the input example count exactly. We discard entire batches if responses are unparseable or contain fewer labels than the number of provided examples. We calculate two scores. Detection Score: After shuffling positive/negative examples, we present up to 8 unformatted text examples per batch to the model. The model predicts activations (1/0) for each example, generating 23 Preprint. You are a meticulous AI researcher conducting an important investigation into patterns found in language. Your task is to analyze text and provide an explanation that thoroughly encapsulates possible patterns found in it. Guidelines: You will be given a list of text examples on which special words are selected and between delimiters like «this». If a sequence of consecutive tokens all are important, the entire sequence of tokens will be contained between delimiters «just like this». How important each token is for the behavior is listed after each example in parentheses. - Your explanation should be a concise STANDALONE PHRASE that describes observed patterns. - Focus on the essence of what patterns, concepts and contexts are present in the examples. - Do NOT mention the texts, examples, activations or the feature itself in your explanation. - Do NOT write "these texts", "feature detects", "the patterns suggest", "activates" or something like that. - Do not write what the feature does, e.g. instead of "detects heart diseases in medical reports" write "heart diseases in medical reports". - Write explanation in the last line exactly after the [EXPLANATION]: Figure 15: System prompt for feature interpretations. Examples of activations: Text: ’ Leno«,» a San Francisco Democrat«, said in a statement.»’ Activations: ’ said (22.74), statement (27.84), in (27.54)’ Text: ’ city spokesman Tyler Gamble« said in an» email.’ Activations: ’ said (2.92), in (12.81), an (14.91)’ Text: ’ towpath at Brentford Lock. «Speaking» on BBC London 94’ Activations: ’Speaking (3.48)’ Text: ’ Michelle, a quadriplegic,« told» DrBicuspid.com’ Activations: ’ told (4.05)’ Text: ’ CEO Yves Carcelle« said in a statement».’ Activations: ’ said (19.64), in (29.09), statement (29.39)’ Figure 16: Example of user prompt passed to LLM. This feature with 16 examples received the interpretation "Structural elements in discourse, including speech attribution, prepositional phrases, and formal contextual markers" with a detection score of 0.84 and fuzzing score of 0.76. 24 Preprint. up to 128 true/predicted label pairs. The score calculates as: score = 1 2 correctly predicted positives total positives + correctly predicted negatives total negatives .(12) Fuzzing Score: We «highlight» activated tokens on sampled examples, from which 50% are correctly labeled positive examples, 25% are mislabeled positive examples, and 25% are randomly labeled negative examples. We present batches of up to 8 examples and the model identifies correct/incorrect labeling, with scoring following Equation 12. EADDITIONAL FEATURE ANALYSIS RESULTS Feature property correlations. Our analysis reveals significant correlations between feature properties and interpretability scores (Figure 17). Notably, token entropy and mean activation show substantial correlations with interpretability scores, suggesting their potential as proxies for assessing feature quality without running the full interpretation pipeline. These findings are based on analysis of the first 3072 features from 32k TopK and KronSAE (m=4, n=4) trained on 24M tokens, warranting further validation with larger-scale studies. Mean score Detection score Fuzzing score Mean activation Token entropy Multitoken ratio Frequency 1.000.930.870.50-0.41-0.18-0.24 0.931.000.630.45-0.34-0.14-0.25 0.870.631.000.48-0.43-0.20-0.17 0.500.450.481.00-0.32-0.15-0.09 -0.41-0.34-0.43-0.321.000.350.23 -0.18-0.14-0.20-0.150.351.000.51 -0.24-0.25-0.17-0.090.230.511.00 TopK - Pearson 1.000.940.880.54-0.39-0.08-0.20 0.941.000.660.49-0.33-0.05-0.20 0.880.661.000.49-0.43-0.13-0.16 0.540.490.491.00-0.34-0.01-0.08 -0.39-0.33-0.43-0.341.000.420.21 -0.08-0.05-0.13-0.010.421.000.35 -0.20-0.20-0.16-0.080.210.351.00 KronSAE Post - Pearson 1.000.930.880.55-0.44-0.39-0.29 0.931.000.650.49-0.34-0.34-0.26 0.880.651.000.53-0.47-0.37-0.22 0.550.490.531.00-0.43-0.39-0.17 -0.44-0.34-0.47-0.431.000.490.29 -0.39-0.34-0.37-0.390.491.000.51 -0.29-0.26-0.22-0.170.290.511.00 KronSAE Pre - Pearson Mean score Detection score Fuzzing score Mean activation Token entropy Multitoken ratio Frequency Mean score Detection score Fuzzing score Mean activation Token entropy Multitoken ratio Frequency 1.000.930.870.61-0.45-0.15-0.27 0.931.000.640.56-0.38-0.10-0.31 0.870.641.000.56-0.46-0.19-0.16 0.610.560.561.00-0.37-0.27-0.31 -0.45-0.38-0.46-0.371.000.350.31 -0.15-0.10-0.19-0.270.351.000.73 -0.27-0.31-0.16-0.310.310.731.00 TopK - Spearman Mean score Detection score Fuzzing score Mean activation Token entropy Multitoken ratio Frequency 1.000.930.880.68-0.43-0.03-0.23 0.931.000.660.62-0.370.03-0.26 0.880.661.000.60-0.47-0.10-0.14 0.680.620.601.00-0.40-0.03-0.25 -0.43-0.37-0.47-0.401.000.380.36 -0.030.03-0.10-0.030.381.000.63 -0.23-0.26-0.14-0.250.360.631.00 KronSAE Post - Spearman Mean score Detection score Fuzzing score Mean activation Token entropy Multitoken ratio Frequency 1.000.930.880.59-0.45-0.38-0.47 0.931.000.650.53-0.37-0.33-0.45 0.880.651.000.56-0.47-0.39-0.41 0.590.530.561.00-0.53-0.57-0.64 -0.45-0.37-0.47-0.531.000.520.46 -0.38-0.33-0.39-0.570.521.000.85 -0.47-0.45-0.41-0.640.460.851.00 KronSAE Pre - Spearman Figure 17: Correlation coefficients (Pearson and Spearman) between properties of TopK and KronSAE latents. Token entropy emerges as a strong predictor of interpretability scores, while higher mean activation and lower frequency also indicate more interpretable features. Pre-latent to post-latent relationships. We investigate how post-latent properties correlate with various combinations of pre-latent properties, including individual values, means, products, and themANDoperation (product followed by square root). Figure 18 demonstrates that post-latent multitoken ratio, token entropy, and frequency show stronger correlations with pre-latent products or mAND values than with individual pre-latent properties or their means. Basis geometry. As noted in Section 5.3, latent embeddings primarily exhibit clustering within their originating groups (head, base, extension). With the support of observations reported in Sections 4.1 and 5.3, we find that models with more heads achieve better reconstruction while producing more diverse basis vectors. This suggests that fine-grained architectures yield more expressive 25 Preprint. 0.25 0.50 0.75 1.00 Pearson r 0.26 0.31 0.38 0.36 0.37 Detection score 0.30 0.32 0.41 0.39 0.40 Fuzzing score 0.60 0.57 0.73 0.75 0.73 Token entropy Base Extension Mean Product mAND 0.25 0.50 0.75 1.00 Pearson r 0.46 0.46 0.60 0.66 0.64 Multitoken ratio Base Extension Mean Product mAND 0.51 0.54 0.67 0.59 0.69 Mean activation Base Extension Mean Product mAND 0.51 0.54 0.65 0.86 0.88 Frequency Figure 18: Correlation patterns between properties of post-latents and pre-latents. representations, although they may also exhibit undesired challenging behavior like feature splitting (Bricken et al., 2023) or absorption (Chanin et al., 2024). Figure 19 visualizes this structure through UMAP projections (n_neighbors=15, min_dist=0.05, met- ric=’cosine’) of decoder weights from the first 8 heads of 32k SAEs with varying m,n configurations. The plots reveal distinct clustering patterns: form < nwe observe tight base-wise clustering with weaker grouping by extension, and for m≥ n extension-wise clustering is stronger. By Head m = 4, n = 4, EV = 0.845m = 4, n = 8, EV = 0.824m = 8, n = 8, EV = 0.815m = 8, n = 16, EV = 0.812m = 16, n = 16, EV = 0.799 By Base By Extension Figure 19: UMAP visualization of post-latent clustering patterns by head, base, and extension group membership. We observe tight clusters by base for m < n and by extension for m≥ n. This asymmetry suggests that pre-latent capacity requirements directly manifest in the embedding geometry - components with lower polysemanticity (extensions when m < n) exhibit greater geometric diversity. We expect symmetric behavior for reciprocal configurations (e.g., m=4,n=8 vs. m=8,n=4), merely swapping the roles of bases and extensions. Interpretability across sparsity regimes.We compare KronSAE and its Matryoshka variant with TopK, Matryoshka and Switch SAE baselines across different sparsity regimes for Gemma 2 2B and Qwen 2.5 1.5B. For each SAE we follow the same pipeline as for results in Section 5.3, but use only 26 Preprint. 18 million tokens for examples collection. For KronSAE and TopK we use first 4096 features, and for other models we sample 4096 features randomly. Scores forF = 2 16 andF = 2 15 are presented in Figures 20 and 21. 0.5 0.6 0.7 0.8 0.9 1.0 Detection score Gemma 2 2BQwen2.5 1.5B 163264128 0 0.5 0.6 0.7 0.8 0.9 1.0 Fuzzing score 163264128 0 TopK Switch Matryoshka KronSAE KronSAE + Matryoshka Figure 20: Interpretability scores forF = 2 16 = 65536. KronSAE shows good consistency across different sparsity regimes, in some cases outperforming the TopK baseline. 0.5 0.6 0.7 0.8 0.9 1.0 Detection score Gemma 2 2BQwen2.5 1.5B 163264128 0 0.5 0.6 0.7 0.8 0.9 1.0 Fuzzing score 163264128 0 TopK Switch Matryoshka KronSAE KronSAE + Matryoshka Figure 21: Interpretability scores forF = 2 15 = 32768. KronSAE shows good consistency across different sparsity regimes, in some cases outperforming the TopK baseline. The result is the same as for 65k dictionary size at the Figure 20, but with slightly lower scores. Results show that TopK and KronSAE are very stable across both models and sparsity regimes, while for other models their architectural and training design significantly affect the interpretability. We attribute those differences between Gemma and Qwen to different capacity of residual stream - Gemma has hidden state size of 2304, while Qwen has only 1536 dimensions (1.5x smaller). Hovewer, differences between training data and architectural choices can also be the cause. Interpretability tradeoff.As shown in Section 4.1 KronSAE introduces tradeoff between computa- tional efficiency, explained variance and interpretability. We can force more features to be correlated in same head by increasing themandn, and this will improve computational efficiency, but at the cost of reconstruction performance; however, improving the explained variance (with smallmand fine-grained structure of groups) goes with the cost of slight reduction in the interpretability. These tradeoffs are expected and presented across variety of SAEs (Karvonen et al., 2025). 27 Preprint. To evaluate how KronSAE behave under differentm,nwe compute autointerpretability scores following the same setup as described in Appendix D for Qwen-2.5 1.5B and Gemma-2 2B models with sparsityℓ 0 ∈ 16, 32, 64, 128. As Figure 22 shows, setupm = 1andh = 4096is less interpretable thanm = 2andm = 4with the same number of heads despite having stronger reconstruction performance. 0.65 0.70 0.75 0.80 0.85 0.90 0.95 Gemma 2 2B Explained variance m=1 m=2 m=4 TopK 0.0 0.1 0.2 0.3 0.4 0.5 Mean absorption fraction score 0.5 0.6 0.7 0.8 0.9 1.0 Detection score (h = 4096) 163264128 0 0.65 0.70 0.75 0.80 0.85 0.90 0.95 Qwen 2.5 1.5B 163264128 0 0.0 0.2 0.4 0.6 163264128 0 0.5 0.6 0.7 0.8 0.9 1.0 Figure 22: EV, absorption score and interpretability forF = 2 16 . For constanth, decreasingm leads to lower interepretability scores and absorption scores, but increase EV - this is expected due to increasing polysemanticity of pre-latents and entanglement of the features. See Section 5.3 for a explanation of the retrieval mechanism. FKRONSAE IN TERMS OF TENSOR DIAGRAM The proposed encoder architecture can be visualized as a tensor diagram (Figure 23). Notably, this formulation draws a connection to quantum mechanics, where|f⟩represents the (unnormalized) state of two disentangled qubits described by|p⟩ and|q⟩. If we were to sum the outputs of the encoder’s heads instead of concatenating them,|f⟩would corre- spond to a separable quantum state. This scenario can be expressed via the Schmidt decomposition: |f⟩ = X h |p h ⟩⊗ K |q h ⟩, where⊗ K denotes the Kronecker product. However, preliminary experiments revealed that this alternative design results in poorer performance compared to the concatenation-based approach. GANALYSIS OF COMPOSITIONAL STRUCTURE Here we analyze more examples of interactions in various heads. Head 3. For this head we have selected all base elements and extension 2, shown in Table 8. Extension element 2 shows moderate interpretability with clear AND-like interactions: with base 1 (semantic inheritance through shared pre-latent semantics) and base 2 (retaining only instrument- related semantics). Notable interactions occur with base 0 (acquiring medical semantics while preserving metric/number aspects) and base 3 (combining instrument semantics with necessity to yield therapy/treatment concepts). The high interpretability scores suggest potential additional encoding mechanisms beyond simple intersection, possibly related to activation magnitude, though dataset or interpretation artifacts cannot be ruled out without further validation. 28 Preprint. Figure 23: For a single head, the KronSAE encoder architecture separates the inputxinto two distinct components,pandq, via matrix multiplicationsP k andQ k accordingly, followed byReLU activation. These components are then combined via the Kronecker productp⊗qand square root operation √ · , resulting in an output vector f . ComponentInterpretationScore Extension 2Scientific instruments, acronyms, and critical numerical values in technical and astronomical contexts 0.71 Base elements and their compositions with extension 2 Base 0Interpretation: Punctuation marks and line breaks serving as structural separators in text. 0.66 Composition: Health-related metrics focusing on survival rates, life ex- pectancy, and longevity. 0.88 Base 1Interpretation: Numerical values, both in digit form and as spelled-out numbers, often accompanied by punctuation like decimals or commas, in contexts of measurements, statistics, or quantitative expressions. 0.80 Composition: Numerical digits and decimal points within quantitative values.0.86 Base 2Interpretation: Nuanced actions and adverbial emphasis in descriptive con- texts. 0.71 Composition: Astronomical instruments and their components, such as space telescopes and their acronyms, in scientific and observational contexts. 0.90 Base 3 Interpretation: Forms of the verb "to have" indicating possession, necessity, or occurrence in diverse contexts. 0.91 Composition: Antiretroviral therapy components, viral infection terms, and medical treatment terminology. 0.87 Table 8: Interactions between extension 2 in head 3 and all base elements in that head. Head 136. This head exhibits higher interpretability in post-latents than pre-latents. Key observa- tions from the Table 9 include: extension 2 with base 0 narrows semantics to Illinois (likely inheriting geographical subsemantics), while interactions with bases 2-3 demonstrate complexity beyond simple intersection, often introducing additional semantics requiring deeper investigation. Head 177. Latents presented in Table 10 emonstrates more consistent AND-like behavior than Heads 3 and 136, closely matching the interaction pattern shown in Figure 2. 29 Preprint. ComponentInterpretationScore Extension 2Hierarchical scopes, geographic references, and spatial dispersal terms0.78 Base elements and their compositions with extension 2 Base 0Interpretation: Numerical decimal digits in quantitative expressions and proper nouns. 0.79 Composition: The state of Illinois in diverse contexts with high significance.0.95 Base 1 Interpretation: The number three and its various representations, including digits, Roman numerals, and related linguistic forms. 0.84 Composition: Geographic place names and their linguistic variations in textual contexts. 0.91 Base 2 Interpretation: Ordinal suffixes and temporal markers in historical or chrono- logical contexts. 0.87 Composition: Terms indicating layers, degrees, or contexts of existence or operation across scientific, organizational, and conceptual domains. 0.82 Base 3 Interpretation: Question formats and topic introductions with specific terms like "What", "is", "of", "the", "Types", "About" in structured text segments. 0.77 Composition: Spatial spread and occurrence of species or phenomena across environments. 0.87 Table 9: Interactions between extension 2 in head 136 and all base elements in that head. ComponentInterpretationScore Extension 1Geographical mapping terminology and institutional names, phrases involv- ing spatial representation and academic/organizational contexts 0.90 Base elements and their compositions with extension 1 Base 0Interpretation: Proper nouns, abbreviations, and specific named entities.0.64 Composition: Geographical or spatial references using the term "map".0.93 Base 1Interpretation: Emphasis on terms indicating feasibility and organizations.0.80 Composition: Specific organizations and societal contexts.0.89 Base 2 Interpretation: Institutional names and academic organizations, particularly those containing "Institute" or its abbreviations, often paired with preposi- tions like "of" or "for" to denote specialization or affiliation. 0.89 Composition: Institutional names containing "Institute" as a core term, often followed by prepositions or additional descriptors. 0.92 Base 3Interpretation: Closure and termination processes, initiating actions.0.79 Composition: Initiating or establishing a state, direction, or foundation through action. 0.85 Table 10: Interactions between extension 1 in head 177 and all base elements in that head. 30 Preprint. HKRONSAE SIMPLIFIED IMPLEMENTATION 1 class KronSAE(n.Module): 2 def __init__(self, config): 3 super().__init__() 4 self.config = config 5 _t = torch.n.init.normal_( 6 torch.empty( 7 self.config.act_size, 8 self.config.h * (self.config.m + self.config.n) 9 ) 10 ) / math.sqrt(self.config.dict_size * 2.0) 11 self.W_enc = n.Parameter(_t) 12 self.b_enc = n.Parameter( 13 torch.zeros(self.config.h * (self.config.m + self.config.n)) 14 ) 15 W_dec_v0 = einops.rearrange( # Initialize decoder weights 16 _t.t().clone(), "(h mn) d -> h mn d", 17 h=self.config.h, mn=self.config.m + self.config.n 18 )[:, :self.config.m] 19 W_dec_v1 = einops.rearrange( 20 _t.t().clone(), "(h mn) d -> h mn d", 21 h=self.config.h, mn=self.config.m + self.config.n 22 )[:, self.config.m:] 23 self.W_dec = n.Parameter(einops.rearrange( 24 W_dec_v0[..., None, :] + W_dec_v1[..., None, :, :], 25 "h m n d -> (h m n) d" 26 )) 27 self.W_dec.data[:] = ( 28 self.W_dec.data / self.W_dec.data.norm(dim=-1, keepdim=True) 29 ) 30 self.b_dec = n.Parameter(torch.zeros(self.config.act_size)) 31 32 def encode(self, x: torch.Tensor) -> torch.Tensor: 33 B, D = x.shape 34 acts = F.relu( 35 x @ self.W_enc + self.b_enc 36 ).view(B, self.h, self.m + self.n) 37 all_scores = torch.sqrt( 38 acts[..., :self.config.m, None] * \ 39 acts[..., self.config.m:, None, :] + 1e-5 40 ).view(B, -1) 41 scores, indices = all_scores.topk( 42 self.config.k, dim=-1, sorted=False 43 ) 44 acts_topk = torch.zeros( 45 (B, self.config.dict_size) 46 ).scatter(-1, indices, scores) 47 return acts_topk 48 49 def forward(self, x): 50 acts_topk = self.encode(x) 51 x_rec = acts_topk @ self.W_dec + self.b_dec 52 output = self.get_loss_dict(x, x_rec) 53 return output 54 55 def get_loss_dict(self, x, x_rec): 56 loss = (x_rec - x.pow(2).mean() 57 pt_l2 = (x_rec - x).pow(2).sum(-1).squeeze() 58 var = (x - x.mean(0)).pow(2).sum(-1).squeeze() 59 ev = (1 - pt_l2 / var).mean() 60 return loss, ev 31 Preprint. IUSAGE OF LARGE LANGUAGE MODELS We have used LLMs as the main tool for conducting the interpretability experiments, as described in section D, and as the instrument for language polishing and word choice. 32