Paper deep dive
DePass: Unified Feature Attributing by Simple Decomposed Forward Pass
Xiangyu Hong, Che Jiang, Kai Tian, Biqing Qi, Youbang Sun, Ning Ding, Bowen Zhou
Models: LLaMA-2-13B, LLaMA-2-7B, LLaMA-3.1-70B, LLaMA-3.1-8B, LLaMA-3.2-1B, LLaMA-3.2-3B, Qwen2-1.5B, Qwen2-7B
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 95%
Last extracted: 3/11/2026, 12:31:13 AM
Summary
DePass is a unified framework for feature attribution in Transformer models that uses a single decomposed forward pass to isolate and trace the contributions of specific components (tokens, attention heads, MLP neurons, or subspaces) to model outputs without requiring auxiliary training or model modifications.
Entities (6)
Relation Signals (3)
DePass → attributes → Transformer
confidence 95% · DePass serves as a foundational tool for broader applications in interpretability of Transformer models.
DePass → evaluatedon → Known_1000
confidence 90% · We evaluate on two benchmarks targeting different reasoning types: Known_1000 and IOI.
DePass → evaluatedon → CounterFact
confidence 90% · We evaluate on two factuality benchmarks: CounterFact and TruthfulQA.
Cypher Suggestions (2)
Find all datasets used to evaluate the DePass framework. · confidence 95% · unvalidated
MATCH (d:Dataset)<-[:EVALUATED_ON]-(f:Framework {name: 'DePass'}) RETURN d.nameIdentify the relationship between the framework and the model architecture. · confidence 90% · unvalidated
MATCH (f:Framework {name: 'DePass'})-[r:ATTRIBUTES]->(m:ModelArchitecture) RETURN type(r), m.nameAbstract
Abstract:Attributing the behavior of Transformer models to internal computations is a central challenge in mechanistic interpretability. We introduce DePass, a unified framework for feature attribution based on a single decomposed forward pass. DePass decomposes hidden states into customized additive components, then propagates them with attention scores and MLP's activations fixed. It achieves faithful, fine-grained attribution without requiring auxiliary training. We validate DePass across token-level, model component-level, and subspace-level attribution tasks, demonstrating its effectiveness and fidelity. Our experiments highlight its potential to attribute information flow between arbitrary components of a Transformer model. We hope DePass serves as a foundational tool for broader applications in interpretability.
Tags
Links
PDF not stored locally. Use the link above to view on the source site.
Full Text
75,725 characters extracted from source content.
Expand or collapse full text
DePass: Unified Feature Attributing by Simple Decomposed Forward Pass Xiangyu Hong 1∗ , Che Jiang 1∗ , Kai Tian 1 , Biqing Qi 2 , Youbang Sun 1 , Ning Ding 1,2 , Bowen Zhou 1,2† 1 Department of Electronic Engineering, Tsinghua University 2 Shanghai AI Laboratory hong-xy22@mails.tsinghua.edu.cn jc23@mails.tsinghua.edu.cn zhoubowen@tsinghua.edu.cn Abstract Attributing the behavior of Transformer models to internal computations is a central challenge in mechanistic interpretability. We introduce DePass, a uni- fied framework for feature attribution based on a single decomposed forward pass. DePass decomposes hidden states into customized additive components, then propagates them with attention scores and MLP’s activations fixed. It achieves faithful, fine-grained attribution without requiring auxiliary training. We validate DePass across token-level, model component-level, and subspace- level attribution tasks, demonstrating its effectiveness and fidelity. Our exper- iments highlight its potential to attribute information flow between arbitrary components of a Transformer model. We hope DePass serves as a founda- tional tool for broader applications in interpretability. Code is available at https://github.com/TsinghuaC3I/Decomposed-Forward-Pass 1 Introduction Mechanistic interpretability is the foundation to monitor, modify, and predict Transformer-based models’ behavior. The first step of such reverse engineering is to decompose the neural network and analyze what contributes to the model’s behavior [1,2]. Researchers continue to develop methods for decomposing this highly complex system. Without modifying or abstracting the network, directly applying noise ablations [3,4] or activation patching [5–7] to all modules is computationally expensive and provides limited insight into intermediate information flow [8]. Gradient-based attribution methods also face theoretical challenges [9]. In contrast, approximating or abstracting the model can partially align with human cognition [10–13], but often fails to reach fine-grained components such as neurons or attention heads. Moreover, non-conservative approximations may compromise the faithfulness of attribution. In this paper, we propose DePass (Decomposed Forward Pass), a direct and unified feature attribution framework that addresses these shortcomings through a single decomposed forward pass. The main concept of DePass is simple: we break down every hidden state into additive components, propagate these components through the remaining layers, and then obtain each component’s exact contribution to the target representation. In the decomposed forward pass, attention scores and MLP activations are fixed, and weighted contributions are assigned based on the decomposed components. This method guarantees several advantages: •Faithfulness and Completeness: By freezing attention scores and MLP activations, DePass elimi- nates second-order effects when propagating forward [8]. As a result, per-component summation reconstructs exactly the hidden state of the original model. ∗ Equal contribution. † Corresponding author. 39th Conference on Neural Information Processing Systems (NeurIPS 2025). arXiv:2510.18462v2 [cs.CL] 24 Oct 2025 •Unification Across Components: DePass provides a unified attribution framework for multiple model granularities—including input tokens, attention heads, neurons, and subspaces of the residual stream. This enables consistent, fine-grained interpretability without modifying the model or using task-specific approximations. •Representation-Level Attribution: Unlike conventional attribution scores [11], DePass tracks how decomposed representations evolve through the forward pass. This facilitates more fine-grained attributions and a natural interface to align with human reasoning and sparse dictionary learning (SDL) such as SAE. We conduct attribution experiments at the token level, model component level, and subspace level to evaluate the effectiveness and generality of DePass. Our results show that DePass enables lossless, additive decomposition of hidden states throughout the forward pass of Transformer models, allowing faithful tracking of information flow according to attribution needs. This unified framework offers a new analytical tool for mechanistic interpretability. 2 Architecture of Transformer-Based Autoregressive Language Models Transformer-based autoregressive language models use the residual stream as the backbone for information processing. In DePass, the decomposed components follow the same pathway as the original hidden states, passing through each decoder layer’s multi-head self-attention (MHSA) sublayer, feedforward network (MLP) sublayer, and layer normalization (LayerNorm) operations. In this section, we establish the notations used in the standard forward pass, laying the groundwork for the formal definition of the DePass method. 2.1 Transformer Decoder Layers LetX ℓ ∈R N×D denote the token representations at layerℓ, whereNis the sequence length andD is the dimension of model’s hidden state. Multi-Head Self-Attention (MHSA) The multi-head self-attention (MHSA) mechanism enables each token to attend to others in the sequence, capturing contextual dependencies. At layerℓ, the input X ℓ−1 ∈R N×D is first normalized by layer normalization, ̃ X ℓ = LN(X ℓ−1 ). For each headj = 1,...,H, queries, keys, and values are computed by linear projections of ̃ X ℓ , and attention scores are obtained via scaled dot-product attention with a causal mask M ℓ,j : A ℓ,j = Softmax Q ℓ,j (K ℓ,j ) ⊤ p D/H + M ℓ,j ! .(1) Letting W ℓ,j V O = W ℓ,j V W ℓ,j O ∈R D×D , the output of all attention heads is: Output ℓ attn = H X j=1 O ℓ,j ,where O ℓ,j = A ℓ,j ̃ X ℓ W ℓ,j V O .(2) Finally, the residual connection is applied: X ℓ attn = X ℓ−1 + Output ℓ attn .(3) Feedforward Network (MLP) After the attention sublayer, hidden states are passed through a position-wise feedforward network to enhance per-token representation capacity. In particular, a LayerNorm is applied to the attention output, and then the MLP operates independently at each position i: Output ℓ ffn = W ℓ D σ W ℓ U LN(X ℓ attn ) ,(4) whereW ℓ U ∈R d ffn ×D ,W ℓ D ∈R D×d ffn , andσ(·)is a nonlinear activation function. MLPs can be regarded as a soft key-value retrieval mechanism [14]. Let W ℓ U = f ℓ 1 . . . f ℓ N ∈R N×D , W ℓ D = v ℓ 1 · v ℓ N ∈R D×N ,(5) 2 where f ℓ k ∈R D and v ℓ k ∈R D represent the k-th subkey and subvalue. The output at position i is: Output ℓ ffn,i = N X k=1 m ℓ i,k v ℓ k , m ℓ i,k = σ f ℓ k · ̃ X ℓ attn,i ,(6) where ̃ X ℓ attn,i = LN(X ℓ attn,i ) acts as a query vector over the subkeys. Finally, the residual connection yields the output for the next layer: X ℓ+1 = X ℓ attn + Output ℓ ffn .(7) 2.2 Language Modeling Head At positioni, the final hidden stateX L i ∈R D is normalized and projected by language modeling head W lm ∈R V×D , whereVis the vocabulary size. The resulting logits yield the next-token distribution: P (w i+1 | w ≤i ) = Softmax W lm LN(X L i ) ,(8) where w i+1 is the predicted token at the next position.. 3 Decomposed Forward Pass We introduce DePass (Decomposed Forward Pass), a method for isolating and tracing representational components through Transformer layers. This section outlines: (1) the initialization of decomposed hidden states; (2) the reformulated forward pass for propagating these components; and (3) the use of decomposed states for decoding and attribution. Throughout this paper, we use the notation X[i 1 ,i 2 ,... ]to index into multidimensional tensors following the NumPy-style indexing convention. 3.1 Initialization of Decomposed Hidden States Given a hidden stateX (ℓ) ∈R N×D at layerℓ, withNtokens and hidden dimensionD, we define its decomposed formX (ℓ) dec ∈R N×M×D , whereMis the number of decomposition components. Each token representation is the sum of its components: X (ℓ) i = M X m=1 X (ℓ) dec [i,m, :].(9) The index m is shared across positions to ensure semantic alignment, and its interpretation depends on task-specific decomposition strategies. 3.2 Propagating Decomposed Hidden States Through Transformer Blocks We describe how decomposed hidden states propagate through LayerNorm, Multi-Head Self-Attention (MHSA), and MLP modules in a Transformer block. LayerNormSince most layer normalization performs feature-wise scaling and shifting, we illustrate here how the more efficient RMSNorm—used in our experimental models—operates on decomposed hidden states. Given token i’s hidden state X (ℓ) i = P M m=1 X (ℓ) dec [i,m, :], RMSNorm computes: RMSNorm(X (ℓ) i ) = γ⊙ X (ℓ) i RMS(X (ℓ) i ) ,where RMS(X (ℓ) i ) = r 1 d X (ℓ) i 2 .(10) The scaling operation in RMSNorm is additive over decomposed components, allowing the scaling factor to be applied independently to each component, so it distributes over components as: ̃ X (ℓ) dec [i,m, :] = A i X (ℓ) dec [i,m, :],where A i = diag γ RMS(X (ℓ) i ) ! .(11) 3 Multi-Head Self-Attention (MHSA) on Decomposed Hidden States LetX (ℓ−1) dec ∈R N×M×D represent the decomposed hidden states from the previous layer. To apply multi-head self-attention in a component-aware manner, we extend the standard attention formulation (see Eq. 2) to operate on each component individually. For each attention headj, the attention output for tokenifrom component m is: o (ℓ,j) i,m = A (ℓ,j) X (ℓ−1) dec [i,m, :]W (ℓ,j) V O ,(12) whereA (ℓ,j) is the original attention scores. The MHSA output is aggregated across all heads and components, combined with residuals: X (ℓ) attn,dec [i,m, :] = X (ℓ−1) dec [i,m, :] + H X j=1 o (ℓ,j) i,m .(13) MLP on Decomposed Hidden States The MLP module processes decomposed representations X (ℓ) attn,dec ∈R N×M×D by distributing each neuron’s output across components. For each tokeni, component m, and neuron k, we compute a relevance score: a i,m,k = f ℓ k ⊤ X (ℓ) attn,dec [i,m, :], α i,m,k = exp(a i,m,k ) P M m ′ =1 exp(a i,m ′ ,k ) ,(14) whereα i,m,k determines how neuronk’s output is apportioned to componentm. Alternative normal- ization methods to softmax are compared in Appendix A. The updated decomposed hidden state is then: X (ℓ+1) dec [i,m, :] = X (ℓ) attn,dec [i,m, :] + D mlp X k=1 α i,m,k · m ℓ i,k · v ℓ k ,(15) wherem ℓ i,k is the neuron activation andv ℓ k ∈R D is the output projection. The original hidden state is recovered by summing over components: X (ℓ+1) i = M X m=1 X (ℓ+1) dec [i,m, :].(16) Given the decomposed hidden states from the previous layer, each component can be further prop- agated through the next layer independently. This maintains disentangled contributions across components and allows exact reconstruction of the original hidden state. All computations are fully parallelizable, incurring minimal overhead. 3.3 LM Head and Attribution Score LetX (L) i = P M m=1 X (L) dec [i,m, :] be the final hidden state of tokeni, and letw y = W LM [y, :]be the LM head vector for target y. The output logit and component-wise attribution are: logits y = w ⊤ y X (L) i ,∆logits y,m = w ⊤ y X (L) dec [i,m, :],(17) with P M m=1 ∆logits y,m = logits y . This yields fine-grained attribution across components. To attribute with respect to a given subspace at a particular layer, we project eachX (ℓ) dec [i,m, :]onto its direction and use the projection values as scores. 4 DePass Attribution across multi-granular Levels 4.1 Token-Wise DePass Token-Wise Decomposed Hidden States Initialization To analyze how each input token con- tributes to the model’s hidden states, we initialize a token-wise decomposition of the hidden states: each token’s hidden state is split intoNadditive components, one per input token. At the embedding layer, the decomposition is defined as: X (0) dec [i,m, :] = X (0) [i, :], if i = m 0,otherwise (18) 4 whereX (0) ∈R N×d is the input embedding andX (0) dec ∈R N×N×d is its token-level decomposition. This structure is propagated layer by layer according to the method in Section 3.2. At any layer, the m-th component of the decomposed hidden state represents the contribution of them-th input token to the overall hidden states. 4.1.1 Token-Level Output Attribution via DePass Problem Definition.Given a modelM, inputx = [x 1 ,...,x n ], and outputˆy =M(x), token-wise attribution assigns each tokenx i a scores i ∈Rindicating its influence onˆy—with higher scores denoting greater impact. DePass-Based Output Attribution.Starting from token-wise decomposition initialization (Eq. 18) and applying the forward pass decomposition (Section 3.2), we use the language modeling head (Section 3.3) on the final decomposed hidden states. This yields token-level attribution scores, quantifying how each input token contributes to the model’s prediction of output y. Experiment Setup. Baselines. We compare against standard attribution methods on a fixed pretrained model. Gradient-based: Input×Gradient [15], Integrated Gradients [16], Gradient SHAP [17]; Attention-based: Mean Attention, Last-layer Attention [18], and Attention Rollout [6]. Tasks. We evaluate on two benchmarks targeting different reasoning types: Known_1000 [4] 3 (factual QA, e.g., "Audible.com is owned by") and IOI [19] (indirect object identification, e.g., "Eleanor and Deanna were thinking about going to the mountain. Eleanor wanted to give a watermelon to"). Evaluation Protocol. For each inputx, we first compute attribution scores for the correct answer via various methods. Based on these scores, we apply token-level interventions: patch top—mask the topK%tokens with highest attribution; recover top—mask the bottom(100−K)%tokens (lowest attribution) and then restore the top K%. The remaining tokens are reassembled into a new prompt and fed to the model. Faithfulness is then evaluated by measuring the change in the predicted probability of the correct answer: Comprehen- siveness: Drop in probability under patch top (higher is better); Sufficiency: Retained probability under recover top (lower is better). We compute the relative change in predicted probability as: ∆p (K) = p(ˆy | x)− p(ˆy | ̃x (K) ) p(ˆy | x) , whereˆydenotes the target token and ̃x (K) denotes the perturbed input under each intervention. The figures report the average ∆p (K) across all data points per dataset at each masking level. To ensure a fair comparison with other token-level baselines, in the experiments corresponding to Figure 1, we perform ranking and ablation at the subword token level. DePass also supports word-level attribution, with examples illustrated in Appendix B.3. Results.As shown in Figure 1, our method yields substantially higher comprehensiveness scores on both Known_1000 and IOI benchmarks using Llama-2-13b-chat-hf, indicating that the tokens identified by DePass are more critical to the model’s prediction. In terms of sufficiency, all methods are comparable. This is partly due to evaluation limits: when keeping only top tokens, most methods recover those key for prediction. Attention-based methods, which preserve broad semantic structure, perform slightly better. A representative case of output attribution is shown in the top of Figure 2. Additional results on other model variants and experiment details are provided in Appendix B. 4.1.2 Token-level Subspace Attribution via DePass While methods such as probing or sparse autoencoders (SAEs) are able to uncover meaningful subspaces, they fall short of directly identifying which input tokens are responsible for activating 3 Dataset can be found at: https://rome.baulab.info/data/dsets/known_1000.json 5 Figure 1: Faithfulness evaluation on Known_1000 and IOI using Llama-2-13b-chat-hf. Our method yields better comprehensiveness and competitive sufficiency. those subspaces. Existing approaches that attempt to link tokens with feature activations [20] often depend on intricate graph construction and pruning, making them computationally expensive and difficult to scale. To address this limitation, we introduce DePass as a framework that attributes subspace activations to specific input tokens in a more direct and efficient manner. DePass-Based Subspace Attribution.Given a semantic directionv∈R D and decomposed hidden states X (ℓ) dec ∈R N×N×D at layer ℓ, the subspace activation at position i is: a i = v ⊤ X (ℓ) i = N X m=1 v ⊤ X (ℓ) dec [i,m, :],(19) where each term quantifies the contribution of tokenx m to the subspace activation at positioni. This enables fine-grained attribution to arbitrary linear subspaces. Experiment: Probing Subspace Attribution via DePassWe evaluate our method in the context of factuality, where prior work has shown that truthfulness can be linearly separated in hidden states [21]. A linear probef (x) = w ⊤ x +bis trained to detect untruthful activations, withwdefining a factuality subspace. Some methods leverage such probes to mask misleading input tokens based on whether their intermediate hidden states align with untruthful directions [22]. In contrast, our method attributes subspace activations directly to specific input tokens, enabling more fine-grained interventions such as selectively masking misleading content. Evaluation Setup. We evaluate on two factuality benchmarks: CounterFact[4] (modified for generation, prompting completions to factual statements) and TruthfulQA[23] (converted to multiple- choice with misleading options). We compare the effectiveness of DePass and TACS[22] in identifying misleading tokens that con- tribute to model errors. Each example is tested under four settings: (1) No Information: given only the question, (2) Misinformation: given wrong information along with the question, (3) Mis- information + TACS Masking [22]: tokens aligned with untruthful directions are masked, and (4) Misinformation + Ours (DePass Masking): tokens contributing most to untruthful activation are masked. For each prompt, both methods mask the same number of tokens based on classifier confidence. Results.As shown in Table 1, DePass-based masking consistently improves factual accuracy across models and datasets. Compared to direct probing-based masking, our method yields stronger gains. For example, on Llama-2-7b-chat-hf, accuracy rises from 10.16% (misinformation) to 43.13% (ours), demonstrating DePass’s ability to isolate and suppress harmful inputs. A representative case of subspace attribution is shown in the bottom of Figure 2. See Appendix C for details on dataset formatting, probe training, prompt examples, and our strategy for selecting masked tokens based on classifier outputs from different layers. We also provide examples of applying DePass for SAE feature attribution to illustrate the potential of combining the two methods (Appendix C.5). 6 Table 1: Accuracy (%) across different input settings on CounterFact and TruthfulQA. Masking is applied to 30% of misleading tokens identified using either prior methods or our DePass-based attribution. ModelDatasetNo InfoMisinformation+ TACS Masking+ Ours (DePass Masking) Llama-2-7b-chat-hf CounterFact (Gen)57.5210.1625.6843.13 TruthfulQA (MC)66.1033.0543.5746.51 Llama-2-13b-chat-hf CounterFact (Gen)60.684.8813.0634.90 TruthfulQA (MC)73.5638.9249.3353.00 Llama-3.1-8b-Instruct CounterFact (Gen)60.633.3016.0359.16 TruthfulQA (MC)83.6070.2671.4876.62 Qwen2-1.5B-Instruct CounterFact (Gen)44.043.5117.7543.32 TruthfulQA (MC)71.2446.1460.1051.90 Qwen2-7B-Instruct CounterFact (Gen)39.436.5423.9129.68 TruthfulQA (MC)77.2347.2568.4264.87 Meta-Llama-3.1-70B-Instruct CounterFact (Gen)72.209.5133.2955.62 TruthfulQA (MC)89.9665.6174.5476.87 Figure 2: Two input-level attribution examples. The top case illustrates contributions to the model prediction "Germany", with higher scores indicating greater influence. The bottom case shows token-wise contributions to the classifier’s prediction of the "truthful" label, where more negative scores support the untruthful classification (label 0). 4.2 Model Component-Wise DePass Model Component-Wise Decomposed Hidden States InitializationTo attribute model behavior to specific architectural components, we perform component-level decomposition of hidden states, targeting attention heads and MLP neurons. Attention Head: For a Transformer layer withHattention heads, we assign one decomposition component to each head and one to the residual connection, resulting inM = H + 1. For tokeni, we define: X (ℓ) dec [i,h, :] = o (ℓ,h) i , h = 1,...,H; X (ℓ) dec [i,H + 1, :] = X (ℓ−1) [i, :](20) where o (ℓ,h) i is the output of the h-th attention head at layer ℓ. MLP Neurons: Similarly, for an MLP block withN MLP hidden neurons, we decompose the output and residual connection into N MLP + 1 components: X (ℓ) dec [i,n, :] = m ℓ i,n v ℓ n , X (ℓ) dec [i,N MLP + 1, :] = X ℓ attn [i, :] where m ℓ i,n v ℓ n ∈R D is the contribution of the n-th MLP neuron to the hidden states of token i. This setup aligns with our decomposition-based forward pass and enables localized interpretability. Component-wise importance scores for a given output are computed as described in §3.3. Experiment: Evaluating Component Importance via Masking. We validate the functional significance of components using masking-based ablations guided by various importance metrics. Importance Scoring Methods. We compare several scoring strategies: Norm:ℓ 2 norm of each component’s activation; Coef: Absolute activations after MLP up-projection and nonlinearity; used only for MLP neurons. AtP [24] (Activation×Gradient): A gradient-based importance measure using first-order Taylor approximation; DePass (Ours): Attribution scores derived from decomposed hidden states; DePass-Abs: Absolute values of DePass scores, capturing both supportive and suppressive contributions. 7 Tasks and Setup.Experiments are conducted on: IOI (Indirect Object Identification): Synthetic reasoning benchmark; CounterFact (QA): Factual knowledge recall task. Only correctly answered examples are used for meaningful attribution. Evaluation Protocol. We perform two complementary masking interventions: Top-kMasking (Comprehensiveness): Mask the top-kcomponents, with a sharp accuracy drop indicating critical components. Bottom-kMasking (Sufficiency): Mask only the bottom-kcomponents; high accuracy retention suggests sufficiency for prediction. Masking is applied structurally: attention heads are ablated by zeroing their output projectionsW (ℓ,h) V O , and MLP neurons by zeroing their activations before projection. We report average accuracy across examples and for a robust attribution quality assessment. ResultsFigure 3 presents the results on Llama-2-7b-chat-hf; additional results for other models are provided in the Appendix D. Our method consistently outperforms baseline attribution techniques across different masking strategies. In particular, it achieves a more pronounced accuracy drop in Top-kMasking and better accuracy retention in Bottom-kMasking, indicating its superior ability to identify and attribute critical components. Figure 3: Performance of our method compared to baseline attribution techniques on Llama-2-7b-chat- hf under various masking strategies. Our approach more accurately identifies critical components, as reflected by the sharper drop in Top-k Masking and stronger performance in Bottom-k Masking. 4.3 Subspace-Level Decomposition Subspace-Wise Decomposition of Hidden States To analyze how specific representational di- rections affect model behavior, we decompose hidden states at layerℓinto components within and orthogonal to a target subspace. Given a projection matrixP t ∈R D×D , each token’s hidden state X (ℓ) [i, :]∈R d is split as: X (ℓ) dec [i, 0, :] = P t X (ℓ) [i, :], X (ℓ) dec [i, 1, :] = (I− P t )X (ℓ) [i, :],(21) whereIis the identity matrix. These components are then independently propagated using DePass (Section 3.2), allowing attribution of model behavior to the subspace. Construction details ofP t are provided in Appendix E.2. Experiment: Interpreting Language Subspace EffectsWe examine whether the decomposition framework can reveal functionally distinct subspaces in multilingual settings. Specifically, we hypothesize that model hidden states can be separated into a language subspace capturing language signals, and a semantic subspace encoding language-invariant meaning[25, 26]. Subspace Construction.We apply DePass to Llama-3.1-8B-Instruct, a multilingual model trained on English, French, German, and more. We train a token-level language classifier on CounterFact [4], translated into multiple languages, to detect the language of each input token. The classifier weights define a language-diagnostic subspace, and its orthogonal complement serves as the semantic subspace. Token hidden states are projected into these two subspaces as described in Section 4.3, and decomposition is propagated forward through the model. 8 Evaluation Setup. To evaluate whether DePass faithfully attributes model behavior to distinct subspaces, we begin by projecting hidden states at an intermediate layer (layer 15) into the language and semantic subspaces. These decomposed components are then independently propagated through the remaining layers of the model, yielding two final representations at the last layer:X dec lang [i] = X (L) dec [i, 0, :] and X dec sem [i] = X (L) dec [i, 1, :]. We decode these representations separately by applying the language modeling head to each of them. Table 2 reports the top-5 tokens generated fromX dec lang andX dec sem for multilingual prompts (e.g., “What is Thomas Joannes Stieltjes’s native language? It is”). Results. Tokens generated from the semantic subspace (X dec sem ) consistently reflect factual content (e.g., “Dutch”, “Holland”), largely invariant to the input language. In contrast, the language subspace (X dec lang ) produces lexical or stylistic tokens (e.g., “né”, “nicht”, “de”) aligned with the prompt’s language. These results show that DePass faithfully propagates and preserves the functional roles of each subspace, enabling clear attribution of linguistic and semantic behavior and highlighting its potential for subspace-level analysis. To further verify that DePass faithfully attributes language-related behavior, we apply t-SNE toX dec lang from multilingual inputs. As shown in Figure 4, the resulting clusters align with language identity, confirming that DePass effectively preserves and propagates language-specific signals. These results highlight its effectiveness for subspace-level attribution. Additional examples are in Appendix E. Figure 4: t-SNE visualization of token- wise projections onto the language subspace. Distinct clusters indicate strong language- specific structure. LanguageLanguage Tokens Englisha, the, an, not, N Frenchné, consid, de, conn, ét Germannicht, keine, eine, die, das LanguageSemantic Tokens EnglishDutch, dut, reported, said, Afrika FrenchDutch, holland, Holland, of, N GermanDutch, Holl, Hol, Holland, also Table 2: Top-5 tokens decoded from thelanguage andsemanticsubspaces for different multilingual prompts. 5 Related Work Decomposition-based Attribution. Without modifying or approximating Transformer modules, decomposing hidden states can partially explain model behavior [27]. Decomposition of MLPs and attention blocks at the logit level enables attribution of their residual contributions to model’s output [28–31]. Additionally, attention scores have been used to construct information flow graphs that trace input tokens’ influence [6,11,32]. Combining decomposition with gradient information can further highlight salient features [15,17,16,33]. DecompX [34] applies similar decomposition ideas in restricted settings; our approach generalizes this to arbitrary modules in generative Transformers with finer-grained decomposition of MLP activations. Compared to these methods, DePass performs a more direct additive decomposition of the computation, avoiding early mapping to scalar saliency scores and thereby preserving both faithfulness and interpretability. Ablation-based Attribution. Another line of work attributes importance by ablating specific compo- nents—such as adding noise or zeroing out input tokens or internal activations—and measuring the resulting change in the output distribution [4,3,35–37]. Since exhaustively ablating all components incurs high computational cost, approximate methods or surrogate models have been proposed to accelerate activation patching [38,12,39]. In contrast, DePass provides more faithful attribution through direct decomposition, avoiding the indirectness and potential artifacts of ablation-based approaches. 9 Sparse-Dictionary-Learning-based Attribution. Sparse Dictionary Learning (SDL) is currently the most popular decomposition methods in interpretability. Based on the superposition hypothesis [40,41], SDL aims to recover more interpretable components than the original feature dimensionality. SAE-based methods supervise the reconstruction of current activations [42–45], while Transcoder [20] targets the reconstruction of next-layer activations, and CrossCoder[46] extends this to jointly reconstruct activations across multiple layers. However, the high training cost [1], annotation effort, and reconstruction errors [47] limit their scalability. Some approaches aim to track feature evolution during forward propagation [48–50], but they often rely on proxies like inter-feature correlation or similarity scores rather than precisely tracing the actual transformations of features. Nevertheless, SAE remains valuable for improving downstream task performance (such as model steering [51]) and can be used effectively in conjunction with DePass to bridge model representations and human- interpretable concepts. 6 Conclusions In this paper, we present DePass, a simple yet efficient framework for interpreting Transformer models via decomposed forward pass. By freezing and allocate attention scores and MLP activations, DePass enables lossless additive decomposition, and can be applied to any Transformer-based architecture. DePass achieves more faithful attribution across different levels of granularity comparing to other methods. We hope DePass serves as a general-purpose tool for mechanistic interpretability and inspires broader adoption and diverse applications across the community. 7 Acknowledgements This work is supported by the National Science and Technology Major Project (2023ZD0121403), Young Elite Scientists Sponsorship Program by CAST (2023QNRC001), National Natural Science Foundation of China (No.62406165), Shanghai Municipal Science and Technology Major Project, and the Beijing Natural Science Foundation Undergraduate “Qiyan Research Program” (No. QY24259). We extend our gratitude to the anonymous reviewers for their insightful feedback, which has greatly contributed to the improvement of this paper. References [1]L. Sharkey, B. Chughtai, J. Batson, J. Lindsey, J. Wu, L. Bushnaq, N. Goldowsky-Dill, S. Heimer- sheim, A. Ortega, J. Bloom et al., “Open problems in mechanistic interpretability,” arXiv preprint arXiv:2501.16496, 2025. [2]J. Ferrando, G. Sarti, A. Bisazza, and M. R. Costa-Jussà, “A primer on the inner workings of transformer-based language models,” arXiv preprint arXiv:2405.00208, 2024. [3]M. R. G. Madani, A. P. Gema, G. Sarti, Y. Zhao, P. Minervini, and A. Passerini, “Noiser: Bounded input perturbations for attributing large language models,” arXiv preprint arXiv:2504.02911, 2025. [4] K. Meng, D. Bau, A. Andonian, and Y. Belinkov, “Locating and editing factual associations in GPT,” Advances in Neural Information Processing Systems, vol. 35, 2022. [5]H. Chen, C. Vondrick, and C. Mao, “Selfie: Self-interpretation of large language model embeddings,” in International Conference on Machine Learning.PMLR, 2024, p. 7373– 7388. [6]S. Abnar and W. Zuidema, “Quantifying attention flow in transformers,” in Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, D. Jurafsky, J. Chai, N. Schluter, and J. Tetreault, Eds.Online: Association for Computational Linguistics, Jul. 2020, p. 4190–4197. [Online]. Available: https://aclanthology.org/2020.acl-main.385/ [7] Y. Zhang, S. Li, J. Liu, P. Yu, Y. R. Fung, J. Li, M. Li, and H. Ji, “Knowledge overshadowing causes amalgamated hallucination in large language models,” arXiv preprint arXiv:2407.08039, 2024. 10 [8]A. Makelov, G. Lange, A. Geiger, and N. Nanda, “Is this the subspace you are looking for? an interpretability illusion for subspace activation patching,” in The Twelfth International Conference on Learning Representations, 2024. [9] B. Bilodeau, N. Jaques, P. W. Koh, and B. Kim, “Impossibility theorems for feature attribution,” Proceedings of the National Academy of Sciences, vol. 121, no. 2, p. e2304406120, 2024. [10]S. Yang, S. Huang, W. Zou, J. Zhang, X. Dai, and J. Chen, “Local interpretation of transformer based on linear decomposition,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, p. 10 270–10 287. [11]J. Ferrando, G. I. Gállego, and M. R. Costa-jussà, “Measuring the mixing of contextual information in the transformer,” in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Y. Goldberg, Z. Kozareva, and Y. Zhang, Eds. Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, Dec. 2022, p. 8698–8714. [Online]. Available: https://aclanthology.org/2022.emnlp-main.595/ [12] B. Cohen-Wang, H. Shah, K. Georgiev, and A. Madry, “Contextcite: Attributing model genera- tion to context,” Advances in Neural Information Processing Systems, vol. 37, p. 95 764–95 807, 2024. [13]E. Ameisen, J. Lindsey, A. Pearce, W. Gurnee, N. L. Turner, B. Chen, C. Citro, D. Abrahams, S. Carter, B. Hosmer, J. Marcus, M. Sklar, A. Templeton, T. Bricken, C. McDougall, H. Cunningham, T. Henighan, A. Jermyn, A. Jones, A. Persic, Z. Qi, T. Ben Thompson, S. Zimmerman, K. Rivoire, T. Conerly, C. Olah, and J. Batson, “Circuit tracing: Revealing computational graphs in language models,” Transformer Circuits Thread, 2025. [Online]. Available: https://transformer-circuits.pub/2025/attribution-graphs/methods.html [14] M. Geva, R. Schuster, J. Berant, and O. Levy, “Transformer feed-forward layers are key-value memories,” in Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing, 2021, p. 5484–5495. [15]M. Denil, A. Demiraj, and N. De Freitas, “Extraction of salient sentences from labelled documents,” arXiv preprint arXiv:1412.6815, 2014. [16]M. Sundararajan, A. Taly, and Q. Yan, “Axiomatic attribution for deep networks,” in Proceedings of the 34th International Conference on Machine Learning-Volume 70, 2017, p. 3319–3328. [17]S. M. Lundberg and S.-I. Lee, “A unified approach to interpreting model predictions,” in Neural Information Processing Systems, 2017. [18]S. Jain, S. Wiegreffe, Y. Pinter, and B. C. Wallace, “Learning to faithfully rationalize by construction,” in Annual Meeting of the Association for Computational Linguistics, 2020. [19]K. R. Wang, A. Variengien, A. Conmy, B. Shlegeris, and J. Steinhardt, “Interpretability in the wild:a circuit for indirect object identification in GPT-2 small,” in The Eleventh International Conference on Learning Representations, 2023. [Online]. Available: https://openreview.net/forum?id=NpsVSN6o4ul [20]J. Dunefsky, P. Chlenski, and N. Nanda, “Transcoders find interpretable llm feature circuits,” arXiv preprint arXiv:2406.11944, 2024. [21] K. Li, O. Patel, F. Viégas, H. Pfister, and M. Wattenberg, “Inference-time intervention: Eliciting truthful answers from a language model,” Advances in Neural Information Processing Systems, vol. 36, p. 41 451–41 530, 2023. [22]T. Yu, S. Zhang, and Y. Feng, “Truth-aware context selection: Mitigating hallucinations of large language models being misled by untruthful contexts,” in Findings of the Association for Computational Linguistics ACL 2024, 2024, p. 10 862–10 884. [23]S. Lin, J. Hilton, and O. Evans, “TruthfulQA: Measuring how models mimic human falsehoods,” in Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), S. Muresan, P. Nakov, and A. Villavicencio, Eds. Dublin, Ireland: Association for Computational Linguistics, May 2022, p. 3214–3252. [Online]. Available: https://aclanthology.org/2022.acl-long.229/ 11 [24]A. Syed, C. Rager, and A. Conmy, “Attribution patching outperforms automated circuit discovery,” in Proceedings of the 7th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, Y. Belinkov, N. Kim, J. Jumelet, H. Mohebbi, A. Mueller, and H. Chen, Eds.Miami, Florida, US: Association for Computational Linguistics, Nov. 2024, p. 407–416. [Online]. Available: https://aclanthology.org/2024.blackboxnlp-1.25/ [25]A. Conneau, G. Kruszewski, G. Lample, L. Barrault, and M. Baroni, “What you can cram into a single $&!#* vector: Probing sentence embeddings for linguistic properties,” in Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), I. Gurevych and Y. Miyao, Eds. Melbourne, Australia: Association for Computational Linguistics, Jul. 2018, p. 2126–2136. [Online]. Available: https://aclanthology.org/P18-1198/ [26]G. Alain and Y. Bengio, “Understanding intermediate layers using linear classifier probes,” arXiv preprint arXiv:1610.01644, 2016. [27]L. Arras, B. Puri, P. Kahardipraja, S. Lapuschkin, and W. Samek, “A close look at decomposition- based xai-methods for transformer language models,” arXiv preprint arXiv:2502.15886, 2025. [28]J. Ferrando, G. I. Gállego, I. Tsiamas, and M. R. Costa-jussà, “Explaining how transformers use context to build predictions,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, p. 5486–5513. [29]M. Geva, A. Caciularu, K. Wang, and Y. Goldberg, “Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space,” in Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Y. Goldberg, Z. Kozareva, and Y. Zhang, Eds.Abu Dhabi, United Arab Emirates: Association for Computational Linguistics, Dec. 2022, p. 30–45. [Online]. Available: https://aclanthology.org/2022.emnlp-main.3/ [30]J. Ferrando and E. Voita, “Information flow routes: Automatically interpreting language models at scale,” in Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, 2024, p. 17 432–17 445. [31]R. Achtibat, S. M. V. Hatefi, M. Dreyer, A. Jain, T. Wiegand, S. Lapuschkin, and W. Samek, “Attnlrp: Attention-aware layer-wise relevance propagation for transformers,” in Forty-first International Conference on Machine Learning, 2024. [32]G. Kobayashi, T. Kuribayashi, S. Yokoi, and K. Inui, “Analyzing feed-forward blocks in transformers through the lens of attention maps,” in 12th International Conference on Learning Representations, ICLR 2024, 2024. [33]A. Ali, T. Schnake, O. Eberle, G. Montavon, K.-R. Müller, and L. Wolf, “Xai for transformers: Better explanations through conservative propagation,” in International conference on machine learning. PMLR, 2022, p. 435–451. [34]A. Modarressi, M. Fayyaz, E. Aghazadeh, Y. Yaghoobzadeh, and M. T. Pilehvar, “Decompx: Explaining transformers decisions by propagating token decomposition,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, p. 2649–2664. [35]H. Mohebbi, W. Zuidema, G. Chrupała, and A. Alishahi, “Quantifying context mixing in transformers,” in Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics, A. Vlachos and I. Augenstein, Eds. Dubrovnik, Croatia: Association for Computational Linguistics, May 2023, p. 3378–3400. [Online]. Available: https://aclanthology.org/2023.eacl-main.245/ [36]A. Syed, C. Rager, and A. Conmy, “Attribution patching outperforms automated circuit discov- ery,” in Proceedings of the 7th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, 2024, p. 407–416. [37]B.-D. Oh and W. Schuler, “Token-wise decomposition of autoregressive language model hidden states for analyzing model predictions,” in Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2023, p. 10 105–10 117. 12 [38]J. Kramár, T. Lieberum, R. Shah, and N. Nanda, “Atp*: An efficient and scalable method for localizing llm behaviour to components,” CoRR, 2024. [39]Y.-S. Chuang, B. Cohen-Wang, S. Z. Shen, Z. Wu, H. Xu, X. V. Lin, J. Glass, S.-W. Li, and W.-t. Yih, “Selfcite: Self-supervised alignment for context attribution in large language models,” arXiv preprint arXiv:2502.09604, 2025. [40]W. Gurnee, N. Nanda, M. Pauly, K. Harvey, D. Troitskii, and D. Bertsimas, “Finding neurons in a haystack: Case studies with sparse probing,” arXiv preprint arXiv:2305.01610, 2023. [41]N. Elhage, T. Hume, C. Olsson, N. Schiefer, T. Henighan, S. Kravec, Z. Hatfield-Dodds, R. Lasenby, D. Drain, C. Chen et al., “Toy models of superposition,” arXiv preprint arXiv:2209.10652, 2022. [42]H. Cunningham, A. Ewart, L. Riggs, R. Huben, and L. Sharkey, “Sparse autoencoders find highly interpretable features in language models,” arXiv preprint arXiv:2309.08600, 2023. [43]L. Gao, T. D. la Tour, H. Tillman, G. Goh, R. Troll, A. Radford, I. Sutskever, J. Leike, and J. Wu, “Scaling and evaluating sparse autoencoders,” arXiv preprint arXiv:2406.04093, 2024. [44] S. Rajamanoharan, A. Conmy, L. Smith, T. Lieberum, V. Varma, J. Kramar, R. Shah, and N. Nanda, “Improving sparse decomposition of language model activations with gated sparse autoencoders,” in The Thirty-eighth Annual Conference on Neural Information Processing Systems, 2024. [45]Z. He, W. Shu, X. Ge, L. Chen, J. Wang, Y. Zhou, F. Liu, Q. Guo, X. Huang, Z. Wu et al., “Llama scope: Extracting millions of features from llama-3.1-8b with sparse autoencoders,” arXiv preprint arXiv:2410.20526, 2024. [46]J. Lindsey, A. Templeton, J. Marcus, T. Conerly, J. Batson, and C. Olah, “Sparse crosscoders for cross-layer features and model diffing,” Transformer Circuits Thread, 2024. [47]J. Engels, L. Riggs, and M. Tegmark, “Decomposing the dark matter of sparse autoencoders,” arXiv preprint arXiv:2410.14670, 2024. [48]N. Balagansky, I. Maksimov, and D. Gavrilov, “Mechanistic permutability: Match features across layers,” arXiv preprint arXiv:2410.07656, 2024. [49] D. Laptev, N. Balagansky, Y. Aksenov, and D. Gavrilov, “Analyze feature flow to enhance interpretation and steering in language models,” arXiv preprint arXiv:2502.03032, 2025. [50]D. Balcells, B. Lerner, M. Oesterle, E. Ucar, and S. Heimersheim, “Evolution of sae features across layers in llms,” arXiv preprint arXiv:2410.08869, 2024. [51] S. Chalnev, M. Siu, and A. Conmy, “Improving steering vectors by targeting sparse autoencoder features,” arXiv preprint arXiv:2411.02193, 2024. [52]N. Elhage, N. Nanda, C. Olsson, T. Henighan, N. Joseph, B. Mann, A. Askell, Y. Bai, A. Chen, T. Conerly et al., “A mathematical framework for transformer circuits,” Transformer Circuits Thread, vol. 1, no. 1, p. 12, 2021. 13 A Empirical Comparison of Softmax and Alternative Functions for MLP Attribution To assess the design choice of using softmax for MLP attribution (Eq. 14), we compare it against two alternative normalization methods: (1) Linear-norm normalization: subtracting the minimum and dividing by the sum; (2) Linear-weighted decomposition: linearly decomposing the contribution to the original activation value for each element. This comparison assesses whether the softmax rule is empirically justified. We report results on theKnown_1000dataset using thellama-2-7b-chat-hfmodel for both patch-top and recover-top token removal strategies. Patch-Top (%)0.10.20.30.40.50.60.70.80.91.0 Softmax80.887.194.096.698.298.898.899.199.299.1 Linear-norm64.873.685.793.296.697.598.298.899.099.1 Linear-weighted59.466.877.586.292.495.096.698.498.699.1 Table 3: Comprehensiveness (patch-top) results for different MLP attribution methodsn. Recover-Top (%)0.10.20.30.40.50.60.70.80.91.0 Softmax97.896.993.989.981.675.568.956.843.50.0 Linear-norm98.397.194.892.387.783.578.268.854.90.0 Linear-weighted98.397.796.695.392.390.486.780.771.60.0 Table 4: Sufficiency (recover-top) results for different MLP attribution methods. Overall, the softmax-based attribution consistently outperforms the alternatives across both token removal strategies, supporting its empirical effectiveness despite its heuristic origin. B More Results and Experiment Details on Token-Level Output Attribution In this appendix section, we provide additional experimental details and results for token-level output attribution using DePass across multiple model variants. Our goal is to assess the faithfulness of token attributions, measured in terms of comprehensiveness and sufficiency, following established evaluation metrics. B.1 Tasks and Dataset Details We evaluate our method on two widely used benchmarks that require distinct types of reasoning: Known_1000The Known_1000 dataset [4] consists of factual question-answering prompts, where each prompt targets a known fact (e.g., “Audible.com is owned by Amazon”). This dataset is designed to probe how factual information is stored and retrieved in the model. IOI (Indirect Object Identification)The IOI task [19] involves syntactic reasoning and coreference resolution. Each instance includes a sentence involving two named entities and a pronoun (e.g., “Eleanor and Deanna were thinking about going to the mountain. Eleanor wanted to give a watermelon to→”). The model must correctly resolve the pronoun to the indirect object. This task is well-suited for analyzing compositional reasoning capabilities and token interactions. B.2 Faithfulness Evaluation on More Models We assess our method by comparing its comprehensiveness and sufficiency scores against standard baselines. Comprehensiveness measures how much the model’s confidence drops when top-attributed 14 tokens are removed, while sufficiency assesses how much confidence is retained when only the top-attributed tokens are kept. Figure 5: Faithfulness evaluation using Llama-2-7b-chat-hf. DePass achieves a sharper drop in comprehensiveness and higher sufficiency retention, indicating better identification of faithful tokens. Figure 6: Faithfulness evaluation on Llama-3.2-1B-Instruct. Despite its small scale, DePass effectively captures key tokens contributing to the output, validating its robustness across model sizes. Figure 7: Evaluation on Qwen-2-7B-Instruct across both tasks. DePass consistently outperforms other attribution techniques in both comprehensiveness and sufficiency, demonstrating generalizability across architectures. Figure 8: Evaluation on Qwen-2-1.5B-Instruct across both tasks. 15 Figure 9: Evaluation on Llama-3.2-3B-Instruct across both tasks. These results reinforce the effectiveness of DePass in identifying faithfully contributing input tokens across a diverse set of models and tasks. B.3 More Cases on Output Attribution We present additional examples illustrating how DePass assigns output-dependent importance scores to input tokens. These cases highlight DePass’s ability to differentiate which parts of the input are most responsible for different model outputs, even under sampling variability. Figure 10: Token-wise output attribution scores from DePass for the prompt "Tom and Amy went to the store. Tom handed a bag to". For two sampled outputs—"Amy" and "a"—DePass produces distinct attribution patterns, correctly identifying which input tokens support each specific continuation. Figure 11: Token-wise output attribution scores for the prompt "The capital of France is". When the model produces "Paris" versus "known" as outputs, DePass assigns higher relevance to different parts of the prompt accordingly. This shows DePass’s sensitivity to output semantics in attributing input importance. Illustration for Multi-Token Words’ Attribution We provide two examples to illustrate how subword tokenization affects DePass scores. For word-level scores, corresponding subword tokens are treated as a single component from the start of the DePass process, rather than a simple sum of individual scores. Example 1: 16 Token<s>Cataloniabelongstothecontinentof Score0.286.161.271.751.161.1513.061.36 Word<s>Cataloniabelongstothecontinentof Score0.286.381.921.371.3813.241.59 Example 2: Token<s>ThemothertongueofDanielleDarrieuxis Score0.300.872.099.691.23-0.011.312.002.691.360.73 Word<s>ThemothertongueofDanielleDarrieusis Score0.301.122.299.851.521.854.121.18 Both token-level and word-level attributions with DePass are effective. While token-level scores may disperse across subword tokens, they remain highly informative for distinguishing important tokens. Higher scores on specific subwords can even indicate triggers for the model’s memory of the full word. DePass naturally supports aggregating subword tokens for desired word-level attribution. Token-Level Attribution Examples for Multiple Attribution Methods The following tables present token-level attribution scores computed by DePass and several baseline methods for illustrative prompts. Each table corresponds to a different target token for the same input, highlighting how DePass distributes relevance across input tokens in a context-sensitive manner. Raw DePass scores are provided alongside normalized values to facilitate comparison with other attribution methods. TokenDePassNormalized DePassAllLastRolloutIntegrated GradientsSignedNorm <s>-0.30-0.020.300.280.350.000.170.40 The1.200.070.140.140.130.010.170.14 capital8.940.520.140.140.130.840.160.17 of1.020.060.140.140.130.140.170.14 France5.060.300.140.150.130.000.160.07 is1.200.070.140.150.130.010.170.09 Table 5: Token-level attribution scores for the prompt “The capital of France is” with target “Paris”. TokenDePassNormalized DePassAllLastRolloutIntegrated GradientsSignedNorm <s>0.090.010.300.280.350.260.181.00 The1.910.150.140.140.130.000.080.00 capital6.000.480.140.140.130.650.200.00 of1.410.110.140.140.130.000.160.00 France1.800.140.140.150.130.000.220.00 is1.410.110.140.150.130.090.170.00 Table 6: Token-level attribution scores for the prompt “The capital of France is” with target “known”. 17 TokenDePassNormalized DePassAllLastRolloutIntegrated GradientsSignedNorm <s>0.320.020.140.120.170.020.070.88 gra1.520.110.070.070.060.000.060.00 pe0.080.010.070.070.060.000.100.00 :0.200.010.070.070.060.000.060.11 pur5.250.360.070.070.060.000.070.00 ple0.140.010.070.070.060.000.080.00 ,0.100.010.070.070.060.000.060.00 ban0.480.030.070.070.060.000.080.00 ana0.120.010.070.070.060.000.050.00 :0.150.010.070.070.060.980.090.00 yellow1.530.110.070.070.060.000.050.00 ,0.190.010.070.070.060.000.080.00 apple4.160.290.070.070.060.000.080.00 :0.170.010.070.070.060.000.060.00 Table 7: Token-level attribution scores for fruit-color completion task with target “red”. DePass reports both raw and normalized attribution scores, with raw values summing exactly to the output logit. This allows direct interpretation of each score’s impact on the prediction. Attribution distributions vary across different targets for the same input, reflecting the model’s context-dependent reasoning. DePass consistently assigns attribution to semantically meaningful input tokens, whereas baseline and gradient-based methods often fail to highlight relevant tokens or exhibit clear semantic alignment. CMore Results and Experiment Details on Token-Wise Subspace Attribution C.1 Prompt Construction C.1.1 CounterFact Prompt Construction The CounterFact dataset [4] is designed to evaluate factual recall in language models. Each instance provides: • a subject entity (e.g., Go Hyeon-jeong), • a target (the correct answer, e.g., Korean), • a target_new (an incorrect, but plausible alternative, e.g., French), • multiple natural languagepromptsquerying the relation (e.g.,“The mother tongue of Go Hyeon-jeong is”). We construct multiple prompt variants per example to simulate different reasoning settings. Specifi- cally, we create three types of prompts: •Initialization Prompt (No Information): A question-only prompt constructed by sampling one of the available templates, without providing supporting information. •Untruthful Prompt: A misleading information context is constructed by inserting the subject along with the incorrect fact (target_new) into a randomly selected prompt template from the provided list. This prompt is then used as external information preceding the same question template, simulating a factual error in the input. Each final prompt is formatted as follows: According to the given information and your knowledge, answer the question. Information: <Inserted Factual or Counterfactual Sentence> (optional) Question: <Sampled Question Template> The answer is: 18 During construction, we ensure diversity by randomly sampling prompt templates for both the question and the information. For each data point, one question prompt is selected and paired with either the correct or incorrect fact from the candidate list. C.1.2 TruthfulQA Prompt Construction The TruthfulQA dataset [23] contains questions designed to probe language models’ susceptibility to producing factually incorrect or misleading answers. For each question, it provides a set of factually correct answers (mc1_targets and mc2_targets) and a pool of plausible but incorrect ones. We construct two types of prompts per example: •Initialization Prompt (No Information): A question-only multiple-choice prompt where the model must choose the best answer from a randomized mix of one correct and one incorrect option, without access to any supporting context. •Untruthful Prompt: A misleading information-bearing prompt, where a randomly sam- pled incorrect statement from the providedmc1_targetsormc2_targetsis inserted as Information. The model must then answer the same question with the same randomized options, now influenced by the incorrect prior. Each prompt is formatted as follows: According to the given information and your knowledge, choose the best choice from the following options. Information: <Random Incorrect Statement> (optional) Question: <Original Question> Options: A: <Incorrect Option> B: <Correct Option> (random order) The answer is: To ensure variation and avoid position bias, we shuffle the order of the correct and incorrect choices. In all cases, the model is required to pick between exactly two answer candidates, enabling us to isolate the effect of misleading information on model predictions. C.2 Truthful Probe Training and Evaluation Probe Training Following [22], we train linear probes on hidden states from transformer-based language models to detect factual consistency. We use the CounterFact and TruthfulQA datasets, each labeled with truthful and untruthful prompts. The datasets are split into training and testing sets with balanced labels to ensure fair evaluation. For each prompt, we extract the hidden states of the final token across all transformer layers. These per-layer activations serve as features for training logistic regression classifiers (one per layer) to distinguish between truthful and untruthful inputs. We use thesagasolver inscikit-learn, with a learning rate of 0.01 and maximum iteration count of 1000. During training, we shuffle and concatenate the hidden states of truthful and untruthful samples and fit each layer-specific classifier independently. After training, we evaluate the classifiers on a held-out test set to assess the layer-wise linear separability of factual information. Probe EvaluationFigure 12 and Figure 14 show the classification accuracy for different Llama and Qwen model variants. We observe a consistent trend across models: classifier accuracy improves with depth, often exceeding 90% in the middle-to-late layers. This suggests that factual signals become increasingly linearly separable in deeper representations. 19 Figure 12: Truthful classifier accuracy on Llama-2-7B-chat-hf and Llama-3.1-8B-Instruct. Figure 13: Truthful classifier accuracy on Llama-2-13b-chat-hf and Meta-Llama-3.1-70B-Instruct Figure 14: Truthful classifier accuracy on Qwen-2-1.5B-Instruct and Qwen-2-7B-Instruct. C.3 Masking Strategy Token selection for masking. For a given prompt, we extract the hidden states of all tokens across all layers. Based on classifier accuracy trends, we consider only classifiers from layer 10 onward, where truthfulness signals are most separable. For each token, we compute its average predicted probability of being untruthful across these layers. Tokens with a mean probability≤ 0.5are retained. For those with probability> 0.5, we apply two masking strategies: 1. Direct masking (TACS)[22]: Remove all tokens predicted as untruthful (p > 0.5) from the input and re-run the model. 2.DePass-based masking: For each token identified as untruthful, we decompose its hidden state to attribute contributions from other input tokens. We average the untruthful tokens’ 20 contribution vectors to compute a global untruthfulness attribution score. Then, we identify the topkcontributing tokens in the input (matching the number of untruthful tokens in direct masking) and remove them. These complementary methods help evaluate whether selectively removing suspected untruthful information can steer model predictions towards higher factuality, while controlling for deletion magnitude. C.4 More Cases on Subspace Attribution We present additional untruthful examples to further illustrate how DePass attributes subspace-level information to input tokens. Figure 15: Additional examples demonstrating DePass’s ability to attribute information from specific subspaces back to input tokens. Attribution scores are computed with respect to the output label0 (untruthful); more negative scores indicate stronger contributions toward untruthful signal. C.5 Case Studies: Potential of Combining DePass with SAE Combining DePass with SAE. While SAEs can uncover meaningful features, they do not inherently indicate which input tokens activate those features. DePass provides a complementary capability: by decomposing activations into additive components, it flexibly connects different elements of a model—tokens, components, and subspaces—with SAE features. This enables fine-grained analysis of how interpretable features are triggered by specific tokens and how they propagate through the model. Below, we present several case studies demonstrating the potential of combining DePass with SAE for token-level attribution. The feature annotations used in these examples are drawn from Neuronpedia 4 , and all cases are based on LLaMA-3.1-8B with the corresponding SAE [45]. Case 1: Climate Change. Prompt: “In exploring the relationship between climate change and urban planning, the research identifies key strategies for sustainable development. It underscores the urgency of integrating environmental considerations into city planning.” Feature: Layer 31, Feature 5226 – “climate change and its associated impacts” We analyze three tokens that activate this feature. DePass identifies that the activation of the SAE feature is primarily driven by the semantic contribution ofclimate (6), which then propagates to other relevant tokens such as change (7) and environmental (27) (Table 8). 4 https://w.neuronpedia.org/ 21 Token analyzed (index)Activation valueSource token (index)Contribution climate (6)35.2500climate (6)26.8750 In (1)3.0781 exploring (2)1.8750 the (3)1.4844 relationship (4)1.4844 between (5)1.4844 change (7)15.6875climate (6)16.5000 relationship (4)0.8594 between (5)0.3984 change (7)0.5586 In (1)-1.9141 environmental (27)7.4688climate (6)9.4375 exploring (2)1.0234 environmental (27)-0.6562 underscores (22)-0.1670 urgency (24)-0.2432 integrating (26)-0.3516 Table 8: DePass token-level attribution of SAE feature 5226 (“climate change and its associated impacts”). Case 2: Simplicity in Context. Prompt: “Avocado toast, a simple yet trendy breakfast option, gains a delightful twist with a sprinkle of chili flakes and a drizzle of honey for that perfect balance of spicy and sweet.” Feature: Layer 31, Feature 26874 – “the context of simplicity in various contexts” DePass shows that the activation of this feature is largely driven by the semantic link fromsimple (6)to tokens such asyet (7)and punctuation, highlighting how contextual modifiers contribute to SAE features (Table 9). Token analyzed (index)Activation valueSource token (index)Contribution yet (7)7.4062simple (6)5.8438 toast (3)1.9219 yet (7)1.1953 , (4)-0.5312 a (5)-0.5898 <bos> (0)-1.3750 ,(11)3.5312simple (6)1.7031 toast (3)0.8125 yet (7)0.7344 <bos> (0)-1.8906 Table 9: DePass token-level attribution of SAE feature 26874 (“simplicity in context”). Case 3: Economic Concepts. Prompt: “In economics, supply and demand is a model for under- standing how prices and quantities are determined in a market system. This concept is foundational in economic theory and affects various market structures.” Feature: Layer 25, Feature 9618 – “references to economic topics and concepts” Here, DePass reveals that the activation of the “economic concepts” feature stems from semantic contributions of economics (2) that propagate to economic (28) and and (30) (Table 10). Across these examples, combining DePass with SAE enables precise token-level attribution of interpretable features. Whereas SAE identifies semantically coherent features, DePass traces their origins and propagation across tokens, yielding a more complete mechanistic account of model representations. 22 Token analyzed (index)Activation valueSource token (index)Contribution economic (28)15.0625economics (2)5.9375 economic (28)4.9062 <bos> (0)0.9805 . (22)0.1348 and (30)7.3438economics (2)4.7500 economic (28)1.9844 <bos> (0)1.1719 concept (24)-0.1025 In (1)-0.6680 Table 10: DePass token-level attribution of SAE feature 9618 (“economic topics and concepts”). D More Results and Experiment Details on Component-Wise DePass D.1 Additional Model Component-wise Attribution Results on Different Models To further evaluate the generality and robustness of DePass across architectures and scales, we present additional results on component-wise attribution. Specifically, we assess the ability of DePass to identify important attention heads and MLP neurons across various model families, including LLaMA-3.2 and Qwen2 at both small and medium scales. We apply Top-kand Bottom-kmasking strategies: masking the top-kmost attributed components should lead to a notable performance drop, while masking the bottom-kshould minimally impact output quality. Across all settings, DePass demonstrates strong alignment between attribution scores and functional importance. Figure 16: Top-kand Bottom-kmasking results for attention head attribution on Llama-3.2-1B- Instruct and Llama-3.2-3B-Instruct. DePass leads to a larger accuracy drop with Top-kmasking and better retention with Bottom-kmasking, highlighting its effectiveness in identifying key attention heads across different model scales. 23 Figure 17: Top-kand Bottom-kmasking evaluation on Qwen2-1.5B-Instruct and Qwen2-7B-Instruct for attention head attribution. DePass consistently achieves greater accuracy drop under Top-k masking and stronger performance retention under Bottom-kmasking, demonstrating its ability to pinpoint influential attention heads across architectures. Figure 18: Top-kand Bottom-kmasking evaluation on Llama-3.2-1B-Instruct and Llama-3.2-3B- Instruct for MLP neuron attribution. DePass outperforms baselines by inducing sharper performance degradation when top neurons are masked and maintaining better accuracy when only low-ranked neurons are ablated. D.2 Efficiency of Neuron-level Attribution To ensure the practicality of fine-grained attribution, we compare the runtime of DePass with the standard ablation approach when attributing neurons in intermediate MLP layers of different LLaMA models. As shown in Table 11, DePass provides a significant speedup, reducing computation time by up to two orders of magnitude, while maintaining attribution fidelity. This efficiency makes DePass especially suitable for scaling to larger models and longer prompts. Table 11: Runtime comparison for neuron-level attributions over an intermediate MLP layer. DePass achieves substantial acceleration over the ablation baseline. ModelIntermediate SizeMethodTime (s) LLaMA-2-7B-Chat11008Ablation321.04 DePass7.22 LLaMA-3.2-3B-Instruct8192Ablation234.77 DePass2.91 LLaMA-3.2-1B-Instruct8192Ablation134.76 DePass2.23 D.3 Important Head and Neuron Distributions We provide a detailed analysis of the internal model components by visualizing the distributions of top-attributed attention heads and MLP neurons, both at the dataset level and for individual prompts. 24 These analyses reveal how different models and attribution methods identify key substructures that contribute to model predictions. D.3.1 Average Distribution over the Dataset To better understand which internal components contribute most to model predictions, we visualize the average importance scores assigned by DePass to attention heads and MLP neurons across two representative datasets: IOI and CounterFact. These aggregated views highlight consistent attribution patterns across layers and architectures. For attention heads, each heatmap aggregates importance scores across all prompts, with rows corresponding to transformer layers and columns representing head indices. For MLP neurons, we group neurons into bins of 100 for ease of visualization, plotting the average importance per bin across datasets. These visualizations provide structural insights into how different model components are utilized in practice, and where important predictive capacity is concentrated across architectures and tasks. Figure 19: Average attention head importance identified by DePass on Llama-2-7b-chat-hf, aggregated across datasets. Rows indicate transformer layers; columns indicate head indices. Figure 20: Average MLP neuron importance identified by DePass on Llama-2-7b-chat-hf, aggregated across datasets. Neurons are grouped in bins of 100 for visualization. Figure 21: Average MLP neuron importance identified by DePass on Llama-3.2-3B-Instruct, aggre- gated across datasets. Neurons are grouped in bins of 100 for visualization. 25 D.3.2 Per-Prompt Case Analysis We further present case-by-case visualizations to analyze the distribution of important attention heads and MLP neurons for individual prompts. These visualizations compare DePass with baseline attribution methods (e.g., AtP, Norm) and aim to identify the components most responsible for producing the correct answer under each input. In contrast to dataset-level views, these examples highlight how attribution patterns vary across prompts and methods. Figure 22: Attention head attribution for the prompt“The mother tongue of Danielle Darrieux is”, “French” using Llama-2-7b-chat-hf. DePass more effectively isolates attention heads that contribute to generating the correct answer compared to baselines. Figure 23: Attention head attribution for the prompt“After the lunch, Mildred and Eleanor went to the market. Eleanor gave a watermelon to” ,“Mildred”using Llama-2-7b- chat-hf. DePass reveals a more focused and task-relevant distribution of important attention heads. Figure 24:MLP neuron attribution for the prompt“The mother tongue of Danielle Darrieux is”, “French” using Llama-2-7b-chat-hf. Neurons are grouped into bins of 100 for visualization. DePass effectively identifies neurons critical for predicting the correct answer. 26 Figure 25:MLP neuron attribution for the prompt“Tarvaris Jackson plays in the position of”, “quarterback”using Llama-2-7b-chat-hf. Neurons are grouped into bins of 100 for visualization. DePass more accurately isolates neurons essential for producing the correct answer compared to baselines. E More Results and Experiment Details on Subspace-Wise DePass E.1 Classifier Training and Evaluation Dataset for Probing Classifier Training To construct a multilingual dataset with identical se- mantics across different languages, we translate a subset of the CounterFact[4] dataset into multiple languages while ensuring that the meaning remains strictly unchanged. This allows us to probe whether the identified important subspaces generalize across linguistic variations that share the same semantic content. An example entry is shown below: Original (English): "prompt": "The mother tongue of Danielle Darrieux is", "answer": "French" French: "prompt": "La langue maternelle de Danielle Darrieux est le", "answer": "français" German: "prompt": "Die Muttersprache von Danielle Darrieux ist", "answer": "Französisch" Italian: "prompt": "La lingua madre di Danielle Darrieux è il", "answer": "francese" Classifier Training Details We then use this multilingual, semantically aligned dataset to train a language classifier. This setup allows us to evaluate whether the subspace directions identified by DePass are stable and predictive across languages, thus offering a rigorous test of attribution generalization. For each input prompt, we extract the hidden state of the final token and use it as the input feature to the classifier. The corresponding language of the prompt serves as the class label. We train a separate multi-class classifier at each transformer layer using the translated subset of the CounterFact dataset, which is evenly split into training and testing sets with balanced label distributions across languages. We use logistic regression implemented via thescikit-learnlibrary, with thesagasolver, a learning rate of0.01, and a maximum iteration count of1000. During training, we concatenate the hidden states across all prompts and shuffle the dataset to ensure robustness and reduce bias. This approach enables us to analyze which layers encode language-distinguishing information and how DePass-selected subspaces contribute to that encoding. 27 Classifier EvaluationThe language classifier achieves high accuracy across both models, indicating that language-specific directions are reliably encoded in the hidden states. Figure 26: Accuracy of the language classifier across different layers on Llama-3.1-8B-Instruct and Qwen2-7B-Instruct. Results show strong separability of language-specific information in hidden representations. E.2 Subspace Projection We describe how to construct the projection matrixP t ∈R d×d used to decompose hidden states into components within and orthogonal to a target subspace. This subspace can be defined by any set of directions of interest, such as the row space of a linear classifier, neuron activations, or gradient-based attribution vectors. Given a matrixW ∈R d×c whose column space spans the target subspace (e.g., the weight matrix of a linear classifier), we compute the singular value decomposition (SVD) of W ⊤ as: W ⊤ = U ΣV ⊤ , and retain the top-rleft singular vectors inU, wherer = rank(W ). These vectors form an orthonor- mal basis U r ∈R d×r for the subspace of interest. The projection matrix onto this subspace is given by: P t = U r U ⊤ r , and its orthogonal complement is I− P t , where I is the identity matrix. For any hidden state x i ∈R D , we obtain the decomposition: x ∥ i = P t x i , x ⊥ i = (I− P t )x i , wherex ∥ i lies in the target subspace andx ⊥ i is orthogonal to it. This decomposition enables precise attribution and intervention by isolating the contribution of specific representational directions to model behavior. E.3 More Results on Qwen2-7B-Instruct We present further analysis demonstrating the subspace attribution capability of DePass on Qwen2- 7B-Instruct. To validate the separation of representational content, we apply t-SNE to the language subspace projectionsX dec lang obtained from multilingual inputs. As shown in Figure 27, the resulting embeddings form distinct clusters based on language identity. This indicates that DePass successfully isolates language-specific patterns within a dedicated subspace, supporting its ability to disentangle and attribute form-related features. We further examine subspace semantics by decoding from both thelanguageandsemanticsub- spaces for a shared English input: “In which city did Charles Alfred Pillsbury’s life end?” As shown in Table 12, tokens from the language subspace are dominated by structural and frequent function words characteristic of each language, while tokens from the semantic subspace correspond to factual entities such as names and cities. This further supports that DePass routes distinct representational factors—form and meaning—into appropriate subspaces. 28 Figure 27: t-SNE visualization of token-wise representations in thelanguagesubspace (X dec lang ) from multilingual prompts in Qwen2-7B-Instruct. The clear clustering by language confirms that DePass accurately attributes language-specific signals to this subspace. LanguageDecoded Tokens from Lan- guage Subspace Decoded Tokens from Se- mantic Subspace EnglishSan, ", __, New, _order, aug, which, Charles, Charles Frenchquoi, chaque, cette, son, lMinneapolis, vec, cause, par- tir, uc Germanund, der, die, San, dasMinneapolis, ámb, Stockholm, Cambridge, St Spanish las, el, ¿, los, laMinneapolis, Bloom, Saint, ven, och Table 12: Tokens decoded from thelanguageandsemanticsubspaces of Qwen2-7B-Instruct. The language subspace captures structural and high-frequency tokens, while the semantic subspace reflects content-specific entities relevant to the input query. F Implementation Details We run all experiments on a cluster of A6000 GPUs. Since DePass operates entirely via forward decomposition without requiring model fine-tuning or backpropagation, our method is highly efficient and incurs minimal memory overhead. All experiments are conducted in PyTorch, using pre-trained models with lightweight modifications to enable decomposition-aware forward passes. G Limitations of DePass To precisely trace information flow within the model, DePass freezes the attention scores during decomposed forward pass. As a result, under the Transformer circuits framework [52], DePass captures only the flow through the output-value (OV) circuit, without explaining query-key (QK) circuit interactions. Decomposing the QK circuit remains a challenging open problem: first, QK and VO circuits are entangled yet structurally distinct, making it difficult to construct human-interpretable representations of their interactions; second, following DePass’s additive decomposition approach, attribution of QK circuit would involve a quadratic complexity explosion. We believe that unwrapping the QK circuit is a key direction for future work. DePass is compatible with most Transformer architectures. In this paper, we validate it on two widely used families—LLaMA and Qwen—across different model sizes. Extending DePass to other architectures primarily requires adapting to differences in LayerNorm implementation and updating module names at the code level. We plan to release a general-purpose toolkit supporting a broad range of Transformer models. 29 While DePass supports one-pass attribution across all model components and, in theory, allows hidden states to be decomposed into an infinite number of parts, practical memory constraints require grouping components and computing their forward passes sequentially. This can lead to increased time costs when a finer-grained analysis is desired. Future work includes optimizing computational efficiency and improving hardware deployment strategies to better support attribution with higher resolution. 30