← Back to papers

Paper deep dive

Fine-Grained Activation Steering: Steering Less, Achieving More

Zijian Feng, Tianjiao Li, Zixiao Zhu, Hanzhang Zhou, Junlang Qian, Li Zhang, Jia Jim Deryl Chua, Lee Onn Mak, Gee Wah Ng, Kezhi Mao

Year: 2026Venue: ICLR 2026Area: Model EditingType: EmpiricalEmbeddings: 79

Models: Gemma2-27B, Gemma2-9B-it, LLaMA2-13B, LLaMA2-7B-Chat, LLaMA3.1-8B, Qwen3-8B

Abstract

Abstract:Activation steering has emerged as a cost-effective paradigm for modifying large language model (LLM) behaviors. Existing methods typically intervene at the block level, steering the bundled activations of selected attention heads, feedforward networks, or residual streams. However, we reveal that block-level activations are inherently heterogeneous, entangling beneficial, irrelevant, and harmful features, thereby rendering block-level steering coarse, inefficient, and intrusive. To investigate the root cause, we decompose block activations into fine-grained atomic unit (AU)-level activations, where each AU-level activation corresponds to a single dimension of the block activation, and each AU denotes a slice of the block weight matrix. Steering an AU-level activation is thus equivalent to steering its associated AU. Our theoretical and empirical analysis show that heterogeneity arises because different AUs or dimensions control distinct token distributions in LLM outputs. Hence, block-level steering inevitably moves helpful and harmful token directions together, which reduces efficiency. Restricting intervention to beneficial AUs yields more precise and effective steering. Building on this insight, we propose AUSteer, a simple and efficient method that operates at a finer granularity of the AU level. AUSteer first identifies discriminative AUs globally by computing activation momenta on contrastive samples. It then assigns adaptive steering strengths tailored to diverse inputs and selected AU activations. Comprehensive experiments on multiple LLMs and tasks show that AUSteer consistently surpasses advanced baselines while steering considerably fewer activations, demonstrating that steering less achieves more.

Tags

ai-safety (imported, 100%)empirical (suggested, 88%)model-editing (suggested, 92%)

Links

PDF not stored locally. Use the link above to view on the source site.

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 93%

Last extracted: 3/11/2026, 1:17:47 AM

Summary

The paper introduces AUSteer, a fine-grained activation steering method for LLMs that decomposes block-level activations into atomic unit (AU)-level activations. By identifying discriminative AUs using a novel 'activation momentum' metric and applying adaptive steering, the method achieves superior performance with significantly fewer interventions compared to traditional block-level steering.

Entities (5)

AUSteer · method · 100%LLM · technology · 100%Activation Momentum · metric · 95%Atomic Unit (AU) · concept · 95%Block-level steering · method · 95%

Relation Signals (3)

Activation Momentum identifies Atomic Unit (AU)

confidence 95% · AUSteer first identifies discriminative AUs globally by computing activation momenta on contrastive samples.

AUSteer uses Atomic Unit (AU)

confidence 95% · AUSteer, a simple and efficient method that operates at a finer granularity of the AU level.

Block-level steering islessefficientthan AUSteer

confidence 90% · AUSteer consistently surpasses advanced baselines while steering considerably fewer activations

Cypher Suggestions (2)

Find all methods related to activation steering · confidence 85% · unvalidated

MATCH (m:Method)-[:USED_FOR]->(t:Task {name: 'Activation Steering'}) RETURN m

Identify the relationship between metrics and entities · confidence 80% · unvalidated

MATCH (m:Metric)-[r]->(e:Entity) RETURN m.name, type(r), e.name

Full Text

79,010 characters extracted from source content.

Expand or collapse full text

Published as a conference paper at ICLR 2026 FINE-GRAINED ACTIVATION STEERING: STEERING LESS, ACHIEVING MORE Zijian Feng 1 Tianjiao Li 1 Zixiao Zhu 1 Hanzhang Zhou 1 Junlang Qian 1 Li Zhang 1 Jia Jim Deryl Chua 2 Lee Onn Mak 2 Gee Wah Ng 2 Kezhi Mao 1, ∗ 1 School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore 2 Home Team Science and Technology Agency (HTX), Singapore feng0119, zixiao001, hanzhang001, junlang001, zhan0735@e.ntu.edu.sg tianjiao.li, ekzmao@ntu.edu.sg deryl chua, makleeonn, nggeewah@htx.gov.sg ABSTRACT Activation steering has emerged as a cost-effective paradigm for modifying large language model (LLM) behaviors. Existing methods typically intervene at the block level, steering the bundled activations of selected attention heads, feedfor- ward networks, or residual streams. However, we reveal that block-level activa- tions are inherently heterogeneous, entangling beneficial, irrelevant, and harmful features, thereby rendering block-level steering coarse, inefficient, and intrusive. To investigate the root cause, we decompose block activations into fine-grained atomic unit (AU)–level activations, where each AU-level activation corresponds to a single dimension of the block activation, and each AU denotes a slice of the block weight matrix. Steering an AU-level activation is thus equivalent to steering its associated AU. Our theoretical and empirical analysis show that heterogeneity arises because different AUs or dimensions control distinct token distributions in LLM outputs. Hence, block-level steering inevitably moves helpful and harm- ful token directions together, which reduces efficiency. Restricting intervention to beneficial AUs yields more precise and effective steering. Building on this insight, we propose AUSteer, a simple and efficient method that operates at a finer gran- ularity of the AU level. AUSteer first identifies discriminative AUs globally by computing activation momenta on contrastive samples. It then assigns adaptive steering strengths tailored to diverse inputs and selected AU activations. Com- prehensive experiments on multiple LLMs and tasks show that AUSteer consis- tently surpasses advanced baselines while steering considerably fewer activations, demonstrating that steering less achieves more 1 . 1INTRODUCTION In the era of large language models (LLMs), activation steering has emerged as a powerful paradigm for modulating model behavior on downstream tasks (Zou et al., 2023; Li et al., 2023b; Rimsky et al., 2024). Unlike reinforcement learning from human feedback (Bai et al., 2022), supervised fine- tuning (Wei et al., 2022), or prompt engineering (Brown et al., 2020), activation steering intervenes directly in the LLM intermediate activations during forward propagation, enabling fine-grained con- trol without additional training. Prior work (Turner et al., 2023; Rimsky et al., 2024; Han et al., 2024; Wang et al., 2025a;b;c) generally builds task-specific steering vectors and injects them at in- ference time as biases or rescaling factors in selected LLM components, thereby steering the model toward the target objective. ∗ Corresponding author. 1 Code: https://github.com/zijian678/AUSteer 1 arXiv:2602.04428v1 [cs.CL] 4 Feb 2026 Published as a conference paper at ICLR 2026 However, a common practice in existing methods is block-level steering, where a “block” denotes the multi-head attention (MHA), the feed-forward network (FFN), or the layer’s residual stream. As shown in Figure 1 (a), the intervention is vector-level: every dimension of the selected block’s acti- vation is bundled and steered simultaneously. One of the main limitations of block-level intervention is that it ignores heterogeneity within block activations. These activations often span hundreds or thousands of dimensions, each indicating a different feature. Some features are beneficial for the task, while others are irrelevant or harmful. As a result, block level steering is (1) too coarse: a block can be decomposed into finer functional units, and treating it as a single entity prevents pre- cise targeting; (2) inefficient: steering the entire block amplifies both useful and harmful signals, which reduces efficiency and risks performance degradation; and (3) overly intrusive: it modifies many dimensions unnecessarily, increasing the intervention footprint. Multi-Head Attention Add & Norm FFN Add & Norm (a) Block-level (b) AU-level Original Activation Steered ActivationSteered Activation 휸 ퟏ 휸 ퟐ 휸 ퟑ ...휸 풏 AU Localization Intervention Vector Intervention Scalar Transformer Layer 휸 ퟑ Figure 1: Comparison of block-level steering (prior work) and AU-level steering (Ours). In greater depth, we empirically and theoretically justify the heterogeneity of block-level activations. We first decompose block-level activations into finer-grained atomic unit (AU) activations, where each AU-level activation corresponds to a single di- mension of the block activation, and each AU de- notes a slice of the block weight matrix. Steering an AU-level activation is thus equivalent to steering its associated AU. As shown in Figure 1 (b), each AU- level intervention targets a single dimension 2 . Both the intervention value and the affected activation are scalars. Empirically, we find that AU-level steer- ing effects vary widely: some dimensions improve performance, some degrade it, and others are neu- tral, confirming heterogeneity. In many cases, steer- ing a single dimension or a small subset outperforms steering the entire block. Our theoretical analysis reveals that the heterogene- ity stems from different AUs modulating distinct output-token distributions. Steering a single di- mension therefore shifts the model’s output distri- bution toward the distribution controlled by that AU. Some AUs favor task-irrelevant or harmful tokens; steering their dimensions degrades performance. This also explains why block-level steering, which mixes helpful and harmful AUs together, can underperform more targeted and precise AU-level steering. Targeting only beneficial AUs can reduce the intervention footprint and improve efficiency, that is, steering less achieves more. Beyond the promise of AU-level steering, these findings also pose challenges: (1) how can we localize the most important AUs for intervention? and (2) how can we ensure adaptive steering across diverse inputs and AUs? To address these challenges, we introduce AUSteer, a simple and efficient method with two com- ponents. First, we propose activation momentum, a new metric that analyzes each activation’s momentum in positive and negative samples to evaluate its discriminative power. This counting- based metric supports global comparison and avoids the issue of increasing activation magnitudes across layer. We then localize the most discriminative AUs or activations for steering. Second, to ensure adaptivity across inputs and AUs, we assign a per sample steering scalar that follows the original activation pattern rather than a constant shift. This makes the update scale with the current activation, and preserves direction. We also assign dynamic steering strength to each AU according to its discriminative power, with important AUs receiving higher strength. We compare AUSteer with state-of-the-art (SOTA) methods that intervene at the block level by steering hundreds to thousands of activations. Using far fewer steered activations (at most 100), AUSteer significantly outperforms these methods across diverse tasks, demonstrating that steering less achieves more. 2 For clarity: a block-level activation is a vector associated with a block (MHA, FFN, or a layer’s residual stream), usually comprising hundreds to thousands of dimensions, whereas an AU-level activation is a scalar corresponding to a single dimension within that block activation. 2 Published as a conference paper at ICLR 2026 The contributions of this work are summarized as follows: • Conceptually, we study the heterogeneity within block-level activations and its root causes, both theoretically and empirically, and propose decomposing block-level intervention into fine-grained AU-level intervention (§3). • Methodologically, we propose AUSteer, a framework that localizes discriminative AUs with acti- vation momenta for steering, and ensures adaptivity across diverse inputs and AUs (§4). • Empirically, we evaluate AUSteer on multiple LLMs of varying sizes across diverse tasks, includ- ing commonsense reasoning, mathematical problem solving, and open-ended generation. With less intrusive intervention, AUSteer significantly outperforms other SOTA activation steering methods, underscoring that steering less achieves more (§5). 2RELATED WORK Activation steering (also known as activation editing) has become a popular and cost-effective ap- proach for modifying LLM behaviors and aligning them with downstream tasks (Turner et al., 2023; Rimsky et al., 2024; Han et al., 2024; Wang et al., 2025c; Soo et al., 2025; Stickland et al., 2024; Li et al., 2023c; Wang et al., 2025a;b; Stolfo et al., 2025). The standard workflow involves ex- tracting steering vectors from prompts or contrastive samples and injecting them into LLMs at in- ference time. Most of these methods intervene at the block level. For instance, at MHA blocks, ITI (Li et al., 2023b) derives steering vectors from contrastive activations in attention blocks and then applies interventions using the extracted vectors to important heads. Bhattacharjee et al. (2024) compute category-specific activations from attention heads to reduce unsafe responses. In residual streams, CAA (Rimsky et al., 2024) extracts vectors from positive and negative samples and applies them to residual streams, while van der Weij et al. (2024) extend this approach to multi-vector steer- ing across residual streams. EAST (Rahn et al., 2024) obtains steering vectors by weighting input prompts with entropy and injects them into the layer outputs. Postmus & Abreu (2024) use multiple steering vectors as a conceptor to redirect behaviors via residual stream activations. Safety methods such as SafeSwitch (Han et al., 2025) and Safety Arithmetic (Hazra et al., 2024) intervene in residual streams to suppress harmful outputs. Konen et al. (2024) extract steering vectors from layer outputs to control emotion and writing style, while AnyEdit (Jiang et al., 2025) updates hidden states and knowledge by steering layer outputs. More recently, Stolfo et al. (2025) steer residual streams to enhance instruction following. Some methods can operate across multiple blocks. For example, SADI (Wang et al., 2025b) computes steering vectors from MHA, FFN, or residual streams, then applies mask-adaptive steering. Notably, STA (Wang et al., 2025a) identifies atoms in pretrained sparse autoencoders (SAEs) (Lieberum et al., 2024; He et al., 2024; Gao et al., 2025) of target LLMs and steers residual streams using these localized units. Although STA uses the term atom, its meaning differs from ours: in STA, an atom is a knowledge unit in an SAE, whereas in our work an atom is a unit in the original LLM weight matrices. Methodologically, STA depends on pretrained SAEs that currently exist for only a few model families such as LLaMA3.1 (Touvron et al., 2023) and Gemma2 (Team et al., 2024), which limits generalization. Moreover, STA’s intervention remains at the block level as the computed vectors are injected into the residual stream. 3HETEROGENEOUS BLOCK ACTIVATIONS: STEERING LESS ACHIEVES MORE 3.1BLOCK DECOMPOSITION We first show how computations within LLM blocks can be decomposed into fine-grained AU cal- culations. The backbone architecture of LLMs is the Transformer, which consists of attention blocks and FFN blocks in every layer. The outputs of these blocks are added to the layer residual stream for forward propagation. In both MHA and FFN, weight matrix computations (Q,K,V,O in MHA and the up projection and down projection in FFN) are linear projections of the form y = Wx, where x is the input activation, W is the weight matrix, and y is the output activation. 3 Published as a conference paper at ICLR 2026 In existing studies, block activations (x and y) are typically treated as indivisible vectors. Steering vectors are calculated and applied at this coarse block level. To decompose blocks into finer-grained units, we reformulate the linear projection as y = Wx = X i x i W :,i .(1) Here, x i denotes the i-th dimension of the input activation x. This formulation allows us to isolate each single-dimensional activation. In this view, every scalar x i serves as the coefficient for the corresponding column W :,i of the weight matrix. We refer to each column W :,i as an Atomic Unit (AU) in our study. 3 In this way, steering the i-th dimension activation x i is equivalent to steering the corresponding i-th AU. To clarify: • x, y: block-level activations, represented as vectors (the standard formulation in prior work). • W :,i : the i-th column of the weight matrix W, representing the i-th AU. • x i : i-th dimension or i-th AU-level activation, which is a scalar and the coefficient for the i-th AU. 3.2HETEROGENEITY IN BLOCK ACTIVATIONS In this section, we examine the heterogeneous effects of AU-level activations within the block acti- vation. To ensure generalizability, we adopt two representative steering methods: the pioneering ITI (Li et al., 2023b) and SOTA SADI (Wang et al., 2025b), applying them to MHA and FFN blocks. We use LLaMA2-7B-Chat (Touvron et al., 2023) as the backbone model and BoolQ (Clark et al., 2019) as the illustrative dataset, where the model answers “yes” or “no” for each question and the accuracy is reported. The experimental setup follows SADI, as described in Appendix B. We first use ITI and SADI to identify important attention heads and FFNs for intervention, then compare six conditions: (1) Baseline, the original model without steering; (2) ITI, block-level intervention on attention head activations; (3) SADI (Wang et al., 2025b), block-level steering on attention heads (128 dimensions) and FFNs (4096 dimensions); (4) Dimension Sweep, steering single dimensions rather than whole blocks, sampling one of every four dimensions in attention heads and one of every 100 in FFNs; (5) Positive Combination, steering a small subset of beneficial dimensions; and (6) Mixed Comb., steering a subset of beneficial and detrimental dimensions . 020406080100120 Dimension Index 68 69 70 71 72 73 74 75 Accuracy Dimension Sweep Baseline ITI SADI Pos. Comb. Mixed Comb. (a) Steering results for the 7th attention output in layer 27. Positive Combination: steering four ben- eficial dimensions (28, 40, 84, 92). Mixed Com- bination: steering those four plus two detrimental dimensions (108, 116). 05001000150020002500300035004000 Dimension Index 66 68 70 72 74 Accuracy Dimension Sweep Baseline SADI Pos. Comb. Mixed Comb. (b) Steering results for the FFN output in layer 20. Positive Combination: steering four beneficial di- mensions (0, 1300, 2300, 3000). Mixed Combi- nation: steering those four plus two detrimental di- mensions (1200, 1700). Figure 2: Heterogeneous steering results for MHA and FFNs. Figure 2a shows the results of interventions on the 7th attention head at the 27th layer. The orig- inal model achieves 70.52% accuracy, ITI reaches 71.56%, and SADI achieves 73.70%. Steering individual AU activations, however, produces highly heterogeneous outcomes: some dimensions 3 Each column ofW corresponds to an AU, while each row corresponds to what is traditionally termed a “neuron.” To ensure rigor and avoid confusion, we adopt the term AU rather than neuron. 4 Published as a conference paper at ICLR 2026 degrade performance, while others improve it. Notably, steering a single dimension can outperform full block steering. For example, steering the 84th dimension alone achieves 74.53%, surpassing the baseline, ITI, and SADI. Furthermore, steering only four positively contributing dimensions (Pos. Comb.) yields even stronger results. While introducing detrimental dimensions (Mixed Comb.), the perform drops. Similar observations hold for FFN blocks in Figure 2b. Additional empirical results for other attention heads and FFNs are provided in Appendix A. These findings indicate that block-level steering is inefficient, as it mixes beneficial and detrimental components. In contrast, fine-grained AU-level steering enables selective amplification of useful features, achieving more effective control. In short, steering less achieves more. 3.3INTERPRETING THE HETEROGENEITY To explain the observed heterogeneity, as discussed above, we treat the block activation as coeffi- cients on an AU basis, so steering a single dimensional activation x i is equivalent to steering its AU. Building on prior theory of interpreting LLMs in the embedding space (Geva et al., 2022; Dar et al., 2023), different AUs may control different token distributions in LLM outputs. Steering task- relevant AUs promotes the probability of task-specific tokens, whereas steering task-irrelevant AUs may increase the probability of uninformative or even harmful tokens. This provides a theoretical justification for the observed heterogeneity. [AU s 44 , AU 44 ] [AU s 84 , AU 84 ] [AU s 44 , AU s 84 ] 10 100 1000 10000 100000 Strength 0.7680.1880.000 0.6920.1160.025 0.1960.0040.058 0.0070.0000.848 0.0000.0001.000 0.00 0.25 0.50 0.75 1.00 KL divergence Figure 3: Pairwise KL divergence when steering different AUs. s means strength. To further validate this, we first examine the con- vergence behavior of AU steering: different AUs govern different output token distributions, and as steering strength increases, the LLM’s output tends to converge to the AU’s token distribution. For the selected 7th attention head at the 27th layer, we scale the AU coefficient from 10 to an extremely large value (100,000) and compute the normalized KL di- vergence between the output at each strength and the output at 100,000. In Figure 3, columns 1 and 2 show these divergences for the 44th AU and the 84th AU. The divergence decreases with strength, indicat- ing convergence. Column 3 shows the pairwise KL divergence between the 44th AU and the 84th AU across strengths. The divergence increases with strength, indicating that the two AUs tend to drive the model toward different output distributions. Probability (high low) AU 44 AU 84 AU 100 '_no' 'NO' '_called' '_No' 'no' '_yes′_Yes' 'Yes' 'yes' '_YES' '_your' '_they' 'your' '_Your' '_all' Decoded top-k tokens Figure 4: Top-k deceode tokens con- trolled by different AUs. The answer to input prompt is “yes”. Figure 4 illustrates this phenomenon by reporting the top-5 output tokens after steering three different AUs with single-dimensional activations. The input prompt is a question from BoolQ dataset with the answer “yes”. Steering x 84 promotes the correct answer token “yes” while suppressing the incorrect “no”, thereby improving accuracy. In contrast, steering dimensions x 44 or x 100 el- evates task-irrelevant or incorrect tokens, resulting in de- graded performance. These observations align with the accuracies shown in Figure 2a. In summary, heterogeneity arises because each AU gov- erns a distinct output-token distribution. Block-level activations inevitably mix beneficial, irrelevant, and harmful AUs, making block-level interventions coarse, inefficient, and intrusive. By contrast, selectively steering only the helpful AUs amplifies the desired distribution and enables more efficient control. 4METHODOLOGY: AUSTEER Breaking block-level interventions into finer-grained AU-level interventions has shown promise for modifying LLM behaviors. Yet AU-level steering faces some fundamental challenges: identifying important AU-level activations for intervention and ensuring adaptability across diverse inputs and AUs. To address these challenges, we propose AUSteer shown as Figure 5. 5 Published as a conference paper at ICLR 2026 Block-Level Activations ...... AU-Level Activations Pos. Samples Pos. Samples Pos. Samples AU-Localization ... 풖 ퟏ ퟏ 풖 ퟐ ퟏ 풖 풊 ퟏ ... 풖 ퟏ 푵 풖 ퟐ 푵 풖 풊 푵 ... Discriminative Score 풔 ퟏ 풔 ퟐ ⋮ 풔 풊 풓 ퟏ 풑풔 > 풓 ퟏ 풏 ... 풓 ퟐ 풑풔 < 풓 ퟐ 풏 풓 풊 풑풔 > 풓 풊 풏 MAX ... ... Ranked AUsActivation Momentum Adaptive Steering Pos. Samples Pos. Samples Neg. Samples : Original AU-level activation : Steered AU-level activation, with color depth indicating strength 푢 푖 푗 : 푖-th AU on 푗-th sample Figure 5: Overview of AUSteer: (1) AU localization using activation momentum and discriminative scores; and (2) Adaptive steering across diverse inputs and AUs. 4.1ATOMIC UNIT LOCALIZATION The first challenge is to identify which AUs and their activations should be steered. Prior work often uses probing (Li et al., 2023b) or activation values (Wang et al., 2025b) of contrastive pairs as impor- tance metrics. Here, a contrastive pair consists of a positive example with a correct or high-quality response and a matched negative example with an incorrect or low-quality response. However, prob- ing requires additional training resources and does not transfer well to single-dimension settings of AUs, while activation magnitudes tend to increase with layer depth, making cross-layer comparisons unreliable. To overcome these limitations, we propose an activation momentum strategy. Given N pairs of contrastive samples 4 , an AU is discriminative if its activation coefficient x i con- sistently separates positives from their matched negatives. Concretely, if x i is systematically higher (or lower) for the positive sample of each pair than for the negative sample, the AU promotes (or suppresses) activation for positives relative to negatives. Such consistency indicates that the AU distinguishes positive from negative cases and is therefore task-relevant. Formally, let u i denote the i-th AU with the activation x i . For the j-th sample pair, we define the activation momentum as m j i = x j,pos i − x j,neg i , where x j,pos i and x j,neg i are the activation values of u i on the j-th positive and negative sample, respectively. Note that both x j,pos i and x j,neg i are one-dimensional scalars as defined in Eq.1. When m j i > 0, the AU exhibits an activation promotion effect for positive samples, whereas m j i < 0 indicates a suppression effect. By counting the occurrences of promotion and suppression across samples, we can assess whether an AU shows a consistent effect on positive or negative cases, thereby quantifying its discriminative power. The proportions of positive and negative momenta are then given by r pos i = 1 N N X j=1 1(m j i > 0), r neg i = 1 N N X j=1 1(m j i < 0).(2) The discriminative score of the i-th AU is defined as: s i = max(r pos i ,r neg i ). This scoring provides a unified scale for cross-layer comparison, allowing us to rank AUs globally and select the most important k AUs for steering. To verify how activation momentum contributes to discriminative causality and the final model outputs, we provide both theoretical and empirical analyses in Appendix H. 4.2ADAPTIVE STEERING The steering of an activation x i should be adaptive in two respects. First, it should adapt to diverse inputs. Different samples produce activations with different magnitudes and semantic contexts. Adding a constant vector ignores this variation, can distort useful directions, and may impair model 4 Details of contrastive sample construction can be found in Appendix B.1 6 Published as a conference paper at ICLR 2026 performance. We therefore obtain the steered activation by ˆx i = x i +γ i x i , which scales the current activation, preserves its sign, and adapts well across varies samples. Second, steering should adapt across AUs. More discriminative AUs receive stronger steering, while less important ones receive weaker intervention. This concentrates changes on useful AUs and limits unnecessary perturbations. To achieve this, we compute γ i as γ i = ( αr pos i , r pos i > r neg i , −αr neg i , otherwise, where α is a global steering strength factor, r pos i and r neg i are the positive and negative discriminative scores of the i-th AU calculated by Eq.2. The steering direction is determined by whether the AU has a promotive or suppressive effect. Finally, for the selected AUs, activations are updated as ˆx i = x i + γ i x i . Applicability of AUSteer. The proposed AUSteer can be applied to all key components of LLMs, including MHA, FFN, and residual streams, as the analysis above holds uniformly across these mod- ules. Unlike previous approaches that operate on entire block-level activations, AUSteer intervenes only on the most important dimensions within each block activation. This yields interventions that are both more efficient and less intrusive, embodying the principle of steering less to achieve more. 5EXPERIMENTS 5.1EXPERIMENTAL SETTINGS Tasks and Evaluation Metrics. We evaluate AUSteer on three types of tasks: • Commonsense reasoning. We use widely adopted datasets including BoolQ (Clark et al., 2019), COPA (Gordon et al., 2012), and WinoGrande (Sakaguchi et al., 2021), and report accuracy of the model’s responses using exact match. • Math problem solving. We experiment with SVAMP (Patel et al., 2021) and MAWPS (Koncel- Kedziorski et al., 2016), where the model is required to solve math questions with or without reasoning. We evaluate accuracy by comparing the predicted answer with the correct number. • Open-ended generation. We employ RealToxicPrompts (Gehman et al., 2020) and BPO (Cheng et al., 2024). For RealToxicPrompts, which contains challenging prompts that often elicit toxic content, we apply different steering methods to reduce toxicity. Automatic evaluation follows prior work (Wang et al., 2025a): detoxification performance, where toxicity is measured using the Perspective API 5 . For BPO, which aligns model outputs with human-preferred behaviors, we adopt the automatic evaluation protocol of Zheng et al. (2023); Liang et al. (2024) and report ∆WR = WR steered − WR original . The win-rates are obtained by using GPT-5-mini (OpenAI, 2025) and prompts from Liang et al. (2024) to compare the original and steered responses. For both datasets, human evaluation is conducted, where 3 annotators assess text quality (fluency, diversity) and alignment with the target objective on a 1–5 scale. Target LLMs. We evaluate AUSteer on a diverse set of LLMs: (1) LLaMA2-7B-Chat (Touvron et al., 2023), which serves as the backbone in many related studies; (2) Gemma2-9B-it (Team et al., 2024), a strong decoder-only model for text generation; and (3) Qwen3-8B (Yang et al., 2025), one of the most recent LLMs. To further assess scalability, we also experiment with other LLMs and larger variants (e.g., 13B, 27B), with results reported in §5.4. Baselines. We compare AUSteer against several competitive activation steering methods: • ITI (Li et al., 2023b), which uses contrastive samples to identify important attention heads, then derives steering vectors from activation differences for intervention. • CAA (Rimsky et al., 2024), which extracts steering vectors from activation differences in residual streams and applies them at the block level. 5 https://w.perspectiveapi.com 7 Published as a conference paper at ICLR 2026 Table 1: Overall results of baseline methods and the proposed AUSteer across seven tasks. “#Acts” denotes the number of intervened activations for each method. k h indicates the number of selected attention heads, ranging from 2 to 64. Results with † are from (Wang et al., 2025a;b). ModelMethod#Acts (↓) Commonsense Reasoning (↑) Math Problem Solving (↑)Avg.Open Generation (↑) BoolQ COPAWinoG.SVAMPMAWPSAcc.Detox.∆WR LLaMA2-7B-Chat Vanilla070.52 † 70.80 † 50.91 † 36.0051.8356.01– ITIk h · 12874.10 † 77.20 † 52.80 † 36.6754.0858.9783.4912.50 CAA409674.98 † 75.20 † 52.64 † 34.3355.2158.4784.5711.00 SADIk h · 12874.35 † 78.80 † 53.04 † 36.3354.9359.4986.3213.50 AUSteer-Head ≤ 10076.2775.4052.8037.6756.0659.6489.6616.00 AUSteer-FFN ≤ 10075.5782.8053.2837.0058.03 61.3489.2422.00 Gemma2-9B-it Vanilla086.6477.4067.2567.6792.1178.21– ITIk h · 25686.8288.4068.1169.0093.52 81.1798.8323.00 CAA358486.8589.4068.3568.0092.9681.11 98.75 † 23.50 SADIk h · 25686.7992.4069.5368.0093.5282.0599.0825.50 STA358487.0391.6069.6168.6792.3981.86 99.33 † 25.00 AUSteer-Head ≤ 10086.7991.0070.7271.0094.6582.8399.0026.50 AUSteer-FFN ≤ 10087.2597.6070.8870.0094.08 83.9699.2530.00 Qwen3-8B Vanilla087.4389.0063.6168.6784.5178.64– ITIk h · 12887.4090.8065.5969.6787.3280.1675.1526.00 CAA409687.8391.2064.0171.6788.4580.6376.2327.00 SADIk h · 12887.6593.2064.2573.0085.9280.8078.0725.50 AUSteer-Head ≤ 10087.7195.4065.3576.3391.2783.2180.9026.00 AUSteer-FFN ≤ 10088.2090.8067.2571.6789.5881.5081.6534.00 • SADI (Wang et al., 2025b), which localizes important attention heads, FFNs, or layers via activa- tion differences, and applies adaptive steering through masking and scaling. We report results for its best-performing variant, SADI-HEAD. • STA (Wang et al., 2025a), which identifies important atoms in sparse autoencoders (SAEs) of the target LLM, then applies steering vectors to residual streams. Since pretrained SAEs are currently available only for LLaMA 3.1 and Gemma2, we report its results only on Gemma2. AUSteer Variants. The proposed method can be applied to any key component of LLMs. Since the two core modules in each Transformer layer are the attention and FFN blocks, we validate the generalizability of AUSteer by implementing two variants: AUSteer-Head, which steers AU-level activations in MHA, and AUSteer-FFN, which steers AU-level activations in FFN. Implementation details for AUSteer and all other baseline methods, including contrastive pair construction, dataset statistics, and prompt templates, are provided in Appendix B. Hyperparameter settings. For baseline methods, we perform hyperparameter sweeps following the recommendations in their papers to ensure a fair comparison. AUSteer introduces two hyper- parameters: (1) k, the number of AU-level activations selected for steering; and (2) α, a global steering-strength factor. To verify the claim that we can steer less to achieve more, we cap the number of steered activations at 100 and then run the sweep. Full details appear in Appendix C. 5.2MAIN RESULTS AUSteer significantly improves commonsense reasoning and math problem solving with min- imal intervention. Table 1 reports overall results on LLaMA2-7B-Chat, Gemma2-9B-it, and Qwen3-8B. Across all five tasks on commonsense reasoning and math questions, either AUSteer- FFN or AUSteer-Head attains the highest average accuracy while steering at most 100 activations, in contrast to SADI’s k h × 128 head interventions and CAA/STA, which modify thousands of acti- vations. Concretely, AUSteer-FFN improves the average over SADI by +1.85 on LLaMA2-7B-Chat (61.34 vs. 59.49), +1.91 on Gemma2-9B-it (83.96 vs. 82.05), and +0.7 on Qwen3-8B. AUSteer- Head is also competitive, exceeding SADI on Qwen3-8B by +2.41 under the same low-budget con- straint. Beyond averages, AUSteer-Head or AUSteer-FFN consistently achieves the best scores on individual tasks across the five commonsense and math benchmarks. 8 Published as a conference paper at ICLR 2026 Table 2: Human evaluation on open-ended generation tasks. LLaMA2-7B-ChatGemma2-9B-itQwen3-8B SADIAUSteerSADISTAAUSteerSADIAUSteer Quality (↑)3.33.44.24.44.34.14.3 Alignment (↑)3.63.84.54.74.73.94.1 AUSteer improves open-ended gen- eration.In automatic evaluation (Table 1), AUSteer significantly in- creases detoxification rates under toxic prompts.Compared with SADI, it yields around 2%-3% higher detoxification on Llama2 and Qwen3. On BPO datasets, AUSteer steers models toward human-preferred responses, improving win-rates (∆WR) by 8.5%, 4.5%, and 7% on the three LLMs, respectively. In human evaluation (Table 2), AUSteer outperforms baselines in most cases on generation quality (fluency and diversity) and on alignment with the generation target. 5.3ABLATION STUDIES Table 3:Ablation study results on Gemma2-9B-it. MethodAvg. Acc AUSteer-FFN83.96 Random Loc.79.08 (-4.88) Act. Diff.83.12 (-0.84) Fixed. Vec.82.05 (-1.91) Fixed. Strength83.04 (-0.92) We evaluate the contribution of each component in AUS- teer: AU localization and adaptive steering. To assess the proposed activation momentum localization, we com- pare it with (1) random localization, which selects acti- vations at random for steering, and (2) activation differ- ence across contrastive samples for localization, as intro- duced in SADI. To assess adaptive steering, we compare it with (3) a fixed steering vector, which replaces γ i x i with the mean activation difference, following ITI, and (4) a fixed steering strength γ, which applies the same strength across all selected AUs. The average accuracy across commonsense reasoning and math questions are shown in Table 3. When using random or activation-difference localization, steering performance drops substantially, verifying the effectiveness of the proposed activation momentum-based localization. Similarly, replacing adaptive steering with a fixed vector or fixed strength reduces performance by 1.91 and 0.92, respectively, demonstrating the importance of adaptivity across diverse inputs and AUs. 5.4SCALABILITY AND GENERALIZABILITY OF AUSTEER Table 4: Experimental results on more LLMs. LLaMA3.1-8B-InstructLLaMA2-13B-ChatGemma2-27B-it BoolQCOPAWinoG.BoolQCOPAWinoG.BoolQCOPAWinoG. Vanilla82.5783.8057.7784.0189.0053.9986.8886.0063.61 AUSteer-Head83.1886.0060.3885.2591.0059.4388.1090.2067.25 AUSteer-FFN83.7986.0061.5685.0291.2058.8888.4189.8066.30 We evaluate AUSTEER on larger and varied LLMs, including LLaMA3.1-8B-Instruct, LLaMA2- 13B-Chat (Touvron et al., 2023), and Gemma2-27B-it (Team et al., 2024), on commonsense rea- soning tasks. Table 4 reports the results. Both AUSTEER-HEAD and AUSTEER-FFN substantially improve the base models, confirming the method’s scalability and generalizability. More results on larger LLMs with diverse structures including Qwen3-30B-A3B and Llama-3.3-70B-Instruct are provided in Appendidx G. 5.5FURTHER ANALYSIS To investigate the internal mechanisms of AUSteer more comprehensively, we provide the following discussions. • Appendix C. We illustrate the hyperparameter sweep for k and α and report their optimal values across tasks. We also provide guidelines for hyperparameter search in both resource-sufficient and resource-constrained settings. 9 Published as a conference paper at ICLR 2026 • Appendix D. We characterize activation momentum for different AUs and analyze the locations of AUs within MHA and FFN. • Appendix E. We present and discuss the overlap of localized AUs across tasks. • Appendix F. We evaluate AUSTEER under varying numbers of contrastive pairs used for AU localization. • Appendix G. We demonstrate AUSteer’s scalability on larger LLMs with diverse architectures, including Qwen3-30B-A3B (a sparse MoE model) and Llama-3.3-70B-Instruct (evaluated in its 4-bit quantized form). • Appendix H. We verify how activation momentum contributes to discriminative causality and final model outputs, providing both theoretical and empirical analyses. • Appendix I. We present a detailed analysis of AUSteer’s efficiency and computational overhead compared with baseline methods. Overhead results on Llama-3.3-70B-Instruct are also included. • Appendix J. We experiment with additional control variants of AUSteer—such as steering all AUs or broader subsets—and confirm that, as with AUSteer, steering should be limited to task-relevant and beneficial AUs rather than blindly steering all or large numbers of units. • Appendix K. To determine whether we should promote useful AUs or suppress unhelpful ones, we conduct both empirical and theoretical analyses and show that promotion consistently outperforms suppression. 6CONCLUSION In this work, we investigate the heterogeneity and its root cause of block-level activations and pro- pose AUSteer, a fine-grained AU-level activation steering method. AUSteer localizes salient AUs via activation momentum and assigns dynamic steering strengths per input and AU. Extensive ex- periments show that, with far fewer intervened activations, AUSteer significantly outperforms state- of-the-art methods across diverse tasks, demonstrating that steering less achieves more. ACKNOWLEDGMENTS We extend our heartfelt gratitude to the reviewers for their insightful and constructive feedback. This research was supported by the Home Team Science and Technology Agency (HTX), Singa- pore under the NTU-HTX collaboration project: Parsimonious Domain Specific Large Language Model Enabled Multimodality Sensemaking. We express our sincere appreciation to HTX for their continued support and collaboration. 10 Published as a conference paper at ICLR 2026 REFERENCES Yuntao Bai, Andy Jones, Kamal Ndousse, Amanda Askell, Anna Chen, Nova DasSarma, Dawn Drain, Stanislav Fort, Deep Ganguli, Tom Henighan, et al. Training a helpful and harmless assistant with reinforcement learning from human feedback. arXiv preprint arXiv:2204.05862, 2022. Amrita Bhattacharjee, Shaona Ghosh, Traian Rebedea, and Christopher Parisien. Towards inference- time category-wise safety steering for large language models. In Neurips Safe Generative AI Workshop 2024, 2024. Tom Brown, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D Kaplan, Prafulla Dhari- wal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, Sandhini Agar- wal, Ariel Herbert-Voss, Gretchen Krueger, Tom Henighan, Rewon Child, Aditya Ramesh, Daniel Ziegler, Jeffrey Wu, Clemens Winter, Chris Hesse, Mark Chen, Eric Sigler, Mateusz Litwin, Scott Gray, Benjamin Chess, Jack Clark, Christopher Berner, Sam McCandlish, Alec Radford, Ilya Sutskever, and Dario Amodei.Language models are few-shot learners.In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin (eds.), Advances in Neu- ral Information Processing Systems, volume 33, p. 1877–1901. Curran Associates, Inc., 2020.URL https://proceedings.neurips.c/paper_files/paper/2020/ file/1457c0d6bfcb4967418bfb8ac142f64a-Paper.pdf. Jiale Cheng, Xiao Liu, Kehan Zheng, Pei Ke, Hongning Wang, Yuxiao Dong, Jie Tang, and Minlie Huang. Black-box prompt optimization: Aligning large language models without model training. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Proceedings of the 62nd Annual Meet- ing of the Association for Computational Linguistics (Volume 1: Long Papers), p. 3201–3219, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/ 2024.acl-long.176. URL https://aclanthology.org/2024.acl-long.176/. Christopher Clark, Kenton Lee, Ming-Wei Chang, Tom Kwiatkowski, Michael Collins, and Kristina Toutanova.BoolQ: Exploring the surprising difficulty of natural yes/no questions.In Jill Burstein, Christy Doran, and Thamar Solorio (eds.), Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Lan- guage Technologies, Volume 1 (Long and Short Papers), p. 2924–2936, Minneapolis, Min- nesota, June 2019. Association for Computational Linguistics. doi: 10.18653/v1/N19-1300. URL https://aclanthology.org/N19-1300/. Guy Dar, Mor Geva, Ankit Gupta, and Jonathan Berant. Analyzing transformers in embedding space. In Anna Rogers, Jordan Boyd-Graber, and Naoaki Okazaki (eds.), Proceedings of the 61st Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pa- pers), p. 16124–16170, Toronto, Canada, July 2023. Association for Computational Linguis- tics. doi: 10.18653/v1/2023.acl-long.893. URL https://aclanthology.org/2023. acl-long.893/. Zijian Feng, Hanzhang Zhou, Zixiao Zhu, Tianjiao Li, Chua Jia Jim Deryl, Mak Lee Onn, Gee Wah Ng, and Kezhi Mao. Restoring pruned large language models via lost component compensation. In The Thirty-ninth Annual Conference on Neural Information Processing Systems, 2025. Leo Gao, Tom Dupre la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders. In The Thirteenth Inter- national Conference on Learning Representations, 2025. Samuel Gehman, Suchin Gururangan, Maarten Sap, Yejin Choi, and Noah A. Smith. RealToxic- ityPrompts: Evaluating neural toxic degeneration in language models. In Trevor Cohn, Yulan He, and Yang Liu (eds.), Findings of the Association for Computational Linguistics: EMNLP 2020, p. 3356–3369, Online, November 2020. Association for Computational Linguistics. doi: 10.18653/v1/2020.findings-emnlp.301. URL https://aclanthology.org/2020. findings-emnlp.301/. Mor Geva, Avi Caciularu, Kevin Wang, and Yoav Goldberg. Transformer feed-forward layers build predictions by promoting concepts in the vocabulary space. In Yoav Goldberg, Zornitsa Kozareva, and Yue Zhang (eds.), Proceedings of the 2022 Conference on Empirical Methods 11 Published as a conference paper at ICLR 2026 in Natural Language Processing, p. 30–45, Abu Dhabi, United Arab Emirates, December 2022. Association for Computational Linguistics. doi: 10.18653/v1/2022.emnlp-main.3. URL https://aclanthology.org/2022.emnlp-main.3/. Andrew Gordon, Zornitsa Kozareva, and Melissa Roemmele. SemEval-2012 task 7: Choice of plau- sible alternatives: An evaluation of commonsense causal reasoning. In Eneko Agirre, Johan Bos, Mona Diab, Suresh Manandhar, Yuval Marton, and Deniz Yuret (eds.), *SEM 2012: The First Joint Conference on Lexical and Computational Semantics – Volume 1: Proceedings of the main conference and the shared task, and Volume 2: Proceedings of the Sixth International Workshop on Semantic Evaluation (SemEval 2012), p. 394–398, Montr ́ eal, Canada, 7-8 June 2012. Asso- ciation for Computational Linguistics. URL https://aclanthology.org/S12-1052/. Chi Han, Jialiang Xu, Manling Li, Yi Fung, Chenkai Sun, Nan Jiang, Tarek Abdelzaher, and Heng Ji. Word embeddings are steers for language models. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Proceedings of the 62nd Annual Meeting of the Association for Com- putational Linguistics (Volume 1: Long Papers), p. 16410–16430, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.acl-long.864. URL https://aclanthology.org/2024.acl-long.864/. Peixuan Han, Cheng Qian, Xiusi Chen, Yuji Zhang, Denghui Zhang, and Heng Ji. Internal activation as the polar star for steering unsafe llm behavior. arXiv preprint arXiv:2502.01042, 2025. Rima Hazra, Sayan Layek, Somnath Banerjee, and Soujanya Poria. Safety arithmetic: A framework for test-time safety alignment of language models by steering parameters and activations. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Proceedings of the 2024 Conference on Empirical Methods in Natural Language Processing, p. 21759–21776, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main. 1212. URL https://aclanthology.org/2024.emnlp-main.1212/. Zhengfu He, Wentao Shu, Xuyang Ge, Lingjie Chen, Junxuan Wang, Yunhua Zhou, Frances Liu, Qipeng Guo, Xuanjing Huang, Zuxuan Wu, et al. Llama scope: Extracting millions of features from llama-3.1-8b with sparse autoencoders. arXiv preprint arXiv:2410.20526, 2024. Houcheng Jiang, Junfeng Fang, Ningyu Zhang, Mingyang Wan, Guojun Ma, Xiang Wang, Xiangnan He, and Tat-Seng Chua. Anyedit: Edit any knowledge encoded in language models. In Forty- second International Conference on Machine Learning, 2025. Shahar Katz, Yonatan Belinkov, Mor Geva, and Lior Wolf.Backward lens: Projecting lan- guage model gradients into the vocabulary space. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Proceedings of the 2024 Conference on Empirical Methods in Nat- ural Language Processing, p. 2390–2422, Miami, Florida, USA, November 2024. Associa- tion for Computational Linguistics. doi: 10.18653/v1/2024.emnlp-main.142. URL https: //aclanthology.org/2024.emnlp-main.142/. Rik Koncel-Kedziorski, Subhro Roy, Aida Amini, Nate Kushman, and Hannaneh Hajishirzi. MAWPS: A math word problem repository. In Kevin Knight, Ani Nenkova, and Owen Rambow (eds.), Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, p. 1152–1157, San Diego, Califor- nia, June 2016. Association for Computational Linguistics. doi: 10.18653/v1/N16-1136. URL https://aclanthology.org/N16-1136/. Kai Konen, Sophie Jentzsch, Diaoul ́ e Diallo, Peer Sch ̈ utt, Oliver Bensch, Roxanne El Baff, Dominik Opitz, and Tobias Hecking. Style vectors for steering generative large language models. In Yvette Graham and Matthew Purver (eds.), Findings of the Association for Computational Linguistics: EACL 2024, p. 782–802, St. Julian’s, Malta, March 2024. Association for Computational Lin- guistics. URL https://aclanthology.org/2024.findings-eacl.52/. Chong Li, Shaonan Wang, Yunhao Zhang, Jiajun Zhang, and Chengqing Zong. Interpreting and exploiting functional specialization in multi-head attention under multi-task learning. In Houda Bouamor, Juan Pino, and Kalika Bali (eds.), Proceedings of the 2023 Conference on Empiri- cal Methods in Natural Language Processing, p. 16460–16476, Singapore, December 2023a. Association for Computational Linguistics. doi: 10.18653/v1/2023.emnlp-main.1026. URL https://aclanthology.org/2023.emnlp-main.1026/. 12 Published as a conference paper at ICLR 2026 Kenneth Li, Oam Patel, Fernanda Vi ́ egas, Hanspeter Pfister, and Martin Wattenberg. Inference- time intervention: Eliciting truthful answers from a language model.In A. Oh, T. Nau- mann, A. Globerson, K. Saenko, M. Hardt, and S. Levine (eds.), Advances in Neural Information Processing Systems, volume 36, p. 41451–41530. Curran Associates, Inc., 2023b.URL https://proceedings.neurips.c/paper_files/paper/2023/ file/81b8390039b7302c909cb769f8b6cd93-Paper-Conference.pdf. Kenneth Li, Oam Patel, Fernanda Vi ́ egas, Hanspeter Pfister, and Martin Wattenberg. Inference-time intervention: Eliciting truthful answers from a language model. Advances in Neural Information Processing Systems, 36:41451–41530, 2023c. Zhaoyi Li, Gangwei Jiang, Hong Xie, Linqi Song, Defu Lian, and Ying Wei. Understanding and patching compositional reasoning in LLMs. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Findings of the Association for Computational Linguistics: ACL 2024, p. 9668–9688, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/ v1/2024.findings-acl.576. URL https://aclanthology.org/2024.findings-acl. 576/. Zihan Liang, Ben Chen, Zhuoran Ran, Zihan Wang, Huangyu Dai, Yufei Ma, Dehong Gao, Xi- aoyan Cai, and Libin Yang. Self-renewal prompt optimizing with implicit reasoning. In Yaser Al-Onaizan, Mohit Bansal, and Yun-Nung Chen (eds.), Findings of the Association for Com- putational Linguistics: EMNLP 2024, p. 3030–3041, Miami, Florida, USA, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.findings-emnlp.171. URL https://aclanthology.org/2024.findings-emnlp.171/. Tom Lieberum, Senthooran Rajamanoharan, Arthur Conmy, Lewis Smith, Nicolas Sonnerat, Vikrant Varma, Janos Kramar, Anca Dragan, Rohin Shah, and Neel Nanda. Gemma scope: Open sparse autoencoders everywhere all at once on gemma 2. In Yonatan Belinkov, Najoung Kim, Jaap Jumelet, Hosein Mohebbi, Aaron Mueller, and Hanjie Chen (eds.), Proceedings of the 7th Black- boxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, p. 278–300, Miami, Florida, US, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024. blackboxnlp-1.19. URL https://aclanthology.org/2024.blackboxnlp-1.19/. Clement Neo, Luke Ong, Philip Torr, Mor Geva, David Krueger, and Fazl Barez. Towards inter- preting visual information processing in vision-language models. In The Thirteenth International Conference on Learning Representations, 2025. OpenAI. Gpt-5 models. https://platform.openai.com/docs/models, 2025. Arkil Patel, Satwik Bhattamishra, and Navin Goyal. Are NLP models really able to solve sim- ple math word problems? In Kristina Toutanova, Anna Rumshisky, Luke Zettlemoyer, Dilek Hakkani-Tur, Iz Beltagy, Steven Bethard, Ryan Cotterell, Tanmoy Chakraborty, and Yichao Zhou (eds.), Proceedings of the 2021 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, p. 2080–2094, Online, June 2021. Association for Computational Linguistics. doi: 10.18653/v1/2021.naacl-main.168. URL https://aclanthology.org/2021.naacl-main.168/. Joris Postmus and Steven Abreu. Steering large language models using conceptors: Improving addition-based activation engineering. arXiv preprint arXiv:2410.16314, 2024. Nate Rahn, Pierluca D’Oro, and Marc G Bellemare. Controlling large language model agents with entropic activation steering. In ICML 2024 Workshop on Mechanistic Interpretability, 2024. Nina Rimsky, Nick Gabrieli, Julian Schulz, Meg Tong, Evan Hubinger, and Alexander Turner. Steering llama 2 via contrastive activation addition.In Lun-Wei Ku, Andre Martins, and Vivek Srikumar (eds.), Proceedings of the 62nd Annual Meeting of the Association for Com- putational Linguistics (Volume 1: Long Papers), p. 15504–15522, Bangkok, Thailand, August 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.acl-long.828. URL https://aclanthology.org/2024.acl-long.828/. Keisuke Sakaguchi, Ronan Le Bras, Chandra Bhagavatula, and Yejin Choi. Winogrande: an adver- sarial winograd schema challenge at scale. Commun. ACM, 64(9):99–106, August 2021. ISSN 0001-0782. doi: 10.1145/3474381. URL https://doi.org/10.1145/3474381. 13 Published as a conference paper at ICLR 2026 Samuel Soo, Wesley Teng, and Chandrasekaran Balaganesh. Steering large language models with feature guided activation additions. arXiv e-prints, p. arXiv–2501, 2025. Asa Cooper Stickland, Alexander Lyzhov, Jacob Pfau, Salsabila Mahdi, and Samuel R Bowman. Steering without side effects: Improving post-deployment control of language models. In Neurips Safe Generative AI Workshop 2024, 2024. Alessandro Stolfo, Vidhisha Balachandran, Safoora Yousefi, Eric Horvitz, and Besmira Nushi. Im- proving instruction-following in language models through activation steering. In The Thirteenth International Conference on Learning Representations, 2025. Gemma Team, Thomas Mesnard, Cassidy Hardin, Robert Dadashi, Surya Bhupatiraju, Shreya Pathak, Laurent Sifre, Morgane Rivi ` ere, Mihir Sanjay Kale, Juliette Love, et al. Gemma: Open models based on gemini research and technology. arXiv preprint arXiv:2403.08295, 2024. Hugo Touvron, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Niko- lay Bashlykov, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, et al. Llama 2: Open founda- tion and fine-tuned chat models. arXiv preprint arXiv:2307.09288, 2023. Alexander Matt Turner, Lisa Thiergart, Gavin Leech, David Udell, Juan J Vazquez, Ulisse Mini, and Monte MacDiarmid. Activation addition: Steering language models without optimization. arXiv e-prints, p. arXiv–2308, 2023. Teun van der Weij, Massimo Poesio, and Nandi Schoots. Extending activation steering to broad skills and multiple behaviours. arXiv preprint arXiv:2403.05767, 2024. Mengru Wang, Ziwen Xu, Shengyu Mao, Shumin Deng, Zhaopeng Tu, Huajun Chen, and Ningyu Zhang. Beyond prompt engineering: Robust behavior control in LLMs via steering target atoms. In Wanxiang Che, Joyce Nabende, Ekaterina Shutova, and Mohammad Taher Pilehvar (eds.), Proceedings of the 63rd Annual Meeting of the Association for Computational Linguistics (Vol- ume 1: Long Papers), p. 23381–23399, Vienna, Austria, July 2025a. Association for Compu- tational Linguistics. ISBN 979-8-89176-251-0. doi: 10.18653/v1/2025.acl-long.1139. URL https://aclanthology.org/2025.acl-long.1139/. Weixuan Wang, JINGYUAN YANG, and Wei Peng. Semantics-adaptive activation intervention for llms via dynamic steering vectors. In The Thirteenth International Conference on Learning Representations, 2025b. Xintong Wang, Jingheng Pan, Liang Ding, Longyue Wang, Longqin Jiang, Xingshan Li, and Chris Biemann. CogSteer: Cognition-inspired selective layer intervention for efficiently steering large language models. In Wanxiang Che, Joyce Nabende, Ekaterina Shutova, and Mohammad Taher Pilehvar (eds.), Findings of the Association for Computational Linguistics: ACL 2025, p. 25507– 25522, Vienna, Austria, July 2025c. Association for Computational Linguistics. ISBN 979-8- 89176-256-5. doi: 10.18653/v1/2025.findings-acl.1308. URL https://aclanthology. org/2025.findings-acl.1308/. Jason Wei, Maarten Bosma, Vincent Zhao, Kelvin Guu, Adams Wei Yu, Brian Lester, Nan Du, An- drew M Dai, and Quoc V Le. Finetuned language models are zero-shot learners. In International Conference on Learning Representations, 2022. An Yang, Anfeng Li, Baosong Yang, Beichen Zhang, Binyuan Hui, Bo Zheng, Bowen Yu, Chang Gao, Chengen Huang, Chenxu Lv, et al.Qwen3 technical report.arXiv preprint arXiv:2505.09388, 2025. Lianmin Zheng, Wei-Lin Chiang, Ying Sheng, Siyuan Zhuang, Zhanghao Wu, Yonghao Zhuang, Zi Lin, Zhuohan Li, Dacheng Li, Eric P. Xing, Hao Zhang, Joseph E. Gonzalez, and Ion Stoica. Judging llm-as-a-judge with mt-bench and chatbot arena. In Proceedings of the 37th International Conference on Neural Information Processing Systems, NIPS ’23, Red Hook, NY, USA, 2023. Curran Associates Inc. Andy Zou, Long Phan, Sarah Chen, James Campbell, Phillip Guo, Richard Ren, Alexander Pan, Xuwang Yin, Mantas Mazeika, Ann-Kathrin Dombrowski, et al. Representation engineering: A top-down approach to ai transparency. arXiv preprint arXiv:2310.01405, 2023. 14 Published as a conference paper at ICLR 2026 AMORE RESULTS ON ACTIVATION HETEROGENEITY We present additional results for attention heads and FFNs in Figures 6 to 8, confirming heterogene- ity in block level activations. 020406080100120 Dimension Index 62 64 66 68 70 72 74 76 Accuracy Dimension Sweep Baseline SADI Pos. Comb. Mixed Comb. Figure 6: Steering results for 26th attention output at 15th layer. Positive Combination: steering four beneficial dimensions (32, 36, 48, 64). Mixed Combination: steering those four plus two detrimental dimensions (44, 88). 020406080100120 Dimension Index 70 71 72 73 74 75 Accuracy Dimension Sweep Baseline SADI Pos. Comb. Mixed Comb. Figure 7: Steering results for 1st attention output at 19th layer. Positive Combination: steering three beneficial dimensions (28, 40, 100). Mixed Combination: steering those three plus two detrimental dimensions (64, 68). 05001000150020002500300035004000 Dimension Index 66 68 70 72 74 Accuracy Dimension Sweep Baseline SADI Pos. Comb. Mixed Comb. Figure 8: Steering results for the FFN output at layer 17. Positive Combination: steering three beneficial dimensions (400, 800, 2400). Mixed Combination: steering those three plus two detri- mental dimensions (2200, 3500). BDETAILED EXPERIMENT SETUP B.1CONTRASTIVE SAMPLE CONSTRUCTION Contrastive sample pairs are required by AUSteer and all baseline methods. To ensure a fair com- parison, we follow SADI (Wang et al., 2025b) and STA (Wang et al., 2025a) to construct same pairs 15 Published as a conference paper at ICLR 2026 for every method. For each sample in commonsense reasoning, we form a positive sample by con- catenating the question with the correct answer, and a negative sample by concatenating the question with a randomly selected incorrect answer. In math problem solving, we use the question plus the correct answer as the positive sample. For the negative sample, we use a sentence encoder to select the most semantically similar incorrect answer from the answer pool and concatenate it with the question. For detoxification, we select entries from RealToxicityPrompts with high toxicity scores as negative prompts. Following STA, the safe response is used as the positive sample. In BPO, we use the original prompt paired with a high-quality (human-preferred) response as the positive sample, and the same prompt paired with a low-quality response from the dataset as the negative sample. We clarify that (1) contrastive samples are required by almost all activation steering methods and are a common practice in prior work (Li et al., 2023b; Rimsky et al., 2024; Wang et al., 2025b;a), rather than a limitation unique to AUSteer; (2) constructing these pairs is generally straightforward based on available samples and easy to implement; and (3) we provide and verify a simple, general, and ready-to-use procedure for constructing contrastive pairs across different and new tasks. (1) Contrastive samples are widely required in activation steering. Existing activation steering methods, including ITI, CAA, SADI, and STA, all rely on contrastive positive–negative samples to localize important components and/or to estimate steering vectors. Thus, the requirement of contrastive pairs is not a limitation specific to AUSteer, but rather a standard and widely adopted practice. For fair comparison, we also ensure that all baseline methods use the same contrastive pairs in our experiments. (2) Constructing contrastive pairs is simple in practice. Following prior work such as SADI and STA, constructing contrastive pairs is straightforward. For commonsense reasoning tasks, the neg- ative sample can be obtained by pairing the question with an incorrect answer. For other datasets, negative samples can be generated by selecting semantically similar responses from a pool of can- didate answers, or by using datasets that already include ready-to-use negative samples. (3) A general solution for new tasks. For tasks not covered in existing studies, we use a general and effective approach. Positive sample: concatenate the question with the correct answer. Negative sample: use a sentence encoder to identify the most semantically similar incorrect answer from the answer pool and concatenate it with the question (Feng et al., 2025). For example, previous studies did not include math tasks, so we constructed contrastive pairs for those tasks using this method. For all other tasks, we use the contrastive pairs provided by prior work to ensure fair comparison. (4) Empirical verification of the general solution. Using the above general construction method, we re-evaluated AUSteer on Llama2-7B-Chat. As shown in Table 5, this simple approach achieves performance comparable to or even slightly better than our original results. Table 5: Results on LLaMA2-7B-Chat with new contrastive pairs. MethodAvg. Acc. (5 tasks)DetoxBPO Vanilla56.01– SADI59.4986.3213.50 AUSteer (previous result)61.3489.2422.00 AUSteer (new solution)61.5389.9922.50 In summary, contrastive pairs are commonly required across activation steering studies and are not a unique limitation of AUSteer. Moreover, constructing them is straightforward, and our general solution is simple, effective, and empirically validated to yield strong performance. We acknowledge that the reliance on contrastive pairs is an inherent limitation of existing activation-steering methods, and we plan to explore approaches that reduce or eliminate this requirement in future work. B.2DATA STATISTICS Following SADI, we use at most 1,000 contrastive pairs per task to identify important MHA and FFN components or to generate steering vectors. For evaluation, we use the full test set of each task. Detailed dataset statistics are provided in Table 6. 16 Published as a conference paper at ICLR 2026 Table 6: The number of contrastive pairs and testing samples for 7 tasks. BoolQCOPAWinoGrandeSVAMPMAWPSDetoxBPO # of contrastive pairs100010001000700100010001000 Test327050012673003551199200 B.3PROMPTS FOR DATASETS AND EVALUATION To ensure a fair comparison, we use identical prompt templates across all methods. For common- sense reasoning tasks, the templates strictly follow SADI (Wang et al., 2025b) and the authors’ released code. For RealToxicityPrompts, the templates follow STA (Wang et al., 2025a). Figure 9 shows the templates for SVAMP and MAWPS. For BPO, we use the prompts provided in the dataset directly. SVAMP Answer the following grade-school math word problem. Reply with only the final answer as a number. Question: question Answer: MAWPS Answer the grade school math word problem below, using step-by-step problem-solving process. Print the final answer after \"####\. Question: question Answer: Figure 9: Prompt templates for math problems. CHYPERPARAMETER SENSITIVITY 5102050100 32 50 64 80 100 k 79.0080.2081.6064.2043.00 87.2091.4094.8095.4094.40 87.6093.6096.2096.4042.40 87.8094.8097.0097.6059.60 88.4093.8096.2096.8042.20 COPA 15102050 8 16 32 50 64 k 92.3992.1192.6893.2493.52 92.3992.3992.6894.0893.24 92.1192.9694.0893.5292.39 92.3993.2493.2492.9692.96 92.1192.9692.9692.1193.52 MAWPS 0.5151020 8 16 32 50 64 k 3.5022.5024.0030.0019.50 15.0013.5011.5011.0014.00 14.5025.0022.00-1.0014.00 26.5010.509.0024.504.50 15.0021.0026.5011.00-3.50 BPO 60 80 92.5 93.0 93.5 94.0 0 10 20 30 Figure 10: Performance heatmaps for COPA, MAWPS, and BPO tasks as functions of α and k. AUSteer introduces two hyperparameters: (1) k, the number of AU-level activations selected for steering; and (2) α, a global steering-strength factor. To verify the claim that we can steer less to achieve more, we cap the number of steered activations at 100 and sweep both k and α from 1 to 100 for main experiments. Figure 10 reports performance across COPA, MAWPS, and BPO. Neighboring settings around the optimal hyperparameters achieve comparable results, indicating robustness. The optimal values vary across tasks to some extent, showing that the hyperparameters are task-specific, a trend consistent with Wang et al. (2025b). To set the hyperparameters for each task, we provide two solutions: (1) under sufficient computing resources, we perform a full hyperparameter sweep, which is consistent with previous studies (Li et al., 2023b; Rimsky et al., 2024; Wang et al., 2025b;a); and (2) in computing-constrained scenarios, we recommend using a very small validation set to conduct a quick hyperparameter sweep. In addition, (3) the optimal hyperparameters used in our experiments are reported in Tables 8 and 9. General hyperparameter sweep (resource-sufficient case). Task-specific hyperparameters are still a common challenge in activation steering, and the standard solution used widely in existing studies is to perform a sweep (Li et al., 2023b; Rimsky et al., 2024; Wang et al., 2025b;a). Following these studies, we perform a full hyperparameter sweep to empirically determine optimal α and k. 17 Published as a conference paper at ICLR 2026 We also run the same sweep for all baseline methods to ensure fair comparison in Table 1. Across tasks, both α and k typically fall within 1–100 and consistently yield strong results. Fast sweep using a small validation set (resource- or time-constrained case). When resources are limited, we recommend sweeping using only 50–100 validation samples. This process is ex- tremely fast (e.g., ∼5 minutes on an H100 GPU for 100 samples for the COPA task). Results using this small-set search are shown in below Table 7. It can be observed that even with only very few samples for hyperparameter selection, our proposed method still significantly outperforms the baseline methods and achieves results comparable to the full search. Table 7: Results of fast sweep on LLaMA2-7B-Chat MethodAvg. Acc. (5 tasks)DetoxBPO Vanilla56.01– SADI59.4986.3213.50 AUSteer (100-sample search)61.0388.4922.00 AUSteer (Full search)61.3489.2422.00 The optimal values of α and k used for each task are reported in Tables 8 and 9. These values were obtained via full sweep, and the same process was applied to baseline methods for fairness. The task-specific variation of hyperparameters aligns with observations from prior work, indicating that different tasks may require different hyperparameter values. However, for any given task, the hyperparameters are stable and robust. For example, an shown in Figure 10, for the COPA task, when 20 ≤ α ≤ 50 and 64 ≤ k ≤ 100, the performance remains stable and varies within only 1.5%, while still significantly outperforming the baseline methods. For the MAWPS task, when 10 ≤ α ≤ 50 and 16 ≤ k ≤ 50, the performance also varies within approximately 1.5%. Therefore, for each specific task, our method is hyperparameter-robust, and within the optimal region, it achieves comparable results with only small variations. Table 8: Optimal α for main experiments. BoolQCOPAWinograndeSVAMPMAWPSDetoxic.BPO 1550100881532 505010010050810 10202010501016 Table 9: Optimal k for main experiments. BoolQCOPAWinograndeSVAMPMAWPSDetoxic.BPO 10016250801616 8806448168 10081001002810 In summary, although the hyperparameters remain robust within an individual task, task-specific hyperparameters are still a common challenge in activation steering. The standard solution used widely in existing studies is to perform a sweep. To further reduce cost, we show that sweeping on a very small validation set is both efficient and highly effective, while still outperforming strong baselines. We will explore more principled approaches to reducing task-dependent hyperparameter sensitivity in future work. DCHARACTERISTICS OF ACTIVATION MOMENTUM AND LOCALIZED AUS Figures 11a and 11b report the discriminative score s i for each AU in both MHA and FFN, computed via activation momentum. We observe pronounced heterogeneity: within attention heads and FFN blocks, some dimensions/AUs are strongly discriminative while others are not. Moreover, most AUs localize to the middle or latter layers, consistent with prior findings (Wang et al., 2025b) that middle layers support reasoning while latter layers are critical for language generation. 18 Published as a conference paper at ICLR 2026 05001000150020002500300035004000 Dimension 0 5 10 15 20 25 30 35 40 Layer > 0.85 0.2 0.4 0.6 0.8 Value (a) AU scores in the MHA of Gemma2-9B-it on the COPA dataset. 0500100015002000250030003500 Dimension 0 5 10 15 20 25 30 35 40 Layer > 0.85 0.5 0.6 0.7 0.8 0.9 Value (b) AU scores in the FFN of Gemma2-9B-it on the COPA dataset. Figure 11: Characteristics of AUs in MHA and FFN. EAU OVERLAP ACROSS TASKS 1234567 1 2 3 4 5 6 7 1.0 0.11.0 0.00.11.0 0.00.00.01.0 0.00.00.00.01.0 0.00.00.00.00.01.0 0.00.00.00.00.00.01.0 (a) Overlap of identified AUs in the MHA of Gemma2-9B-it across different tasks. 1234567 1 2 3 4 5 6 7 1.0 0.01.0 0.00.11.0 0.00.00.01.0 0.00.00.00.01.0 0.00.00.00.00.01.0 0.00.00.00.00.00.01.0 (b) Overlap of identified AUs in the FFN of Gemma2-9B-it across different tasks. Figure 12: Overlap of localized AUs across tasks. Tasks 1–7 correspond to BoolQ, COPA, Wino- Grande, SVAMP, MAWPS, Detoxification, and BPO. We visualize the overlap of localized AUs across tasks in Figures 12a and 12b. Only very few AUs are shared between tasks, indicating that the AUs supporting different functions are highly specialized. This pattern is consistent with prior studies (Li et al., 2023a; Wang et al., 2025b). FSTEERING STABILITY WITH VARYING DATA SIZE Following prior work, AUSteer uses contrastive sample pairs to localize important AUs. We evaluate how its accuracy varies with the number of contrastive pairs. As shown in Figure 13, the accuracy on Gemma2-9B-it improves as the dataset grows. Notably, with 300–500 pairs, AUSteer achieves per- formance comparable to using 1,000 pairs. This demonstrates its effectiveness in low-data regimes. GSCALABILITY ON MORE LLMS To further verify the generalizability and scalability of AUSteer, we evaluate it on two representative large models with diverse structures: (1) Qwen3-30B-A3B, a 30B-scale sparse MoE model; and (2) Llama-3.3-70B-Instruct, where we use the 4-bit quantized version to enable evaluation on a consumer GPU and to test AUSteer’s compatibility with heavily quantized LLMs. The results are shown in Table 10. In most cases, AUSteer improves performance by 1%–3%, confirming its effectiveness and scalability across larger, structurally diverse and heavily quantized LLMs. 19 Published as a conference paper at ICLR 2026 2004006008001000 Data Size 77.5 80.0 82.5 85.0 87.5 90.0 92.5 95.0 97.5 Accuracy (%) AUSteer-Attn on COPA AUSteer-FFN on COPA AUSteer-Attn on BoolQ AUSteer-FFN on BoolQ Figure 13: Relationship between accuracy and the number of contrastive pairs Table 10: More results of diverse and larger LLMs. ModelMethodBoolQCOPAWinoG Qwen3-30B-A3BVanilla86.8293.4065.98 AUSteer88.6997.8067.17 Llama-3.3-70B-InstructVanilla89.5498.6078.14 AUSteer90.6799.2079.95 HANALYZING ACTIVATION MOMENTUM IN DISCRIMINATIVE CAUSALITY We explain the connection between activation momentum and causality based on both theoretical justification and empirical evidence. Theoretical Justification. Prior work in LLM interpretability (Dar et al., 2023; Geva et al., 2022; Li et al., 2024; Katz et al., 2024; Neo et al., 2025) shows that intermediate hidden states x in LLMs can be directly projected to the output logits through the LM head. This projection directly affects the model’s final next-token distribution. Formally, the LM headM computes: o =Mx, where o is the vector of output logits. This aligns with our observations in Section 3.3: different AUs govern different output token distri- butions, and as steering strength increases, the LLM’s output tends to converge to the AU’s token distribution. For a contrastive pair, the logit difference caused by the two inputs is: ∆o = o pos − o neg . For AU u i and contrastive pair j, define the activation momentum: m j i = x pos i − x neg i . Based on o =Mx, we apply a first-order Taylor expansion around x neg i : o(x pos i )≈ o(x neg i ) + ∂o ∂x i x pos i − x neg i . Rearranging gives: ∆o j i = o pos − o neg ≈ ∂o ∂x i m j i . 20 Published as a conference paper at ICLR 2026 This equation shows that the change in activation of AU u i directly causes a proportional change in the output logits. Thus: • m j i > 0 tends to increase the logit difference favoring the positive sample. • m j i < 0 tends to favor the negative sample. • If m j i is consistent across many pairs, then the AU u i has a stable discriminative causal effect on the output logits. This provides the theoretical grounding for activation momentum. Empirical Evidence To further validate the effectiveness of activation momentum, we compare it against two alternatives: (i) randomly selected AUs and (i) the activation-difference method used in SADI. On Gemma2-9B-it, the performance follows the order: 83.96 (activation momentum, ours) > 83.12 (activation difference by SADI) > 79.08 (random selection). These results demonstrate the superior performance of activation momentum. Additional experimental details are provided in Section 5.3. To summarize, we establish the connection between activation momentum and discriminative output causality through both theoretical analysis and empirical validation, thereby grounding and verifying our method. ICOMPUTATION OVERHEAD ANALYSIS We conducted a detailed efficiency and computation analysis from two perspectives: (1) smaller steering footprint, and (2) the actual computational overhead measured in practice, including activation-momentum computation time, inference-time cost, latency, and stability. Our results show that AUSteer requires less overhead and fewer interventions while achieving better performance than baseline methods. Smaller steering footprint. As shown in Table 1, baseline methods typically require intervening on 3,000–4,000 activations (or k h × 128), whereas AUSteer requires at most 100 intervened activations while still achieving the best results on most tasks. Detailed overhead analysis. We examine the computational overhead of our method and all base- lines at each stage of the method. In the preparation phase, AUSteer extracts activations from con- trastive pairs to compute activation momentum, whereas baseline methods usually require compo- nent localization and steering-vector estimation. During inference, each method applies its corre- sponding intervention, and we compare the resulting overhead across methods. It is worth men- tioning that activation momentum calculation only requires a single forward pass over a small set of contrastive examples. No backward pass, gradient computation, model modification, or training is needed. Extracting activations simply involves reading intermediate hidden states. Therefore, for any LLM size, activation momentum can always be computed using the same GPU memory required for standard inference, since both perform identical forward passes. Table 11 below compares the computation cost of AUSteer with ITI and SADI across six metrics: (1) GPU memory for contrastive samples of all tasks, (2) total runtime on all contrastive samples, (3) GPU memory during inference, (4) inference time over seven tasks, (5) latency, and (6) latency stability (std from five repeated trials). It is noted that all methods rely on contrastive samples to compute the necessary steering signals—whether for activation differences, localization, activation momentum (ours), or steering-vector estimation (other methods). The backbone LLM is Gemma2- 9B-it (batch size = 1, GPU = NVIDIA H100). Compared to other activation-steering baselines, AUSteer has the lowest overhead while achieving the best results. Specifically, AUSteer requires only∼15 minutes to compute activation momentum and localization, no additional GPU memory beyond inference, and exhibits lower overhead than ITI and SADI while achieving better performance, demonstrating its computational efficiency. During inference (steering), AUSteer also requires slightly less time than the baseline methods, further demonstrating its efficiency in runtime overhead. 21 Published as a conference paper at ICLR 2026 Table 11: Computation Overhead Comparison. MethodGPU Memory (contrastive)Time (contrastive)GPU Memory (Inference)Inference TimeLatencyStability Vanilla LLM–18 GB53 min 12 sec0.45 s/sample∆0.005 ITI18 GB18 min 39 sec18 GB59 min 29 sec0.50 s/sample∆0.01 SADI18 GB14 min 41 sec18 GB55 min 05 sec0.47 s/sample∆0.005 AUSteer (Ours)18 GB14 min 41 sec18 GB54 min 41 sec0.46 s/sample∆0.007 For the computational cost on larger LLMs such as 4-bit Llama-3.3-70B-Instruct, taking COPA as an example, we report both the preparation (activation-momentum computation) and inference overhead. During the activation-momentum computation stage, using 1000 contrastive pairs, AUS- teer requires 40 GB of GPU memory and around 15 minutes. During inference, the vanilla LLM requires 40 GB of GPU memory and 3 min 46 sec to run all test samples, while AUSteer requires 40 GB and 3 min 54 sec. These empirical results show that activation momentum scales successfully to large LLMs and remains far from computationally intensive, even on a 70B LLM. To summarize, our proposed method requires the least intervention footprint and lowest com- putational overhead, while achieving the best performance on most tasks. This provides clear empirical evidence supporting our argument that a smaller steering footprint can achieve improved efficiency. JBROADER CONTROL VARIANTS OF AUSTEER We conducted additional experiments on broader steering variants and found that, contrary to the assumption that “steering more AUs should be better,” precise partial AU control is the correct strategy. It offers clear advantages over steering a large portion—or all—of the AUs. Steering all AUs leads to consistent performance degradation. To test whether AUSteer is merely a constrained version of a more general “steer-all-units” method, we applied AUSteer-style dynamic weights to all AUs (e.g., 32×4096 = 131,072 AUs in LLaMA2-7B-Chat). After extensive hyperpa- rameter sweeps, steering all AUs still failed to outperform the vanilla model (without any steering). This matches our analysis in Section 3.3: different AUs regulate different output distributions, and only a small subset is task-relevant. Steering all AUs inevitably introduces strong task-irrelevant signals, effectively injecting noise into the model outputs. In contrast, partial AU steering focuses only on useful and task-relevant subspaces, yielding meaningful and targeted interventions. Broader AU steering does not guarantee better performance. We further tested variants that steer increasingly large subsets of AUs. Table 12 (COPA, LLaMA2-7B-Chat) shows that steering more than 5,000 AUs results in worse performance than the vanilla model. This again confirms that broader steering introduces many task-irrelevant or harmful output distributions, degrading performance. These findings also align with our results in Section 3.2, where steering certain AUs leads to negative effects. Table 12: Experimental results on steering broader AUs using COPA and LLaMA2-7B-Chat. There are 32× 4096 = 131,072 AUs in total in LLaMA2-7B-Chat. # of AUs0 (vanilla) < 10020050010003000500010000 Accuracy (%)70.882.877.273.270.870.670.470.4 Overall, our experiments demonstrate that AUSteer should only steer task-relevant or beneficial AUs, rather than steering a broad or full set of units. Partial AU control is therefore not a restricted version of a more general steering method—it is the correct and uniquely effective strategy for activation steering in LLMs. KPROMOTION VERSUS SUPPRESSION To determine whether we should promote useful AUs or suppress unhelpful ones, we conduct both empirical and theoretical analyses and show that promotion consistently outperforms suppression. 22 Published as a conference paper at ICLR 2026 Empirical evidence. To evaluate the “suppression” strategy, we use AU importance scores to iden- tify the least important AUs and apply a decreasing factor to suppress their activations. We vary the number of suppressed AUs from 0% to 99.95%, search decreasing factors from 0.05 to 0.99, and report the best results in Table 13. Experiments are conducted on LLaMA2-7B-Chat using three commonsense reasoning datasets. The results show that although suppression can yield im- provements over the vanilla model, it consistently underperforms compared to the promotion-based steering used in AUSteer. Table 13: Experimental results of suppressing AUs. MethodBoolQCOPAWinoG. Vanilla70.5270.850.91 Suppression73.3671.653.12 Promotion (AUSteer, ours)75.5782.853.28 Theoretical explanation. Prior work (Geva et al., 2022; Dar et al., 2023) shows that LLMs update predictions primarily through a promotion mechanism, where top-candidate tokens are driven by dominant positive sub-updates rather than by suppressing irrelevant ones. Consequently, directly promoting task-relevant AUs aligns better with the model’s intrinsic update dynamics, producing stronger and more targeted effects than suppression. 23