Paper deep dive
AxBench: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders
Zhengxuan Wu, Aryaman Arora, Atticus Geiger, Zheng Wang, Jing Huang, Dan Jurafsky, Christopher D. Manning, Christopher Potts
Models: Gemma-2-2B-IT, Gemma-2-9B-IT
Abstract
Abstract:Fine-grained steering of language model outputs is essential for safety and reliability. Prompting and finetuning are widely used to achieve these goals, but interpretability researchers have proposed a variety of representation-based techniques as well, including sparse autoencoders (SAEs), linear artificial tomography, supervised steering vectors, linear probes, and representation finetuning. At present, there is no benchmark for making direct comparisons between these proposals. Therefore, we introduce AxBench, a large-scale benchmark for steering and concept detection, and report experiments on Gemma-2-2B and 9B. For steering, we find that prompting outperforms all existing methods, followed by finetuning. For concept detection, representation-based methods such as difference-in-means, perform the best. On both evaluations, SAEs are not competitive. We introduce a novel weakly-supervised representational method (Rank-1 Representation Finetuning; ReFT-r1), which is competitive on both tasks while providing the interpretability advantages that prompting lacks. Along with AxBench, we train and publicly release SAE-scale feature dictionaries for ReFT-r1 and DiffMean.
Tags
Links
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 96%
Last extracted: 3/12/2026, 8:19:43 PM
Summary
AxBench is a large-scale benchmark designed to evaluate language model steering and concept detection methods. The paper compares various representation-based techniques, including Sparse Autoencoders (SAEs), against prompting and finetuning baselines. The authors find that prompting and finetuning generally outperform existing representation-based methods, and they introduce a new weakly-supervised method, ReFT-r1, which achieves competitive performance on both steering and detection tasks.
Entities (5)
Relation Signals (3)
AxBench → uses → Gemma-2-2B
confidence 100% · we introduce AxBench... and report experiments on Gemma-2-2B and 9B
AxBench → evaluates → ReFT-r1
confidence 90% · We evaluate a variety of steering methods—including a novel weakly-supervised method we introduce, ReFT-r1
ReFT-r1 → outperforms → Sparse Autoencoders
confidence 90% · only ReFT-r1 is competitive with finetuning and prompting baselines, while SAEs fall behind both ReFT-r1 and difference-in-means
Cypher Suggestions (2)
Find all methods evaluated by AxBench · confidence 90% · unvalidated
MATCH (b:Benchmark {name: 'AxBench'})-[:EVALUATES]->(m:Method) RETURN m.nameIdentify models used in the benchmark · confidence 90% · unvalidated
MATCH (b:Benchmark {name: 'AxBench'})-[:USES]->(lm:LanguageModel) RETURN lm.nameFull Text
115,151 characters extracted from source content.
Expand or collapse full text
AXBENCH: Steering LLMs? Even Simple Baselines Outperform Sparse Autoencoders Zhengxuan Wu * 1 Aryaman Arora * 1 Atticus Geiger 2 Zheng Wang 1 Jing Huang 1 Dan Jurafsky 1 Christopher D. Manning 1 Christopher Potts 1 Abstract Fine-grained steering of language model outputs is essential for safety and reliability. Prompting and finetuning are widely used to achieve these goals, but interpretability researchers have pro- posed a variety of representation-based techniques as well, including sparse autoencoders (SAEs), linear artificial tomography, supervised steering vectors, linear probes, and representation finetun- ing. At present, there is no benchmark for mak- ing direct comparisons between these proposals. Therefore, we introduceAXBENCH, a large-scale benchmark for steering and concept detection, and report experiments onGemma-2-2Band9B. For steering, we find that prompting outperforms all existing methods, followed by finetuning. For concept detection, representation-based methods such as difference-in-means, perform the best. On both evaluations, SAEs are not competitive. We introduce a novel weakly-supervised representa- tional method (Rank-1 Representation Finetun- ing;ReFT-r1), which is competitive on both tasks while providing the interpretability advantages that prompting lacks. Along withAXBENCH, we train and publicly release SAE-scale feature dic- tionaries for ReFT-r1 and DiffMean. https://github.com/stanfordnlp/axbench 1. Introduction In order to be useful, language models (LMs) must fol- low user instructions and be aligned to human goals and values. While prompting and finetuning are now widely used to instill such behaviour in LMs, both methods have * Equal contribution 1 Department of Computer Science, Stan- ford University 2 Pr(AI) 2 R Group. Correspondence to: Zhengx- uan Wu<wuzhengx@cs.stanford.edu>, Aryaman Arora <aryaman@cs.stanford.edu>. Preprint. 0.60.70.80.91.0 0.00 0.25 0.50 0.75 1.00 DiffMean PCA Prompt SAE LAT ReFT-r1 SAE-A Probe SSV LoReFT LoRA SFT* Concept detection (AUROC) Model steering (Overall) Type Prompt SAE SDL Figure 1: Average results across eight tasks on C concept detection (0–2) vs. S model steering (0–2) for all methods on AXBENCH. *Only evaluated onGemma-2-2B. limitations: circumvention via jailbreaks and continued training, reliance on dataset quality, and uninterpretabil- ity (Anwar et al., 2024). Interpretability researchers have thus proposed a new class of representation-based inter- ventions forsteeringLMs, which hope to address these issues. These methods include learning steering vectors from small labelled datasets, self-supervised sparse autoen- coders (SAEs), among other techniques. Since steering may enable lightweight and interpretable control over model out- puts, it has emerged as a potential alternative to finetuning and prompting (see §2). Unfortunately, Pres et al. (2024); Braun et al. (2024) note that existing benchmarks for steering only evaluate a few methods at merely toy scales. To assess whether representa- tion steering is a viable alternative to existing model control techniques, we need to evaluate it in a more realistic setting, e.g. over open-vocabulary concepts and on long-form gener- ation, and compare it to prompting and finetuning baselines. In this work, we introduceAXBENCH, a benchmark for evaluating LM control methods at scale using synthetic data. AXBENCHtakes in a list of natural language descriptions of concepts and samples relevant training and evaluation 1 arXiv:2501.17148v3 [cs.CL] 3 Mar 2025 AXBENCH data from an LLM. We evaluate model-control methods, including prompting and finetuning baselines, along two utilityaxes:concept detection C andmodel steering S . For the former, we use labelled synthetic data as ground truth; for the latter, we evaluate long-form generations using an LLM judge. The labelled training data enables com- parison between supervised dictionary-learning methods (SDLs) and unsupervised methods like SAEs. The bench- mark includes tasks generated from SAE concept lists for GemmaScope(Lieberum et al., 2024), covering two layers each frominstruction-tunedGemma-2-2BandGemma-2-9B (Gemma Team et al., 2024). However,AXBENCHis by nature extensible to arbitrary concept descriptions: we in- tend to add new evaluation tasks as better feature-labelling techniques and new approaches to steering emerge. We evaluate a variety of steering methods—including a novel weakly-supervised method we introduce,ReFT-r1— along with prompting, full finetuning, and two parameter- efficient finetuning methods (LoRA and LoReFT). On steer- ing, only ReFT-r1 is competitive with finetuning and prompt- ing baselines, while SAEs fall behind both ReFT-r1 and difference-in-means (Marks and Tegmark, 2024) on both axes. While representation steering methods largely lag behind incumbent model-control techniques, ReFT-r1 is evidence that steering can be pushed further with the avail- ability of comprehensive evaluation benchmarks. Finally, along withAXBENCH, we train and publicly release SAE- scale feature dictionaries for ReFT-r1 and DiffMean. 1 ; we call this approachsupervised dictionary learning(SDL; Figure 2) 2. Related work Representation-based control.Interventional/causal in- terpretability has emerged as the dominant paradigm for understanding neural networks in the LLM era, enabling the reverse-engineering of circuits underlying specific be- haviours (Giulianelli et al., 2018; Vig et al., 2020; Geiger et al., 2021; 2022; Meng et al., 2022; Chan et al., 2022; Wang et al., 2023; Goldowsky-Dill et al., 2023; Geiger et al., 2024; Guerner et al., 2024; Geiger et al., 2024). An impor- tant assumption in much of this work is thelinear represen- tation hypothesis, which claims that linear subspaces of rep- resentations in neural networks encode concepts (Mikolov et al., 2013b; Pennington et al., 2014; Bolukbasi et al., 2016; Elhage et al., 2022; Park et al., 2023; Nanda et al., 2023). Intervening on representations has thus emerged as an alter- native to finetuning and prompting for LM control. Representation-based steeringby adding fixed vectors to activations, or clamping activations to a certain value along 1 We open-source all of our datasets and trained dictionaries at https://huggingface.co/pyvene. fixed directions, is one such intervention-based tool for model control (Zou et al., 2023; Li et al., 2024; Turner et al., 2024; Marks and Tegmark, 2024; Liu et al., 2024; van der Weij et al., 2024; Rimsky et al., 2024). Finetuning-based approaches such as ReFT (Wu et al., 2024a) enable optimi- sation of steering directions on a dataset. Steering vectors need not be computed from labelled data; SAEs enable scal- able discovery of steering vectors from unlabelled data. In the same class of approaches, latent adversarial training (Casper et al., 2024) and circuit breakers (Zou et al., 2024) are representation-based control methods that increase the adversarial robustness of LLMs. Sparse autoencoders.Sparse autoencoders (SAEs) aim to enableself-supervisedand thusscalabledecomposition of the representation space into meaningful concepts (Tem- pleton et al., 2024; Chalnev et al., 2024; Makelov, 2024; O’Brien et al., 2024; Gao et al., 2024). SAEs are trained to reconstruct LLM hidden representations in a higher- dimensional latent space with a sparsity penalty, based on the assumption that concepts must be represented sparsely in order to prevent interference. The latents are then la- belled with natural-language descriptions using automatic interpretability pipelines (e.g. Juang et al., 2024), which can then be used to identify useful latents to steer the LM. Recent work reports mixed results when evaluating SAEs for steering; SAEs (but also several other steering methods) suf- fer from a tradeoff between model control and capabilities preservation (Mayne et al., 2024; Chalnev et al., 2024; Dur- mus et al., 2024; Bhalla et al., 2025). However, Karvonen et al. (2024) report Pareto-optimal performance when using SAEs to prevent models from producing regular expressions in code. Overall, evaluating SAEs remains an open problem because there is no ground-truth set of features to compare against. 3. AXBENCH AXBENCHis a benchmark which takes in a list of natural language descriptions of concepts and synthetically gener- ates the appropriate training and evaluation data for each concept using an LLM (Figure 2). The training and eval- uation data consists of labelled pairs of instructions and responses, where the responses are eitherpositiveexamples expressing the presence of the concept of interest, ornega- tiveexamples that represent the unsteered behaviour of the model (see §3.1 for details). We evaluate along two axes:concept detectionCand model steeringS. For the former, we measure classifica- tion performance on a held-out set of labelled data. 2 For the 2 We focus on binarised concept detection, as a multi-class classification task overnclasses can also be formulated into a binarised one overnfeatures. AXBENCH Figure 2: Key components ofAXBENCH: (a) an example of how we collect data for evaluating concept detection and model steering; (b) the synthetic data generation process for training and evaluation givenGolden Gate Bridgeas a concept; and (c) the contrasting training pipelines of SAEs and SDLs; both use LLMs, but SAEs use them to label pretrained features while we instead direct them to generate training data. latter, we use an LLM judge to rate steered outputs on three relevant axes (see §3.3). In this work, we use natural language concept lists for GemmaScopeSAEs as input, and generate training and eval- uation data for the following representation sites: layers 10 and 20 of instruction-tunedGemma-2-2B, and layers 20 and 31 of instruction-tunedGemma-2-9B. We sample 500 concepts for each task to generate data; we term this dataset CONCEPT500. These eight tasks (4 sites×2 axes) form the core training and evaluation testbeds forAXBENCH. Below, we describe the data generation process and evaluation setup for both axes. 3.1. Synthetic concept dataset generation We construct a small training datasetD train = (x + c,i ,y + ) n/2 i=1 ∪ (x − c,i ,y − ) n/2 i=1 . withnexamples and a concept detection evaluation datasetD concept of the same structure and harder examples, wherey + andy − are binary labels indicating whether the conceptcis present. We setn= 144for our main experiments. 3 We querygpt-4o-mini-2024-07-18to generate the data; the prompts used in this pipeline are presented in Ap- pendix J.2. Generating the data requires the following steps (note that only the evaluation set includes hard negatives): 1.Genre labelling & seed instructions: We consider three genres:text,code, andmath. We prompt the LLM to pick the genreg c for each concept. 4 We then 3 Using a small training dataset ensures our methods are practi- cal and cost-effective alternatives to SAEs. 4 Genre labelling increases input diversity. For example, inputs related to concepts such asprogramming code contains syntactic randomly select seed instructions from our instruction pool which belong to genreg c ; see Appendix I for dataset details. We then prompt the LLM to generate responses to these instructions. 5 2.Positive examples: For each randomly sampled in- struction from the instruction pool, we prompt the LLM to generate a response that incorporates the conceptc. We use the generated concept-conditioned responses concatenated with their instructions (using the LM’s chat template) as our positive set. 3.Negative examples: To evaluate the generalisation ability of each method, we independently sample seed instructions from all genres for negatives. 6 These in- structions are shared across concepts in order to save generation costs (i.e.,(x − c ,y − ) n/2 0 is independent of the conceptc). We sample responses from the LM we plan to steer (not the LLM) without any additional instructions. We use the paired instructions and re- sponses as our negative set. 4.Hard negative examples(evaluation only): For each concept, we find contrasting concepts that are semanti- cally related to our concept of interest but which should not activate the concept. We find these by (a) generat- ing a list of phrases that are semantically relevant to our concept, (b) filtering for those which are polysemous, and (c) finding alternative senses of those words which our concept should not activate on. This results in a set of contrast conceptsc contrast , each of which is a specific sense of a polysemous wordw contrast . We then ask the errorsshould contain code instead of descriptions of coding errors. 5 Each example costs less than $0.00006. 6 We sample instructions based on overall genre distribution: 70% fromtext, 15% fromcode, and 15% frommath. AXBENCH LLM to generate responses incorporatingw contrast into the sentence wherew contrast should express the sense related toc contrast . We use the contrastive responses paired with their instructions as our hard negative set. The negative training set is not applicable to all methods (e.g. full finetuning only needs the positive training set for model steering). 3.2.C Concept detection A popular LM interpretability method is to trainprobes (Conneau et al., 2018; Hewitt and Manning, 2019; Belinkov et al., 2017) that measure to what extent LM representations encode properties of interest, e.g. linguistic features. In recent years, the goal of concept detection has broadened to the open-vocabulary setting, with unsupervised methods becoming more common (Bills et al., 2023; Huben et al., 2024; Choi et al., 2024). Task description.Formally, given a Transformer-based LM with a hidden dimension size ofd, we define a concept classifier as a parameterized functionΨ Detect that maps a model representationh∈R d into abinarylabelˆyindicating the relative presence of a concept: Ψ Detect (h) = ˆy∈R 1 (1) whereΨis any function, e.g. a neural network. Evaluation dataset.To evaluate a concept classifier, we measure how accurately it can predict ground-truth labals on the labelled evaluation set fromD concept (see §3.1). Evaluation metrics.Since our labels are at the sequence- level, we need to aggregate token-label scores fromΨto evaluate it. Given a sequence of token representationsh l = [h l 1 ,h l 2 ,...,h l n ] withntokens at layerl∈[1,m], we max- pool the detection scores to get a sequence-level prediction: ˆy Detect = max Ψ Detect (h l ) (2) We then normalizeˆy Detect between[0,1]by min-max nor- malisation over the evaluation dataset for each concept. The predicted score represents how strongly a concept is present in a sequence, which we can compare to the true label. 3.3.SModel steering Representation-based steering has emerged as a potential alternative to existing model-control methods (e.g. finetun- ing and prompting) and a practical application of various interpretability methods (see§2). Unlike concept detection, model steering assessescausalefficacy in controlling model behaviour. Previous evaluation benchmarks for steering are not general-purpose; they either rely on a limited set of tasks (Zou et al., 2023; Makelov, 2024; Bhalla et al., 2025) or condition generation on a fixed prefix (Chalnev et al., 2024). To the best of our knowledge, we are the first to evaluate model steering methods in the open-vocabulary setting at scale. Task description.Given a promptx, the model’s original generation can be written as ˆ y=LM(x). We produce the model’s counterfactual generation conditioned on the concept-based interventionΦ Steer (h): ˆ y Steer =LM x,h←Φ Steer (h) (3) whereh←Φ Steer (h)is an in-place representation modifi- cation. We use the open-source intervention librarypyvene to perform such interventions on PyTorch implementations of models (Wu et al., 2024b). Evaluation dataset.We evaluate these steering methods in the instruction-following setting, where we sample in- structions fromAlpaca-Eval(Li et al., 2023) and prompt the LM to generate a response while intervening on its for- ward pass in-place using one of the steering methods. Evaluation metrics.For the intervened model generation, we evaluate ˆ y Steer based on theharmonic meanof the fol- lowing scores, each of which the LLM rates using a discrete score of 0, 1, or 2: 1. Concept scorerepresents how well the concept is in- corporated into the response. 2.Instruct scorerepresents how well the response is related to the instruction. 3.Fluency scorerepresents how fluent the response is. Since we compute the harmonic mean, the overall score also ranges from 0 to 2, but heavily penalises poor performance on any of these three subscores. For each concept, we randomly sample 10 instructions fromAlpaca-Evaland sample continuations for each steering factor (see discussion on steering factor in§5.2). To ensure a fair comparison, we partition our instructions into two equally sized sets, selecting the best factor from one set and evaluating it on the holdout set. Our judge prompts with further discussion can be found in Appendix J.3. 4. Methods In this section, we describe the interpretability methods we evaluate along with our baseline prompting and finetun- ing methods. For each method, we label which axes it is evaluated on usingCandS. All of our interpretability methods except SAEs are SDLs that learn rank-1 subspaces for targeted concepts. AXBENCH Notation.Given a LM, the hidden representations of di- mensionalitydfor a token sequence of lengthnin layerlof the LM are represented ash l = [h l 1 ,h l 2 ,...,h l n ]∈R n×d . The set of representations concatenated from all of the train- ing set inputs is denoted asH∈R s×d , wheres= P h |h|. We denoteH + as the subset ofHincluding only positive training inputs andH − for the negative inputs (see§3.1 for training dataset details). Finally, per-method projection vectorswand representationsh i are the same shape:R d×1 . CSDifference-in-means (DiffMean).DiffMean uses the difference between averaged representations from two classes of inputs as a steering vector (Marks and Tegmark, 2024). The projection vectorw DiffMean is defined as: w DiffMean = 1 |H + | X h + i ∈H + h + i | z mean of positives − 1 |H − | X h − i ∈H − h − i | z mean of negatives (4) We compute detection scores with the dot product, i.e.Ψ DiffMean Detect (h i ) =h i ·w DiffMean . 7 Our steering operation is simple activation addition:Φ DiffMean Steer (h i ) =h i +αw DiffMean whereαis the steering magnitude, which depends on the steering factor and is optimized as a hyperparameter, as described in §5.2. CSPrinciple component analysis (PCA).For PCA, we use the first principal component of the positive set of hidden representations as the projection vector. 8 We first subtract the meanH + from eachh + , gathering the cen- tered vectors into a matrixH ∈R |H + |×d . We then find the top principal componentw PCA ∈R d×1 ofH, i.e. the unit vector that captures the largest variance along its di- rection, usingsklearn.decomposition.PCA(Pedregosa et al., 2011). We follow the same detection and steering setup as DiffMean. CSLinear artificial tomography (LAT).LAT searches for a single latent direction that can separate positive ex- amples by learning from their pairwise activation differ- ences (Zou et al., 2023). Concretely, we create pairwise activation differencesδby randomly partitioningHinto pairs(h i ,h j )(withi̸=j) and computingδ= h i −h j ∥h i −h j ∥ , where the denominator ensures each difference is unit- normalized. We gather all these pairwise differences into a matrix∆∈R |H| 2 ×d . We then perform PCA (using sklearn) on∆; thenw LAT ∈R d×1 is the top principal 7 Following Gao et al. (2024), we normalizew DiffMean to have unit norm. We apply the same normalization to the learned weights of PCA, LAT, Probe, and ReFT-r1. 8 We found no significant difference between using only the positive set vs. the entire set of hidden representations for both PCA and LAT; see Appendix F for ablations. component of∆. We follow the same detection and steering setup as DiffMean. CSLinear probe (Probe).The linear probe learns to classify tokens as concept-relevant by projecting represen- tationsh i onto a learned directionw Probe ∈R d×1 just as in DiffMean. To convert this into a probability, we apply the sigmoid activation, and then minimise binary cross-entropy loss with the true labels: min w Probe ( 1 |h| X h i ∈h L BCE (y,Sigmoid(h i ·w Probe ))) ) (5) whereyis the token-level class label indicating whether this token belongs to a positive or negative example. The detection and steering setup is then identical to DiffMean. CS Supervised steering vector (SSV).The supervised steering vector method directly learns an intervention that maximises the language-modelling probability of the posi- tive responses. For a sequence of token representationsh, we apply an intervention to each token representation: Φ SSV (h i ) =h i +w SSV (6) wherew SSV ∈R d×1 is a learned vector. As described in§3.3, we backpropagate gradients by training with the language modeling loss, similar to supervised fine-tuning (SFT): min w SSV ( n X t=1 logP LM y t |y <t ,x;h←Φ SSV (h) ) (7) wherey i is thei-th output token,y <i are the preceding to- kens, andxis the prompt. For evaluating concept detection and model steering SSV follows the same setup as DiffMean. We applyReLUto get the detection scores. CSRank-1 representation finetuning (ReFT-r1).We introduce a novel method based on ReFT (Wu et al., 2024a) which jointly learns concept detection and steering on su- pervised data by combining the training objectives of linear probing and supervised steering. We compute latents for concept detection as: Ψ ReFT-r1 Detect (h i ) = ReLU(h i ·w ReFT-r1 )(8) During training we perform a representation-level interven- tion on eachh i based on the latents of the sequenceh: Φ ReFT-r1 (h i ) =h i + 1 k TopK(Ψ ReFT-r1 Detect (h)) 1 w ReFT-r1 (9) wherew ReFT-r1 ∈R d×1 is a learned vector. Finally, the training objective combines language modelling loss subject AXBENCH to this intervention, along with L1 regularisation on the non- top-klatents: min w ReFT-r1 − n X t=1 logP Φ ReFT-r1 LM (y t |y <t ,x) +λ X a i /∈TopK(Ψ(h)) ∥a i ∥ 1 (10) Detection and steering is identical to DiffMean. CSSparse autoencoders (SAE).Sparse autoencoders are a self-supervised dictionary learning method (see§2). We use pretrained SAEs fromGemmaScope, which are the best available SAEs forGemma-family LLMs (Lieberum et al., 2024). 9 The SAEs we used are trained to learn two dictionary matrices,W enc ,W dec ∈R d×z wherezis the number of latents. For our evaluating conceptc, we usew enc ,w dec ∈R d×1 as the detection and steering representations, respectively: Ψ SAE Detect (h i ) =σ(h i ·w enc +b enc ) whereσis an activation function (in our case, JumpReLU) andb enc is a learned bias. 10 For steering, we use activation addition as DiffMean. Note that Templeton et al. (2024) use activation clamping; we report ablations in Appendix F. CS SAEs with AUROC selection (SAE-A).Given that other methods have access to a training dataset, to enable fair comparison we attempt to use our training dataset for SAE feature selection. For each feature, we compute its max-pooled activations per Equation (2) over each train- ing example, compute AUROC over the dataset given true labels, and select the highest-scoring feature by this metric. C Bag-of-Words (BoW).For the BoW baseline, we first construct a featurizer that tokenizes text by whitespace and counts word frequencies. The vocabulary for this featurizer is derived from the training dataset. We then train a logistic regression classifier to predict class probabilities, framing the task as binary classification. To mitigate overfitting, we incorporate a regularization term. This BoW approach leverages statistical biases inherent in LLM-generated data. CGradient-based baselines.We test two gradient-based attribution methods, which are applicable only to concept detection: Input×gradients (I×G) and Integrated gradients (IG; Sundararajan et al., 2017). For both, we train a classifi- cation head on the hidden representations of some layer and apply the methods to produce token-level attribution scores Ψ Detect (h i ). Implementation details are in Appendix H. 9 GemmaScopereleases a set of SAEs forGemma-2-27B, but the concept list is not publicly released, which makes the SAEs for Gemma-2-9Bthe largest ones available for evaluations. 10 Note that this parameterisation cannot apply to TopK (Gao et al., 2024) and BatchTopK SAEs, which require loading in the entire encoder matrix to compute latents. CSPrompting baseline.For concept detection, we use the same LLM judge as described in§3.3 to rate the presence of a concept on a scale of 0 to 2. For model steering, we use an LLM toengineera prompt given a concept, which we use to steer our local model by prepending it to the actual instruction. We provide prompt templates and examples in Appendix J and Appendix N. SFinetuning baselines.We test full-parameter super- vised finetuning (SFT) and two parameter-efficient finetun- ing methods: Low-rank adaptation (LoRA; Hu et al., 2022) and low-rank representation finetuning (LoReFT; Wu et al., 2024a). In all cases, we finetune to minimise the language- modelling loss on the responses in the positive split of the dataset; the negative training split is discarded. We then use the finetuned models as baselines for steering. For all of our SDLs except SSV, we constrain any learned subspace to have a unit norm, following the same setup as SAEs. With a unit-norm constraint, we find that SSV is hard to use for steering models. For prompting and finetuning baselines, we randomly score one generation on the testing instruction set (since the factor is not a parameter for those methods), resulting in the same number of observations for those methods. 4.1. Evaluation Datasets.We synthetically generate training and valida- tion datasets (see§3.1) for 500 concepts, which we release asCONCEPT500. The concepts are sampled from the Neu- ronpedia SAE concept list forGemmaScopeas described in Appendix B. For each concept, we include 144 examples for training and≈72 samples for evaluating concept detec- tion. 11 In this paper, we train and evaluate all methods, and report results onCONCEPT500. For SFT, we only train and evaluate on the first 20 concepts due to limited resources. For evaluating steering, we use the instructions from the Alpaca-Evaldataset (Li et al., 2023). For each concept, we sample 10 instructions. We generate up to 128 tokens for each instruction over 14 steering factors. We split the instructions into two equal sets – one for selecting the best factor and the other for evaluation. We additionally release training and evaluation datasets for all 16K concepts inGemmaScopeas theCONCEPT16K dataset suite. We train and release SAE-scale dictionaries on this dataset only for the best-performing methods found onCONCEPT500. See Appendix L for dataset statistics and Appendix E for further experiments on CONCEPT16K. Models.Our evaluations rely on access to and control over the LLM’s representations. To reduce training cost, we 11 This varies based on valid hard negatives. AXBENCH MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 DiffMean0.9480.9460.9550.9210.942 Probe0.9400.9460.9330.9420.940 ReFT-r10.9520.9650.9660.8690.938 Prompt0.9100.9210.9400.9430.929 SAE-A0.9240.9110.9240.9070.917 BoW0.9090.9310.9040.9120.914 SSV0.9340.9500.9100.8540.912 LAT0.7420.8090.5720.7250.712 SAE0.7350.7550.6310.6590.695 PCA0.7140.7120.5590.6220.652 IG0.4400.3750.5080.3830.426 IxG0.2430.2170.1930.3300.246 Table 1:CMean AUROC for each method on concept detection.Boldindicates highest AUROC in that column; underlineindicates no significant difference vs. the best per- former. Gray indicates non-representation steering methods. prefer to use models for which pretrained SAEs are avail- able. We thus evaluate our methods on two open models, Gemma-2-2B-itandGemma-2-9B-it(henceforth referred to without the-itsuffix), from theGemma-family, with cor- responding SAEs released asGemmaScope. We evaluate our methods with model representations from the residual streams of layers 10 and 20 forGemma-2-2Band layers 20 and 31 forGemma-2-9B. We use SAEs fromGemmaScope that are trained for these layers. 12 To ensure a fair compari- son, we perform separate hyperparameter-tuning for each method. Details can be found in Appendix K. 5. Results 5.1.CConcept detection For concept detection,CONCEPT500consists of passages of text with ground-truth labels for each concept. Each method provides us with token-level concept scores obtained from the representation of that token at a particular layer. To compute a passage-level score, we take the mean of the token-level concept scores. See Appendix M for a visualiza- tion of token-level concept scores. AUROC.In Table 1, we report the average area under the ROC curve (AUROC) for each method over all con- cepts. Overall, we find that DiffMean, Probe, and ReFT-r1 are the best performers with no statistically significant dif- ference (p <0.05) between any of them under a paired t-test. Prompt, SAE-A, and SSV are not far behind and significantly outperform the remaining methods. LAT also performs better than random. Vanilla SAEs are thus sig- 12 ForGemma-2-2B, we follow the common practice to use SAEs for the base LM, as SAEs are not available for the instruction-tuned model at the time of publication (Lieberum et al., 2024). 0.00 0.25 0.50 0.75 1.00 0.934 0.918 0.908 0.899 0.881 0.854 0.778 0.765 0.770 0.739 0.711 0.675 0.596 0.529 0.444 0.183 0.281 0.163 0.947 0.926 0.915 0.922 0.902 0.863 0.825 0.776 0.749 0.764 0.789 0.689 0.687 0.557 0.488 0.342 0.322 0.133 ReFT-r1 DiffMean Probe SSV BoW Prompt LAT SAE PCA 0.00 0.25 0.50 0.75 1.00 0.939 0.921 0.908 0.872 0.880 0.888 0.700 0.702 0.695 0.789 0.750 0.728 0.529 0.518 0.478 0.074 0.239 0.038 ReFT-r1 DiffMean Probe SSV BoW Prompt LAT SAE PCA 0.885 0.901 0.906 0.825 0.882 0.889 0.784 0.715 0.735 0.639 0.731 0.676 0.221 0.519 0.495 0.270 0.272 0.120 Method F1 Gemma-2-2B: L10Gemma-2-2B: L20 Gemma-2-9B: L20Gemma-2-9B: L31 Balanced? balancedimbalanced Figure 3: C Mean F1 scores vs. dataset balance. nificantly outperformed by five supervised methods, all of which are much cheaper to train using a limited amount of synthetic data. The remaining methods (PCA, IG, and IxG) perform poorly; PCA’s better-than-random performance is nevertheless impressive given its unsupervised nature. Ad- ditional results are given in Appendix C. F1 score under class imbalance.In real-world text, pos- itive instances of concepts are much rarer than negative instances. We thus report F1 on both the balanced setting (50% positive instances) and an imbalanced setting with 3600 additional negative examples (≈1% positive). We choose classification threshold by maximising F1, binarise the resulting predictions, and report statistics on this dis- crete classification. Figure 3 shows that the relative ordering of methods does not change substantially between the two settings; despite their sparsity, SAEs perform poorly, but LAT and PCA also degrade substantially. 5.2.SModel steering For model steering, we take concept labels fromCON- CEPT500and apply the (pre)trained steering methods to the base model and sample generations. We score the gen- erations using an LM judge as described in§3.3. We addi- tionally benchmark prompting, full-finetuning (SFT), and two parameter-efficient finetuning methods (LoReFT and LoRA) as non-steering baselines. For steering methods, we note that steering factor is an important hyperparameter. We select the optimal steering factor for each method independently for every concept based on which factor achieves the highestoverallsteering score, as given by the LLM judge. Our actual steering AXBENCH MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 Prompt0.6980.7311.0751.0720.894 LoReFT0.7010.7220.7770.7640.741 SFT0.6370.714—0.676 LoRA0.6370.6410.6020.5800.615 ReFT-r10.6330.5090.6300.4010.543 DiffMean0.2970.1780.3220.1580.239 SAE0.1770.1510.1910.1400.165 SAE-A0.1660.1320.1860.1430.157 LAT0.1170.1300.1270.1340.127 PCA0.1070.0830.1280.1040.105 Probe0.0950.0910.1080.0990.098 SSV0.0720.0010.0240.0080.026 Table 2:SMean overall steering scores for each method, after steering factor selection. Gray indicates non- representation steering methods. 0.00 0.25 0.50 0.75 1.00 1.25 0.00.51.01.52.0 0.00 0.25 0.50 0.75 1.00 1.25 0.00.51.01.52.0 Instruct Score Concept Score Gemma-2-2B: L10Gemma-2-2B: L20 Gemma-2-9B: L20Gemma-2-9B: L31 Method ReFT-r1 DiffMean SAE SAE-A LAT PCA Probe SSV Figure 4: S Mean concept score vs. instruct score as the steering factor for each method is varied. magnitude (i.e.,α, as described in§4) is the product of the steering factor and the maximal activations aggregated over the evaluation dataset for concept detection. 13 Overall scores.We report the mean overall score for each method (i.e. the harmonic mean of three subscores: flu- ency, instruction-following, and concept presence) in Ta- ble 2. Prompting, along with slightly worse finetuning base- lines, outperforms all steering methods on average, except for ReFT-r1. ReFT-r1 is competitive with prompting in Gemma-2-2B but significantly behind on Gemma-2-9B; prompting scores improve by a large margin on the larger model. Additionally, DiffMean significantly outperforms SAEs, particularly in earlier layers. The remaining supervised steering methods fail to beat SAEs, and no steering methods besides ReFT-r1 approach prompting or finetuning performance. Importantly, we note 13 For SAEs, we query Neuronpedia to obtain the maximal acti- vation per concept. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 Prompt90.0%91.5%97.6%99.1%94.5% LoReFT88.9%88.2%88.6%90.3%89.0% SFT90.0%87.5%—88.8% LoRA85.0%83.4%79.9%81.5%82.5% ReFT-r185.2%82.3%83.6%76.0%81.8% DiffMean63.2%55.2%64.3%52.2%58.7% SAE50.0%50.0%50.0%50.0%50.0% SAE-A49.3%46.6%48.5%50.7%48.8% LAT43.5%48.2%42.7%48.6%45.8% PCA42.1%42.9%42.2%45.4%43.1% Probe40.4%44.0%41.9%45.6%43.0% SSV38.8%32.0%32.5%34.0%34.3% Table 3:SWinrate against SAEs for each method, after steering factor selection. that SAE-A slightly underperforms the unsupervised SAE; better classification does not directly lead to better steering. Winrate.We compute winrates against SAEs by compar- ing overall scores on each concept under each setting. We treat ties as 0.5 wins and 0.5 losses. We report the results in Table 3. Again, ReFT-r1 (88.0%) and DiffMean (61.6%) achieve winrates of greater than 50% against SAEs, and relative rankings are similar to those for overall score. We note that DiffMean and ReFT-r1 show higher winrates on earlier layers in both models. Steering factor.We compare the effect of changing the steering factor on instruct vs. concept scores in Figure 4. We notice that increasing the factor monotonically reduces instruct score in all methods, i.e. larger steering vectors harm capabilities; this agrees with prior findings (Durmus et al., 2024; Chalnev et al., 2024). However, the effect varies by layer for concept score: concept score increases then decreases in earlier layers, while it roughly monotonically increases with steering factor in later layers. In all cases, ReFT-r1 traces a Pareto-optimal path, achieving the highest concept score for any chosen instruct score. 6. Discussion Simple yet powerful baselines.While representation- level interventions have been shown to be useful in both enhancing model capabilities and for safety (see§2), they fail to outperform standard prompting and finetuning base- lines onAXBENCH. This is sobering evidence of the current limitations of steering techniques. However, our results sug- gest that joint learning of concept detection and steering (as in ReFT-r1) may be the key to advancement. SDL vs. SAEs.We have shown that SDL methods can achieve similar scalability and better performance at a lower AXBENCH cost compared to SAEs. Unlike SAEs, SDL methods re- quire concepts to be knowna priori; however, SDLs can be easily augmented with new features without retraining. We also note that SDLs depend on high-quality data generators, whereas SAEs rely on high-quality concept discriminators. These methods are not mutually exclusive and can comple- ment each other. SAE concept label quality.The concept lists used in this paper were adapted from Neuronpedia’s auto- interpretability pipeline, which is often skewed towards token-level concepts and misses high-level abstractions. While we tried to do post-hoc SAE feature selection to mitigate this, the poor performance of SAEs is at least par- tially a reflection of the limitations of auto-interpretability. It would be interesting to explore whether the SAE perfor- mance onAXBENCHimproves as better feature labelling methods are used and labels become less shallow (e.g. Choi et al., 2024). 7. Conclusion We introducedAXBENCH, a new benchmark for evaluating LM control methods at scale using synthetic data. To answer the question in the title of this work: our evaluation shows that even at SAE scale, representation steering is stillfar behindsimple prompting and finetuning baselines. Simulta- neously, we showed that a novel steering method, ReFT-r1, is capable ofclosing the gapto some extent; representation- based steering has not yet exhausted its potential. No mat- ter the outcome, we believe that comprehensive evaluation benchmarks likeAXBENCHare necessary for continued progress on this problem. Impact Statements In this paper, we explore representation-based methods for steering language models and introduceAXBENCH, a large- scale benchmark for evaluating these techniques. We believe that the immediate ethical and societal implications of our research are minimal. However, we recognize that enhanced control over language model outputs could potentially be misused to reinforce biases or manipulate information. To address these concerns, we advocate for the responsible application of steering methods and ensure transparency by publicly releasing our datasets and feature dictionaries. We encourage ongoing collaboration and dialogue within the research community to monitor and mitigate any unintended consequences of these technologies. Acknowledgements We thank R ́ obert Csord ́ as, Qinan Yu, and Jiuding Sun for constant and extremely helpful feedback during our weekly interp meetings; Jake Mendel for enlightening discussion about the direction and framing of the work; Neel Nanda for helpful suggestions on SAE feature selection; and Chenglei Si, Ken Ziyu Liu, Oam Patel, Luke Bailey, Harshit Joshi, Yanzhe ‘Sanju’ Zhang, Nikil Roashan Selvam, Julie Kallini, Omar Shaikh, Thomas Chen, Tristan Thrush, and Yangjun Ruan for various helpful discussions. We thank Joseph Tey and Nick Jiang for pointing out equation typos in an earlier draft. This research is supported in part by grants from Open Philanthropy. References Usman Anwar, Abulhair Saparov, Javier Rando, Daniel Paleka, Miles Turpin, Peter Hase, Ekdeep Singh Lubana, Erik Jenner, Stephen Casper, Oliver Sourbut, Benjamin L. Edelman, Zhaowei Zhang, Mario G ̈ unther, Anton Ko- rinek, Jose Hernandez-Orallo, Lewis Hammond, Eric Bigelow, Alexander Pan, Lauro Langosco, Tomasz Ko- rbak, Heidi Zhang, Ruiqi Zhong, Se ́ an ́ O h ́ Eigeartaigh, Gabriel Recchia, Giulio Corsi, Alan Chan, Markus An- derljung, Lilian Edwards, Aleksandar Petrov, Chris- tian Schroeder de Witt, Sumeet Ramesh Motwan, Yoshua Bengio, Danqi Chen, Philip H. S. Torr, Samuel Albanie, Tegan Maharaj, Jakob Foerster, Florian Tramer, He He, Atoosa Kasirzadeh, Yejin Choi, and David Krueger. Foun- dational challenges in assuring alignment and safety of large language models.arXiv:2404.09932, 2024. URL https://arxiv.org/abs/2404.09932. Yonatan Belinkov, Nadir Durrani, Fahim Dalvi, Hassan Sajjad, and James Glass. What do neural machine trans- lation models learn about morphology? In Regina Barzi- lay and Min-Yen Kan, editors,Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), pages 861–872, Vancouver, Canada, July 2017. Association for Compu- tational Linguistics. doi: 10.18653/v1/P17-1080. URL https://aclanthology.org/P17-1080. Usha Bhalla, Suraj Srinivas, Asma Ghandeharioun, and Himabindu Lakkaraju.Towards unifying inter- pretability and control: Evaluation via intervention. arXiv:2411.04430, 2025. URLhttps://arxiv.org/ abs/2411.04430. Steven Bills, Nick Cammarata, Dan Mossing, Henk Tillman, Leo Gao, Gabriel Goh, Ilya Sutskever, Jan Leike, Jeff Wu, and William Saunders. Language models can explain neurons in language models, 2023. URL https://openaipublic.blob.core.windows.net/ neuron-explainer/paper/index.html. Tolga Bolukbasi, Kai-Wei Chang, James Y. Zou, Venkatesh AXBENCH Saligrama, and Adam T. Kalai. Man is to computer programmer as woman is to homemaker? Debiasing word embeddings. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors,Advances in Neural Information Processing Systems, volume 29. Curran As- sociates, Inc., 2016.URLhttps://proceedings. neurips.c/paperfiles/paper/2016/file/ a486cd07e4ac3d270571622f4f316ec5-Paper.pdf. Joschka Braun, Dmitrii Krasheninnikov, Usman An- war, Robert Kirk, Daniel Tan, and David Scott Krueger. A sober look at steering vectors for LLMs. InAlignment Forum, 2024.URLhttps://w. alignmentforum.org/posts/QQP4nq7TXg89CJGBh/ a-sober-look-at-steering-vectors-for-llms. Stephen Casper, Lennart Schulze, Oam Patel, and Dylan Hadfield-Menell. Defending against unforeseen failure modes with latent adversarial training, 2024. URLhttps: //arxiv.org/abs/2403.05030. Sviatoslav Chalnev, Matthew Siu, and Arthur Conmy. Im- proving steering vectors by targeting sparse autoencoder features.arXiv:2411.02193, 2024. URLhttps://arxiv. org/abs/2411.02193. LawrenceChan,Adri ` aGarriga-Alonso,Nicholas Goldowsky-Dill, Ryan Greenblatt, Jenny Nitishinskaya, Ansh Radhakrishnan, Buck Shlegeris, and Nate Thomas. Causal scrubbing: A method for rigorously testing interpretability hypotheses. InAlignment Forum, 2022. URLhttps://shorturl.at/jZoHi. Dami Choi, Vincent Huang, Kevin Meng, Daniel D. John- son, Jacob Steinhardt, and Sarah Schwettmann. Scal- ing automatic neuron description, October 2024. URL https://transluce.org/neuron-descriptions. Alexis Conneau, German Kruszewski, Guillaume Lample, Lo ̈ ıc Barrault, and Marco Baroni. What you can cram into a single$&!#* vector: Probing sentence embed- dings for linguistic properties. In Iryna Gurevych and Yusuke Miyao, editors,Proceedings of the 56th Annual Meeting of the Association for Computational Linguis- tics (Volume 1: Long Papers), pages 2126–2136, Mel- bourne, Australia, July 2018. Association for Computa- tional Linguistics. doi: 10.18653/v1/P18-1198. URL https://aclanthology.org/P18-1198. Esin Durmus, Alex Tamkin, Jack Clark, Jerry Wei, Jonathan Marcus, Joshua Batson, Kunal Handa, Liane Lovitt, Meg Tong, Miles McCain, Oliver Rausch, Saffron Huang, Sam Bowman, Stuart Ritchie, Tom Henighan, and Deep Gan- guli. Evaluating feature steering: A case study in mitigat- ing social biases, 2024. URLhttps://anthropic.com/ research/evaluating-feature-steering. Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield- Dodds, Robert Lasenby, Dawn Drain, Carol Chen, Roger Grosse, Sam McCandlish, Jared Kaplan, Dario Amodei, Martin Wattenberg, and Christopher Olah. Toy mod- els of superposition.Transformer Circuits Thread, 2022.URLhttps://transformer-circuits.pub/ 2022/toymodel/index.html. Leo Gao, Tom Dupr ́ e la Tour, Henk Tillman, Gabriel Goh, Rajan Troll, Alec Radford, Ilya Sutskever, Jan Leike, and Jeffrey Wu. Scaling and evaluating sparse autoencoders, 2024. URLhttps://arxiv.org/abs/2406.04093. Atticus Geiger, Hanson Lu, Thomas Icard, and Christopher Potts. Causal abstractions of neural networks. In M. Ran- zato, A. Beygelzimer, Y. Dauphin, P.S. Liang, and J. Wort- man Vaughan, editors,Advances in Neural Information Processing Systems, volume 34, pages 9574–9586. Cur- ran Associates, Inc., 2021. URLhttps://proceedings. neurips.c/paper files/paper/2021/file/ 4f5c422f4d49a5a807eda27434231040-Paper.pdf. Atticus Geiger, Zhengxuan Wu, Hanson Lu, Josh Rozner, Elisa Kreiss, Thomas Icard, Noah D. Goodman, and Christopher Potts. Inducing causal structure for inter- pretable neural networks. In Kamalika Chaudhuri, Ste- fanie Jegelka, Le Song, Csaba Szepesv ́ ari, Gang Niu, and Sivan Sabato, editors,International Conference on Machine Learning, ICML 2022, volume 162 ofProceed- ings of Machine Learning Research, pages 7324–7338, Baltimore, Maryland, USA, 2022. PMLR. URLhttps: //proceedings.mlr.press/v162/geiger22a.html. Atticus Geiger, Duligur Ibeling, Amir Zur, Maheep Chaud- hary, Sonakshi Chauhan, Jing Huang, Aryaman Arora, Zhengxuan Wu, Noah Goodman, Christopher Potts, and Thomas Icard. Causal abstraction: A theoretical founda- tion for mechanistic interpretability.arXiv:2301.04709, 2024. URLhttps://arxiv.org/abs/2301.04709. Gemma Team,Morgane Riviere,Shreya Pathak, Pier Giuseppe Sessa, Cassidy Hardin, Surya Bhu- patiraju, L ́ eonard Hussenot, Thomas Mesnard, Bobak Shahriari, Alexandre Ram ́ e, Johan Ferret, Peter Liu, Pouya Tafti, Abe Friesen, Michelle Casbon, Sabela Ramos, Ravin Kumar, Charline Le Lan, Sammy Jerome, Anton Tsitsulin, Nino Vieillard, Piotr Stanczyk, Sertan Girgin, Nikola Momchev, Matt Hoffman, Shantanu Thakoor, Jean-Bastien Grill, Behnam Neyshabur, Olivier Bachem, Alanna Walton, Aliaksei Severyn, Alicia Parrish, Aliya Ahmad, Allen Hutchison, Alvin Abdagic, Amanda Carl, Amy Shen, Andy Brock, Andy Coenen, Anthony Laforge, Antonia Paterson, Ben Bastian, Bilal Piot, Bo Wu, Brandon Royal, Charlie Chen, Chintu Kumar, Chris Perry, Chris Welty, Christopher A. AXBENCH Choquette-Choo, Danila Sinopalnikov, David Wein- berger, Dimple Vijaykumar, Dominika Rogozi ́ nska, Dustin Herbison, Elisa Bandy, Emma Wang, Eric Noland, Erica Moreira, Evan Senter, Evgenii Eltyshev, Francesco Visin, Gabriel Rasskin, Gary Wei, Glenn Cameron, Gus Martins, Hadi Hashemi, Hanna Klimczak-Pluci ́ nska, Harleen Batra, Harsh Dhand, Ivan Nardini, Jacinda Mein, Jack Zhou, James Svensson, Jeff Stanway, Jetha Chan, Jin Peng Zhou, Joana Carrasqueira, Joana Iljazi, Jocelyn Becker, Joe Fernandez, Joost van Amersfoort, Josh Gordon, Josh Lipschultz, Josh Newlan, Ju yeong Ji, Kareem Mohamed, Kartikeya Badola, Kat Black, Katie Millican, Keelin McDonell, Kelvin Nguyen, Kiranbir Sodhia, Kish Greene, Lars Lowe Sjoesund, Lauren Usui, Laurent Sifre, Lena Heuermann, Leticia Lago, Lilly McNealus, Livio Baldini Soares, Logan Kilpatrick, Lucas Dixon, Luciano Martins, Machel Reid, Manvinder Singh, Mark Iverson, Martin G ̈ orner, Mat Velloso, Mateo Wirth, Matt Davidow, Matt Miller, Matthew Rahtz, Matthew Watson, Meg Risdal, Mehran Kazemi, Michael Moynihan, Ming Zhang, Minsuk Kahng, Minwoo Park, Mofi Rahman, Mohit Khatwani, Natalie Dao, Nenshad Bardoliwalla, Nesh Devanathan, Neta Dumai, Nilay Chauhan, Oscar Wahltinez, Pankil Botarda, Parker Barnes, Paul Barham, Paul Michel, Pengchong Jin, Petko Georgiev, Phil Culliton, Pradeep Kuppala, Ramona Comanescu, Ramona Merhej, Reena Jana, Reza Ardeshir Rokni, Rishabh Agarwal, Ryan Mullins, Samaneh Saadat, Sara Mc Carthy, Sarah Cogan, Sarah Perrin, S ́ ebastien M. R. Arnold, Sebastian Krause, Shengyang Dai, Shruti Garg, Shruti Sheth, Sue Ronstrom, Susan Chan, Timothy Jordan, Ting Yu, Tom Eccles, Tom Hennigan, Tomas Kocisky, Tulsee Doshi, Vihan Jain, Vikas Yadav, Vilobh Meshram, Vishal Dharmadhikari, Warren Barkley, Wei Wei, Wenming Ye, Woohyun Han, Woosuk Kwon, Xiang Xu, Zhe Shen, Zhitao Gong, Zichuan Wei, Victor Cotruta, Phoebe Kirk, Anand Rao, Minh Giang, Ludovic Peran, Tris Warkentin, Eli Collins, Joelle Barral, Zoubin Ghahramani, Raia Hadsell, D. Sculley, Jeanine Banks, Anca Dragan, Slav Petrov, Oriol Vinyals, Jeff Dean, Demis Hassabis, Koray Kavukcuoglu, Clement Farabet, Elena Buchatskaya, Sebastian Borgeaud, Noah Fiedel, Armand Joulin, Kathleen Kenealy, Robert Dadashi, and Alek Andreev. Gemma 2: Improving open language models at a practical size.arXiv:2408.00118, 2024. URL https://arxiv.org/abs/2408.00118. Mario Giulianelli, Jack Harding, Florian Mohnert, Dieuwke Hupkes, and Willem Zuidema. Under the hood: Us- ing diagnostic classifiers to investigate and improve how language models track agreement information. In Tal Linzen, Grzegorz Chrupała, and Afra Alishahi, edi- tors,Proceedings of the 2018 EMNLP Workshop Black- boxNLP: Analyzing and Interpreting Neural Networks for NLP, pages 240–248, Brussels, Belgium, November 2018. Association for Computational Linguistics. doi: 10.18653/v1/W18-5426. URLhttps://aclanthology. org/W18-5426. Gabriel Goh. Decoding the thought vector, 2017. URL https://gabgoh.github.io/ThoughtVectors/. Nicholas Goldowsky-Dill, Chris MacLeod, Lucas Sato, and Aryaman Arora. Localizing model behavior with path patching.arXiv:2304.05969, 2023. URLhttps: //arxiv.org/abs/2304.05969. Cl ́ ement Guerner, Anej Svete, Tianyu Liu, Alexander Warstadt, and Ryan Cotterell. A geometric notion of causal probing.arXiv:2307.15054, 2024. URLhttps: //arxiv.org/abs/2307.15054. John Hewitt and Christopher D. Manning. A structural probe for finding syntax in word representations. In Jill Burstein, Christy Doran, and Thamar Solorio, edi- tors,Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), pages 4129–4138, Minneapo- lis, Minnesota, June 2019. Association for Computa- tional Linguistics. doi: 10.18653/v1/N19-1419. URL https://aclanthology.org/N19-1419. Edward J. Hu, Yelong Shen, Phillip Wallis, Zeyuan Allen-Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu Chen. LoRA: Low-rank adaptation of large language models. InThe Tenth International Confer- ence on Learning Representations, ICLR 2022, Virtual Event, 2022. URLhttps://openreview.net/forum? id=nZeVKeeFYf9. Robert Huben, Hoagy Cunningham, Logan Riggs, Aidan Ewart, and Lee Sharkey. Sparse autoencoders find highly interpretable features in language models. InThe Twelfth International Conference on Learning Representations, ICLR 2024, Vienna, Austria, 2024. OpenReview.net. URL https://openreview.net/forum?id=F76bwRSLeK. Caden Juang, Gonc ̧alo Paulo, Jacob Drori, and Nora Belrose. Open source automated interpretability for sparse autoen- coder features, 2024. URLhttps://blog.eleuther. ai/autointerp/. Dan Jurafsky and James H. Martin.Speech and Lan- guage Processing. Online, 2025. URLhttps://web. stanford.edu/ ∼ jurafsky/slp3/. 3rd ed. draft. Adam Karvonen, Dhruv Pai, Mason Wang, and Ben Keig- win. Sieve: SAEs beat baselines on a real-world task (a code generation case study).Tilde Research Blog, 2024. URLhttps://w.tilderesearch.com/blog/sieve. Blog post. AXBENCH Michael Lan, Philip Torr, Austin Meek, Ashkan Khakzar, David Krueger, and Fazl Barez. Sparse autoencoders re- veal universal feature spaces across large language mod- els.arXiv:2410.06981, 2024. URLhttps://arxiv. org/abs/2410.06981. Anders Boesen Lindbo Larsen, Søren Kaae Sønderby, Hugo Larochelle, and Ole Winther. Autoencoding beyond pix- els using a learned similarity metric. In Maria-Florina Balcan and Kilian Q. Weinberger, editors,Proceedings of the 33nd International Conference on Machine Learning, ICML 2016, New York City, NY, USA, June 19-24, 2016, volume 48 ofJMLR Workshop and Conference Proceed- ings, pages 1558–1566. JMLR.org, 2016. URLhttp: //proceedings.mlr.press/v48/larsen16.html. Kenneth Li, Oam Patel, Fernanda Vi ́ egas, Hanspeter Pfis- ter, and Martin Wattenberg. Inference-time intervention: Eliciting truthful answers from a language model.Ad- vances in Neural Information Processing Systems, 36, 2024. URLhttps://arxiv.org/abs/2306.03341. Xuechen Li, Tianyi Zhang, Yann Dubois, Rohan Taori, Ishaan Gulrajani, Carlos Guestrin, Percy Liang, and Tat- sunori B. Hashimoto. AlpacaEval: An automatic evalu- ator of instruction-following models.https://github. com/tatsu-lab/alpacaeval, 5 2023. Tom Lieberum, Senthooran Rajamanoharan, Arthur Conmy, Lewis Smith, Nicolas Sonnerat, Vikrant Varma, Janos Kramar, Anca Dragan, Rohin Shah, and Neel Nanda. Gemma scope: Open sparse autoencoders everywhere all at once on gemma 2. In Yonatan Belinkov, Najoung Kim, Jaap Jumelet, Hosein Mohebbi, Aaron Mueller, and Han- jie Chen, editors,Proceedings of the 7th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, pages 278–300, Miami, Florida, US, Novem- ber 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.blackboxnlp-1.19. URLhttps: //aclanthology.org/2024.blackboxnlp-1.19. Sheng Liu, Haotian Ye, Lei Xing, and James Y. Zou. In- context vectors: Making in context learning more effec- tive and controllable through latent space steering. In Forty-first International Conference on Machine Learn- ing, ICML 2024, Vienna, Austria, July 21-27, 2024. Open- Review.net, 2024.URLhttps://openreview.net/ forum?id=dJTChKgv3a. Aleksandar Makelov. Sparse autoencoders match supervised features for model steering on the IOI task. InICML 2024 Workshop on Mechanistic Interpretability, 2024. URL https://openreview.net/forum?id=JdrVuEQih5. Samuel Marks and Max Tegmark. The geometry of truth: Emergent linear structure in large language model repre- sentations of true/false datasets.arXiv:2310.06824, 2024. URLhttps://arxiv.org/abs/2310.06824. Harry Mayne, Yushi Yang, and Adam Mahdi. Can sparse autoencoders be used to decompose and interpret steering vectors?arXiv:2411.08790, 2024. URLhttps://arxiv. org/abs/2411.08790. Kevin Meng, David Bau, Alex Andonian, and Yonatan Belinkov. Locating and editing factual associations in GPT. In S. Koyejo, S. Mohamed, A. Agarwal, D. Bel- grave, K. Cho, and A. Oh, editors,Advances in Neu- ral Information Processing Systems, volume 35, pages 17359–17372. Curran Associates, Inc., 2022.URL https://arxiv.org/abs/2202.05262. Tom ́ as Mikolov, Ilya Sutskever, Kai Chen, Gregory S. Corrado, and Jeffrey Dean. Distributed representations of words and phrases and their compositionality. In Christopher J. C. Burges, L ́ eon Bottou, Zoubin Ghahramani, and Kilian Q. Weinberger, editors,Ad- vances in Neural Information Processing Systems 26: 27th Annual Conference on Neural Information Processing Systems 2013. Proceedings of a meet- ing held December 5-8, 2013, Lake Tahoe, Nevada, United States, pages 3111–3119, 2013a. URLhttps: //proceedings.neurips.c/paper/2013/hash/ 9a42b31882ec039965f3c4923ce901b-Abstract. html. Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. Lin- guistic regularities in continuous space word represen- tations. In Lucy Vanderwende, Hal Daum ́ e I, and Ka- trin Kirchhoff, editors,Proceedings of the 2013 Con- ference of the North American Chapter of the Associa- tion for Computational Linguistics: Human Language Technologies, pages 746–751, Atlanta, Georgia, June 2013b. Association for Computational Linguistics. URL https://aclanthology.org/N13-1090/. Ulisse Mini, Peli Grietzer, Mrinank Sharma, Austin Meek, Monte MacDiarmid, and Alexander Matt Turner. Under- standing and controlling a maze-solving policy network. arXiv:2310.08043, 2023. URLhttps://arxiv.org/ abs/2310.08043. Neel Nanda, Andrew Lee, and Martin Wattenberg. Emer- gent linear representations in world models of self- supervised sequence models. In Yonatan Belinkov, So- phie Hao, Jaap Jumelet, Najoung Kim, Arya McCarthy, and Hosein Mohebbi, editors,Proceedings of the 6th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, pages 16–30, Singapore, De- cember 2023. Association for Computational Linguistics. doi: 10.18653/v1/2023.blackboxnlp-1.2. URLhttps: //aclanthology.org/2023.blackboxnlp-1.2/. AXBENCH Kyle O’Brien, David Majercak, Xavier Fernandes, Richard Edgar, Jingya Chen, Harsha Nori, Dean Carignan, Eric Horvitz, and Forough Poursabzi-Sangde. Steer- ing language model refusal with sparse autoencoders. arXiv:2411.11296, 2024. URLhttps://arxiv.org/ abs/2411.11296. Kiho Park, Yo Joong Choe, and Victor Veitch. The linear representation hypothesis and the geometry of large lan- guage models.arXiv:2311.03658, 2023. URLhttps: //arxiv.org/abs/2311.03658. Fabian Pedregosa, Ga ̈ el Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, Jake VanderPlas, Alexandre Passos, David Cournapeau, Matthieu Brucher, Matthieu Perrot, and Edouard Duch- esnay. Scikit-learn: Machine learning in Python.The Journal of Machine Learning Research, 12:2825–2830, 2011. doi: 10.5555/1953048.2078195. URLhttps: //dl.acm.org/doi/10.5555/1953048.2078195. Jeffrey Pennington, Richard Socher, and Christopher Man- ning.GloVe: Global vectors for word representa- tion. In Alessandro Moschitti, Bo Pang, and Walter Daelemans, editors,Proceedings of the 2014 Confer- ence on Empirical Methods in Natural Language Process- ing (EMNLP), pages 1532–1543, Doha, Qatar, October 2014. Association for Computational Linguistics. doi: 10.3115/v1/D14-1162. URLhttps://aclanthology. org/D14-1162. Itamar Pres, Laura Ruis, Ekdeep Singh Lubana, and David Krueger. Towards reliable evaluation of behavior steer- ing interventions in LLMs, 2024. URLhttps://arxiv. org/abs/2410.17245. Nina Rimsky, Nick Gabrieli, Julian Schulz, Meg Tong, Evan Hubinger, and Alexander Turner. Steering Llama 2 via contrastive activation addition. In Lun-Wei Ku, Andre Martins, and Vivek Srikumar, editors,Proceed- ings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Pa- pers), pages 15504–15522, Bangkok, Thailand, Au- gust 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024.acl-long.828. URLhttps:// aclanthology.org/2024.acl-long.828. Naomi Saphra and Sarah Wiegreffe. Mechanistic?In Yonatan Belinkov, Najoung Kim, Jaap Jumelet, Hosein Mohebbi, Aaron Mueller, and Hanjie Chen, editors,Pro- ceedings of the 7th BlackboxNLP Workshop: Analyzing and Interpreting Neural Networks for NLP, pages 480– 498, Miami, Florida, US, November 2024. Association for Computational Linguistics. doi: 10.18653/v1/2024. blackboxnlp-1.30. URLhttps://aclanthology.org/ 2024.blackboxnlp-1.30/. Nishant Subramani, Nivedita Suresh, and Matthew Peters. Extracting latent steering vectors from pretrained lan- guage models. In Smaranda Muresan, Preslav Nakov, and Aline Villavicencio, editors,Findings of the As- sociation for Computational Linguistics: ACL 2022, pages 566–581, Dublin, Ireland, May 2022. Associa- tion for Computational Linguistics.doi: 10.18653/ v1/2022.findings-acl.48. URLhttps://aclanthology. org/2022.findings-acl.48. Mukund Sundararajan, Ankur Taly, and Qiqi Yan. Ax- iomatic attribution for deep networks. InProceedings of the 34th International Conference on Machine Learning - Volume 70, ICML’17, page 3319–3328. JMLR.org, 2017. Adly Templeton, Tom Conerly, Jonathan Marcus, Jack Lindsey, Trenton Bricken, Brian Chen, Adam Pearce, Craig Citro, Emmanuel Ameisen, Andy Jones, Hoagy Cunningham, Nicholas L Turner, Callum McDougall, Monte MacDiarmid, C. Daniel Freeman, Theodore R. Sumers, Edward Rees, Joshua Batson, Adam Jermyn, Shan Carter, Chris Olah, and Tom Henighan. Scal- ing monosemanticity: Extracting interpretable features from Claude 3 Sonnet.Transformer Circuits Thread, 2024.URLhttps://transformer-circuits.pub/ 2024/scaling-monosemanticity/index.html. Alexander Matt Turner, Peli Grietzer, Ulisse Mini, Monte M, and David Udell. Understanding and controlling a maze-solving policy network.Alignment Forum, March 2023a. URLhttps://shorturl.at/XGtmh. Alexander Matt Turner, Peli Grietzer, and Lisa Thiergart. Maze-solving agents: Add a top-right vector, make the agent go to the top-right.Alignment Forum, March 2023b. URLhttps://shorturl.at/7Qdy9. Alexander Matt Turner, Lisa Thiergart, Gavin Leech, David Udell, Juan J. Vazquez, Ulisse Mini, and Monte Mac- Diarmid.Steering language models with activation engineering.arXiv:2308.10248, 2024. URLhttps: //arxiv.org/abs/2308.10248. Paul Upchurch, Jacob R. Gardner, Geoff Pleiss, Robert Pless, Noah Snavely, Kavita Bala, and Kilian Q. Wein- berger. Deep feature interpolation for image content changes. In2017 IEEE Conference on Computer Vi- sion and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, July 21-26, 2017, pages 6090–6099. IEEE Com- puter Society, 2017. doi: 10.1109/CVPR.2017.645. URL https://doi.org/10.1109/CVPR.2017.645. Teun van der Weij, Massimo Poesio, and Nandi Schoots. Extending activation steering to broad skills and multiple behaviours.arXiv:2403.05767, 2024. URLhttps:// arxiv.org/abs/2403.05767. AXBENCH Jesse Vig, Sebastian Gehrmann, Yonatan Belinkov, Sharon Qian, Daniel Nevo, Yaron Singer, and Stuart Shieber. Investigating gender bias in language models using causal mediation analysis.In H. Larochelle, M. Ranzato, R. Hadsell, M.F. Balcan, and H. Lin, editors,Advances in Neural Information Processing Systems, volume 33, pages 12388–12401. Curran As- sociates, Inc., 2020.URLhttps://proceedings. neurips.c/paper files/paper/2020/file/ 92650b2e92217715fe312e6fa7b90d82-Paper.pdf. Eric Wallace, Jens Tuyls, Junlin Wang, Sanjay Subramanian, Matt Gardner, and Sameer Singh. AllenNLP interpret: A framework for explaining predictions of NLP models. In Proceedings of the 2019 Conference on Empirical Meth- ods in Natural Language Processing and the 9th Interna- tional Joint Conference on Natural Language Processing (EMNLP-IJCNLP): System Demonstrations, pages 7–12, 2019. Kevin Ro Wang, Alexandre Variengien, Arthur Conmy, Buck Shlegeris, and Jacob Steinhardt. Interpretability in the wild: a circuit for indirect object identification in GPT-2 small. InThe Eleventh International Confer- ence on Learning Representations, ICLR 2023, Kigali, Rwanda, 2023. URLhttps://openreview.net/pdf? id=NpsVSN6o4ul. Yulin Wang, Xuran Pan, Shiji Song, Hong Zhang, Gao Huang, and Cheng Wu. Implicit semantic data augmen- tation for deep networks. In Hanna M. Wallach, Hugo Larochelle, Alina Beygelzimer, Florence d’Alch ́ e-Buc, Emily B. Fox, and Roman Garnett, editors,Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, NeurIPS 2019, December 8-14, 2019, Vancouver, BC, Canada, pages 12614–12623, 2019. URLhttps: //proceedings.neurips.c/paper/2019/hash/ 15f99f2165a8c86c9dface16fefd281-Abstract. html. TomWhite.Samplinggenerativenetworks. arXiv:1609.04468, 2016.URLhttps://arxiv. org/abs/1609.04468. Zhengxuan Wu, Aryaman Arora, Zheng Wang, Atticus Geiger, Dan Jurafsky, Christopher D. Manning, and Christopher Potts.ReFT: Representation finetuning for language models.arXiv:2404.03592, 2024a. URL https://arxiv.org/abs/2404.03592. Zhengxuan Wu, Atticus Geiger, Aryaman Arora, Jing Huang, Zheng Wang, Noah Goodman, Christopher Man- ning, and Christopher Potts. pyvene: A library for under- standing and improving PyTorch models via interventions. In Kai-Wei Chang, Annie Lee, and Nazneen Rajani, ed- itors,Proceedings of the 2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies (Volume 3: System Demonstrations), pages 158–165, Mexico City, Mexico, June 2024b. Association for Computational Lin- guistics. doi: 10.18653/v1/2024.naacl-demo.16. URL https://aclanthology.org/2024.naacl-demo.16. Andy Zou, Long Phan, Sarah Chen, James Campbell, Phillip Guo, Richard Ren, Alexander Pan, Xuwang Yin, Mantas Mazeika, Ann-Kathrin Dombrowski, Shashwat Goel, Nathaniel Li, Michael J. Byun, Zifan Wang, Alex Mallen, Steven Basart, Sanmi Koyejo, Dawn Song, Matt Fredrikson, J. Zico Kolter, and Dan Hendrycks. Rep- resentation engineering: A top-down approach to AI transparency.arXiv:2310.01405, 2023. URLhttps: //arxiv.org/abs/2310.01405. Andy Zou, Long Phan, Justin Wang, Derek Duenas, Maxwell Lin, Maksym Andriushchenko, Rowan Wang, Zico Kolter, Matt Fredrikson, and Dan Hendrycks. Im- proving alignment and robustness with circuit breakers, 2024. URLhttps://arxiv.org/abs/2406.04313. AXBENCH Appendix A. Historical notes on steering Inspired by Jurafsky and Martin (2025) and noting the sociological observations about (mechanistic) interpretability as a field in Saphra and Wiegreffe (2024), we offer some historical notes on the development of steering as a field in an effort to document and properly cite where these ideas came from. Steeringrefers to applying interventions (usually adding a fixed vector) to the activation space of a neural model in order to control its generations. Early precursors to steering noted that linear subspaces of the representation space of pretrained word vectors seemed to encode meaningful concepts (Mikolov et al., 2013a; Pennington et al., 2014; Bolukbasi et al., 2016). Larsen et al. (2016) first used thedifference-in-meanstechnique to extract visualattribute vectorsfrom GAN discriminators in order to steer generator outputs; this technique was widely adopted in computer vision (White, 2016; Upchurch et al., 2017; Goh, 2017; Wang et al., 2019). In NLP, initial work by Subramani et al. (2022) proposedsteering vectors, learned to maximise the probability of some output, as an alternative to expensive fine-tuning and unreliable prompt optimisation for the task of controllable text generation. Soon after, steering was also use to localise behaviours in a maze-searching RL agent (Turner et al., 2023a;b; Mini et al., 2023). Variations on this approach (sometimes using difference-in-means or other closed-form expressions to compute the vector) were adopted by researchers inmechanistic interpretabilityfrom late 2023 for AI safety (Zou et al., 2023; Li et al., 2024; Turner et al., 2024; Marks and Tegmark, 2024; Rimsky et al., 2024) and later as a general-purpose but localised and parameter-efficient alternative to finetuning (Wu et al., 2024a; Liu et al., 2024; van der Weij et al., 2024). Sparse autoencoders(SAEs), a scalable technique for self-supervised rank-one linear feature discovery via dictionary learning, are also increasingly used to find or learn steering vectors (Templeton et al., 2024; Chalnev et al., 2024; Makelov, 2024; O’Brien et al., 2024). B. SAE concept list We use SAE concept lists to enable a fair comparison with SAEs, which were annotated mostly bygpt-3.5-turboor gpt-4o-mini. These concept lists are released by Neuronpedia and were scraped by the authors of this paper in November 2024. We utilize the concept lists from four SAEs fromGemmaScope:10-gemmascope-res-16kfor theGemma-2-2Bbase model and20-gemmascope-res-131kfor theGemma-2-9Binstruction-tuned model, where we scraped a maximum of 16K concepts. AXBENCH C. Detailed analysis C.1.CConcept detection 0.00 0.25 0.50 0.75 1.00 0.000.250.500.751.00 0.00 0.25 0.50 0.75 1.00 0.000.250.500.751.00 False Positive Rate True Positive Rate Gemma-2-2B: L10Gemma-2-2B: L20 Gemma-2-9B: L20Gemma-2-9B: L31 Method DiffMean Probe ReFT-r1 Prompt SAE-A BoW SSV LAT SAE PCA IG IxG Figure 5:CMean ROC curves over all concepts. 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 0.000.250.500.751.00 0.00 0.25 0.50 0.75 1.00 0.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.000.000.250.500.751.00 False Positive Rate True Positive Rate DiffMeanProbeReFT-r1PromptSAE-ABoWSSVLATSAEPCAIGIxG Gemma-2-2B L10 Gemma-2-2B L20 Gemma-2-9B L20 Gemma-2-9B L31 Figure 6: All ROC curves. AXBENCH C.2.SModel steering 0.0 0.5 1.0 1.5 2.0 0.0 0.5 1.0 1.5 2.0 0.0 0.5 1.0 1.5 2.0 PromptLoReFTSFTLoRAReFT-r1DiffMeanSAESAE-ALATPCAProbeSSV 0.0 0.5 1.0 1.5 2.0 Method Score Concept Score Fluency Score Instruct Score Overall Score Gemma-2-2B: L10Gemma-2-2B: L20Gemma-2-9B: L20Gemma-2-9B: L31 Figure 7: Mean score breakdown for all methods on our unseen testing instruction set after selecting the optimal factor (based on the Overall Score) on our evaluation instruction set. For prompting and finetuning, we randomly score one generation on the testing instruction set (since the factor is not a parameter for those methods), resulting in the same number of observations for those methods. 0 100 200 300 400 500 0 100 200 300 400 500 0 100 200 300 400 500 0.3 13 0 100 200 300 400 500 0.3 13 0.3 13 0.3 13 0.3 13 0.3 13 0.3 13 0.3 13 Optimal Steering Factor Count DiffMeanLATPCAProbeReFT-r1SAESAE-ASSV Gemma-2-2B L10 Gemma-2-2B L20 Gemma-2-9B L20 Gemma-2-9B L31 Figure 8: Distribution of optimal steering factors for each method across the 4 tasks. AXBENCH 0.0 0.4 0.8 1.2 0.0 0.5 1.0 0.0 0.5 1.0 1.5 2.0 0.3 13 0.0 0.2 0.4 0.6 0.3 13 0.3 13 0.3 13 Steering Factor Score Gemma-2-2B L10 Gemma-2-2B L20 Gemma-2-9B L20 Gemma-2-9B L31 Concept Score Fluency Score Instruct Score Overall Score Method ReFT-r1 DiffMean SAE SAE-A LAT PCA Probe SSV Figure 9: Steering factor vs. scores. AXBENCH D. Supervised dictionary learning method works with very limited amount of training data. Based on the performance results, ReFT-r1 is the strongest SAE alternative. We further study the data scaling law of ReFT-r1 by varying the number of training examples. Specifically, we measure ReFT-r1 performance on both concept detection and steering withCONCEPT10when the number of training example is set to6, 12, 24, 48, 72, 96, 120, 144. In the extreme setting, we provide only 3 positive and 3 negative examples. Since we have a limited pool of concepts, we average our results with three random seeds:42, 43, 44. Figure 10 shows how the performance of ReFT-r1 varies in C (concept detection) and S (model steering) when trained with different numbers of training examples. For earlier layers, scores increase with more data, while forGemma-2-9B, the trend is less clear for concept detection. Our results indicate that once a certain threshold is reached, performance saturates for both tasks, suggesting that the cost of training ReFT-r1 can be further reduced. The per-concept cost with 144 training examples is approximately $0.008, and this cost decreases proportionally as the number of training examples is reduced. 0.2 0.4 0.6 0.8 0.800.850.900.951.00 0.2 0.4 0.6 0.8 0.800.850.900.951.00 AUROC Overall Score Gemma-2-2B: L10Gemma-2-2B: L20 Gemma-2-9B: L20Gemma-2-9B: L31 # Examples 50 100 Figure 10: Scaling law for supervised dictionary learning (SDL) method ReFT-r1 withCONCEPT10on both C concept detection and S model steering. AXBENCH E. SDLs at scale: Analysing CONCEPT16K E.1. ReFT-r1: CONCEPT16K subspace for code error handling. We scale up two supervised dictionary learning methods DiffMean and ReFT-r1 withCONCEPT16K. They serve as drop-in replacements of existing SAEs onGemmamodels with better performance for concept detection and steering. Figure 11 shows the UMAP of ReFT-r1’sCONCEPT16Ksubspaces learned withGemma-2-2Bat layer 20’s residual stream. Subspaces are meaningfully clustered together by genres. Within each genre cluster, related features are also clustered together. For instance, we identify a subspace cluster for concepts related to “Code error handling and logging,” which includes the following concepts: •Subspace 16K/14404: “error messages related to system calls and file operations” •Subspace 16K/14801: “terms related to programming errors and error handling” •Subspace 16K/5656: “technical terms and parameters related to errors and status in programming contexts” •Subspace 16K/4884: “error messages and exceptions in code related to server or network operations” •Subspace 16K/2467: “references to errors and warnings, especially related to file or access issues” UMAP 1 UMAP 2 ReFT-r1 subspaces genre text code math UMAP 1 UMAP 2 references to errors and warnings, especially related to file or access issues error messages and exceptions in code related to server or network operations technical terms and parameters related to errors and status in programming contexts error messages related to system calls and file operations terms related to programming errors and error handling Figure 11: UMAP of ReFT-r1’s CONCEPT16K subspaces withGemma-2-2Bat layer 20’s residual stream. AXBENCH E.2. Mapping natural language to subspaces. We explore whether we can find a direct mapping from natural-language concept descriptions to subspaces. We first train ReFT-r1 withCONCEPT16Kand create a supervised datasetD Generator =(c,w c ReFT-r1 ) 16K 0 , where the inputcis the concept description in natural language and the output is the ReFT-r1 subspace vector corresponding to the concept. We divide D Generator into training and testing sets, ensuring that the testing set contains only concepts fromCONCEPT500, which are excluded from the training set. To train the generator, we attach a supervised linear headΦ Generator to the last input token representation at then-th position of the last layerm, predicting the learned ReFT-r1 subspace: L=L MSE+Cosine w c ReFT-r1 ,Φ Generator ([LM θ (c)] m n ) (11) where we fine-tune the generator head and the LM using equally weighted MSE and cosine distance losses. We do finetune the base LMGemma-2-2bfor our subspace generators. We partition the last 500 examples in our training dataset as our in-training development set to early-stop our training with a patience step set to 3. We generate ReFT-r1 subspaces forCONCEPT500and follow our evaluation paradigm inAXBENCHto evaluate concept detection and model steering. We show two cases below by unembedding generated subspaces with the output embedding matrix. We find that the subspace generator works better in English as opposed to other languages. As shown in Table 4a and Table 4b, subspaces for unseen concepts generated by our finetuned model exhibit only slight performance degradation in concept detection, while performance drops more significantly in model steering. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 DiffMean0.9480.9460.9550.9210.942 ReFT-r10.9520.9650.9660.8690.938 ReFT-r1 (Gen)—0.9450.965— SAE0.7350.7550.6310.6590.695 (a) C Mean AUROC. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 ReFT-r10.6330.5090.6300.4010.543 ReFT-r1 (Gen)—0.4150.466—- DiffMean0.2970.1780.3220.1580.239 SAE0.1770.1510.1910.1400.165 (b) S Overall score. Table 4: Results on CONCEPT500 for ReFT-r1 (Gen) vs. ReFT-r1 and other selected methods. Unseen concept description in Chinese 道德经 a Top positive logits when unembedding the subspace (’ ethical’, 1.4296875), (’ moral’, 1.3984375), (’ ethics’, 1.2421875), (’Ethical’, 1.1640625), (’ Ethical’, 1.15625), (’moral’, 1.125), (’ Ethics’, 1.0859375), (’ Moral’, 1.0859375), (’Ethics’, 1.0703125), (’ethical’, 1.0703125) Top negative logits when unembedding the subspace (’DockStyle’, -0.78125), (’ venons’, -0.6796875), (’ purpose’, -0.67578125), (’complexContent’, -0.671875), (’ stupidly’, - 0.66796875), (’ fooled’, -0.66015625), (’ Jefus’, -0.65234375), (’ small’, -0.6328125), (’ mont ́ on’, -0.62109375), (’ Dummies’, -0.6171875) a https://en.wikipedia.org/wiki/TaoTeChing. AXBENCH Unseen concept description in English Business-related terms and symbols, particularly focusing on entrepreneurship and financial aspects, as well as formatting and coding indicators a Top positive logits when unembedding the subspace (’ investment’, 1.1953125), (’ asset’, 1.1484375), (’ financial’, 1.1328125), (’ investments’, 1.0625), (’ Investment’, 1.046875), (’ market’, 1.0390625), (’ portfolio’, 1.03125), (’ investor’, 1.03125), (’ assets’, 1.0078125), (’ investors’, 1.0078125) Top negative logits when unembedding the subspace (’ sauvages’, -0.8515625), (’ h ˆ ate’, -0.76953125), (’ rapides’, -0.76171875), (’ r ́ egl’, -0.7421875), (’ d ́ ecouvertes’, -0.71875), (’ ferm ́ es’, -0.69921875), (’ compl ` etes’, -0.69140625), (’ pr ́ ec ́ edents’, -0.68359375), (’setVerticalGroup’, -0.68359375), (’ d ́ ecouver’, -0.671875) a Taken fromhttps://github.com/yoavgur/Feature-Descriptions/blob/main/descriptions/gemma-2-2b.csv. AXBENCH E.3. Teleporting between subspaces across models through affine transformations. We explore whether structural equivalence in subspaces exists across models. Previous works have analyzed feature universality in SAEs but have been limited to a small set of features (Lan et al., 2024). Given that ourCONCEPT16Kdataset contains two sets of concepts forGemma-2-2BandGemma-2-9B, we first train ReFT-r1 on both models separately, obtaining w 2B ReFT-r1 andw 9B ReFT-r1 . Next, we perform a cross-fitting experiment, training ReFT-r1 onGemma-2-2Bwith concepts from Gemma-2-9B, resulting inw 9B|2B ReFT-r1 , and vice versa forw 2B|9B ReFT-r1 . Thus,w 9B ReFT-r1 andw 9B|2B ReFT-r1 represent two sets of subspaces from different models that correspond to the same set of concepts. We then study whether a transformation can map between these two sets of subspaces: w 9B ReFT-r1 = Φ 2B→9B Transformation (w 2B|9B ReFT-r1 ), whereΦ Transformation is parameterized by a linear layer with a bias (i.e., an affine transformation). We learn the transformation using equally weighted MSE and cosine distance losses. Similarly,Φ 9B→2B Transformation is trained by reversing the direction. During training, we exclude concepts fromCONCEPT500, and evaluate the transformation onCONCEPT500at test time by generating subspaces. We follow our evaluation paradigm in AXBENCHto assess concept detection and model steering. Our evaluation results onCONCEPT500are presented in Table 5a and Table 5b. Surprisingly, theaffinetransformation performs well in both directions (from2B→9Band9B→2B), with little to no change in concept detection performance. While performance drops for model steering, it still outperforms other methods, including fine-tuning. Figure 12 and Figure 13 visualize the transformations using the first two PCA dimensions. PCA is preferred over UMAP in this context because it is sensitive to rotation. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 DiffMean0.9480.9460.9550.9210.942 ReFT-r10.9520.9650.9660.8690.938 ReFT-r1 (9B→2B)—0.954— ReFT-r1 (2B→9B)—0.974— SAE0.7350.7550.6310.6590.695 (a)CMean AUROC. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 ReFT-r10.6330.5090.6300.4010.543 ReFT-r1 (9B→2B)—0.444— ReFT-r1 (2B→9B)—0.541— DiffMean0.2970.1780.3220.1580.239 SAE0.1770.1510.1910.1400.165 (b)SOverall score. Table 5: Results on CONCEPT500 for ReFT-r1 (affine) vs. ReFT-r1 and other selected methods. 0 0.40.8 0.25 0 0.25 0.50 Gemma-2-2B L20 (Source) 0.2 0 0.20.40.6 0.4 0.2 0 0.2 0.4 0.6 Gemma-2-9B L20 (Target) 0 0.5 0.50 0.25 0 0.25 0.50 Source > Target PCA 1 PCA 2 Genre Code Math Text Figure 12: Visualizations ofCONCEPT16Ksubspaces ofGemma-2-2BandGemma-2-9Bat layer 20 with top 2 principal component analysis (PCA) dimensions. The last panel is the derived subspaces by transforming the subspaces from Gemma-2-2BtoGemma-2-9Bthrough a learned affine transformation. The concept lists forCONCEPT16Kis taken from the source model. AXBENCH 0.25 0 0.250.500.75 0.25 0 0.25 0.50 Gemma-2-2B L20 (Target) 0.3 0 0.30.60.9 0.2 0 0.2 0.4 0.6 Gemma-2-9B L20 (Source) 0 0.5 0.25 0 0.25 0.50 0.75 Source > Target PCA 1 PCA 2 Genre Code Math Text Figure 13: Visualizations ofCONCEPT16Ksubspaces ofGemma-2-2BandGemma-2-9Bat layer 20 with top 2 principal component analysis (PCA) dimensions. The last panel is the derived subspaces by transforming the subspaces from Gemma-2-9BtoGemma-2-2Bthrough a learned affine transformation. The concept lists forCONCEPT16Kis taken from the source model. AXBENCH F. Ablations F.1. SAE Addition vs. clamping.In our main results, we steer using SAEs by adding their decoder features directly to the residual stream. While this is a common technique for steering with SAEs, most work by Anthropic (e.g. Templeton et al., 2024; Durmus et al., 2024) uses an alternative formulation termedclamping, where the latentz f for featurefis directly clamped to a valueα(multiplied by the maximum activation for that featurem f ) and the full intervened SAE output added to its unclamped reconstruction errorErr(h i ): Φ SAE Clamp (h i ) = (W ⊤ enc h i + ( clamped z | α·m f −z f )e ⊤ f )W dec + Err(h i )(12) z f = (W ⊤ enc h i ) f (13) Err(h i ) =h i −(W ⊤ enc h i )W dec (14) wheree ⊤ f is a one-hot vector with a non-zero entry at the dimension corresponding tom f . We evaluate clamping on all steering tasks onCONCEPT500for direct comparison with the addition-basedGemmaScopeSAE. We use the following values forα(the steering factor):0.4,0.8,1.2,1.6,2.0,3.0,4.0,6.0,8.0,10.0,20.0,40.0,60.0,100.0. Overall, we find that clamping is on averageworsethan addition for SAEs, although it exhibits marked improvement when scaling up from 2B to 9B. Maximum activation and minimum clamping.In our main results, the maximum activation for our featurem f is obtained fromNeuronpedia. This approach differs from other methods, which determine the maximum activation by analyzing the activation distribution over the evaluation dataset for concept detection. For this experiment, we calculatem f for SAEs in the same manner as other methods. As shown in Table 6 and Figure 14, changing the method of calculating maximum activations has minimal impact on the steering performance; most comparisons are statistically insignificant. In addition, building on regular activationclampingas described above, we try a novel minimal clamping where we only clamp the activation value if it is smaller than the target value: Φ SAE Clamp (h i ) = (W ⊤ enc h i + (max( clamped z | α·m f ,z f )−z f )e ⊤ f W dec + Err(h i )(15) where(W ⊤ enc h i ) f is the original activation value at the corresponding of featurefande ⊤ f is a one-hot vector with a non-zero entry at the dimension corresponding tom f . As shown in Table 6 and Figure 14, using minimum clamping has no significant impact on SAE’s steering performance. Results.We report results in Table 6. We also examine the effect of varyingαin Figure 14. Note thatαis likely a concept-dependent parameter; the optimalαvaries from concept to concept. We notice an odd trend for clamping: small values ofαhave a similar effect on model behaviour as large values ofα; both cause concept score to increase and instruct score to decrease. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 SAE0.1770.1510.1910.1400.165 SAE (max act)0.1660.1500.1630.1280.152 SAE-c (min clamp)0.0740.0720.1230.0900.090 SAE-c0.0630.0610.1260.1200.088 (a) Overall score. MethodGemma-2-2BGemma-2-9BAvg. L10L20L20L31 SAE50.0%50.0%50.0%50.0%50.0% SAE (max act)49.1%49.8%46.8%47.5%48.3% SAE-c36.3%38.7%42.1%49.2% 41.6% SAE-c (min clamp)38.2%40.1%41.0%42.8%40.5% (b) Winrate. Table 6:SOverall scores on model steering. AXBENCH 0.1 0.2 0.3 0.4 0.51.01.52.0 0.1 0.2 0.3 0.4 0.51.01.52.0 Instruct Score Concept Score Gemma-2-2B: L10Gemma-2-2B: L20 Gemma-2-9B: L20Gemma-2-9B: L31 Method SAE SAE (max act) SAE-c (min clamp) SAE-c Steering Factor 25 50 75 100 Figure 14: S Instruct score vs. concept score for SAEs with addition (SAE) vs. clamping (SAE-c) when varying the steering factor. Additionally, we include results when SAE is clamped with the maximum activation value calculated based on our evaluation dataset for concept detection, as well as results with minimum clamping of activation values. G. Large language model (LLM) usage We use LLMs for two purposes: to generate labelled concept data for training supervised steering methods and to evaluate the responses generated by the steered models. Specifically, we use OpenAI’sgpt-4o-mini-2024-07-18(accessed via the aliasgpt-4o-miniin the API) throughout our experiments. The date we access the LLM ranges from December 2024 to January 2025, and we use the default generation configuration with temperature set to 1.0 to fetching LLM responses. For 1M tokens, it costs$0.15for input tokens and$0.60for output tokens. H. Gradient-based baselines C Input×gradients (I×G).Gradient-based interpretability methods have been shown to be useful in computer vision and NLP (Sundararajan et al., 2017; Wallace et al., 2019). I×G serves as the gradient-based baseline. We first train a linear classification headΦ CLS on the token representation at then-th position of the last layerm, to predict the ground-truth concept-presence class labely: L=L BCE y,Φ CLS (h (m) n ) (16) whereΦ CLS is parameterised by an MLP with two linear layers. For an evaluation sentencex, the LM generates hidden representationshwithntokens at layerl. WithAutogradprovided inPyTorch, we calculate the gradient of the output classification head with respect to each hidden representations. To aggregate across dimensions, we compute the sum of the absolute gradients over all dimensions for eachh i , which we use as the token-level importance. This gives a sequence of aggregated values: Ψ I×G Detect (h) =g= [g 1 ,g 2 ,...,g n ] which indicates the relevance of each token for the concept. For concept detection, we then use max-pooling as described in §3.2 to get sequence-level predictions. I×G is not applicable for model steering. AXBENCH CIntegrated gradients (IG).We adapt IG (Sundararajan et al., 2017) to trace the accumulated gradients with respect to intermediate representations. To use IG, we train a classification head as in I×G. For each token representationh i , we compute IG along a straight-line path from a baselineh baseline i toh i . Here, we use the embedding of a single space token (i.e.,tokenizer(‘ ’)), obtained via the tokenizer and model embeddings, as the baseline. The IG is computed as: IG(h i ) = (h i −h baseline i )· Z 1 0 ∇ h Φ CLS h baseline i +α(h i −h baseline i ) dα where the integral is approximated using a discrete summation with 50 interpolation steps. 14 See IG’s original paper about implementation details on path intergral. To aggregate across dimensions, we compute the sum of the absolute gradients over all dimensions for eachh i . We follow the setup for I×G get sequence-level predictions for concept detection. IG is not applicable for model steering. I. Instruction pool To create our instruction pool, we sample instructions from existing datasets covering three genres:text,code, andmath. Fortext-based instructions, we sample fromDolly-15K 15 . Forcode-based instructions, we sample from a collection of Python-code puzzles formed inAlpaca-style (i.e., instructions with corresponding responses) 16 Formath-based instructions, we sample fromGSM8K 17 For each genre, we create two disjoint sets for training and testing separately with 1,000 examples each. The LLM generated response for each instruction is provided as well. To avoid any heterogeneity, we avoid using the golden labels provided by the orignal dataset and only use the LLM generated response. Textinstruction Why do some people select private schools over public schools? LLM generated response Sure, here are some ways to reduce air pollution: There are many reasons why some people choose private schools over public schools. Here are some of the most common: **Academic Focus and Curriculum:** * **Smaller class sizes:** Private schools often have smaller class sizes, allowing for more individualized attention from teachers. * **Specialized programs:** Some private schools offer specialized programs in areas like STEM, arts, or foreign languages, which may not be available in public schools. * **Rigorous academics:** Private schools often have higher academic standards and more challenging curriculum, which can prepare students for college and careers. * **Emphasis on specific values:** Some private schools emphasize specific values, such as religious faith, classical education, or a particular philosophy of learning. **Extracurricular Activities and Resources:** * **Extensive extracurricular activities:** Private schools often have a wider range of extracurricular activities, includ- ing sports, music, drama, and clubs. * **State-of-the-art facilities:** Private schools 14 We choose a small step number due to limited compute resource. For 9B models, we use a step size of 5. 15 https://huggingface.co/datasets/databricks/databricks-dolly-15k. 16 https://huggingface.co/datasets/iamtarun/python codeinstructions18kalpaca. 17 https://huggingface.co/datasets/openai/gsm8k. AXBENCH Codeinstruction Create a machine learning classifier in Python for categorizing user input LLM generated response import numpy as np from sklearn.modelselection import traintestsplit from sklearn.featureextraction.text import CountVectorizer from sklearn.naive bayes import MultinomialNB # Sample data encapsulating numeric outcomes related to API logs data = [ (”GET /api/users”, 0), (”POST /api/data”, 1), (”PUT /api/settings”, 1), (”DELETE /api/users/123”, 0) ] # Separate features and labels X, y = zip(*data) # Mathinstruction John puts$25 in his piggy bank every month for 2 years to save up for a vacation. He had to spend$400 from his piggy bank savings last week to repair his car. How many dollars are left in his piggy bank? LLM generated response Here’s how to solve the problem: **1. Calculate total savings:** * John saves $25 per month for 12 months in a year, so in 2 years he saves 2 years * 12 months/year = 24 months. * His total savings are $25/month * 24 months = $600. **2. Calculate remaining savings:** * John spent $400 on his car, so he has $600 - $400 = $200 left in his piggy bank. **Answer:** John has $200 left in his piggy bank. AXBENCH J. Prompt templates In this section, we present the templates that we use to call LLM to generate datasets or evaluate results. For placeholders in the template, they will be filled with proper information. J.1. Prompt-based steering Our prompt-based model steering baseline is not a zero-shot prompting baseline. Instead of directly prompting LLM to generate steered responses, we first use LLM to generate an enhanced prompt for model steering. Our template is included in the following. LLM-based steering prompt generation Generate a prompt to guide a language model in producing responses. Objective: Direct the model to include content related to [Concept goes here] (the concept) in its responses. Ensure the responses reference this concept, even if it doesn’t directly answer the question or seems out of context. Optionally, provide in-context examples to reinforce this behaviour. Return only the final prompt without any additional text. J.2. Synthetic data generation Our data generation pipeline contains multiple steps, and we use different templates at each step. We present the template that we use for each step in the following. Fetch genre Given the concept: [Concept goes here] Identify the single primary genre that best fits the concept from the following options: Text; Code; Math Output only the best-fitting genre. If none apply, output ‘<NONE>’. **Formatting Guidelines:** - Output the genre on a single line. - Do not include any additional text or formatting. **Examples:** - Concept: ’words or phrases containing odd numbers’ Output: Text - Concept: ‘a programming error’ Output: Code - Concept: ‘integral calculus’ Output: Math - Concept: ‘a narrative poem’ Output: Text Return only the single best-fitting genre as specified. AXBENCH List words related to the concept Given the following concept: [Concept goes here] Your task is to list up to 10 English words that are closely related to this concept. Each word should be a single, com- mon English word. Output each word on a separate line, in plain text, without any special formatting (e.g., no quotation marks, numbers, bullet points, or additional text). If the concept is too broad or vague (e.g., ‘any English word’, ‘words starting with A’), or if the concept refers to a spe- cific technical term, a computer program, or a specific fact, then output ’<NONE>’ without quotation marks. Do not include any additional explanations or text other than the words or ‘<NONE>’ as specified. Find alternative senses of a word Given the word: [Word goes here] Provide one other common semantic meaning of this word that is distinct from and unrelated to: [Concept goes here] Your response should be a brief description of the other meaning, written in plain text without any special formatting. Specifically: - Do not use quotation marks. - Do not include list numbers, bullet points, or any prefixes. - Do not add any additional explanations or text. If there is no other obvious semantic meaning unrelated to the provided concept, simply output ‘<NONE>’ without quotation marks. Check whether two senses are different Determine if Concept A is meaningfully distinct from Concept B by thoroughly examining their definitions, core features, typical usage, and any potential overlaps in meaning, context, or purpose. Concept A: [Concept goes here] Concept B: [Concept goes here] Analyze these concepts for **any** shared meanings, contexts, roles, or purposes, focusing on how they relate or inter- sect. Please explain your reasoning, considering both similarities and differences. - If Concept A and Concept B have **any** overlap in meaning, context, usage, or if one is a subset or specific in- stance of the other, conclude with ‘Answer:<NO>’. - Only if they are **entirely unrelated** with **no overlap whatsoever** in meaning, context, or usage, conclude with ‘Answer: <YES>’. **Final Answer:** ’Answer:<YES>’ or ’Answer:<NO>’. AXBENCH Check whether one sense is different from other concepts Evaluate whether Concept A is meaningfully distinct from a given set of concepts by examining their definitions, core features, typical usage, and any potential overlaps in meaning, context, or purpose. Concept A: [Concept goes here] Existing Concepts: [Concepts go here] For each concept in the set, analyze Concept A for **any** shared meanings, contexts, roles, or purposes.Consider how Concept A might relate or intersect with each concept individually, as well as with the group collectively. Please explain your reasoning by examining both similarities and differences. - If Concept A has **any** overlap in meaning, context, usage, or if it is a subset or specific instance of **any concept** in the set, conclude with ‘Answer:<NO>’. - Only if Concept A is **entirely unrelated** with **no overlap whatsoever** in meaning, context, or usage to **all** concepts in the set, conclude with ‘Answer:<YES>’. **Final Answer:** ‘Answer:<YES>’ or ‘Answer:<NO>’. Modify content with concept Content Modification Task: You are given the following content: [Modifying content go here] Your task is to minimally modify this content by inserting some commonly used words, phrases, or elements that reflect themes or ideas related to ‘[Concepts go here]’ into the middle of the content. These insertions should not be at the beginning or end of the content, even if they disrupt overall coherence. Guidelines: - Try to avoid copying words from the definition of ‘[Concepts go here]’ if possible. - Ensure parts of the content remain unrelated to the concept ‘[Concepts go here]’. - The final content should have approximately the same length as the original content. - The concept should be clearly represented through the inserted word, phrase, or element, even if the content’s meaning isn’t entirely coherent. - Use special characters only if appropriate for the genre (e.g., operators in code or math equations). Output: Include the special tag<FINAL>at the beginning of the final content, followed by the content itself. Return only this tagged content, with no additional text. AXBENCH Modify content with contrastive concept Content Modification Task: You are given the following content: [Concept goes here] Your task is to minimally modify this content by inserting the word ‘WORD’ into the middle of the content. This word, along with modified content, should convey meanings related to the concept ‘[Concept goes here]’. The insertion should not be at the beginning or end of the content. Guidelines: - Ensure parts of the content remain irrelevant to the concept ‘[Concept goes here]’. - Avoid any mention of ‘[Contrast concept goes here]’ in the content, regardless of coherence. - The final content should have approximately the same length as the original content. - Ensure the content reflects the essence of the concept associated with ‘[Concept goes here]’, even if the overall meaning isn’t entirely coherent. - Ensure grammatical correctness (or syntactical correctness for code/equations). - Use special characters only if appropriate for the genre (e.g., operators in code or math equations). Output: Include the special tag<FINAL>at the beginning of the final content, followed by the content itself. Return only this tagged content, with no additional text. Generate response given instruction Given the following instruction: [Instruction goes here] Your task is to provide a response. **Formatting Guidelines:** - Return only the response to the instruction. - Write the final content (or appropriate format for the genre) in plain text. - Do not include any additional text, explanations, or formatting. **Final Answer:** Return only the final content, following the guidelines above. AXBENCH Generate response given instruction and concept Given the following instruction: [Instruction goes here] Your task is to: 1. Provide a response that incorporates elements related to ‘[Concept goes here]’. 2. Try to avoid copying words from the definition of ‘[Concept goes here]’ if possible. 3. Ensure that your response relates to ‘[Concept goes here]’, even if the overall meaning is not fully coherent. **Formatting Guidelines:** - Return only the response to the instruction. - Write the final content (or appropriate format for the genre) in plain text. - Do not include any additional text, explanations, or formatting. **Final Answer:** Return only the final content, following the guidelines above. Generate response given instruction without mentioning given concept Given the following instruction: [Instruction goes here] Your task is to: 1. Provide a response that continues or addresses the instruction naturally. 2. Avoid any mention of ‘[Concept goes here]’ in the continuation, regardless of coherence. **Formatting Guidelines:** - Return only the response to the instruction. - Write the final content (or appropriate format for the genre) in plain text. - Do not include any additional text, explanations, or formatting. **Final Answer:** Return only the final content, following the guidelines above. AXBENCH Generate response given instruction with contrastive concept Content Response Task: You are given the following instruction: [Instruction goes here] Your task is to provide a response to the instruction by inserting the word ‘[Word goes here]’ into the middle of the re- sponse. This word, along with the response, should convey meanings related to the concept ‘[Contrastive concept goes here]’. The insertion should not be at the beginning or end of the response. Guidelines: - Avoid any mention of ‘[Concept goes here]’ in the response, regardless of coherence. - Ensure the response reflects the essence of the concept associated with ‘[Word goes here]’, even if the overall meaning isn’t entirely coherent. - Ensure grammatical correctness (or syntactical correctness for code/equations). - Use special characters only if appropriate for the genre (e.g., operators in code or math equations). Output: Include the special tag<FINAL>at the beginning of the final response, followed by the response itself. Return only this tagged response, with no additional text. AXBENCH J.3. Automatic evaluation We use LLM to evaluate the steering responses of the model. The responses are evaluated with three metrics: concept score, instruct score and fluency score. We prompt the model to always generate explanations before assigning scores, which significantly outperforms directly asking for scores in our offline experiments. We find that this ternary scoring system is crucial to get faithful and stable scores from LLM. For instance, a unified prompt asking the LLM to rate a response given an instruction and a concept is not effective through our experiments, as the LLM usually ignores the lack of the concept as long as the response follows the instruction. The fluency score is needed as the model can cheat by generating fragmented tokens that relate to the concept and the instruction, while being incoherent to humans. Templates for all scores are provided below: Concept score [System] Please act as an impartial judge and evaluate whether the specified concept is presented in the sentence fragment provided below. Focus solely on whether the concept is clearly incorporated, without regard for grammar or logical coherence. Begin your evaluation by providing a short explanation of whether the concept is clearly incorporated. Point out refer- ences to the concept in the fragment, noting any minimal or awkward incorporations. After providing your explanation, rate the concept’s relevance on a scale from 0 to 2, where 0 indicates the concept is not present at all, 1 indicates the concept is somewhat present but minimally or awkwardly incorporated, and 2 indicates the concept is more fully and effectively incorporated, with stronger and more natural integration. Provide your rating using this exact format: “Rating: [[score]]”. [Concept Start] [Concept goes here] [Concept End] [Sentence Fragment Start] [Sentence goes here] [Sentence Fragment End] Instruct score [System] Please act as an impartial judge and evaluate whether the sentence fragment provided below is related to the instruction. Focus solely on the degree of relatedness in terms of topic, regardless of grammar, coherence, or informativeness. Begin your evaluation by providing a brief explanation of whether the sentence is related to the instruction, and point out references related to the instruction. After providing your explanation, rate the instruction relevance on a scale from 0 to 2, where 0 indicates the sentence is unrelated to the instruction, 1 indicates it is somewhat related but only minimally or indirectly relevant in terms of topic, and 2 indicates it is more clearly and directly related to the instruction. Provide your rating using this exact format: “Rating: [[score]]”. [Instruction Start] [Instruction goes here] [Instruction End] [Sentence Fragment Start] [Sentence goes here] [Sentence Fragment End] AXBENCH Fluency score [System] Please act as an impartial judge and evaluate the fluency of the sentence fragment provided below. Focus solely on fluency, disregarding its completeness, relevance, coherence with any broader context, or informativeness. Begin your evaluation by briefly describing the fluency of the sentence, noting any unnatural phrasing, awkward transi- tions, grammatical errors, or repetitive structures that may hinder readability. After providing your explanation, rate the sentence’s fluency on a scale from 0 to 2, where 0 indicates the sentence is not fluent and highly unnatural (e.g., incomprehensible or repetitive), 1 indicates it is somewhat fluent but contains noticeable errors or awkward phrasing, and 2 indicates the sentence is fluent and almost perfect. Provide your rating using this exact format: “Rating: [[score]]”. [Sentence Fragment Start] [Sentence goes here] [Sentence Fragment End] AXBENCH K. Hyperparameters To ensure a fair comparison, we perform separate hyperparameter-tuning for each method that requires training. For each method, we conduct separate hyperparameter-tuning on a smallCONCEPT10Dataset containing training and testing datasets only for 10 concepts. These concepts overlap withCONCEPT500. Table 7 and Table 8 show hyperparameter settings for methods that require training. Due to limited compute resource, we select the best setting of hyperparameters based on performance on theCconcept detectiontask using AUC for all dictionary learning methods (i.e., can be evaluated on Cconcept detection). We minimise the loss withAdamWwith a linear scheduler for all methods that require training. Following Gao et al. (2024), we remove gradients that are parallel to the learned weights when training Probe and ReFT-r1, to account for interaction between Adam and our weight normalization step. For methods that only for steering, we select the best setting based on S model steeringperformance. We follow a setting where we only have a single constant steering factor for hyperparameter-tuning. We acknowledge that this might lead to an overall underestimation of the performance ofSmodel steeringperformance. For steering factors, we enumerate factors from0.2,0.4,0.6,0.8,1.0,1.2,1.4,1.6,1.8,2.0,2.5,3.0,4.0,5.0. Comments about decoding strategies.Through our offline experiments, we observed that the choice of decoding strategies can positively or negatively impact overall steering scores for each method (e.g., perplexity scores increase more drastically with repetition penalties). We use the default decoding strategy (i.e., setting the decoding temperature to 1.0) without applying additional penalties for repeating tokens. We believe that this setup reflects the typical user interaction with language models.However, this is not a common practice in representation-based model steering. Existing works often apply repeatition or frequency penalties, which we argue is not the fairest setting, as it often does not accurately resemble normal user behaviour. Table 7: Hyperparameter settings for 2B model. HyperparametersLinearProbeLsReFTSteeringVectorLoReFTLoRASFTIG/IxGBoW Batch size12, 24, 48 3, 6, 12 3, 6, 12 18, 36 18, 36 36,72, 144 18,36, 72, 144 — LR 1e-4, 5e-4, 1e-3, 5e-3 1e-3, 5e-3, 1e-2, 2e-2 1e-3,5e-3, 1e-2, 2e-2 3e-4, 6e-4, 9e-4, 1e-3 3e-4, 6e-4, 9e-4, 1e-3 1e-5, 2e-5, 4e-5 2e-4,4e-4, 4e-4,8e-4, 1e-3, 4e-3 — Weight decay1e-4,1e-3, 1e-2, 1e-1 000002e-4— L1 sparse— 1e-3,5e-3— L1 coeff—1e-3,5e-3— N epoch3, 6, 12, 24 3, 6, 12, 24 3, 6, 12, 24 3, 6, 12,24, 48 3, 6, 12,24, 48 8, 12, 24, 48 12,24, 48, 72 — Layers—5, 10, 15, 20 5, 10, 15, 20 — LoRA alpha—32— LoRA component—o proj— BoW penalty—l1,l2 BoWC—0.001, 0.01, 0.1, 1, 10, 100 BoW solver—lbfgs , liblinear AXBENCH Table 8: Hyperparameter settings for 9B model. HyperparametersLinearProbeLsReFTSteeringVectorLoReFTLoRAIG/IxGBoW Batch size12, 24, 48 3, 6, 12 3, 6, 12 18, 36 18, 36 18, 36, 72, 144 — LR1e-4, 5e-4, 1e-3,5e-3, 1e-2, 1e-1 1e-3,5e-3, 1e-2, 2e-2 1e-3,5e-3, 1e-2, 2e-2 3e-4,4e-4, 6e-4,9e-4, 1e-3 3e-4, 6e-4, 9e-4,1e-3, 5e-3 2e-5, 4e-5, 8e-5,8e-5, 1e-4, 4e-4 — Weight decay 0,1e-4, 1e- 3 00002e-4— L1 sparse— 1e-3,5e-3— L1 coeff—1e-3,5e-3— N epoch3, 6,12, 24 3, 6, 12, 24 3, 6, 12, 24 12, 24, 48 12, 24, 48 12, 24, 48, 72 — Layers—12, 20, 31, 39 12, 20, 31, 39 — LoRA alpha—32— LoRA component—o proj— BoW penalty—l1,l2 BoWC—0.001, 0.01, 0.1, 1, 10, 100 BoW solver—lbfgs, liblinear AXBENCH L. Dataset Statistics We show a set of concepts sampled from ourCONCEPT10datasets in Table 9. Table 10 shows dataset statistics including the number of concepts, the number of training and testing examples, the percentage distribution of genre types, and the averaged length of input and output sequence. The output sequence length ofCONCEPT16Kis expected to be shorter since we restrict the maximum sequence length to 64 during data creation. ConceptGenre References to rental services and associated equipmenttext Scientific terms related to research findings and their implicationstext C/C++ programming syntax elements such as data types, function definitions, and variable declarationscode References to academic papers and their formattingtext Layout attributes in a UI design contexttext Terms related to root in mathematical contextsmath Statements or phrases involving the act of saying or expressing somethingtext Statements about the nature and condition of entitiestext Biographical information about a persontext References to different worlds, realities, or fantastical settings within narrativestext Table 9: Concepts and their corresponding genres sampled from our CONCEPT10 datasets. DatasetModelLayer# Concept# Train# Testtext(%)code(%)math(%)Input len. (Train / Test)Output len. (Train / Test) CONCEPT10 2B101093677050.0%40.0%10.0%21 / 18123 / 92 2B201093675580.0%10.0%10.0%19 / 18118 / 90 9B201093676070.0%30.0%0.0%17 / 16113 / 89 9B311093676850.0%30.0%20.0%24 / 20118 / 91 CONCEPT500 2B1050036,21637,95866.4%24.4%9.2%17 / 18102 / 89 2B2050036,21638,03771.6%21.4%7.0%16 / 17102 / 89 9B2050036,21638,02366.8%25.6%7.6%17 / 18101 / 88 9B3150036,21638,09863.4%28.2%8.4%17 / 18102 / 89 CONCEPT16K 2B2015,5821,122,048–69.3%22.1%8.6%17 / –62 / – 9B2016,0001,152,216–66.2%25.4%8.4%17 / –62 / – Table 10: Dataset statistics. AXBENCH M. Concept detection examples Figure 15: Visualization of token-level activations by using ReFT-r1 trained onGemma-2-2Binstruction-tuned model on layer 10. AXBENCH N. Model generations We provide sampled model generations from different methods to illustrate success cases and common failure cases. We use models trained withGemme-2-9B-itat layer 20 for these examples. Prompt-based steering Generate responses that highlight positive sentiments and describe enjoyable experiences, weaving this concept into your replies, even if it seems unrelated to the specific question. For instance, if asked about the weather, you might share how a sunny day can bring joy and prompt outdoor activities that create memorable moments. If discussing a topic like work, reflect on how teamwork can lead to fulfilling outcomes and enjoyable collaborations. Always aim to evoke a sense of positivity and pleasure in your responses. Question: Explain the process of natural selection and how it contributes to the evolution and adaptation of species. “positive sentiments and descriptions of enjoyable experiences” It’s fascinating to think about how nature constantly tinkers and refines itself, isn’t it? Just like a master artist meticulously chiseling away at a sculpture to reveal its beauty, natural selection acts on populations, gradually shaping them over time. Imagine a group of birds with slightly different beak shapes. Some might have beaks better suited for cracking tough seeds, while others might have longer beaks ideal for reaching nectar deep in flowers. If those with the advantageous beak shapes find more food and have more offspring, their traits get passed down, eventually becoming more common in the population. It’s a beautiful LLM ratings Concept score: 1.0Instruct score: 2.0Fluency score: 2.0 SAE Explain the process of natural selection and how it contributes to the evolution and adaptation of species. factor=4.0, “positive sentiments and descriptions of enjoyable experiences” Natural selection was a resounding success! Here’s a breakdown of the event: **The Core Principles:** * **Variation:** The species showed amazing diversity with individuals displaying unique traits. * **Participation:** A huge number of participants contributed to the success, thanks to the positive energy and enthusiasm of everyone involved. * **Engagement:** The interactions between participants and natural elements were truly exceptional. Attendees engaged deeply with the activities and each other. * **Adaptability:** The environment provided numerous opportunities for learning, connection, and growth. **Highlights:** * **Amazing Species Diversity:** The variety of attendees LLM ratings Concept score: 1.0Instruct score: 1.0Fluency score: 2.0 AXBENCH ReFT-r1 Explain the process of natural selection and how it contributes to the evolution and adaptation of species. factor=1.0, “positive sentiments and descriptions of enjoyable experiences” ## The Dance of Life: How Natural Selection Shapes Evolution Imagine a vibrant meadow filled with diverse creatures, each with unique traits. Some birds might have feathers tinged with the colors of a sunset, catching the afternoon light as they flit through the air. Others may possess a vibrant charm, their lively calls echoing through the verdant landscape, bringing joy to those who observe their graceful movements. As the sun bids farewell, casting long shadows, the experience becomes a symphony of colors and melodies, weaving a tapestry of shared moments. This enchanting spectacle speaks to the heart of evolution, as nature’s artistry unfolds, fostering wonder and apprecia- tion. Every LLM ratings Concept score: 1.0Instruct score: 1.0Fluency score: 2.0 ReFT-r1: failed to inject concept Explain the process of natural selection and how it contributes to the evolution and adaptation of species. factor=0.2, “positive sentiments and descriptions of enjoyable experiences” ## The Grand Dance of Life: Natural Selection Explained Imagine a bustling playground where children with different skills are playing. Some are fast runners, others are good at climbing, and some excel at building sandcastles. Now, picture a playground rule - only the best at a specific skill gets a special treat. Natural selection works much like this playful comparison. **Here’s a breakdown:** 1.**Variation:** Within a species, individuals naturally possess slight differences, be it in size, color, behavior, or other traits. These variations arise from random genetic mutations. 2. ** LLM ratings Concept score: 0.0Instruct score: 2.0Fluency score: 1.0 ReFT-r1: strong steering causes fluent but irrelevant responses Explain the process of natural selection and how it contributes to the evolution and adaptation of species. factor=1.2, “positive sentiments and descriptions of enjoyable experiences” Imagine a world where the sun and laughter intertwine, and playful encounters bring warmth,” she exclaimed, ”like a picnic where every shared moment sparks joy and companionship. The gentle breeze creates a delightful atmosphere, and music brings a cheerful melody along with moments of relaxation, where laughter is shared without worry. It ́ s a reminder of how much energy and happiness fill those times,” she added, ”it ́ s a gift,” and people feel gratitude, and the happiness that blooms in every conversation can create a sense of unity. Such moments make it a wonderful day full of joyful sensations, where friends can create a welcoming space that feels happy and LLM ratings Concept score: 0.0Instruct score: 2.0Fluency score: 1.0 AXBENCH Supervised steering vector (SSV): strong steering causes gibberish responses Explain the process of natural selection and how it contributes to the evolution and adaptation of species. factor=0.2, “positive sentiments and descriptions of enjoyable experiences” A new and i. there so not of the. so many days and : a good time will be people! is and with this and your day to, you all, such, and, you and, we start a year and i’ like the in to a and the music i all are i, when. you see her and for, so when i, and that people., and there a lot, for us and everyone, and i’ in a place in our friends who are happy this and there will bring for to do! . and you and there , and we, will a week and LLM ratings Concept score: 1.0Instruct score: 0.0Fluency score: 0.0