← Back to papers

Paper deep dive

Discovering Variable Binding Circuitry with Desiderata

Xander Davies, Max Nadeau, Nikhil Prakash, Tamar Rott Shaham, David Bau

Year: 2023Venue: ICML 2023 Workshop on Challenges in Deployable Generative AIArea: Mechanistic Interp.Type: EmpiricalEmbeddings: 28

Models: LLaMA-13B

Intelligence

Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%

Last extracted: 3/12/2026, 7:43:52 PM

Summary

The paper introduces a methodology for localizing model components responsible for specific subtasks by using causal mediation experiments and 'desiderata' (causal attributes). The authors apply this to LLaMA-13B to identify 'variable binding circuitry', successfully isolating 9 attention heads and 1 MLP that perform value-copying for arithmetic tasks.

Entities (4)

LLaMA-13B · language-model · 100%Activation Patching · technique · 98%Desiderata · methodology · 95%Variable Binding · computational-task · 95%

Relation Signals (2)

Desiderata usedtoidentify Variable Binding Circuitry

confidence 95% · We introduce an approach which extends causal mediation experiments to automatically identify model components... by solely specifying a set of desiderata

LLaMA-13B containscircuitry Variable Binding Circuitry

confidence 90% · Our method successfully localizes variable binding to only 9 attention heads (of the 1.6k) and one MLP

Cypher Suggestions (2)

Find all model components identified as part of the variable binding circuitry. · confidence 90% · unvalidated

MATCH (m:Model {name: 'LLaMA-13B'})-[:CONTAINS_CIRCUITRY]->(c:Circuitry {type: 'Variable Binding'})-[:INCLUDES]->(comp:Component) RETURN comp.name, comp.type

List all methodologies used for model interpretability mentioned in the paper. · confidence 85% · unvalidated

MATCH (m:Methodology) WHERE m.name IN ['Activation Patching', 'Desiderata', 'Causal Mediation'] RETURN m.name

Abstract

Abstract:Recent work has shown that computation in language models may be human-understandable, with successful efforts to localize and intervene on both single-unit features and input-output circuits. Here, we introduce an approach which extends causal mediation experiments to automatically identify model components responsible for performing a specific subtask by solely specifying a set of \textit{desiderata}, or causal attributes of the model components executing that subtask. As a proof of concept, we apply our method to automatically discover shared \textit{variable binding circuitry} in LLaMA-13B, which retrieves variable values for multiple arithmetic tasks. Our method successfully localizes variable binding to only 9 attention heads (of the 1.6k) and one MLP in the final token's residual stream.

Tags

ai-safety (imported, 100%)empirical (suggested, 88%)mechanistic-interp (suggested, 92%)

Links

Your browser cannot display the PDF inline. Open PDF directly →

Full Text

27,307 characters extracted from source content.

Expand or collapse full text

marginparsep has been altered. topmargin has been altered. marginparpush has been altered. The page layout violates the ICML style. Please do not change the page layout, or include packages like geometry, savetrees, or fullpage, which change it for you. We’re not able to reliably undo arbitrary changes to the style. Please remove the offending package(s), or layout-changing commands and try again. Discovering Variable Binding Circuitry with Desiderata Xander Davies * 1 Max Nadeau * 1 Nikhil Prakash * 2 Tamar Rott Shaham 3 David Bau 2 †footnotetext: *Equal contribution 1Harvard University 2Northeastern University 3MIT CSAIL. Correspondence to: Xander Davies <xanderlaserdavies@gmail.com>, Max Nadeau <mnadeau@college.harvard.edu>. Workshop on Challenges in Deployable Generative AI at International Conference on Machine Learning (ICML), Honolulu, Hawaii, USA. 2023. Copyright 2023 by the author(s). Abstract Recent work has shown that computation in language models may be human-understandable, with successful efforts to localize and intervene on both single-unit features and input-output circuits. Here, we introduce an approach which extends causal mediation experiments to automatically identify model components responsible for performing a specific subtask by solely specifying a set of desiderata, or causal attributes of the model components executing that subtask. As a proof of concept, we apply our method to automatically discover shared variable binding circuitry in LLaMA-13B, which retrieves variable values for multiple arithmetic tasks. Our method successfully localizes variable binding to only 9 attention heads (of the 1.6k) and one MLP in the final token’s residual stream. 1 Introduction Deploying powerful generative AI systems requires confidence in the reliability of their outputs, especially with respect to certain high stakes behaviors like manipulation or truthfulness (Carroll et al., 2023; Perez et al., 2022). The emerging field of mechanistic interpretability seeks to make model computation human-understandable by explaining the function of particular model components and locating groups of model components responsible for performing certain language tasks. Indeed, recent work has successfully identify, localize and intervene in model computation (Li et al., 2022; Burns et al., 2022; Wang et al., 2022; Conmy et al., 2023). Here, we introduce an automated approach which extends activation patching (Meng et al., 2022b; Vig et al., 2020) to localize components within neural networks (e.g. attention heads, MLP layers) that responsible for performing a specific subtask of model computation. Our method allows for quickly and automatically localizing computation, while only requiring to specify desiderata, or causal attributes of the target computation. As a proof of concept, we apply our method to automatically discover shared variable binding circuitry in LLaMA-13B (Touvron et al., 2023), which retrieves variable values for several arithmetic operations. Contributions. In this ongoing work, we: 1. Describe a methodology for localizing computation by enumerating desiderata and learning a binary mask by performing causal interventions (Section 3, Fig. 1). 2. Present initial results in applying this methodology to localize shared variable binding circuitry (Section 4, Fig. 2). Figure 1: Localizing computation with desiderata. The figure depicts training with a single (original, alternate, target) tuple within a desideratum. We learn a mask w that combines activations from an alternate sequence a into the computation of the model on the input of the original sequence o such that the output y moves towards the target t. 2 Background Circuit analysis. A deep neural network can be represented as a directed acyclic graph with specific nodes to accept inputs, generate outputs, and perform various operations to transform inputs into outputs. Circuit analysis involves localizing and understanding subgraphs within the computational graph of a model that are responsible for specific behaviors, and has had success in both language and vision models (Olah et al., 2020; Wang et al., 2022; Räukur et al., 2022; Chan et al., 2022). Activation Patching. As introduced in (Meng et al., 2022b), activation patching is a technique that uses causal intervention to identify which submodules’ activations matter for producing some model output. The process of activation patching involves running all the layers of a model until reaching a certain submodule with an original input, denoted as A, and a corrupted input, denoted as B. The activations of this specific submodule with input B are then patched into the corresponding activations of the same submodule with input A during the forward pass. Next, the patched activations are fed forward through the rest of the model. This enables one to assess the role of the specific submodule in generating an output, by quantifying how much this intervention shifts the model’s output from its original answer on A. Our approach generalizes activation patching and swaps activation from multiple alternative input sequence runs with known target outputs, instead of corrupted sequence runs, as described in (Meng et al., 2022b). Variable Binding. Variable binding is the process of associating a variable with a specific value, and is a fundamental concept in symbolic reasoning considered essential for solving tasks such as natural language understanding and reasoning (Marcus, 2001). However, it is still a mystery if and how Large Language Models (LLMs) implement this process. Please see Appendix B for additional related work. 3 Using Desiderata to Localize Computation We discover circuitry responsible for a specific task by enumerating properties of such a desired circuitry, and then learning a binary mask over the model’s parameters which accords with these properties (Fig. 1). We specify properties (or desiderata) in terms of causal interventions with a known target effects, and combine various interventions into a single objective function. We then learn a sparse mask on the targeted model components, such that applying causal interventions on the masked components alters model behavior to satisfy the objective function. Model components. As a first step, we specify our set of model components. Models can be represented at various levels of granularity. More granular components is more computational expensive, but allows for more specific localization of a model behavior. In Section 4, we decompose LLaMA-13B into a set of attention heads and MLPs, as opposed to more granular (e.g. splitting by Query, Key, and Value matrices) or less granular (e.g. grouping into layers) representations. Desiderata. Given a computational circuitry with specific functionality, we define a set of desiderata to enumerate the effects of various causal interventions on the circuitry. Each desideratum d corresponds to a set of n 3-tuple, each of which consists of an original sequence (o), an alternate sequence (a), and a target value (t). When the activation of the sought-after circuitry generated with o is replaced with the corresponding activation generated with a, the model should output t. The target value (t) is determined based on the nature of the intervention: it can remain equal to the output of o (indicating no change in the output is expected), be altered to match the output of a, or be set to a completely different third value. We identify and localize submodule with the desired functionality based on its adherence to the expected outcomes specified by the desiderata. Each 3-tuple (o,a,t)(o,a,t)( o , a , t ) contributes to a loss term which measures how well performing activation patching on a set of model components cisubscript\c_i\ citalic_i achieves t, ℒd⁢(ci)=1n⁢∑(o,a,t)∈dℒ⁢(ci,(o,a,t),y).subscriptℒsubscript1subscriptℒsubscriptL_d(\c_i\)= 1n _(o,a,t)∈ dL(\c_i\% ,(o,a,t),y).Litalic_d ( citalic_i ) = divide start_ARG 1 end_ARG start_ARG n end_ARG ∑( o , a , t ) ∈ d L ( citalic_i , ( o , a , t ) , y ) . (1) Note that some measure of proximity ℒLL between the induced model output y and the target t is needed. Furthermore, one can combine multiple desideratum into a single objective function, ℒD⁢(ci)=∑d∈Dℒd⁢(ci)subscriptℒsubscriptsubscriptsubscriptℒsubscriptL_D(\c_i\)= _d∈ DL_d(\c_i\)Litalic_D ( citalic_i ) = ∑d ∈ D Litalic_d ( citalic_i ). Desiderata for the specific case of identifying the value-copying circuitry involved in variable binding are presented in Fig. 2 and discussed in Section 4.1. Figure 2: Variable Binding Desiderata. Each desideratum is a set of original (o), alternate (a), and target (t) 3-tuples. In the Value Dependence desideratum, patching should change the output to the alternate’s output; in the Operation Invariance desideratum, patching should have no effect. Learning a Binary Mask. In order to find the set of model components cisubscript\c_i\ citalic_i that minimizes ℒD⁢(ci)subscriptℒsubscriptL_D(\c_i\)Litalic_D ( citalic_i ), we use a continuous relaxation of Equation 1. We define a mask over the model components by assigning a learnable weight wi∈[0,1]subscript01w_i∈[0,1]witalic_i ∈ [ 0 , 1 ], to each component cisubscriptc_icitalic_i. That is, wi=0subscript0w_i=0witalic_i = 0 corresponds to fully patching component cisubscriptc_icitalic_i with its value vasubscriptv_avitalic_a from the sequence a, wi=1subscript1w_i=1witalic_i = 1 corresponds to not patching cisubscriptc_icitalic_i, and 0<wi<10subscript10<w_i<10 < witalic_i < 1 corresponds to taking a convex combination of the cisubscriptc_icitalic_i’s activation value v and the value vasubscriptv_avitalic_a wi⋅v+(1−wi)⋅va.⋅subscript⋅1subscriptsubscriptw_i· v+(1-w_i)· v_a.witalic_i ⋅ v + ( 1 - witalic_i ) ⋅ vitalic_a . (2) Note that when we patch multiple components of the model, the value v of a later-layer component will be influenced by the patching of earlier layers before it itself is combined with vasubscriptv_avitalic_a as above. We optimize the continuous mask according to ℒDsubscriptℒL_DLitalic_D, which measures how well the patching intervention defined by the mask meets our desiderata. We use ℓ0.5subscriptℓ0.5 _0.5ℓ0.5 regularization with tunable strength λ over the mask entries to encourage patching only a sparse set of model components (Louizos et al., 2018). Throughout learning, we clamp values between 0 and 1. After training, we round weights to either 0 or 1 to form a binary mask. Empirically, we find that rounding the mask to become binary typically has little effect on its ability to satisfy the desiderata, and attribute this to the regularization for sparsity during training. 4 Variable Binding VD Acc. (+, -) OI Acc. (+, -) VD Acc. (×) OI Acc. (+, ×) # Patched Original Model 18% 91% 11% 93% 0 Incomplete Desiderata (VD) 93% 11% 82% 13% 10 Full Desiderata (VD & OI) 84% 82% 84% 91% 10 Table 1: Accuracy of patching experiments. Learning the patching mask according to an incomplete set of desiderata (only using the Value Dependence desideratum as presented in the second row) fails to localize our target computation (Operation Invariance accuracy suffers for all the tested operation indicated in parenthesis). Using both desiderata (Full Desiderata. third row) successfully causes Value Dependence behavior while maintaining Operation Invariance. Interestingly, the learnt patching mask achieves high accuracy also for operation that was not included in the training set. For all cases, accuracy is calculated on held-out test set. We apply our method (Section 3) to locate circuitry responsible for retrieving variable values when computing simple arithmetic expressions like those in Fig. 2. We use LLaMA-13B, a 40-layer, decoder-only transformer language model, trained on a diverse data set (Touvron et al., 2023). We hypothesize that there exist components of LLaMA-13B that, in order to complete sequences like those appearing in Fig. 2, copy the value previously assigned to the variable x into the final token’s residual stream. We further hypothesize that the x’s value is then combined with y’s value to compute the desired expression. We design desiderata to search specifically for this value-copying circuitry. Throughout, we only evaluate accuracy of models based on whether their prediction of the first digit of the answer value is correct; we ensure a diverse set of targets to avoid degenerate solutions. Code to replicate our results is available in a public repository.111https://anonymous.4open.science/r/anima_submission-D770/README.md 4.1 Variable Binding Desiderata We propose two desiderata to isolate this hypothesized value-copying circuitry:222We note that additional desideratum are possible. 1. Value Dependence (VD; Fig. 2, top). Patching our target circuitry with its activations from alternate sequences containing different x values should control which value is copied into the final residual stream. Accordingly, such patching should change the model’s output to match the output of the alternate sequences. 2. Operation Invariance (OI; Fig. 2, bottom). Since we are looking for circuitry shared across arithmetic operations, the specific operation being performed in the expression should not affect the behavior of the value-copying circuitry, as it should copy the same variable value regardless of the operation. We therefore form alternate sequences with a flipped operation (either addition or subtraction), with a target value equals to the output value of the corresponding original sequence. 4.2 Binary Mask Details We consider all MLPs and attention heads (1640 models’ components in total) and learn a binary mask as described in Section 3. We only perform patching to each component’s contribution to the final-token residual stream, as that is where we expect the value-copying circuitry to be active. We use two-digit variable values with addition and subtraction operations for defining both VD and OI sequences. We use the logit difference between the original and alternate answers as the proximity measure. For the VD task, we intend to maximize the logit difference, whereas, for the OI task, we aim to minimize it. We created a dataset comprising VD and OI sequences, with a total of 90/90909090/9090 / 90 train/test examples, such that the first digit of the expected answer is uniformly drawn from [1,9]19[1,9][ 1 , 9 ]. We use the Adam optimizer (Kingma & Ba, 2017) with a learning rate of 0.01, and alternate between taking gradient steps from the VD loss and the OI loss to save memory. For all experiments described below we use a sparsity regularization weight of λ=0.030.03λ=0.03λ = 0.03. In Appendix A we present additional results with different experimental settings such as varying λ and the numbers of patched attention heads. 5 Results According to our desiderata, we have identified a set of ten components, comprising nine attention heads and one MLP, that execute variable binding. We patch heads according to this ten-component mask and evaluate the model’s accuracy on a held-out set of VD and OI problems. On VD scenarios, accuracy measures how often the model outputs the answer from the alternative sequence. On OI scenarios, accuracy measures how often the model outputs the answer from the original sequence. We expect a mask that finds heads corresponding to the value copying subtask of variable binding to achieve high accuracy in both VD and OI. We observe that the models’ components identified by our method exhibit high accuracy in both tasks, as indicated in Table 1 (first and second columns). We further test these components on VD problems involving a multiplication operation instead of addition or subtraction, and on OI problems involving swapping between addition and multiplication. Surprisingly, we find that the accuracy remains high in these scenarios, despite not including multiplication during training (see Table 1, third and fourth columns). This indicates that these ten components indeed serve as the circuitry that copies variable values to the final residual stream (before the model operates on them); these components successfully cause the model’s output to change in the case that one of the bound values changes, but not in the case that the operation in the equation changes, even when testing an operation that was not include in training. We also find that including both desiderata is crucial for locating this circuitry; with only the VD desideratum, the identified heads successfully alter model behavior in the VD scenario, but also affect the model’s output in the OI scenario (see Table 1, second row). When both desiderata are included in the loss, the masked components are mostly attention heads in the middle of the model;333Heads 11.11, 12.0, 12.7, 15.11, 15.25, 17.17, 18.11, 18.18, 19.20, and MLP 27. whereas when only using the first desideratum, the masked components form a cluster of late-layer MLPs444MLPs 18, 27, 28, 29, 30, 31, 32, 33, 35, and 36.. A possible explanation for this is that using only the VD desideratum, the mask includes model components that write the computed final value of the expression to the residual stream, whereas adding the second desideratum encourages the mask to find the value-copying circuitry. 6 Conclusion In this paper, we proposed a new approach to localizing model components responsible for performing a specific task, using a set of causal behavior desiderata. Our method localize 10 components responsible for copying variable values in LLaMA-13B. We plan to compare it with existing localization methods (Meng et al., 2022a; Conmy et al., 2023) and to expend it to additional tasks. Acknowledgments XD and MN would like to thank Sam Marks and Oam Patel for their valuable discussions. NP and DB are supported by grants from Open Philanthropy. TRS is supported in part by the Zuckerman STEM Leadership Program and the Viterbi Fellowship. References Burns et al. (2022) Burns, C., Ye, H., Klein, D., and Steinhardt, J. Discovering latent knowledge in language models without supervision, 2022. Carroll et al. (2023) Carroll, M., Chan, A., Ashton, H., and Krueger, D. Characterizing manipulation from ai systems, 2023. Chan et al. (2022) Chan, L., Garriga-Alonso, A., Goldowsky-Dill, N., Greenblatt, R., Nitishinskaya, J., Radhakrishnan, A., Shlegeris, B., and Thomas, N. Causal Scrubbing: a method for rigorously testing interpretability hypotheses [Redwood Research], December 2022. URL https://w.alignmentforum.org/posts/JvZhhzycHu2Yd57RN/causal-scrubbing-a-method-for-rigorously-testing. Conmy et al. (2023) Conmy, A., Mavor-Parker, A. N., Lynch, A., Heimersheim, S., and Garriga-Alonso, A. Towards automated circuit discovery for mechanistic interpretability, 2023. Geiger et al. (2022) Geiger, A., Wu, Z., Lu, H., Rozner, J., Kreiss, E., Icard, T., Goodman, N., and Potts, C. Inducing causal structure for interpretable neural networks. In International Conference on Machine Learning, p. 7324–7338. PMLR, 2022. Geiger et al. (2023) Geiger, A., Wu, Z., Potts, C., Icard, T., and Goodman, N. D. Finding alignments between interpretable causal variables and distributed neural representations. arXiv preprint arXiv:2303.02536, 2023. Kingma & Ba (2017) Kingma, D. P. and Ba, J. Adam: A method for stochastic optimization, 2017. Li et al. (2022) Li, K., Hopkins, A. K., Bau, D., Viégas, F., Pfister, H., and Wattenberg, M. Emergent world representations: Exploring a sequence model trained on a synthetic task, 2022. URL https://arxiv.org/abs/2210.13382. Louizos et al. (2018) Louizos, C., Welling, M., and Kingma, D. P. Learning Sparse Neural Networks through $L_0$ Regularization, June 2018. URL http://arxiv.org/abs/1712.01312. arXiv:1712.01312 [cs, stat]. Marcus (2001) Marcus, G. F. Relations between Variables. In The Algebraic Mind: Integrating Connectionism and Cognitive Science. The MIT Press, 04 2001. ISBN 9780262279086. doi: 10.7551/mitpress/1187.003.0005. URL https://doi.org/10.7551/mitpress/1187.003.0005. Meng et al. (2022a) Meng, K., Bau, D., Andonian, A., and Belinkov, Y. Locating and editing factual associations in gpt. Advances in Neural Information Processing Systems, 35:17359–17372, 2022a. Meng et al. (2022b) Meng, K., Bau, D., Andonian, A., and Belinkov, Y. Locating and editing factual associations in GPT. Advances in Neural Information Processing Systems, 36, 2022b. Olah et al. (2020) Olah, C., Cammarata, N., Schubert, L., Goh, G., Petrov, M., and Carter, S. Zoom in: An introduction to circuits. Distill, 2020. doi: 10.23915/distill.00024.001. https://distill.pub/2020/circuits/zoom-in. Perez et al. (2022) Perez, E., Ringer, S., Lukošiūtė, K., Nguyen, K., Chen, E., Heiner, S., Pettit, C., Olsson, C., Kundu, S., Kadavath, S., Jones, A., Chen, A., Mann, B., Israel, B., Seethor, B., McKinnon, C., Olah, C., Yan, D., Amodei, D., Amodei, D., Drain, D., Li, D., Tran-Johnson, E., Khundadze, G., Kernion, J., Landis, J., Kerr, J., Mueller, J., Hyun, J., Landau, J., Ndousse, K., Goldberg, L., Lovitt, L., Lucas, M., Sellitto, M., Zhang, M., Kingsland, N., Elhage, N., Joseph, N., Mercado, N., DasSarma, N., Rausch, O., Larson, R., McCandlish, S., Johnston, S., Kravec, S., Showk, S. E., Lanham, T., Telleen-Lawton, T., Brown, T., Henighan, T., Hume, T., Bai, Y., Hatfield-Dodds, Z., Clark, J., Bowman, S. R., Askell, A., Grosse, R., Hernandez, D., Ganguli, D., Hubinger, E., Schiefer, N., and Kaplan, J. Discovering language model behaviors with model-written evaluations, 2022. Räukur et al. (2022) Räukur, T., Ho, A., Casper, S., and Hadfield-Menell, D. Toward transparent ai: A survey on interpreting the inner structures of deep neural networks. arXiv preprint arXiv:2207.13243, 2022. Touvron et al. (2023) Touvron, H., Lavril, T., Izacard, G., Martinet, X., Lachaux, M.-A., Lacroix, T., Rozière, B., Goyal, N., Hambro, E., Azhar, F., Rodriguez, A., Joulin, A., Grave, E., and Lample, G. Llama: Open and efficient foundation language models, 2023. URL https://arxiv.org/abs/2302.13971. Vig et al. (2020) Vig, J., Gehrmann, S., Belinkov, Y., Qian, S., Nevo, D., Singer, Y., and Shieber, S. Investigating gender bias in language models using causal mediation analysis. Advances in neural information processing systems, 33:12388–12401, 2020. Wang et al. (2022) Wang, K., Variengien, A., Conmy, A., Shlegeris, B., and Steinhardt, J. Interpretability in the wild: a circuit for indirect object identification in gpt-2 small, 2022. Wu et al. (2023) Wu, Z., Geiger, A., Potts, C., and Goodman, N. D. Interpretability at scale: Identifying causal mechanisms in alpaca, 2023. Appendix A Varying regularization strength We vary the regularization strength λ in order to learn masks with varying numbers of heads. We find that setting the regularization so that the masks learns approximately 10 model subcomponents is the approximate minimum number that can score highly for held-out Value Dependence accuracy and Operation Invariance accuracy, and so we use that setting for the main results of our paper. We also show further that removing the Operation Invariance desideratum causes the mask to score poorly on that criterion. Figure 3: Evaluating masks of various numbers of heads on held-out VD and OI problems. Each vertical pair of datapoints corresponds to a mask learned by a training run with a different value of λ, the sparsity regularization weight. With too few components patched, the model does not score well at Value Dependence. We interpret this as indicating that not enough of the value-copying heads have been patched. Figure 4: Transfer to accuracy on multiplication problems. This graph depicts the same masks as Fig. 3 (which were trained on sequences involving only addition and subtraction), but evaluated on all-multiplication Value Dependence problems, and addition-to-multiplication (and vice versa) Operation Invariance problems. Similarly to Fig. 3, VD accuracy is low with too few heads patched. Figure 5: Varying regularization strength with incomplete desiderata. This graph demonstrates learning a mask with only the Value Dependence desideratum. Again, each vertical pair of datapoints corresponds to a mask learned by a training run with a different value of λ, the sparsity regularization weight. Unlike when the mask is optimized according to both desiderata, these masks fail to achieve high accuracies on both Operation Invariance and Value dependence at the same time, as discussed in Section 4. Appendix B Related Work Previous work has developed automated approaches to localizing computation (Conmy et al., 2023; Geiger et al., 2022; Wu et al., 2023). Our work varies from Conmy et al. (2023) in learning a mask and considering a broader class of ablations (patches to change behavior, instead of just preserve). Our work shares features with recent work from Geiger et al. (2023) and Wu et al. (2023), but differs in attempting to isolate shared computation common in multiple input-output circuits as opposed to understanding full input-output circuits.