Paper deep dive
Towards Worst-Case Guarantees with Scale-Aware Interpretability
Lauren Greenspan, David Berman, Aryeh Brill, Ro Jefferson, Artemy Kolchinsky, Jennifer Lin, Andrew Mack, Anindita Maiti, Fernando E. Rosas, Alexander Stapleton, Lucas Teixeira, Dmitry Vaintrob
Abstract
Abstract:Neural networks organize information according to the hierarchical, multi-scale structure of natural data. Methods to interpret model internals should be similarly scale-aware, explicitly tracking how features compose across resolutions and guaranteeing bounds on the influence of fine-grained structure that is discarded as irrelevant noise. We posit that the renormalisation framework from physics can meet this need by offering technical tools that can overcome limitations of current methods. Moreover, relevant work from adjacent fields has now matured to a point where scattered research threads can be synthesized into practical, theory-informed tools. To combine these threads in an AI safety context, we propose a unifying research agenda -- \emph{scale-aware interpretability} -- to develop formal machinery and interpretability tools that have robustness and faithfulness properties supported by statistical physics.
Tags
Links
- Source: https://arxiv.org/abs/2602.05184
- Canonical: https://arxiv.org/abs/2602.05184
PDF not stored locally. Use the link above to view on the source site.
Intelligence
Status: succeeded | Model: google/gemini-3.1-flash-lite-preview | Prompt: intel-v1 | Confidence: 94%
Last extracted: 3/11/2026, 12:41:27 AM
Summary
The paper proposes a research agenda called 'scale-aware interpretability,' which leverages the renormalization framework from statistical physics to create robust and faithful interpretability tools for neural networks. By treating neural networks as systems with hierarchical, multi-scale structures, the authors argue that renormalization techniques—specifically coarse-graining and rescaling—can help identify relevant features, define model-natural scales, and provide guarantees regarding the influence of fine-grained noise.
Entities (5)
Relation Signals (3)
Renormalization → informs → Scale-Aware Interpretability
confidence 95% · we propose a unifying research agenda -- emph{scale-aware interpretability} -- to develop formal machinery and interpretability tools that have robustness and faithfulness properties supported by statistical physics.
Neural Networks → exhibits → Multi-scale Structure
confidence 92% · Neural networks organize information according to the hierarchical, multi-scale structure of natural data.
Scale-Aware Interpretability → aimstoimprove → AI Safety
confidence 90% · To combine these threads in an AI safety context, we propose a unifying research agenda -- emph{scale-aware interpretability}
Cypher Suggestions (2)
Map the relationship between AI systems and their properties. · confidence 90% · unvalidated
MATCH (s:System)-[r:EXHIBITS]->(p:Property) RETURN s.name, r.type, p.name
Find all interpretability tools mentioned in the context of renormalization. · confidence 85% · unvalidated
MATCH (t:Tool)-[:USED_IN]->(f:Framework {name: 'Renormalization'}) RETURN t.nameFull Text
159,125 characters extracted from source content.
Expand or collapse full text
Towards Worst-Case Guarantees with Scale-Aware Interpretability Lauren Greenspan 1 , David Berman 2 , Aryeh Brill 1 , Ro Jefferson 3 , Artemy Kolchinsky 4 , Jennifer Lin 1 , Andrew Mack 1 , Anindita Maiti 5 , Fernando E. Rosas 6,7,8 , Alexander Stapleton 2 , Lucas Teixeira 1 , and Dmitry Vaintrob 1 1 Principles of Intelligence, USA 2 Centre for Theoretical Physics, Queen Mary University of London, Mile End Road, London, E1 4NS, United Kingdom 5 Perimeter Institute for Theoretical Physics, Waterloo ON, Canada 3 Institute for Theoretical Physics, and Department of Information and Computing Sciences, Utrecht University, Princetonplein 5, 3584 C Utrecht, The Netherlands 4 Universitat Pompeu Fabra, Barcelona, Spain 6 Department of Informatics, University of Sussex, Brighton, United Kingdom 7 Department of Brain Sciences, Imperial College London, London, United Kingdom 8 Centre for Eudaimonia and human flourishing, University of Oxford, Oxford, United Kingdom Abstract Neural networks organize information according to the hierarchical, multi-scale structure of natural data. Methods to interpret model internals should be similarly scale-aware, explicitly tracking how features compose across resolutions and guaranteeing bounds on the influence of fine-grained structure that is discarded as irrelevant noise. We posit that the renormalisation framework from physics can meet this need by offering technical tools that can overcome limitations of current methods. Moreover, relevant work from adjacent fields has now matured to a point where scattered research threads can be synthesized into practical, theory-informed tools. To combine these threads in an AI safety context, we propose a unifying research agenda – scale-aware interpretability – to develop formal machinery and interpretability tools that have robustness and faithfulness properties supported by statistical physics. 1 Introduction By probing the internal structure of neural networks (NNs), the growing field of AI interpretability aims to improve our understanding of, trust in, and ability to audit AI systems. The field relies heavily on tools that open the black box and yield testable hypotheses about the mechanisms and patterns that govern AI behavior [1–3]. Importantly, these tools – though engineering artefacts – should be bolstered by theoretical and empirical support. For example, sparse auto-encoders (SAEs) grew out of a theoretical hypothesis about feature representations in NNs that was empirically verified in a toy setting [4] and then scaled into a practical analysis tool [5,6]. This trajectory – from theory and experiment to scalable instrumentation – offers an attractive scientific pipeline for ambitious interpretability, upon which we hope to iterate 1 . We argue that the renormalisation framework from statistical physics offers a useful language and set of design constraints for ambitious interpretability: it can help formalize (i) multiscale regularities learned from data and (i) the construction of abstractions that are stable under changes of resolution [7–9]. Our goal is not to claim that modern NNs are renormalisable in a strict field-theoretic sense, but that renormalisation suggests concrete questions and failure modes for interpretability methods. Moreover, we contend that the field is now ready for coalescence, and call for coordination across the physics, neuroscience, computer science, and AI safety communities to achieve the goals set out in this paper. Concretely, renormalisation formalizes three key aspects of NNs that current methods handle poorly: the importance of scale (granularity or resolution), relevance (which degrees of freedom matter at a particular scale), and coarse-graining 1 For more detail about this pipeline and the discovery of SAEs, see Appendix A) 1 arXiv:2602.05184v1 [hep-th] 5 Feb 2026 (how irrelevant degrees of freedom are systematically ignored). We sketch the necessary background in Section 2. In Section 3, we outline an interpretability framework that: 1. Finds natural scales in NNs for coarse-graining features during training and inference. 2. Extracts a principled, scale-dependent notion of feature relevance. 3. Supports guarantees—or at least diagnostics—for when fine-grained fluctuations can be ignored at a chosen level of abstraction, a property we refer to as separation of scales. Note that ‘renormalization’ is not a single prescriptive procedure, but a general framework for understanding how physical theories depend on scale [10–12]. It has evolved to include a variety of techniques capable of making reliable predictions across contexts, from quantum materials to collider physics. We aim to build a similarly adaptable framework in NNs, an approach we’re calling Scale-Aware Interpretability. We consider coarse-graining over tokens, weights, activations, or the data space equally within scope. (Surveyed in Section 4; see also Appendix B.) In Section 4, we sketch the current state of the field. Finally, in Section 5, we propose research objectives to guide interdisciplinary work going forward. The details of the renormalisation framework we depend on throughout this draft will be sparsely cited, and can be found in any introductory textbook on the topic (e.g., [12–14]). We aim to be as heuristic and falsifiable as possible while crafting our position in a N setting, without sticking to any particular paper or theoretical premise, and point the reader to Appendix B for supporting work and additional references. 2 Background 2.1 Renormalisation: A Primer In a physics context, renormalisation plays two main roles: to describe physical systems by effective theories at different scales of observation, and to decouple their degrees of freedom into a hierarchy of approximately local interactions by systematically identifying the relevant parameters as the scale is varied, so that fine-grained details are discarded while coarse-grained behavior is preserved. Starting with a theoretical description of a statistical system – for example, a Hamiltonian, or action - specifying interacting degrees of freedom (e.g., spins, fields, or N components 2 ) and couplings that determine their interaction strengths, a renormalisation step can be thought of as an operation of two parts: 1. Coarse-graining: Eliminate (by averaging, marginalizing, or integrating out) degrees of freedom above a chosen resolution up to a cutoff scale 3 to obtain an effective description in terms of coarser variables. This requires choosing a notion of resolution (e.g., momentum, distance, graph scale, or another problem-dependent ordering). 2.Rescaling: Re-express the fields and couplings so that the effective theoretical description – and the observables, or measurable quantities, it predicts – recover the correct macroscopic measurements 4 . One can view this as an operator acting on probability distributions (or parameterized models), mapping them to progressively coarser descriptions along a chosen notion of scale. Iterating this procedure results in a renormalisation group (RG) flow: the couplings evolve between the microscopic and macroscopic scales, leading to a chain of effective field theories 5 . The map between these theories is generally nonlinear; integrating out high-resolution modes generates effective couplings that are complicated functions of the original ones. In continuum RG, the flow is often summarized by aβ-function, e.g.β(g) = ∂g/∂ lnμ, wheregdenotes a coupling (or vector of couplings) andμa scale parameter. Linearizing the flow around a point gives a Jacobian on coupling space, with eigendirections that are relevant (grow 2 Because some of the literature refers to quantum or statistical field theory, we may refer to these components as ‘field-like’, despite differences between discrete N parameters, continuous fields, and the meaning of locality in each setting. 3 This is often referred to as the UV, or ultraviolet, cutoff, as this corresponds with a high-energy limit. 4 This is known as the IR, or infrared. 5 We will use ‘renormalisation’ and ‘RG’ interchangeably in a N context. This is a useful convention, and not a statement about (semi-)group-like structure within NNs. 2 under coarse-graining), irrelevant (shrink), or marginal (approximately stable). This classification is sharpest near fixed points, but local linearization can still be a useful heuristic away from them. In practice, there are two common viewpoints: Wilsonian or Polchinski RG [10,15], which operates in momentum- space, acting on functions and tracking couplings as they flow along a continuous parameter, and real-space RG which is often applied to discrete systems and explicitly removes degrees of freedom in a stepwise fashion (for example, via Kadanoff decimation on a lattice)[16]. Though a lot of existing work takes a Wilsonian view, we will not discount real-space methods, which may still prove useful for coarse-graining in certain ML settings (see, for example, Section B.2.4). We guide the interested reader to Appendix C for a glossary of key terms. 2.2 Multi-scale Structure in AI systems It is often said that NNs are grown rather than built; they exhibit behaviors more similar to systems studied by statistical physics rather than those that are human-engineered [7,17–19]. This, in turn, suggests that they are amenable to renormalisation techniques. We stress that thinking of neural networks as coarse-graining information in various ways (i.e., across layers) is not new; our goal is to weave together various threads to serve our AI safety agenda. For NNs, we find it useful to distinguish between implicit or explicit renormalisation. We define these below, and provide heuristic examples of N settings that appear to exhibit scale-dependent or coarse-graining behavior. 2.2.1 Implicit renormalisation During training and inference, NNs coarse-grain information about the data into ‘model-natural’ structures that reflect the inductive biases informed by a network’s architecture, data distribution, and training details. Implicit renormalisations schemes describe this process, with scale and coarse graining arising from the model’s own training dynamics. We do not expect there to be a single, canonical implicit scheme; different theoretical descriptions can track different scales and coarse-grainings (e.g., across depth, noise level, or context length, during training or inference). We consider diffusion models[20] to be a prime example of implicit renormalisation. Here, what we call ‘scale’ would be the diffusion time or noise level. As for coarse-graining, the forward noising process gradually adds noise to clean data, washing out fine details to model stable, coarse patterns (e.g., global composition, abstract shapes). A denoising step, implemented by a neural network, reverses this process, progressively resolving finer-scale features (e.g., textures, edges) by predicting an image at a slightly less noisy scale. Here, the architecture and training objective explicitly impose a scale hierarchy as an ordered sequence of noise levels, but the way information is represented and reused at each scale is learned implicitly during training. This suggests that relevant features of the data are those that remain predictive across many noise levels. Diffusion is one of the clearest empirical examples of renormalisation-like behavior in AI systems, but that does not mean the parallels with statistical physics are complete or exact. Instead of a flow of an equilibrium ensemble as its scale of interaction changes, diffusion is a trained, approximately invertible noising–denoising process on samples. We treat it as RG-like because noise level defines a scale hierarchy and organizes coarser versus finer descriptions of the data distribution. In a language model setting, there are many ways we think about coarse-graining. Candidate Scales include data granularity (e.g., context length) or eigenmodes of latent features (e.g., correlations of activations or the NTK, see Section B.1.1). Language models implicitly organize information across scales, tracking slowly varying context variables (setting, speaker identity, topic, emotional tone) that remain stable across long stretches of text, alongside finer-grained details like grammar, syntax, and token-level semantics (see, for example, references in Section B.2.1 for formal models of hierarchical input structure, including PCFGs and the random hierarchy model. That they learn such multi-scale features suggests that internal activations and attention patterns are sensitive to a text scale, and that a latent hierarchy of features emerges during training and adapts during inference. Unlike diffusion, language models are not explicitly trained to coarse-grain over the data scale; the latent hierarchy need not coincide with, for example, spectral scales defined by kernel eigenmodes. Heuristically, this points to multiple model-natural scales that could organize features in terms of their relevance to a given task. Relating or combining these is a core problem for the type of scale-aware interpretability we have in mind. [21] gives an analogy between transformers and the 1-dimensional Potts model, and [22] introduces the diffusion transformer, a robust architecture that puts phenomena related to renormalisation and lattice models in the forefront. 3 2.2.2 Explicit renormalisation There is much less relevant literature for explicit renormalisation, so we paint a picture here of what we have in mind. The goal of interpretability is to design post-hoc tools that are both faithful to the model (‘model-natural’) and human-interpretable. With explicit renormalisation, this goal is achieved by relating faithfulness with an implicit renormalisation hypothesis. Resulting tools use explicit scale parameters and coarse-graining rules to make sense of model internals (e.g., weights or activations), leading to a multi-scale model of interpretations that reflects their learned, multi-scale structure. For example, increasing the hidden dimension in SAEs often yields finer-grained semantic features—a coarse ‘scientist’ feature splits into Einstein, Turing, Goeppert-Mayer. This resembles RG flows where previously degenerate observables separate at finer resolution, suggesting that such tools are capable of probing data- model interactions at a tunable scale. This scale is likely more complicated than a single hyperparameter; understanding when such explicit decompositions respect, distort, or completely miss the model’s implicit scales is one of the problems that a renormalisation-guided perspective is meant to clarify. Matryoshka SAEs [23] were designed partially to combat this problem (see also Section B.2.6 on compressibility and multi-scale structure in SAEs). Other heuristics for explicit renormalisation methods include: clustering or pruning neurons, truncating spectral decompositions (e.g.., keeping only the top-k kernel eigenmodes), or applying information-theoretic compression schemes. Each of these methods defines its own notion of scale (e.g., number of clusters or sparsity, eigenvalue cutoff) and relevance (which features are treated as important for explaining behavior). 3 Renormalisation for Interpretability We imagine modeling NNs and the data they represent via a hierarchical decomposition of components (e.g., features) that depends on a coarse-graining resolution, or scale. This is an intentionally broad operationalization, meant to keep our focus on the interpretability goal rather than any particular renormalisation scheme. For each RG-like scheme developed in an N context, there are three important questions to address: 1.How do we coarse grain? What defines an RG-like coarse-graining scheme for NNs? Which model-natural notion of scale does it track, what is the metric for relevance, and in which empirical settings is that scheme robust? 2.What are the inputs to, and properties of, an effective description that comes out of a particular coarse-graining scheme? Given a choice of cutoff, what are the relevant features for a given computation, and which can really be treated as irrelevant-small fluctuations? How far is an effective description from a critical point? 3.Why does this matter for interpretability? What experimental signals correspond with safety-relevant observables? Under what conditions does separation of scales hold? To what extent do universality classes meaningfully constrain or describe a model’s behavior? We develop each of these points in what follows. The current literature does not split cleanly into the categories we define here, so we leave supporting references in B. 3.1 How? Defining RG-like Coarse-Grainings in Neural Networks In physics, there is a clear notion of scale (distance, energy, or momentum) attached to a physical hierarchy of particle interactions. This guides the development of an RG scheme: spins on a spatial lattice are grouped into blocks in real space while momentum shell methods like Wilsonian RG or Polchinski RG are more natural for a particle system in space-time. Each scheme depends on a cutoff scale which determines how much to coarse grain at each point along an RG flow. In lattice models, there is both a maximum distance (the system size) and a minimum distance (the lattice spacing), which effectively grows with the number of coarse-graining steps. Other RG schemes systematically remove high-momentum modes for particle hierarchies organized according to this scale. In either setting, fine-grained degrees of freedom beyond the cutoff scale are ignored because they don’t contribute to phenomena of interest, marking a physical limit of the effective theoretical description 6 . 6 A cutoff can also mark an epistemic limit beyond which a new, more complete theory is needed to describe physics beyond that scale. The standard model, for example, is widely regarded as an effective truncation of a theory of quantum gravity. 4 Keeping with the distinction we made earlier, we define two types of cutoffs for N renormalisation. Implicit cutoffs arise during training and reflect finite resources of the model. Given a particular dataset, architecture, and optimization procedure, the trained network only represents a restricted subset of possible functions, resolving structure in the data up to some effective resolution. We conjecture that this structure reflects an implicit feature hierarchy along some model-natural notion of scale 7 . There are empirical indications of this, including spectral gaps in kernel eigenvalues and low-variance directions in activation space. While these patterns likely reflect partial aspects of a true data feature hierarchy, they indicate that, for a given task, many directions are treated by the model as irrelevant small fluctuations (see the kernel-regime taxonomy in Appendix B.1). In contrast, explicit cutoffs are imposed on the system by interpretability or analysis tools. PCA, SAEs, clustering methods, or graph-based coarse-graining procedures all introduce additional scales, defining relevance according to, for example, sparsity, dictionary size, variance explained thresholds, or graph resolution. These choices determine the finest scale at which we attempt to resolve features in an explicit decomposition. Beyond that scale, additional components may be too fine-grained for the description we want (as in the Ising model), unstable under small changes, or not meaningfully interpretable (as in the standard model). In a renormalisation-guided framework, a good explicit cutoff should respect the implicit cutoff of the trained model; the resolution of our interpretation should line up with the finest scales at which the network has actually learned structured features, rather than artificially abstract finer details (though it may cut off fine details for a coarser description). We aim for future renormalisation-based tools to provide a practical and safety relevant cutoff on interpretability resolution (see Section B.2). In addition to a cutoff, any RG-like construction rests on an implicit notion of locality, which determines the range of interactions and gives a measure for closeness between degrees of freedom. In physics, most interactions are short-ranged, or local, in space-time, though non-local interactions do arise, particularly in effective field theories. Importantly, renormalisation is constrained such that locality is approximately preserved 8 . In NNs, ‘nearby’ depends on context; in input space it could be neighboring pixels or tokens, in data space it could be according to some graph structure of natural language (see Section B.2), and in representation space it could be a kernel-induced distance (see Section B.1). A good explicit renormalisation scheme should probe and preserve an implicit notion of locality (see Sections B.2.1 and B.2.2 for disucssions of locality in the input space and data space, respectively). If our understanding of scale and locality is badly misaligned with how the model actually couples its degrees of freedom, coarse-graining could produce cryptic or highly nonlocal effective descriptions, undermining the interpretability guarantees we seek. Finally, an RG-like scheme requires a direction along which degrees of freedom can be collectively summarized. These are not as obvious in NNs as in nature 9 , but just as power laws of operators do in physics, these can be guided by candidate measures of local interactions that exhibit power-law behavior (see Section B.2.2). We stress that individual hyperparameters (e.g., width, learning rate, dataset size, initialization variance, SAE dimension) are not what we mean here, though these help shape the inductive biases that give rise to a model’s implicit hierarchy. These quantities are better thought of as control parameters that select broad regimes or basins of attraction in model space (e.g., NTK-like v. feature learning regimes, see Sections B.1.1, B.1.2, and B.1.3), rather than directions for coarse-graining. We see model-natural scales as something to be discovered and justified as the renormalisation framework is developed. A single, model-natural coarse-graining scale, if it exists, is likely to be a combination of the various scales we can currently track, such as ordering kernel eigenmodes by eigenvalue, evolving diffusion time or noise level, tracking text granularity, or organizing representations across layers by some summary statistic (see Sections B.2.4 for an information-theoretic view of scale, B.2.5 for a discussion about scale and relevance framed in terms of statistical inference, and B.2.6 for a discussion on compressibility and interpretability). Different choices will typically induce different flows, though some may converge to similar effective descriptions. In physical systems, many different notions of scale and RG flow lead to the same renormalised theory; in some sense, the space of RG flows with the same limit is open in a suitable topology, and quite forgiving in practice 10 . 7 Some candidate examples of scales in specific settings were given in section 2.2. 8 Coarse-graining may generate longer range interactions, but under a well-defined RG scheme these are suppressed at large scales. 9 To first order, the local approximation of theβ-function yields a linear rescaling of the units used to make measurements, ensuring dimensional consistency. In NNs, where everything is dimensionless, analogous rescalings must be based on other – currently unknown – principles. 10 This is particularly visible in the critical Ising model and its Kramers–Wannier dual: defining a Gaussian convolution RG in the fermionic representation of the critical Ising model yields a natural, local smoothing procedure. However, mapping this RG explicitly back to spin variables produces a well-defined but inherently nonlocal RG scheme. Thus, while the duality preserves the universality and consistency of this RG flow, its interpretation and locality properties differ dramatically between the dual descriptions. 5 3.2 What? Effective Degrees of Freedom and Relevance Once specified, an RG scheme is a transformation from one effective description to another – it takes in a messy, high-dimensional model and reduces it to a smaller set of effective features whose behavior is enough to predict the phenomena we care about, up to a specified error. A useful way of thinking about this is through spectra. In a standard Wilsonian picture, as we change scale, the spectrum of operators reorganizes so that a small set of low-dimension operators carries most of the weight for long-distance observables, while the rest are effectively decoupled. In an analogous picture for neural networks, a handful of modes in a kernel spectrum would acquire large eigenvalues as we change scale or training regime, while the remaining modes collapse into a low-variance tail. For a given observable and error budget, only the top modes are genuinely relevant, and the tail can be safely neglected. More generally, a renormalisation-like coarse-graining scheme should take in: •A set of features. In interpretability research, the term ‘feature’ is used in two related ways: as a ground-truth property of a data distribution that a model can learn, and of the model component (e.g., activation) that represents that property. Similarly, we define this term loosely to mean any model-natural effective component that captures meaningful structure in data and representations, i.e., a derived variable computed from model internals (model inputs, weights, activations, or gradients) that can be tracked under coarse-graining and related to observables. These can be functions of a single-input (e.g., activations, circuits) or a pair of inputs (i.e., kernel eigenmodes). •Depending on the setting – whether we are considering renormalisation in data space, input space, or activation space – and type of RG scheme (implicit or explicit), features can be built from inputs, weights, activations, or gradients. Examples include directions in activation space, eigenmodes of a kernel or covariance operator, and components of a hierarchical data model (see Appendix B for examples). •A model-natural notion of scale and cutoff that separates ‘fine’ from ‘coarse’ features. In an implicit RG scheme, the cutoff is an input to the theoretical description that can be tuned to match the empirical structure that training and inference dynamics have imprinted on representations and kernels (e.g., an observed eigenvalue cliff). In explicit RG, it is an input to a post-hoc tool used to compress that structure (e.g., a constraint on the number of features per neuron). •One or more long-range observables and an associated error budget. These are measurable quantities or quantifiable behaviors that should remain approximately invariant under coarse-graining, up to an acceptable tolerance for error. These could be related to performance on a specific task (e.g., next token prediction), probes of safety relevant behavior, or structural properties of internal activations (e.g., response functions [24]) that we wish to preserve. As in physics, we expect many observables to be global or structural quantities guided by empirical scaling laws [25]. Given these, the map should return: • A reduced set of effective features. These constitute a valid description up to the cutoff. While this should be smaller than the original set of features, it may still be high-dimensional. • A relevance ordering on those features, determined by their contribution to large-scale observables, depending on the cutoff. For implicit RG, this could result in some spectral separation for a downstream task. In explicit RG, this is heavily reliant on the architecture and optimization metric of the tool at hand, but should also reflect the implicit structure. •Bounds on the influence of neglected components. For worst-case guarantees, a minimally rigorous argument would be of the form ‘conditional on an effective coarse description, components above the cutoff have at most small, quantitatively bounded influence on the observables’. 11 11 This is a probabilistic statement about the joint distribution of degrees of freedom across scales that we think can be reasonably achieved. Ideally, we’d like a stronger separation of scales argument that strictly bounds the impact of all components above a relevance threshold. This has implications on the kinds of worst-case guarantees we can make, which will be discussed in the next section. 6 Crucially, relevance is not an absolute notion. It is defined relative to the scale at which we observe or intervene on the system, as this is where we measure observables. In physics, these are often defined by special points 12 in the – sometimes high-dimensional – space of couplings that control the dynamics of nearby effective theories. Many microscopically distinct theories can flow toward the same macroscopic fixed points; they share the same long-distance observables, a property known as universality 13 . If we did not set a long-range scale at which to stop coarse-graining, continued iteration on NNs would lead to meaningless noise similar to a trivial fixed point. Instead, we tie the endpoint of the RG flow to the scale at which we measure things like task performance, elicited behaviors, or training dynamics. We can view effective degrees of freedom as local coordinates adapted to the flow near a given scale. Around each point in theory space, some directions grow (are relevant) under coarse-graining, while others shrink (are irrelevant) or stay the same (are marginal). We expect a similar picture for NNs. A feature can be relevant at one scale and irrelevant at another, or relevant for one class of behaviors and not for another. During coarse-graining, the effective description can change qualitatively: different combinations of features become important for the observables we track, while others collapse into noise. We refer to these points where features become (ir)relevant as crossovers 14 (see Section B.1.4 for examples from the current literature involving phase transitions and scaling laws). 3.3 Why? What We Want out of a renormalisation Framework for Interpretability The power of renormalisation comes from its ability to link microscopic couplings, which depend on somewhat arbitrary choices of cutoff and renormalisation scheme, to macroscopic observables– quantities that can be measured and tracked across scales. A correct application of the renormalisation toolkit to AI interpretability could serve two goals: as a microscope with a dial to hierarchically decouple different scales, and as a diagnostic that reveals interesting large-scale phenomena such as phase transitions and the emergence of new structures (see, for example, Section B.1.5). This would let us faithfully summarize microscopic model details without having to track every parameter individually, allowing for both ambitious interpretability that doesn’t depend on the definition of an arbitrary atomic unit. 3.3.1 Separation of Scales and Hierarchical Emergent Structure Not every coarse-graining scheme has the properties we want. Many formally valid coarse-graining flows produce effective theories that are non-local in undesirable ways or else lose information about the observables we care about 15 . As with any tool or method, conditions, caveats, and regimes of validity are all part of the evaluation process. For AI safety, a key property we’d like to impose on a good renormalisation scheme is separation of scales: microscopic details (e.g., individual parameters or finer features) can vary within some range without materially changing the coarse behavior of a chosen macroscopic description (e.g., a downstream behavior or set of aggregate features). Simply put, we aim to produce interpretability tools with renormalisation-theoretic guarantees, which cleanly separate relevant from irrelevant variables in a context of interest, with bounded error. Worries about bleed-in from other scales and contexts are frequently observed in the AI safety literature. These include: • The use of steganography in chain-of-thought, where information is hidden in apparently random tokens[26], • Bayesian assumptions about independence or Gaussian noise terms that collapse under distributional shift [27] • Causal feature graphs that track statistically relevant circuits in some contexts but are sensitive to slight changes in the input distribution[28]. This is not merely a reframing of existing desiderata. Current tools optimize for reconstruction accuracy or human interpretability, but are incapable of making rigorous claims about causal necessity or insufficiency, let alone make guarantees about what they miss. A useful separation of scales argument would take the form: ‘conditional on an 12 These can be fixed points or critical points that are stable, unstable, or saddle-like of the RG flow. At a fixed point, theβ-function vanishes, and continued coarse-graining leaves the effective theory unchanged. 13 We will discuss the implications for universality of NNs in the next section. 14 Though crossovers can be critical points marking phase transitions, this is not a requirement. They can simply be a point between two effective theories in the same phase, with different feature partitions. 15 For example, a naive spin decimation RG scheme for the two-dimensional Ising model yields an exact but highly non-local effective theory with long-range correlations. One can truncate these by hand to recover locality, but then the coarse-graining no longer preserves observables exactly. 7 effective description, (potentially catastrophic, low probability) perturbations confined to the irrelevant subspace cannot change observableXmore thanε. It is our view that a lot of the work in section 4 is closer to being turned into a theory-driven, SOTA-scale interpretability tool than one might expect. This property is crucial for the last two RG outputs we discussed in the last section: i) the identification of effective features that contribute most to a safety-relevant observable and i) the empirical or theoretical bounds on how much neglected, fine-grained fluctuations can impact that observable. We refer to this last point as hierarchical conditional independence, a Markov-like property of an RG flow that preserves the hierarchy of effective theories along some scale. This means that the effective theories, and the interactions they describe, can really be treated as independent since coarse variables at a given scale act as sufficient summary statistics for finer variables at the next, more microscopic scale 16 . Physicists often impose this requirement by ensuring that RG schemes preserve locality across scales, which puts quantitative bounds on short-range fluctuations generating anomalous large-scale behavior. Our goal is to make similarly rigorous statements for NNs. In this setting, only coarse-graining procedures that robustly preserve a sensible hierarchy of feature interactions, rather than scrambling them, will be capable of making useful guarantees for AI safety. Coarse variables might be effective features at some resolution (e.g., top directions of a kernel or Fisher information matrix, or top activating SAE feature). Conditioned on these, hierarchical conditional independence means that, for a given observable, fine-grained activity in a local region – such as low-variance directions or SAE features that have undergone a degree of feature splitting – can be safely discared since they don’t couple strongly across scales 17 . The effective decomposition of a complex system into independent hierarchical modules with this separation of scales property is similar to causal mechanistic decomposition, which factorizes a system’s behavior in a way that preserves causal counterfactuals at multiple levels of abstraction [29]. A renormalisation framework built on this property – which would add a scale, a flow between abstractions, and quantitative bounds on when we expect it to hold – would represent a significant leap in provable interpretability research. Even an RG scheme for which scale separation is weakly or conjecturally satisfied could be extremely useful in bounding the extent to which potentially dangerous large-scale behaviors can hide in small-scale mechanistic interactions 18 . For example, if we can show that, for a particular model and observable, behavior is controlled by a tractable set of effective features, and that the influence of remaining features is bounded, we can show that any adversarial fluctuations known to exist in the irrelevant subspace cannot impact the observables by a known tolerance. We stress that this separation of scales property is non-trivial. While we may formally introduce any coarse-graining scheme on any statistical system, only some can be compressed in a way that shields – in a way that holds across scales – the long-range effect of a small set of relevant parameters from short-range details. In physics, such systems are said to be renormalizable. A proof of separation of scales is possible in many high energy or idealized condensed matter theories. In other systems, it can nevertheless be shown to hold empirically (for example, by examining scaling laws)[31]. Which of RG’s many theories and techniques – including separation of scales – can be imported directly, and which should be adapted for NNs, is still an open question (see B). We expect coarse-graining schemes with a rigorous separation of scales property to hold in toy models and under restricted conditions, and aim to use these to build up our understanding toward worst-case guarantees that hold in realistic settings. Empirically, we can check that the bounded influence of perturbations confined to irrelevant directions (e.g., spectral tails) move our observables by at most some small tolerance. An evaluation criteria for this desiderata is faithfulness to these observables (and corresponding macroscopic description): does the scheme make correct predictions about how these transform under coarse-graining, and does it preserve them within the stated error budget? We may also ask if our RG scheme preserves locality, such that a coarse-graining does not generate uncontrolled long-range interactions. This is a more ambiguous desideratum for NNs than physics, but we can check whether coarse-grained features retain a semblance of a ground-truth feature hierarchy in toy settings 19 . 16 This property is robustly held in momentum-space RG, but is perhaps easiest to see in real-space RG on a lattice, where block spins are literal aggregates of component spins whose effects are effectively screened off. 17 In other words, they are suppressed by some N analog of locality and correlation length so that they only change coarse observables by a bounded, small amount. 18 By ‘dangerous’, we mean the class of behaviors that significantly violate specified safety desiderata, such as deception. More broadly, we could potentially place bounds on any anomalous or low-probability behavior [30]. 19 We could also empirically track the correlation length. 8 3.3.2 Universality: Bounding our expectations in NNs Separation of scales is a core safety-relevant property for which we want to evaluate a potential renormalisation-like scheme. However, NNs are significantly different from physical systems, and we do not expect all aspects of a physics’ renormalisation – like universality and criticality – to hold in a generic N setting. For one, NNs have extremely large data and representation spaces; high-dimensional feature substances may be needed even at macroscopic scales. Nevertheless, any method to significantly reduce the number of dimensions in the data or representation would be practically valuable for interpretability, and it is worth thinking through what extra machinery a renormalisation framework could bring with it. In physics, renormalisation is often tied to a discussion of universality: many short-range theories flow to the same fixed point under an RG flow. These are typically characterized by a small number of relevant parameters and exhibit the same large-scale behavior, forming families of microscopically different descriptions known as universality classes. Interpretations of NNs are similarly many-to-one, leading to questions about what universality could mean in this setting and how it relates to a renormalisation framework we build (see Section B.1.2). Can we make statements about whole classes of AI behavior based on a finite number of relevant directions? Under what conditions (if any) do these depend on a small number of features? Any stable fixed point defines a universality class in the sense that many different microscopic models can share the long- distance behavior defined there. However, critical points—fixed points controlling continuous phase transitions—are special. There the correlation length becomes very large, and the scaling dimensions of relevant operators show up as critical exponents governing non-analytic, power-law behavior of observables over many scales. This makes universality more dramatic very different materials or lattice models exhibit the same scaling laws near criticality. Generic fixed points still organize phases and irrelevant operators, but the associated scaling structure is typically less striking, with finite correlation lengths and more analytic dependence on parameters. In addition to describing interesting physical phenomena, the theory at a critical point often becomes more tractable. In NNs, critical phenomena may manifest as sharp crossovers in which features dominate as the scale is varied, signaling qualitative changes in how the model represents information, similar to a phase transition (see Section B.1.4). Identifying such regimes can help identify regions where model behavior is particularly fragile, or where microscopic changes can have amplified effects on safety-relevant observables. Smooth crossovers – though harder to diagnose – may be equally important for understanding N behavior. If it is shown to hold, a universality-like property could guarantee that effective features and their relevance order should not change wildly under small changes in data, initialization, or architecture, for a fixed model family. These could be detected by asking: Do effective descriptions and relevance orderings change smoothly along a proposed scale direction, or do small perturbations lead to radical changes? However, there is no fundamental reason a N theory space must have this fixed point structure, and it is unclear i) how close a given representation must be to a fixed point to place it within a certain universality class and i) how stable these are to changes in hyperparameters (or continued training). If NNs do exhibit genuine universality in specific settings (e.g., the Gaussian Process of infinite-width limit, or specific scaling regimes, see Section B.1), these could offer theoretically tractable regimes that can be interpolated or expanded into more realistic settings. 4 A Lens for Viewing Existing Work Much of the existing literature provides examples of implicit renormalisation, though our position is that there is high potential for application to explicit renormalisation. We organize work according to how effective degrees of freedom (features) are defined: i) as kernel components and i) directly in the dataspace. While many of the cases we consider apply in an idealized or toy model of data or inference, we strive for the more general application of these ideas. We note that many of the works considered here use different terminology, and aim to be explicit about this. This section surveys work examining renormalisation-like phenomena found in kernel structure and the data-space. We emphasize that this is a selective overview of directions we find promising, not a comprehensive review of the research landscape. 9 4.1 Kernel renormalisation We broadly view a kernel as a model-natural covariance operator on functions over inputs which organizes features as eigenfunctions (kernel modes). Kernels impart a notion of similarity to features in an input-dependent way, and offer a window into N structure – the function spaces kernels have access to and the dynamics by which they evolve. A review of kernel approaches can be found in Appendix B.1. Two key kernels representing complementary perspectives dominate the literature: the Neural Network Gaussian Process (NNGP) kernel, which captures the prior distribution over functions at initialization, and the Neural Tangent Kernel (NTK), which captures how that function evolves during gradient descent. In certain limits (infinite width with appropriate scaling), these kernels fully determine network behavior. While these kernels have been studied extensively in this limit (also known as the infinite-width or ‘lazy learning’ limit), they have since been studied both empirically and theoretically in more general and expressive regimes 20 . Existing literature has grown rapidly, and differences in terminology and framing abound, with papers often using incompatible or context-specific notation and assumptions 21 . In an effort to standardize work going forward in a way that is useful for AI safety, we paint the following picture: kernel eigenfunctions correspond to features for the NNGP and feature tangent directions for the NTK. Coarse-graining then ranks features by relevance (to first order) according to their associated eigenvalues (prior variance for the NNGP and rate at which features are learned for the NTK). Choosing a UV cutoff is often like spectral truncation – considering modes below a scale that are relevant, for example, for a certain downstream task is like saying that nearby inputs (in the kernel-induced geometry) above this scale become indistinguishable at coarser resolution (see Sections B.1.1 and B.1.2 for some examples). Even if it does not mention ‘renormalisation’ specifically, work from the physics community relating N behavior with field theory has the potential to shed light on natural scales and notions of similarity and relevance in N coarse-graining schemes. Future work should stress-test the limits of existing kernels and their underlying assumptions to understand when certain theoretical regimes (defined by initialization and hyperparameter choices) break down, and how to extend beyond them. Translating largely theoretical insights into practical interpretability tools would also benefit from greater understanding of the relationship between kernel features and SAE features. Could kernel features provide a renormalisation-based fix for SAE pathologies by capturing structure SAEs cannot? We also aim to connect work being done from the physics community with empirical studies of kernel eigenmodes within AI safety, like influence functions[32], providing a theoretical bridge. 4.2 Data-space renormalisation Much of the renormalisation idea hinges on how a N represents the rich structure of datasets, from coarse regularities (long-range correlations) to fine idiosyncrasies (short-range). Data-space approaches consider: 1. How structure should inform our understanding of model-natural scales and coarse-graining schemes and the closure they afford at each scale (e.g., statistical, informational, causal, computational). 2. How renormalisation-like schemes map onto the way data features are compressed. 3.How compressed representations can be reliably interpreted up to a scale of abstraction resonant with the data structure. Although they are less developed than kernel methods, data-space approaches may have more direct engineering applications. We expect synthetic data models with a known hierarchical ground truth to be particularly important in formulating a framework for both implicit and explicit renormalisation; capable of testing theoretical hypotheses and benchmarking new tools (like, e.g., Matryoshka SAEs [23]). Future work could also connect kernel renormalisation to data-space renormalisation, for example RG-based approaches like real-space mutual information (RSMI) [33]) and causal states inference [34, 35]. A review of the relevant literature can be found in Appendix B.2. 20 In the lazy learning regime, kernel regression with the frozen NTK is mathematically equivalent to linear regression on a fixed set of feature functions corresponding to the components of the Jacobian at initialization. The complement to this is sometimes called the ‘feature learning’ regime. 21 For example, ‘field’ in a usual field theoretic sense corresponds with a function on a continuous domain, but in lattice models physicists sometimes define a field on discrete sites. Similarly, AI researchers tend to use ‘feature’ haphazardly. These are two examples of concepts that can be murky for anyone without inside expertise. 10 5 A Call to Action: Towards Interpretability with Worst-Case Guarantees As a framework, renormalisation is a cornerstone in the explanatory powerhouse of modern theoretical physics, capable of capturing the essential empirical aspects of a system and how its theoretical description fits in with the space of possible theories. In spite of mounting research (see section 4), the fields of physics and AI safety have yet to work together on the same RG-supported, AI safety-driven goal. We think this is a missed opportunity to put one of physics’ most flexible foundational frameworks to work on one of today’s most pressing problems. Having motivated the relevance of renormalisation to AI interpretability, we now present a call to action. In Appendix A, we proposed a path to impact by examining a recent case study. In this section, we advocate for a pair of research artifacts around which researchers from across disciplines can focus their efforts. These are analogous to i) toy models of superposition (TMS) and i) sparse autoencoders (SAEs), but come with additional properties like scale separation. We sketch these below, and leave their development for future work. 5.1 Artifact 1: Toy Model of renormalisation (TMR) We aim to develop a model organism of renormalisation analogous to what TMS was for sparsity. Rather than a model for how features interfere, a TMR should generate a hypothesis for how they compose, coarsen, and depend on scale. This artifact maps onto implicit renormalisation, which focuses on the development of robust theoretical descriptions for empirical phenomena. A moonshot result would develop a renormalisation-inspired hypothesis capable of describing training and inference in a way that provably bounds the influence of fine-grained components on safety-relevant behaviors. Along the way, we will piece together rigorous clues to explain phenomena in individual settings (e.g., across tokens, in information space, in various kernels, or in the loss landscape) through the lens of scale separation, operator relevance, or effective degrees of freedom. It is likely that this will involve designing synthetic data distributions with ground truth hierarchical structure and studying the learned representations of models trained on them; such settings can also act as empirical validation methods for desirable properties in interpretability tools (see Section B.2 for more details). We advocate for scalable insights that are well-contextualized in terms of their assumptions and regimes of validity. 5.2 Artifact 2: General renormalisation Tool (GRT) Mapping onto explicit renormalisation, our second goal is to build a general-purpose tool – analogous to SAEs – that extracts interpretable, multiscale structure from real models. Following the logic of Appendix A, GRT should reflect how TMR models the data, weights, or other associated statistical artifact. For example, it could apply lattice RG to activation graphs (see Section B.2.6), construct coarse-grained feature maps from causal states (see Section B.1.5), or a flow defined by Polchinski-style equations[15]. The computational analogy is not “find atomic parts and interpret them” but “find effective descriptions that represent meaningful, interpretable structures as a function of a model-natural scale” (for a directional example on Bayesian renormalisation, see Section B.2.5). In this case, the holy grail, aligned with a sweeping renormalisation hypothesis, would be a completely novel tool capable of outperforming SAEs on all desiderata like canonicity and faithfulness. Instead of coming up with this out of the gate, we imagine an iterative process where theory and practice approach each other incrementally, starting by formalizing renormalisation-like behavior in existing tools. 5.3 Evaluation Protocol and Limitations We acknowledge significant uncertainty about which RG-inspired methods will transfer. First, identifying model-natural scales is largely empirical. Second, renormalisation workhorses like criticality may not hold in NNs, or may hold only in restricted regimes. It is also possible that the computational cost of GRT may limit their scalability to SOTA models, and worst-case guarantees that hold in idealized cases may weaken in messier, real-world settings. We expect that similarities between physical and AI systems will lead to a rich analogy rather than a strict correspondence, underscoring the need for care in selecting research directions with the potential for impact in AI safety and effectively communicating this safety relevance, as well as any physics jargon. 11 In evaluating an RG scheme, we consider the following failure modes. If discarding small-scale components significantly degrades performance or otherwise impacts long-range observables, it may indicate (see section 3 for more details): • The scheme assumes a wrong scaling direction that does not reflect the model-natural hierarchy. •The chosen cutoff is not tuned to the resolution at which model components (e.g., effective features) are observed. • Our notion of relevance is misaligned with the level of detail a computation needs. • We’ve encountered a dangerously irrelevant feature that shrinks under coarse-graining but whose fluctuations has a definitive impact on large-scale behavior 22 [11]. This highlights the importance of choosing observables that properly constraint the feature class. A scheme that fails along one of these axes signals a mismatch between our assumptions and the model’s true structure. Identifying which failure mode applies in an empirical setting can guide theoretical refinements to the coarse-graining procedure or, potentially, reveal regimes where worst-case guarantees are inherently out of reach. We think that even qualified success of the proposed agenda has the potential to build a varied, adaptive framework for AI safety. 6 Discussion This paper argues that renormalisation offers a productive lens for ambitious interpretability, and that the field has matured to the point where scattered research threads can be synthesized into practical tools capable of making worst-case guarantees. Several developments since we began this work suggest growing momentum in this direction as a result of intentional coordination efforts. For kernel renormalisation, these include less idealized pictures of renormalisation and universality in NNs [36] and a hypothesis for feature identification using the eNTK [37]. For data-space renormalisation, [38] presents a code repository for generating synthetic datasets that encode hierarchical data structure and [39] presents a candidate model-natural scale based on tokenization. There has also been work, to appear soon, developing probabilistic SAEs that leverage the hierarchical structure of DAGs [40]. We are not alone in thinking that building better tools goes hand-in-hand with better theory development, or that insights from physics can make progress in AI safety. We hope that the work discussed here will provide productive points of contact with related efforts to close the theory-practice gap within the field. These include the mechanistic modeling of belief states in a transformer residual stream, using insight from computational neuroscience and theoretical machinery from computational mechanics [34], the study of how learning dynamics and generalization depend on data, inspired by singular learning theory [41], and work aiming to formally explain neural network behavior to detect potentially harmful anomalous or low-probability behaviors, rather than focus on the average-case scenario [42]. While renormalisation-inspired methods are a plausible remedy to the problems laid out in this draft, they remain scattered across several disciplines and under-formalised in an AI safety context. The work surveyed here spans computer science, physics, biology, and complex systems science – fields with different vocabularies, tools, and publication venues. We do not advocate for importing physics wholesale, but for the careful translation of concepts – scale, relevance, and separation of scales – so that they can make robust guarantees for real-world AI systems. We believe a deliberate effort to bridge these communities, by sketching the problem (from AI safety) and potential solutions (from renormalisation) in a more neutral language, will accelerate progress so that it can keep pace with threats from AI systems. References [1]Dong Shu, Xuansheng Wu, Haiyan Zhao, Daking Rai, Ziyu Yao, Ninghao Liu, and Mengnan Du. A survey on sparse autoencoders: Interpreting the internal mechanisms of large language models, 2025. URLhttps: //arxiv.org/abs/2503.05613. 22 These may be irrelevant with respect to the fixed point of interest, but relevant with respect to nearby fixed points. Setting these operators to zero can therefore alter the overall structural properties of the RG flow. They also have implications on universality arguments and the calculations of critical exponents. 12 [2]Wes Gurnee, Neel Nanda, Matthew Pauly, Katherine Harvey, Dmitrii Troitskii, and Dimitris Bertsimas. Finding neurons in a haystack: Case studies with sparse probing, 2023. URLhttps://arxiv.org/abs/2305.0 1610. [3]Alexander Matt Turner, Lisa Thiergart, Gavin Leech, David Udell, Juan J. Vazquez, Ulisse Mini, and Monte MacDiarmid. Steering language models with activation engineering, 2024. URLhttps://arxiv.org/ab s/2308.10248. [4]Nelson Elhage, Tristan Hume, Catherine Olsson, Nicholas Schiefer, Tom Henighan, Shauna Kravec, Zac Hatfield-Dodds, Robert Lasenby, Dawn Drain, Carol Chen, Roger Grosse, Sam McCandlish, Jared Kaplan, Dario Amodei, Martin Wattenberg, and Christopher Olah. Toy models of superposition, 2022. URLhttps: //arxiv.org/abs/2209.10652. [5] Hoagy Cunningham, Aidan Ewart, Logan Riggs, Robert Huben, and Lee Sharkey. Sparse autoencoders find highly interpretable features in language models, 2023. URL https://arxiv.org/abs/2309.08600. [6]Trenton Bricken, Adly Templeton, Joshua Batson, Brian Chen, Adam Jermyn, Tom Conerly, Nick Turner, Cem Anil, Carson Denison, Amanda Askell, Robert Lasenby, Yifan Wu, Shauna Kravec, Nicholas Schiefer, Tim Maxwell, Nicholas Joseph, Zac Hatfield-Dodds, Alex Tamkin, Karina Nguyen, Brayden McLean, Josiah E Burke, Tristan Hume, Shan Carter, Tom Henighan, and Christopher Olah. Towards monosemanticity: De- composing language models with dictionary learning. Transformer Circuits Thread, 2023. https://transformer- circuits.pub/2023/monosemantic-features/index.html. [7] Yasaman Bahri, Jonathan Kadmon, Jeffrey Pennington, Samuel S. Schoenholz, Jascha Narain Sohl-Dickstein, and Surya Ganguli. Statistical mechanics of deep learning. Journal of Statistical Mechanics: Theory and Experiment, 2024, 2024. URL https://api.semanticscholar.org/CorpusID:213069049. [8] Daniel A. Roberts, Sho Yaida, and Boris Hanin. The Principles of Deep Learning Theory: An Effective Theory Approach to Understanding Neural Networks. Cambridge University Press, May 2022. ISBN 9781316519332. doi: 10.1017/9781009023405. URL http://dx.doi.org/10.1017/9781009023405. [9]Pankaj Mehta and David J. Schwab. An exact mapping between the variational renormalization group and deep learning, 2014. URL https://arxiv.org/abs/1410.3831. [10]Kenneth G. Wilson and J. B. Kogut. The renormalization group and the epsilon expansion. Physics Reports, 12: 75–200, 1973. URL https://api.semanticscholar.org/CorpusID:205106813. [11]John L. Cardy. Scaling and Renormalization in Statistical Physics. Cambridge University Press, Cambridge, UK, 1996. [12]Michael E. Peskin and Daniel V. Schroeder. An Introduction to quantum field theory. Addison-Wesley, Reading, USA, 1995. ISBN 978-0-201-50397-5, 978-0-429-50355-9, 978-0-429-49417-8. doi: 10.1201/9780429503559. [13]Daniel J. Amit and Victor Mart ́ ıN-Mayor. Field Theory, the Renormalization Group, and Critical Phenomena: Graphs to Computers (3RD Edition), 2005. [14] David R. Nelson and Grace H. Zhang. The Renormalization Group and Condensed Matter Physics. Princeton University Press, 2025. ISBN 978-0691257907. [15]Joseph Polchinski. Renormalization and Effective Lagrangians. Nucl. Phys. B, 231:269–295, 1984. doi: 10.1016/0550-3213(84)90287-6. [16] Leo P. Kadanoff. Real-space renormalization in statistical mechanics. Reviews of Modern Physics, 86:647–667, 2014. doi: 10.1103/RevModPhys.86.647. [17]Zeyuan Allen-Zhu and Yuanzhi Li. Physics of language models: Part 1, learning hierarchical language structures, 2025. URL https://arxiv.org/abs/2305.13673. [18] Greg Yang. Tensor programs i: Wide feedforward or recurrent neural networks of any architecture are gaussian processes, 2021. URL https://arxiv.org/abs/1910.12478. 13 [19]Zohar Ringel, Noa Rubin, Edo Mor, Moritz Helias, and Inbar Seroussi. Applications of statistical field theory in deep learning, 2025. URL https://arxiv.org/abs/2502.18553. [20]Jonathan Ho, Ajay Jain, and Pieter Abbeel. Denoising diffusion probabilistic models, 2020. URLhttps: //arxiv.org/abs/2006.11239. [21] Youran Sun and Babak Haghighat. Phase transitions in large language models and the o(n) model, 2025. URL https://arxiv.org/abs/2501.16241. [22]William Peebles and Saining Xie. Scalable diffusion models with transformers, 2023. URLhttps://arxiv. org/abs/2212.09748. [23]Bart Bussmann, Noa Nabeshima, Adam Karvonen, and Neel Nanda. Learning multi-level features with matryoshka sparse autoencoders, 2025. URL https://arxiv.org/abs/2503.17547. [24]Garrett Baker, George Wang, Jesse Hoogland, and Daniel Murfet. Structural inference: Interpreting small language models with susceptibilities, 2025. URL https://arxiv.org/abs/2504.18274. [25]Ari Brill. Neural scaling laws rooted in the data distribution, 2024. URLhttps://arxiv.org/abs/2412 .07942. [26] Artem Karpov, Tinuade Adeleke, Seong Hah Cho, and Natalia Perez-Campanero. The steganographic potentials of language models, 2025. URL https://arxiv.org/abs/2505.03439. [27] Paul Christiano, Eric Neyman, and Mark Xu. Formalizing the presumption of independence, 2022. URL https://arxiv.org/abs/2211.06738. [28]Samuel Marks, Can Rager, Eric J. Michaud, Yonatan Belinkov, David Bau, and Aaron Mueller. Sparse feature circuits: Discovering and editing interpretable causal graphs in language models, 2025. URLhttps: //arxiv.org/abs/2403.19647. [29]Atticus Geiger, Duligur Ibeling, Amir Zur, Maheep Chaudhary, Sonakshi Chauhan, Jing Huang, Aryaman Arora, Zhengxuan Wu, Noah Goodman, Christopher Potts, and Thomas Icard. Causal abstraction: A theoretical foundation for mechanistic interpretability, 2025. URL https://arxiv.org/abs/2301.04709. [30]Gabriel Wu and Jacob Hilton. Estimating the probabilities of rare outputs in language models, 2025. URL https://arxiv.org/abs/2410.13211. [31]M. Cassandro and E. Olivieri. Renormalization group and analyticity in one dimension: A proof of dobrushin’s theorem, 1981. URL https://doi.org/10.1007/BF01213013. [32]Philipp Alexander Kreer, Wilson Wu, Maxwell Adam, Zach Furman, and Jesse Hoogland. Bayesian influence functions for hessian-free data attribution, 2025. URL https://arxiv.org/abs/2509.26544. [33]Maciej Koch-Janusz and Zohar Ringel. Mutual information, neural networks and the renormalization group. Nature Physics, 14(6):578–582, March 2018. ISSN 1745-2481. doi: 10.1038/s41567-018-0081-4. URL http://dx.doi.org/10.1038/s41567-018-0081-4. [34] Adam S. Shai, Sarah E. Marzen, Lucas Teixeira, Alexander Gietelink Oldenziel, and Paul M. Riechers. Trans- formers represent belief state geometry in their residual stream, 2025. URLhttps://arxiv.org/abs/24 05.15943. [35] Fernando E. Rosas, Bernhard C. Geiger, Andrea I Luppi, Anil K. Seth, Daniel Polani, Michael Gastpar, and Pedro A. M. Mediano. Software in the natural world: A computational approach to hierarchical emergence, 2024. URL https://arxiv.org/abs/2402.09090. [36]Gorka Peraza Coppola, Moritz Helias, and Zohar Ringel. Renormalization group for deep neural networks: Universality of learning and scaling laws, 2026. URL https://arxiv.org/abs/2510.25553. [37]Jennifer Lin. Feature identification via the empirical ntk, 2025. URLhttps://arxiv.org/abs/2510.0 0468. 14 [38]Aryeh Brill. percolation-synthetic-data.https://github.com/aribrill/percolation-synthe tic-data, 2026. [39]David S. Berman and Alexander G. Stapleton. A path to natural language through tokenisation and transformers, 2026. URL https://arxiv.org/abs/2601.03368. [40]Andrew Mack, Theodore Ehrenborg, Patrick Leask, Lauren Greenspan, and Lucas Teixeira. Probabilistic sparse auto-encoders learn hierarchical features. Under Review, forthcoming. [41]Maxwell Adam, Zach Furman, and Jesse Hoogland. The loss kernel: A geometric probe for deep learning interpretability, 2025. URL https://arxiv.org/abs/2509.26537. [42] Gabriel Wu and Jacob Hilton. Estimating the probabilities of rare outputs in language models, 2025. URL https://arxiv.org/abs/2410.13211. [43] D.L. Donoho. Compressed sensing. IEEE Transactions on Information Theory, 52(4):1289–1306, 2006. doi: 10.1109/TIT.2006.871582. [44] Andreas M. Tillmann. On the computational intractability of exact and approximate dictionary learning. IEEE Signal Processing Letters, 22(1):45–49, January 2015. ISSN 1558-2361. doi: 10.1109/lsp.2014.2345761. URL http://dx.doi.org/10.1109/LSP.2014.2345761. [45]Neel Nanda, Andrew Lee, and Martin Wattenberg. Emergent linear representations in world models of self- supervised sequence models, 2023. URL https://arxiv.org/abs/2309.00941. [46]Tomas Mikolov, Wen-tau Yih, and Geoffrey Zweig. Linguistic regularities in continuous space word repre- sentations. In Lucy Vanderwende, Hal Daum ́ e I, and Katrin Kirchhoff, editors, Proceedings of the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, pages 746–751, Atlanta, Georgia, June 2013. Association for Computational Linguistics. URL https://aclanthology.org/N13-1090/. [47]Joshua Engels, Logan Riggs, and Max Tegmark. Decomposing the dark matter of sparse autoencoders, 2025. URL https://arxiv.org/abs/2410.14670. [48]David Donoho and Jared Tanner. Observed universality of phase transitions in high-dimensional geometry, with implications for modern data analysis and signal processing. Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, 367(1906):4273–4293, November 2009. ISSN 1471-2962. doi: 10.1098/rsta.2009.0152. URL http://dx.doi.org/10.1098/rsta.2009.0152. [49] Adam Karvonen, Can Rager, Johnny Lin, Curt Tigges, Joseph Bloom, David Chanin, Yeu-Tong Lau, Eoin Farrell, Callum McDougall, Kola Ayonrinde, Demian Till, Matthew Wearden, Arthur Conmy, Samuel Marks, and Neel Nanda. Saebench: A comprehensive benchmark for sparse autoencoders in language model interpretability, 2025. URL https://arxiv.org/abs/2503.09532. [50]Guillaume Alain and Yoshua Bengio. Understanding intermediate layers using linear classifier probes, 2018. URL https://arxiv.org/abs/1610.01644. [51]Gonc ̧alo Paulo, Alex Mallen, Caden Juang, and Nora Belrose. Automatically interpreting millions of features in large language models, 2025. URL https://arxiv.org/abs/2410.13928. [52]Johnny Lin. Neuronpedia: Interactive reference and tooling for analyzing neural networks, 2023. URL https://w.neuronpedia.org. Software available from neuronpedia.org. [53] Neel Nanda, Lawrence Chan, Tom Lieberum, Jess Smith, and Jacob Steinhardt. Progress measures for grokking via mechanistic interpretability, 2023. URL https://arxiv.org/abs/2301.05217. [54] Joshua Engels, Eric J. Michaud, Isaac Liao, Wes Gurnee, and Max Tegmark. Not all language model features are one-dimensionally linear, 2025. URL https://arxiv.org/abs/2405.14860. [55]R ́ obert Csord ́ as, Christopher Potts, Christopher D. Manning, and Atticus Geiger. Recurrent neural networks learn to store and generate sequences using non-linear representations, 2024. URLhttps://arxiv.org/ abs/2408.10920. 15 [56]Sai Sumedh R. Hindupur, Ekdeep Singh Lubana, Thomas Fel, and Demba Ba. Projecting assumptions: The duality between sparse autoencoders and concept geometry, 2025. URLhttps://arxiv.org/abs/2503 .01822. [57]Gonc ̧alo Paulo and Nora Belrose. Sparse autoencoders trained on the same data learn different features, 2025. URL https://arxiv.org/abs/2501.16615. [58]David Chanin, James Wilken-Smith, Tom ́ a ˇ s Dulka, Hardik Bhatnagar, Satvik Golechha, and Joseph Bloom. A is for absorption: Studying feature splitting and absorption in sparse autoencoders, 2025. URLhttps: //arxiv.org/abs/2409.14507. [59]Patrick Leask, Bart Bussmann, Michael Pearce, Joseph Bloom, Curt Tigges, Noura Al Moubayed, Lee Sharkey, and Neel Nanda. Sparse autoencoders do not find canonical units of analysis, 2025. URLhttps://arxiv. org/abs/2502.04878. [60]Lawrence Chan. Causal scrubbing: A method for rigorously testing interpretability hypotheses.https: //w.lesswrong.com/s/h95ayYYwMebGEYN5y, 2022. Accessed: 2023-01-23. [61]Kevin T. Grosvenor and Ro Jefferson. The edge of chaos: quantum field theory and deep neural networks, 2022. URL https://arxiv.org/abs/2109.13247. [62] Greg Yang and Edward J. Hu. Feature learning in infinite-width neural networks, 2022. URLhttps: //arxiv.org/abs/2011.14522. [63]Zohar Ringel, Noa Rubin, Edo Mor, Moritz Helias, and Inbar Seroussi. Applications of statistical field theory in deep learning, 2025. URL https://arxiv.org/abs/2502.18553. [64]Max Staats, Matthias Thamm, and Bernd Rosenow. Small singular values matter: A random matrix analysis of transformer models, 2025. URL https://arxiv.org/abs/2410.17770. [65]Charles H. Martin and Michael W. Mahoney. Implicit self-regularization in deep neural networks: Evidence from random matrix theory and implications for learning, 2018. URLhttps://arxiv.org/abs/1810.01075. [66]James Halverson, Anindita Maiti, and Keegan Stoner. Neural networks and quantum field theory. Machine Learning: Science and Technology, 2(3):035002, April 2021. ISSN 2632-2153. doi: 10.1088/2632-2153/abeca3. URL http://dx.doi.org/10.1088/2632-2153/abeca3. [67]Jessica N Howard, Ro Jefferson, Anindita Maiti, and Zohar Ringel. Wilsonian renormalization of neural network gaussian processes*. Machine Learning: Science and Technology, 6(2):025038, May 2025. ISSN 2632-2153. doi: 10.1088/2632-2153/adc8fc. URL http://dx.doi.org/10.1088/2632-2153/adc8fc. [68]David S Berman, Marc S Klinger, and Alexander G Stapleton. Bayesian renormalization. Machine Learning: Science and Technology, 4(4):045011, October 2023. ISSN 2632-2153. doi: 10.1088/2632-2153/ad0102. URL http://dx.doi.org/10.1088/2632-2153/ad0102. [69] Zhenfeng Tu, Santiago Aranguri, and Arthur Jacot. Mixed dynamics in linear networks: Unifying the lazy and active regimes, 2024. URL https://arxiv.org/abs/2405.17580. [70]Gadi Naveh and Zohar Ringel. A self consistent theory of gaussian processes captures feature learning effects in finite cnns, 2021. URL https://arxiv.org/abs/2106.04110. [71]Blake Bordelon and Cengiz Pehlevan. Self-consistent dynamical field theory of kernel evolution in wide neural networks, 2022. URL https://arxiv.org/abs/2205.09653. [72] Blake Bordelon and Cengiz Pehlevan. Dynamics of finite width kernel and prediction fluctuations in mean field neural networks, 2023. URL https://arxiv.org/abs/2304.03408. [73]Kirsten Fischer, Javed Lindner, David Dahmen, Zohar Ringel, Michael Kr ̈ amer, and Moritz Helias. Critical feature learning in deep neural networks, 2024. URL https://arxiv.org/abs/2405.10761. [74]Noa Rubin, Kirsten Fischer, Javed Lindner, David Dahmen, Inbar Seroussi, Zohar Ringel, Michael Kr ̈ amer, and Moritz Helias. From kernels to features: A multi-scale adaptive theory of feature learning, 2025. URL https://arxiv.org/abs/2502.03210. 16 [75]Noa Rubin, Orit Davidovich, and Zohar Ringel. Mitigating the curse of detail: Scaling arguments for feature learning and sample complexity, 2025. URL https://arxiv.org/abs/2512.04165. [76]Blake Bordelon, Alexander Atanasov, and Cengiz Pehlevan. A dynamical model of neural scaling laws, 2024. URL https://arxiv.org/abs/2402.01092. [77]Noa Rubin, Inbar Seroussi, and Zohar Ringel. Grokking as a first order phase transition in two layer networks, 2024. URL https://arxiv.org/abs/2310.03789. [78] Philip W Anderson. More is different: broken symmetry and the nature of the hierarchical structure of science. Science, 177(4047):393–396, 1972. [79] Fernando E Rosas, Pedro AM Mediano, Michael Gastpar, and Henrik J Jensen. Quantifying high-order interdependencies via multivariate extensions of the mutual information. Physical Review E, 100(3):032305, 2019. [80] Daniel Kunin, Giovanni Luca Marchetti, Feng Chen, Dhruva Karkada, James B Simon, Michael R DeWeese, Surya Ganguli, and Nina Miolane. Alternating gradient flows: A theory of feature learning in two-layer neural networks. arXiv preprint arXiv:2506.06489, 2025. [81] Andrew M Saxe, James L McClelland, and Surya Ganguli. Exact solutions to the nonlinear dynamics of learning in deep linear neural networks. arXiv preprint arXiv:1312.6120, 2013. [82]Cl ́ ementine CJ Domin ́ e, Nicolas Anguita, Alexandra M Proca, Lukas Braun, Daniel Kunin, Pedro AM Mediano, and Andrew M Saxe. From lazy to rich: Exact learning dynamics in deep linear networks. arXiv preprint arXiv:2409.14623, 2024. [83]Daniel Kunin, Allan Ravent ́ os, Cl ́ ementine Domin ́ e, Feng Chen, David Klindt, Andrew Saxe, and Surya Ganguli. Get rich quick: exact solutions reveal how unbalanced initializations promote rapid feature learning. Advances in Neural Information Processing Systems, 37:81157–81203, 2024. [84]Feng Chen, Daniel Kunin, Atsushi Yamamura, and Surya Ganguli. Stochastic collapse: How gradient noise attracts sgd dynamics towards simpler subnetworks. Advances in Neural Information Processing Systems, 36: 35027–35063, 2023. [85]Jesse Hoogland, George Wang, Matthew Farrugia-Roberts, Liam Carroll, Susan Wei, and Daniel Murfet. Loss landscape degeneracy and stagewise development in transformers. Transactions on Machine Learning Research, 2025. [86] Sumio Watanabe. Algebraic geometry and statistical learning theory, volume 25. Cambridge university press, 2009. [87]Tanishq Kumar, Blake Bordelon, Samuel J. Gershman, and Cengiz Pehlevan. Grokking as the transition from lazy to rich training dynamics, 2024. URL https://arxiv.org/abs/2310.06110. [88]Sebastian Lee, Sebastian Goldt, and Andrew Saxe. Continual learning in the teacher-student setup: Impact of task similarity. In International Conference on Machine Learning, pages 6109–6119. PMLR, 2021. [89]Elchanan Mossel. Deep learning and hierarchal generative models, 2018. URLhttps://arxiv.org/abs/ 1612.09057. [90] Eran Malach and Shai Shalev-Shwartz. A provably correct algorithm for deep learning that actually works, 2018. URL https://arxiv.org/abs/1803.09522. [91]Abhinav Menon, Manish Shrivastava, David Krueger, and Ekdeep Singh Lubana. Analyzing (in)abilities of saes via formal languages, 2025. URL https://arxiv.org/abs/2410.11767. [92] Ekdeep Singh Lubana, Kyogo Kawaguchi, Robert P. Dick, and Hidenori Tanaka. A percolation model of emergence: Analyzing transformers trained on a formal language, 2024. URLhttps://arxiv.org/abs/ 2408.12578. 17 [93]Francesco Cagnetta, Leonardo Petrini, Umberto M. Tomasini, Alessandro Favero, and Matthieu Wyart. How Deep Neural Networks Learn Compositional Data: The Random Hierarchy Model. Physical Review X, 14(3): 031001, July 2024. doi: 10.1103/PhysRevX.14.031001. [94]Francesco Cagnetta and Matthieu Wyart. Towards a theory of how the structure of language is acquired by deep neural networks, 2024. URL https://arxiv.org/abs/2406.00048. [95] Francesco Cagnetta, Hyunmo Kang, and Matthieu Wyart. Learning curves theory for hierarchically compositional data with power-law distributed features, 2025. URL https://arxiv.org/abs/2505.07067. [96]Antonio Sclocchi, Alessandro Favero, and Matthieu Wyart. A phase transition in diffusion models reveals the hierarchical nature of data, 2024. URL https://arxiv.org/abs/2402.16991. [97]Antonio Sclocchi, Alessandro Favero, Noam Itzhak Levi, and Matthieu Wyart. Probing the latent hierarchical structure of data via diffusion models, 2025. URL https://arxiv.org/abs/2410.13770. [98] Ryo Nakamura, Ryu Tadokoro, Ryosuke Yamada, Yuki M. Asano, Iro Laina, Christian Rupprecht, Nakamasa Inoue, Rio Yokota, and Hirokatsu Kataoka. Scaling backwards: Minimal synthetic pre-training?, 2024. URL https://arxiv.org/abs/2408.00677. [99] Manel Baradad, Jonas Wulff, Tongzhou Wang, Phillip Isola, and Antonio Torralba. Learning to see by looking at noise, 2022. URL https://arxiv.org/abs/2106.05963. [100]Hirokatsu Kataoka, Kazushige Okayasu, Asato Matsumoto, Eisuke Yamagata, Ryosuke Yamada, Nakamasa Inoue, Akio Nakamura, and Yutaka Satoh. Pre-training without natural images, 2021. URLhttps://arxiv. org/abs/2101.08515. [101] Benoit B Mandelbrot. The Fractal Geometry of Nature. W.H. Freeman, San Francisco, US, 1983. [102] Peter Bloem and Steven de Rooij. An expectation-maximization algorithm for the fractal inverse problem, 2017. URL https://arxiv.org/abs/1706.03149. [103]Eran Malach and Shai Shalev-Shwartz. Is deeper better only when shallow is good?, 2019. URLhttps: //arxiv.org/abs/1903.03488. [104] Peter Grassberger and Itamar Procaccia. Measuring the strangeness of strange attractors. PHYSICA D, 9(1-2): 189–208, October 1983. ISSN 0167-2789. doi: 10.1016/0167-2789(83)90298-1. [105]Aryeh Brill. Representation learning on a random lattice, 2025. URLhttps://arxiv.org/abs/2504.2 0197. [106]Aryeh Brill. A model for scaling laws of general intelligence. In Submitted to ILIAD 2: ODYSSEY, 2025. URL https://openreview.net/forum?id=9mAX9GZK5e. under review. [107]Tomaso Poggio, Hrushikesh Mhaskar, Lorenzo Rosasco, Brando Miranda, and Qianli Liao. Why and when can deep – but not shallow – networks avoid the curse of dimensionality: a review, 2017. URLhttps: //arxiv.org/abs/1611.00740. [108]David A. Danhofer, Davide D’Ascenzo, Rafael Dubach, and Tomaso Poggio. Position: A theory of deep learning must include compositional sparsity, 2025. URL https://arxiv.org/abs/2507.02550. [109] Henry W. Lin, Max Tegmark, and David Rolnick. Why does deep and cheap learning work so well? Journal of Statistical Physics, 168(6):1223–1247, July 2017. ISSN 1572-9613. doi: 10.1007/s10955-017-1836-5. URL http://dx.doi.org/10.1007/s10955-017-1836-5. [110]Emmanuel Abbe, Enric Boix-Adsera, Matthew Brennan, Guy Bresler, and Dheeraj Nagaraj. The staircase property: How hierarchical structure can guide deep learning, 2021. URLhttps://arxiv.org/abs/21 08.10573. [111]Emmanuel Abbe, Enric Boix-Adsera, and Theodor Misiakiewicz. Sgd learning on neural networks: leap complexity and saddle-to-saddle dynamics, 2023. URL https://arxiv.org/abs/2302.11055. 18 [112]Leo P. Kadanoff, Anthony Houghton, and Mehmet C. Yalabik. Variational approximations for renormalization group transformations. Journal of Statistical Physics, 1976. doi: 10.1007/BF01011765. [113]David J. Schwab and Pankaj Mehta. Comment on ”why does deep and cheap learning work so well?” [arxiv:1608.08225], 2016. URL https://arxiv.org/abs/1609.03541. [114]Amit Gordon, Aditya Banerjee, Maciej Koch-Janusz, and Zohar Ringel. Relevance in the renormalization group and in information theory. Physical Review Letters, 126(24), June 2021. ISSN 1079-7114. doi: 10.1103/physrevlett.126.240601. URLhttp://dx.doi.org/10.1103/PhysRevLett.126.240601. [115]Doruk Efe G ̈ okmen, Zohar Ringel, Sebastian D. Huber, and Maciej Koch-Janusz. Symmetries and phase diagrams with real-space mutual information neural estimation. Phys. Rev. E, 104:064106, Dec 2021. doi: 10.1103/PhysRevE.104.064106. URLhttps://link.aps.org/doi/10.1103/PhysRevE.104.06 4106. [116]Doruk Efe G ̈ okmen, Zohar Ringel, Sebastian D. Huber, and Maciej Koch-Janusz. Statistical physics through the lens of real-space mutual information. Physical Review Letters, 127(24), December 2021. ISSN 1079-7114. doi: 10.1103/physrevlett.127.240603. URLhttp://dx.doi.org/10.1103/PhysRevLett.127.240603. [117]Doruk Efe G ̈ okmen, Sounak Biswas, Sebastian D. Huber, Zohar Ringel, Felix Flicker, and Maciej Koch-Janusz. Compression theory for inhomogeneous systems. Nature Commun., 15(1):10214, 2024. doi: 10.1038/s41467- 024-54341-8. [118]Naftali Tishby and Noga Zaslavsky. Deep learning and the information bottleneck principle, 2015. URL https://arxiv.org/abs/1503.02406. [119]Artemy Kolchinsky, Brendan D. Tracey, and Steven Van Kuyk. Caveats for information bottleneck in deterministic scenarios, 2019. URL https://arxiv.org/abs/1808.07593. [120] Andrew Michael Saxe, Yamini Bansal, Joel Dapello, Madhu Advani, Artemy Kolchinsky, Brendan Daniel Tracey, and David Daniel Cox. On the information bottleneck theory of deep learning. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=ry_WPG-A-. [121]Matthew S. Schmitt, Maciej Koch-Janusz, Michel Fruchart, Daniel S. Seara, Michael Rust, and Vincenzo Vitelli. Information theory for data-driven model reduction in physics and biology, 2025. URLhttps: //arxiv.org/abs/2312.06608. [122]Fernando E. Rosas, Bernhard C. Geiger, Andrea I Luppi, Anil K. Seth, Daniel Polani, Michael Gastpar, and Pedro A. M. Mediano. Software in the natural world: A computational approach to hierarchical emergence, 2024. URL https://arxiv.org/abs/2402.09090. [123]O. Pfante, N. Bertschinger, E. Olbrich, N. Ay, and J. Jost. Comparison between different methods of level identification. Advances in Complex Systems, 17(2), 2014. [124] Cosma Rohilla Shalizi and Cristopher Moore. What is a macrostate? subjective observations and objective dynamics. Foundations of Physics, 55(1), December 2024. ISSN 1572-9516. doi: 10.1007/s10701-024-00814-1. URL http://dx.doi.org/10.1007/s10701-024-00814-1. [125] Olof G ̈ ornerup and Martin Nilsson Jacobi. A method for inferring hierarchical dynamics in stochastic processes. Adv. Complex Syst., 11:1–16, 2007. URLhttps://api.semanticscholar.org/CorpusID: 18797454. [126]Cosma Rohilla Shalizi and James P Crutchfield. Computational mechanics: Pattern and prediction, structure and simplicity. Journal of statistical physics, 104(3):817–879, 2001. [127]Fernando E Rosas. Symmetries at the origin of hierarchical emergence. arXiv preprint arXiv:2512.00984, 2025. [128]Jessica N. Howard, Marc Klinger, Anindita Maiti, and Alexander G. Stapleton. Bayesian rg flow in neural network field theories. SciPost Physics Core, 8(1), March 2025. ISSN 2666-9366. doi: 10.21468/scipostphyscore.8.1.027. URL http://dx.doi.org/10.21468/SciPostPhysCore.8.1.027. 19 [129]D S Berman, M S Klinger, and A G Stapleton. Ncoder—a quantum field theory approach to encoding data. Machine Learning: Science and Technology, 6(2):025059, June 2025. ISSN 2632-2153. doi: 10.1088/2632- 2153/ade04c. URL http://dx.doi.org/10.1088/2632-2153/ade04c. [130]Dmitry Manning-Coe, Jacopo Gliozzi, Alexander G. Stapleton, Edward Hirst, Giuseppe De Tomasi, Barry Bradlyn, and David S. Berman. Grokking vs. learning: Same features, different encodings, 2025. URL https://arxiv.org/abs/2502.01739. [131] Val ́ erie Costa, Thomas Fel, Ekdeep Singh Lubana, Bahareh Tolooshams, and Demba Ba. From flat to hierarchical: Extracting sparse representations with matching pursuit. In Advances in Neural Information Processing Systems, volume 38, 2025. 20 A Route to Impact: A Case Study We are inspired by the recent history of how an exploration of the superposition hypothesis – the idea that models represent more features than directions in activation space – led to the development of sparse autoencoders as a tool to find interpretable features in transformer-based models. We present this story as evidence for the potential route to impact of a novel theoretical perspective. In this case, AI safety researchers imported Compressed Sensing [43], an idea from applied math, along with its workhorse, dictionary learning [44], to make the story come together. It offers a compelling case study in mechanistic interpretability in which empirical observations and well-posed ideas led to a measurable phase change in how we frame, analyze, and interpret NNs. Its progression is marked by two key outputs. A.1 Output 1: Model Organism of Superposition Neural networks are polysemantic – their constituents (neurons, attention heads) are observed to activate in multiple distinct contexts. In Toy Models of Superposition (TMS) [4], their seminal work on the topic, Anthropic explores superposition as a hypothesis for this phenomenon by studying how models trained on data with ground-truth features can represent more of those features when they interfere. Central to TMS are two ideas: 1.Sparsity. Each input is a linear combination of features, defined here as vectors in the input space. Superposition occurs when there are more features with significant statistical weight than there are neurons. TMS encourages this by tuning the sparsity in the input space; when sparsity is high, each feature appears in fewer training examples. To minimize the loss, the model compresses multiple features into overlapping directions in activation space. 2.TMS also relies on the assumption that features are linearly represented, known as the linear representation hypothesis (LRH) [45,46]. This essentially defines a ‘feature’ as a direction in activation space, one that ideally preserves the structure of the input space – this is how they defined a feature in the toy setting. Though directionally useful, this approximation may break down in more realistic settings, making the LRH is at best a useful approximation [47]. TMS is essentially a model organism for superposition – a toy neural network setting designed to encourage superposition implicitly in its internal representation. This simple picture allowed researchers to refine and probe the superposition hypothesis and what it means for N training, inference, and feature geometry. Compressed sensing offers useful guides in their analysis, including bounds on the number of features in superposition [48]. A.2 Output 2: Sparse Autoencoders (SAEs) Underlying the superposition hypothesis and its relation to polysemanticity is that there exists a collection of linear directions in activation space (features) which are monosemantic, i.e., correspond to single atomic concepts. If this is true, and if those directions could be found, it would be a boon for interpretability. Enter the SAE: a tool built on dictionary learning – a compressed sensing technique – which learns an overcomplete feature basis by sparsely reconstructing activation data [1, 6]. The addition of a sparsity constraint to the loss function tweaks the LRH to a new hypothesis which supposes that there exist linear, monosemantic feature directions that explain the activation data, and that these features are activated sparsely by each activation. Validating this hypothesis involved generating two kinds of (empirical) evidence: 1. Statistical competitiveness: Earlier unsupervised attempts to rotate the neuron basis, like PCA, revealed some important directions but failed to produce monosemantic units except for some cherrypicked cases. SAEs were shown to outperform these baselines (noteably, in transformers). More precisely, activations could be compressed more efficiently by assuming sparsity in the SAE basis than in PCA or randomized control bases with matched second moments. While this does not guarantee interpretability, it provides strong evidence that the hypothesis is a good statistical fit. We refer the reader to [49] for a review of SAE benchmarks, a discussion of their limitations. 2.Semantic Interpretability: Even if sparse decompositions exist, they are not guaranteed to be monosemantic in a way that is human-interpretable. But SAE features turned out to be remarkably so, at least compared to any other 21 unsupervised way to get linear features [5]. This was checked in a number of ways, including using probes [2,50] to select for dataset examples that strongly activate a feature and activation addition [3] (e.g., adding a feature to a hidden layer and observing how it affects the model output). These methods revealed that many learned directions correspond to apparently monosemantic concepts, as measured by an automated (LLM-evaluated) interp score and by eye [51, 52]. A.3 The Takeaway While the progression from TMS to SAEs was productive for AI safety, it points to several limitations that motivate the need for better hypotheses, tools, and techniques: 1. The LRH is approximate. TMS assumes that features correspond to directions in activation space, which reflect linear features in the data. While this provides a tractable mathematical setting, it likely misses contextual, hierarchical, or nonlinear structure present in real data (for example, modular addition [53]). SAEs build on this view by using linear dictionaries trained under sparsity constraints to reconstruct N activations, which can be informative but are not guaranteed to reflect the model’s own (potentially nonlinear) internal structure [54–56]. 2. “Feature” is ambiguous. The notion of a “feature” is underdefined and used inconsistently in AI safety. • In TMS, features are (in this case, ground-truth) directions in the input space. The model learns to represent these in activation space, but whether these internal directions match the true data features depends on training regime and inductive bias—it’s not guaranteed. •In SAEs, features are dictionary directions in activation space that reconstruct model activations under a sparsity constraint. These are assumed to correspond to atomic, monosemantic units, but this is typically justified in an ad hoc or unprincipled way. This definition is further muddied by phenomena (like feature splitting) discussed in the ‘limitations of current SAE methods’ below. 3.‘Interpretable’ is ambiguous. Monosemanticity is often taken as a stand-in for interpretability, but are these definitions synchronous? While TMS observes a sharp phase transition between monosemantic and superposed regimes, a monosemantic feature might correspond to a fuzzy, uninterpretable internal concept even if it encodes a statistically pure direction. Conversely, an interpretable feature might be polysemantic in structure but contextually well-understood. 4.Current SAE Methods are limited. While SAE-inspired approaches are currently the most scalable and principled tools for feature decomposition, they fall short of the criteria we consider essential for a robust interpretability toolkit: • Canonicity, incorporating: – Uniqueness: Different SAEs (datasets, hyperparameters, training runs) can produce different decompo- sitions(e.g., [57]). – Completeness: They may omit significant but entangled features (e.g., [56]). –Atomicity: The features are not guaranteed to be irreducible (feature splitting, absorption [58]). Theoretically, SAEs are too weak to capture certain important atomic behaviors(e.g., [59]). •Faithfulness: There’s no guarantee that the features reflect the model’s causal structure or computations. Causal scrubbing and other techniques for validating decompositions reward loss-preserving approximations, not faithful reconstructions of the model’s actual internal mechanisms [29,60]. This raises the risk that tools like SAEs may capture statistically predictive proxies rather than the true causal structure of model computation or reasoning. Additionally, SAEs optimize for a trade-off between reconstruction accuracy and compactness (description length) of an explanation. The extent to which a sparsity prior selects for efficiency over canonicity or faithfulness is a core question in interpretability work [49]. In spite of its limitations, we think that this story exposes a model for research that, if duplicated, could narrow the theory-practice gap and drive progress in ambitious interpretability. Key components of this model include: 22 •A tight theory–experiment loop: The superposition hypothesis led to testable predictions in TMS, which then catalyzed the SAE method for probing real models—a quick and tightly connected pathway from concept to tool. •Cross-Disciplinary Agility: TMS successfully adapted techniques from compressed sensing to the problem of feature decomposition, sharpening both the theoretical framing and the empirical setup, and exemplifying how ideas from applied math and signal processing can be translated into AI safety contexts. Knowing which aspects of an external field are relevant, which are not, and how best to translate them to a new application, is a core but worthwhile challenge in cross-disciplinary collaborations. •Rapid iteration: The time from the initial TMS paper to the adoption of SAEs as a tool in mainstream inter- pretability research was under a year. This reflects a nimble, experimental style of science: start with a tractable toy model, iterate quickly, and scale up promising insights. • Clarity of purpose: The work was animated by clear goals: understanding polysemanticity via superposition and finding interpretable monosemantic models. Each step in the pipeline was framed by these guiding questions. •Aiming for tools that scale: SAEs are computationally tractable and more or less efficient, relatively easy to apply and scale across layers and models. The value of toy models disappears if insights fail to transfer in realistic settings. Using this model, our goal is to develop new, testable hypotheses that lead to interpretability tools without the failure modes described in the last section. Statistical physics, and renormalisation in particular, could provide a theoretical and empirical framework for defining theory and employing tools that leverage meaningful structure that current tools lack. For example, we aim to work with a definition of ‘feature’ that is not predicated on the LRH, capable of handling non-linear, compositional, and hierarchical structure. This should be both model-natural and interpretable, able to capture partial, contextual, or compositional structure. In classifying or interpreting features, we aim to prioritize relevance over canonicity, seeking effective descriptions that summarize behavior at a given scale of abstraction. We hope that additional properties that a renormalisation framework adds – like separation of scales – will improve the faithfulness of a given set of features. B Review B.1 Kernel-Space Renormalization There are many axes along which we can categorize model complexity and behavior. We find it useful to think of the literature according to the following descriptors, noting that this is a partial list and that many references will cut across them: • Whether it can be modeled by a static (equilibrium) or dynamic (training-time dependent) statistical theory. This often aligns with a specific asymptotic kernel (NNGP covariance vs. NTK), and dictates which observables are natural and how we interpret relevance. •Whether training is well-approximated by linearized dynamics (lazy/kernel learning regime) or has substantial kernel drift (feature learning regime). •Whether the theory is free or interacting. In parameter regimes where the 1/width (or another appropriate scaling parameter) is not small, the second order term is no longer sufficient to tell the whole story. This is important for understanding to what degree existing theory approximates real-world behavior. •What the background is. A non-zero mean effectively changes which kernel modes couple to one another, impacting our choice of cutoff and determination of feature relevance. In the dynamic case, the mean also evolves during training, leading to further kernel drift. The physics literature primarily uses language from statistical and quantum field theory, which expands a field with many correlated degrees of freedom in order of its fluctuations (correlation functions, which encode the ways in which degrees of freedom can interact) around its background (or vacuum expectation) value. Importantly, the second order 23 term in this expansion is the covariance matrix or propagator (the kernel) encoding quadratic interactions. In many cases, this expansion is controlled when the width is large (in the limit, infinite), which makes higher order interactions subleading. This is often, though inaccurately, referred to in the ML literature as mean field scaling (or mean-field parametrization), which essentially limits the cumulant hierarchy to second order in 1/width (properly, the Gaussian or free field limit). Tools from perturbation theory and mean field theory can then be used to relax this limit, allowing one to capture finite-width behaviour of real-world models in a controlled manner (e.g., [8,61]). Other scalings (e.g.,μP, adaptive or kernel scaling [62]) aim to capture the non-perturbative behavior by effectively removing the dependence on width. We point the interested reader to this review [63] of statistical physics applications to NNs for more information. While most of the work considered in this section comes from the physics community, we hope to connect it to related work from other fields. We think work on random matrix theory (e.g., [64,65]) will be a particular promising point of intersection. B.1.1 Physics Perspective: The NTK In [8] The Principles of Deep Learning Theory (PDLT), the authors derive neural tangent kernel regression as the leading approximation obtained by performing a linear expansion in the weights and accumulating the resulting first-order updates across gradient descent steps. In this regime, the NTK stays effectively fixed during training. This linearized approximation holds for models with a small depth-to-width aspect ratio and sufficiently many samples relative to the width, whose weights are initialized with a 1/sqrt(width) scaling. Finally, they extend the Taylor expansion to next-to-leading order in the weights and identify conditions under which a quadratic correction dominates higher-order terms. The resulting formula is well-approximated by kernel regression with the NTK at initialization replaced by the average of itself and the empirical NTK (eNTK). These results can be connected to renormalisation in two distinct ways: •The kernel eigendirections define an intra-model RG. The NTK (or eNTK) and the function it approximates, can be further approximated by truncating eigendirections whose eigenvalues are below a certain cutoff. For different cutoffs, this yields a sequence of increasingly coarse-grained approximations to the function learned by the N, via kernel regression with the truncated kernel. Such truncation can be a natural thing to do when the kernel spectrum itself is highly anisotropic, with large fall-offs in its ordered eigenvalues setting effective emergent scales. •Alternatively, from the point of view of an inter-model RG in a space of hyperparameters (width, depth, scalings of weight initializations with respect to them . . . ), the NTK corresponds to a free universality class. During training, the same (data, training task) tuple can move along different trajectories (parameterizations) in model space depending on their hyperparameters and initializations, ending in a basin of attraction that characterizes its behavior. In the regime discussed here, which we can call the NTK basin of attraction, the models exhibiting no representation learning in the sense that they are nearly mathematically equivalent at the end of training to linear regression with a large number of fixed feature functions. The authors of PDLT also invoke a third RG analogy in their book, at a structural (i.e., static) level. In early chapters, they study how, at initialization, layer-wise statistics within a given model (e.g., correlations of neural activations across different inputs) evolve with layer depth. For different fixed activation functions, they show that different choices of the c-number coefficients for the weight and bias variances at initialization separate an “ordered phase” where different model inputs become indistinguishable at late layers, from a “chaotic phase” where small input differences rapidly decorrelate with model depth. In practice one would like to tune these coefficients to lie on a critical line/manifold between the phases, so that inputs can propagate to the output layer in a reasonable way. In short, this version of the RG analogy guides select hyperparameter choices to avoid the problem of vanishing/exploding gradients. Recent work suggests that this perspective is relevant for interpretability. Borrowing from statistical mechanics, where susceptibilities measure an observable’s linear reponse to external perturbations, the authors of [24] develop susceptibility-style probes to understand how observables localized on individual model components (for example, the per-component loss) respond to infinitessimal perturbations of the data distribution. This framework is conceptually close to probing sensitivity of kernel eigenmodes, and offers a principled way to quantify relevance at a scale defined by the chosen perturbation: components with high susceptibility contribute disproportionately to observable behavior, while those with low susceptibility can be considered irrelevant for those distributional shifts. 24 Complementary to this, [32] develop Bayesian influence functions that provide a Hessian-free (i.e., more scalable) approach to data attribution. Influence functions measure how much a particular training example affects model predictions, which naturally aligns with the eNTK picture at finite width. In an RG sense, this work offers a tool for identifying which data points are ‘relevant’ in the sense that their removal would significantly alter the effective description of the learned function. B.1.2 Physics Perspective: Field-Theoretical Approaches to Network Structure (N-QFT) ‘Neural Networks and Quantum Field Theory’ introduces one of the earliest frameworks in “N-QFT” or “NNFT” correspondence [66]. Based on a mapping between N architectures, whether at initialization or at any time-step during training, and statistical field theories (an Euclidean or thermal version of quantum field theories – the backbone of particle physics and string theory), this work primarily builds on field theoretic interpretability of NNs and data. It connects to both [61] and [61], in leveraging theoretical physics for new tools in mechanistic interpretability. NNFT correspondence models any N’s output functional distribution as a statistical field theory “action” (model log-likelihood analog). For NNGP, the action contains a single term quadratic in the model output, which is generally diagonalizable on the basis of N inputs (features); to borrow theoretical physics jargon – “the NNGP action is local in input space”. Deviations from NNGP introduce additional terms in the model action that are higher-than-quadratic order polynomials in output, known as “interactions”, leading to a series sum with infinite such corrections parametrized by1/N, whereNis the width of the hidden layer(s). Independently and identically distributed (i.i.d.) parameters, e.g., at initialization or an NNGP trained via stochastic gradient descent withL 2 loss, constrain the quadratic term in the action asN-independent, while terms higher-than-quadratic order are accompanied by increasing powers of 1/N; as a concrete example, arth polynomial of output in action is accompanied with a prefactor1/N r/2−1 . Non-i.i.d. parameters, such as feature learning limits, introduce mixed scalings of1/Nand⃗αin action, where⃗αinduces statistical correlations among model parameters. The1/Nscaling in action offers a natural hierarchy over non-local mixings in N input (feature) space, as different polynomial-order interactions in action increasingly capture the interplays of model outputs as functions of data at varying spatial or information-geometric separations. This work further introduces a Wilsonian RG scheme over the space of model inputs (features or frequency or momenta, if Fourier-transformed): via an explicit cutoffΛon the data space.Λmay be determined by finite dataset sizes or resolution scales: either case leads to approximations of the model log-likelihood. NNFT correspondence generates N outputs as field configurations – any loss in input-level information transforms the field interactions systematically. More specifically, the shift from infinite to finite data is analogous to coarse-graining over low frequency (momentum) field modes, whereas changes in resolution scale act as coarse-graining over high frequency (momentum) modes, respectively framed as RG of IR and UV field theories. The first case leads to an “effective action” oblivious to explicit roles of long-range data (feature) interactions, whereas the latter case leads to an effective action oblivious to explicit short-range (UV) interactions. This work builds on the IR case, where coarse-grained data modes are originally incapable of deprecating dominant statistical moments of the model output distribution; therefore, RG simplifies model analysis while retaining output quality and safety metrics. Qualitatively, data-related approximations are propagated into analytically tractable RG flows andβ-functions of field “couplings” in interaction terms of action. For example, infinitesimal transformations of the dataset boundary lead to infinitesimal shifts in coupling, creating a continuous flow with potential critical points that may indicate phase transitions in model output space. More broadly, keeping all model parameters and hyperparameters unchanged, any coarse-graining over dataset boundaries or resolution scales show up as rescalings of model action; NNs at large width may have interactions in the action that become increasingly relevant, irrelevant, or stay unchanged, in response to transformations in dataset size, boundaries, and resolution scales. ‘The edge of chaos: quantum field theory and deep neural networks’ [61] is the second of two primary directions that goes under the name “N-QFT correspondence”. At the most basic level, the most obvious difference in these directions is that this approach is concerned solely with the structural properties of network internals, and explicitly constructs the dual statistical or quantum field theory; in contrast, [66] applies to the network output (treated as a functional), and writes down a model action with suitable1/Ncouplings to match observed statistics. At the level of RG analogies, this approach is closely related to that of [8] in that it describes layerwise statistics in the large-but-finite-width limit, in which the ratio of depth to width plays the roll of the perturbative parameter controlling the cumulant expansion. Notably however, while the authors of [8] were careful to avoid the use of the term “field” due to the lack of any meaningful notion of distance within a given network layer, this approach obtains a bona fide field theory by taking the limit in the depth, where distance is meaningfully defined. This is physically interesting because it reveals a close 25 parallel between deep (as well as recurrent) neural networks and well-studied systems in physics known as large-N vector models. Practically, it is interesting because it allows a controlled computation of finite-width corrections relevant for real-world models via a rigorous application of perturbation theory (in contrast, [66] takes a more phenomenological approach to modelling the output functional (which plays the role of the field variable there), while [8] relies on a layer-by-layer coarse-graining procedure to obtain truncated layer statistics; here, a path integral is constructed directly from the structure equation of the network). In short, this allows the authors to compute the cumulants describing fluctuations, i.e., finite-width effects, in internal network statistics. On one hand, this has advantages in that it provides a fine-grained description of network internals at initialisation that allows, e.g., computation of finite-width corrections to the critical regime at which network trainability is optimised. More abstractly, from the perspective of physics, it is promising because it provides a mathematical duality between the structure of deep neural networks and a statistical field theory in0 + 1-dimensions, and hence can be used as a starting ground for more principled studies of renormalisation, universality classes, and critical behaviour in this context. As it provides a fine-grained path integral for the network, it also allows for the explicit computation of interneuron correlation functions (i.e., observables). On the other hand, it has the disadvantage of being agnostic about the output of the network itself, and current versions are purely static: they describe networks at initialisation and do not account for the evolution of trainable parameters under stochastic gradient descent. Additionally, it is unclear how to immediately generalise the approach (which has been developed only for vanilla DNNs and RNNs) to SOTA models such as transformers. More work is needed to cross these gaps. ‘Wilsonian renormalisation of Neural Network Gaussian Processes’ [67] applies RG to the Gaussian Process (GP) kernel, systematically coarse-graining over unlearnable modes to obtain a flow of the ridge parameter (i.e., the variance on the observed noise on the regression target) in which the data sets the cutoff scale. Here, learnability is defined relative to the ratio of the ridge parameter over the average number of datapoints: feature modes are eigenfunctions of the GP kernel, with eigenvaluesλ; when the latter is significantly smaller than this ratio, the modes effectively decouple from the learning process (they behave as though the noise were infinite, or equivalently the number of samples were zero). Thus, they can be integrated out to obtain an effective theory (in the Wilsonian sense) on the learnable (i.e., low-energy or IR) modes, 23 where each RG step consists of integrating out a momentum shell of higher feature modes to obtain an infinitesimal change to the ridge parameter. The data sets the cutoff scale in the sense that this process naturally halts at the first learnable mode, whereupon we can relate the bare (unrenormalized, starting) ridge parameter to the effective ridge parameter that more accurately describes the Gaussian Process. At a practical level, the authors show that this approach can predict the power-law scaling of the mean-squared error (MSE) loss obtained with many real-world data sets. More abstractly however, this is appealing because it goes beyond earlier structural analogies between NNs and RG, and establishes a practical link between RG and learning. In doing so, it provides a concrete stepping-stone towards establishing notions of universality in deep learning. B.1.3 Physics Perspective: Connecting different Kernel Regimes A growing body of work aims to unify the lazy and feature-learning regimes for a more complete picture of how kernels evolve during training. ‘Mixed Dynamics In Linear Networks: Unifying the Lazy and Active Regimes’ [69] show theoretically that for a two-layer linear networkA = W 2 W 1 , one can derive an approximation to how the model changes under each step of gradient descent that contains lazy and feature-learning dynamics in different limits of the weight initializations and the model width. Their formula reveals a mixed regime where some singular values of A evolve lazily, while others are active. The authors also empirically locate a mixed regime in a phase diagram for a noisy matrix reconstruction task. This suggests that, in the context of the “inter-model” RG picture described above, it may be more natural to distinguish between lazy and feature-learning basins at the level of individual singular values of a matrix model (or in nonlinear networks, individual kernel eigendirections), instead of only at a model-wide level. [70] extends the NNGP correspondence to the feature-learning regime by retaining non-Gaussian statistics at finite- width in two-layer CNNs. Feature learning effects emerge through a shift of the GP’s target function; their theoretical framework involves solving a non-linear self-consistency equation for higher-order cumulants of the network, which act as effective couplings beyond the GP. Empirically, they validate the approach for a linear network in a teacher-student setup, including evidence of a phase transition between lazy and feature learning regimes. In this example, the second- 23 Note that the role of the IR here is the opposite of that in [68] mentioned elsewhere in this review, which considers a complementary application of RG in which the IR theory consists of a random DNN with unlearnable modes. 26 order NNGP kernel at initialization is the leading term in an effective description and does not change structurally during training, with feature learning effects entering as a data-dependent shift of the effective regression target (a non-centered GP prior). A complementary line of work makes the kernel itself a dynamical object during training. In [71], the authors develop a dynamical mean-field theory (DMFT) to describe feature learning in infinite-width networks, where interactions between layers are decoupled. Unlike the NTK/kernel regression picture, this formalism implements full gradient- descent parameterized by a feature-learning scale that smoothly interpolates between lazy and rich regimes in this limit. In this setting, the kernel is not just an initialization statistic, but a generator of order parameters – inner products of layer-dependent activations and gradients at pairs of time points. Kernel evolution is governed by self-consistent saddle-point equations (the DMFT equations), resulting in an effective flow, with the fixed point given by a non-trivial function of hyperparameters. The same authors extend this framework to include finite-width fluctuations over random initializations that are non-perturbative in the feature-learning scale [72]. They find kernels are effectively static in the lazy limit, while in rich regimes kernels and prediction fluctuations become dynamically coupled, with a variance governed by the DMFT. They use this framework to analyze when and how width, learning rate (including edge-of-chaos phenomena), depth, training set size, and linearity, control the onset and endpoint of feature learning in both toy settings and CIFAR-10. [73] connect feature learning to classical edge-of-chaos theory in deep nonlinear networks. They show the Bayesian prior can be written as an ensemble of GPs, and interpret feature learning as a reweighting of the components based on training data, framing kernel adaptation as arising from fluctuations in the prior. The capacity for adaptation (i.e., the flexibility afforded by these fluctuations) is maximized at the critical point separating ordered and chaotic phases, where the relevant response functions become large. This provides a theoretical link between criticality, response functions, and feature scale: networks initialized near criticality have the strongest capacity for kernel adaptation. Interpreting this through our lens, it suggests that feature relevance is sensitive to scale and the model’s effective regime (IR basin) determined by training. B.1.4 Physics Perspective: Scaling Laws and Phase Transitions A useful distinction when conceptualizing feature learning is between kernel rescaling – where training changes the overall magnitude of the kernel from initialization, leading to GP-like generalization – and kernel adaptation – where training induces directional, data-dependent changes to the effective kernel. This rescaling v. adaptive axis asks ‘what changes about the kernel?’ during training. A second, largely orthogonal axis concerns different scaling regimes, in particular mean-field scaling (saddle-point approximation suffices) and standard scaling (where addition corrections are required). [74] bridge these perspectives by rewriting the posterior over network outputs as a variational problem, and show that different choices of the dimensionality – an order parameter of this problem – recover either rescaling (low dimensionality) or adaptive (high dimensionality) theories. Separately, a systematic expansion of the output distribution adds the second axis to this picture. For the mean network output in the special case of linear networks, they find that kernel adaptation can be reduced to an effective rescaling—explaining why some feature-learning phenomena do not appear in rescaling-only theories. However, even in this case the framework captures adapative effects (i.e., of the output covariance) that cannot be captured by rescaling alone. This distinction aligns with our agenda: rescaling corresponds to a change in the scale of effective description, while adaptation corresponds to a change in its content. As evidenced by the many approaches in this review, the abundance of factors that contribute to learning make our efforts to develop a comprehensive theory exceedingly difficult. In a renormalisation framing, this ‘curse of detail’ makes defining a single coarse-graining scale similarly hard to do. Through the lens of sample complexity, [75] propose heuristic scaling arguments for predicting when different patterns of feature learning emerge. Rather than solving full DMFT equations numerically, they develop dimensional arguments that reproduce known scaling exponents and extend to complex architectures including three-layer nonlinear networks and attention heads. This provides estimates for the data and width scales at which transitions between lazy and feature-learning regimes occur—directly relevant to sample complexity in interpretability. If a task requires a threshold amount of data before features become relevant, interpretability tools assuming stable features may fail in the low-data regime. In [76], the authors develop a solvable random-feature model to describe various scaling laws – in time, model size, and dataset size – and their various regimes of validity 24 . Relating to earlier work, they find a DMFT description of the large-width asymptotics, and solve the corresponding response functions exactly to find that the loss scales more quickly in larger models. For power-law 24 We list just a few of these here; an expanded list can be found in the paper. 27 distributed features, they predict that performance scaling with training time and model size follows different power-law exponents, leading to an asymmetric strategy for achieving compute-optimality. Importantly, they empirically verify that while random feature (fixed-kernel) networks follow linearized scaling laws, they fall short of the compute-optimal learning curves set by feature learning networks, suggesting that true understanding of this frontier will depend on a more ‘mechanistic theory of feature learning.’ Finally, [77] apply an adaptive kernel approach to study grokking to understand if this phenomenon – a sudden increase in test accuracy before generalization emerges – is really out of reach of lazy/ GP theories of learning. Analyzing teacher-student models on two tasks (cubic-polynomial and modular-addition), they demonstrate a mapping between grokking and the theory of first-order phase transitions. Before grokking, the network is well-described by Gaussian feature learning, marked by the smooth adaptation of pre-activation covariances to the target directions in which their fluctuations – though peaked in the relevant directions – remain Gaussian. During grokking, features emerge discontinuously, yielding a mixed phase with pre-activation statistics described by a mixture of Gaussians. After the transition, the latent kernels develop entirely new features aligned with the teacher that alter sample complexity relative to GP limits. B.1.5A Mixed Persepctive: Saddle-to-saddle dynamics, different kinds of emergence, and the ultra-rich regime One can roughly trace two (non-mutually-exclusive) views of the emergence of structure in complex systems: the thermodynamic approach and the compositional, or mechanism-based, approach. This distinction is most famously made in Anderson’s “More is Different” [78] (see also [79] for a formal approach). The thermodynamic picture studies emergent macroscopic order in high-dimensional systems, usually viewed as composed of many identical microscopic particles (e.g., neurons in an ML context). Structure often appears at a phase transition and as emergent macroscopic order arising from complex microscopic interactions. The mechanism-based approach instead focuses on highly symmetric or ordered systems where low-dimensional pieces emerge locally (often via symmetry-breaking along single neurons or linear directions) and build up precise mechanistic structures, like clockwork components combining into a timepiece. Most of the work that has been presented here so far is in line with the thermodynamic picture. The mechanism-based view is perhaps a useful frame on the ideal output of an interpretation, perhaps obtained by extracting the relevant high-level mechanism after integrating out microscopic structure. In many idealized or highly symmetric settings, learning can even be fully understood in a mechanism-based setting. This framing is particularly suited to interpreting deep linear networks in the ultra-rich learning regime, and the related “saddle-to-saddle” view of learning dynamics. The ultra-rich regime corresponds to initialization and target scaling that starts close to an unstable saddle-point where all weights are zero, and early learning proceeds by symmetry breaking (neurons or features “rolling away” from the unstable equilibrium). This dynamical picture is neatly captured in [80] on alternating gradient flows. It finds that in a modular addition model with small initializations, single neurons spontaneously “jump” from being close to zero to emerging as macroscopic pieces of the known mechanism for learning the modular addition task (a variant of [53]) associated with Fourier modes. These large-scale neuron jumps alternate via “fast” but microscopic lazy learning-flavored re-adjustments of the small neurons as they react to a changing macroscopic target. In the weight landscape, this corresponds to the macroscopic parameters jumping between unstable saddles with better and better loss, with microscopic parameters quickly adjusting to metastable minima. Saxe et al. [81], and a cluster of related work, considers a similar phenomenon in the deep linear setting. Here a very small initialization in the ultra-rich regime bakes in two kinds of symmetry: 1.Localization to the task-aligned manifold (stable in idealized or simple settings). The training curve at each layer mostly stays within a lower-dimensional space of target-aligned matrices which are diagonal in the singular value basis of the target (a linear subspace of weight space in the case of deep linear models). 2. Scale degeneracy (unstable and trained away). At initialization, the singular values of the weight matrices are approximately zero (if they were exactly zero, this would be an unstable fixed point). In [81], the approximate localization symmetry is reinforced (preserved) over training, while scale degeneracy gets broken as singular values learn the directions of the target. The paper derives a formal model for this that also predicts, and empirically confirms, the exact rate of learning of each singular value. They find that large singular values get learned early and small singular values get learned late. 28 Domin ́ e et al. [82] studies the 2-layer linear network case in a more general context where the initialization is not assumed to be very small (i.e., outside the ultra-rich regime). By manually enforcing the basis-alignment property (item 1 above), they model the dependence of training curves on initialization in more detail and trace the difference between rich learning (where learning singular values exhibits symmetry breaking-like emergence) from lazy learning (where it does not, and learning is mostly described by adapting a single layer). They find that in addition to the initialization scale of each layer, an important parameter is the difference between initialization scales of the two layers (if one is too large, the asymmetry may result in only one layer effectively learning). This result partially extends to the case without basis-alignment, where exact learning dynamics is harder to predict but heuristically similar patterns hold. In “Get Rich Quick” (Kunin et al. [83]), the exact shape of transition from lazy to rich learning is theoretically fleshed out for a tiny width-one setting (the opposite of the usual large-width picture), and a series of conservations laws that control the dynamics in this case are found. They find that nonlinearity and depth accelerate the lazy-to-rich transition. The paper [84] theoretically formalizes an analog of the property of “symmetry restoration” (item 1 above), where small initializations bias towards lower-dimensional symmetric “mechanism-like” solutions in simple tasks — in this case, the relevant property is equivalent to sparsity in the neuron basis. It finds evidence of this in training vision models. Hoogland et al. [85] finds evidence of approximate saddle-to-saddle structure in early learning of text transformers. The paper also tracks local thermodynamic properties of the loss landscape, and interprets them by analogy with the singular learning formalism of Watanabe [86]. Singular learning formalism extends the study of low-dimensional symmetry breaking due to algebraic symmetry or Hessian degeneracy to a larger setting of geometric symmetry breaking due to analytic degeneracies of the loss landscape. The saddle-to-saddle perspective on learning is most directly applicable in early learning and simple or low-dimensional settings. In larger and more complex settings, exact saddle-to-saddle phenomena (such as meachanistic explanations built out of single-neuron structures) become enmeshed with high-dimensional and noisy phenomena. For example, Kumar et al. [87] studies modular addition in a rich (rather than ultra-rich) regime and a (roughly) mean-field setting. Here, the precision-tuned single-neuron jumps are replaced by high-dimensional statistical physics effects that explain grokking as a high-dimensional phase transition rather than a saddle-to-saddle phenomenon. Even settings that start ultra-rich can lead to more general settings with high-dimensional phenomena distinct from saddle-to-saddle replacing or interacting with the low-dimensional saddle-to-saddle structure. A few papers starting from the setting of Saxe et al. [81] look at a learning process that starts in the ultra-rich regime but ends up in the mean-field or other regimes, usually in the context of task reversal and continual learning. For example Lee et al. [88] studies a continual learning context where the target is explicitly changed during learning in a way that mimics implicit training realignment related to feature learning phenomena in large models. This paper finds exact ODE descriptions of the resulting non-saddlepoint regime with new phenomena (like non-monotonic forgetting). Despite the fact that the pure saddle-to-saddle picture of emergent low-dimensional transitions is insufficient to describe complex N learning completely, it provides a fully-tractable platform for understanding systems with significant complexity and interesting learning behavior. It is likely that experiments and intuitions from this simplified setting extend to nontrivial insights about late-training and non-ultra-rich settings, especially when studied at appropriate scales and renormalisation settings. B.2 Data-Space Renormalisation Work in this direction does not cleanly separate by discipline. Instead, we consider that data can possess at least three categories of hierarchical structure. This structure may exist among the input features, on the data distribution as a whole, or within the target function. With input-space hierarchical structure, each input is iteratively decomposable into parts and subparts that represent increasingly fine-grained details. In this case, a hierarchical compositional relationship exists among the features or dimensions of each sampled input. With data-space hierarchical structure, the data distribution is iteratively decomposable into increasingly fine-grained subdistributions or clusters of data points. In this case, the hierarchical relationship exists among groups of data points. Finally, for data with functional hierarchical structure, a hierarchical relationship exists among the target function’s components, such as terms in its series expansion. A given dataset may possess all, some, or none of these three complementary hierarchical structures. 29 B.2.1 A Mixed Perspective: Hierarchical Structure in Input Space Hierarchical structure in input space can have different interpretations, depending on the modality. In general, classification tasks in which the input features are related by a regular compositional structure can be efficiently learned by deep neural networks [89,90]. Several recent works analyzing the capabilities of transformer-based language models use a probabilistic context-free grammar (PCFG) as a synthetic model of hierarchical structure present in natural language [17,91,92]. A PCFG consists of a set of terminal symbols, a set of nonterminal symbols including a start symbol, and a set of probabilistically weighted production rules transforming a nonterminal symbol into a string of terminal and/or nonterminal symbols. A string generated by recursively applying a PCFG’s production rules has a hierarchical structure that imposes long-range correlations among the observed terminal symbols or tokens. A particular PCFG-based data model that has been applied in both language and image contexts is the random hierarchy model (RHM) [93]. The RHM is defined to have a regular structure and no ambiguities between the possible production rules that can yield an allowed string, making it an especially tractable model of hierarchical compositional data. A sequence generated by the RHM is represented by an s-regular syntax tree withLlevels. Each level has a distinct set of v possible nonterminal symbols, or terminal symbols for the final level. There aremrandomly chosen and frozen production rules per nonterminal symbol, which are constrained to generate unambiguous productions. In the basic RHM, all production rules are weighted with uniform probability. The RHM can be applied as a data model for classification tasks by predicting the highest-level latent (nonterminal) symbol as a label, or for next-token prediction tasks by masking and predicting the rightmost observable (terminal) symbol. The RHM has been further studied in a number of follow-up works. In [94] is acquired by deep neural networks, the authors analyze token-token correlations in the RHM, showing that a finite training set has the effect of imposing an effective context window that limits the range of correlations to which the learned model is sensitive. Because learning additional correlations reduces the loss, leading to a power-law scaling law for next-token prediction. In [95], the analysis of scaling laws is extended to the case in which the production rules have a power-law distribution at one level of the hierarchy. In this case, power-law scaling occurs for classification, while the exponent for next-token prediction is unchanged. In [96,97], the RHM is employed as a synthetic data model to analyze diffusion for both image and text data. By analyzing the RHM, the authors predict and verify on real data a phase transition in forward-backward diffusion experiments, where at a critical time scale the probability of reconstructing the true class drops to 0. Another data model with hierarchical input-space structure that has been applied for image data is an iterated function system (IFS) that generates a self-similar fractal. IFS fractals have been used to generate synthetic data for pretraining image models [98–100]. The self-similarity of IFS fractals as a model for image data is suggested by the fractal appearance of natural structures produced by physical processes [101]. B.2.2 A Mixed Perspective: Hierarchical Structure in Data Space In comparison to input-space structure, hierarchical structure in data space appears to be less studied. One approach is to use an IFS fractal model of the data distribution rather than of input images. Machine Learning and Fractal Geometry reviews various approaches for developing classical machine learning algorithms that incorporate an inductive bias for fitting data distributions with fractal geometry. [102] presents an expectation-maximization algorithm for fitting an IFS fractal model to data. [103] shows that data distributions with fractal structure can be expressed efficiently by deep networks, but not with shallow ones. However, to learn a classification task on such a distribution using either a deep or shallow network, the negative examples must be coarsely distributed rather than concentrated on the distribution’s fine structure. In addition, fractal-based methods are a well-studied technique for estimating the intrinsic dimension of data [104]. Another approach that models a data distribution as a stochastically self-similar fractal is based on percolation theory on a hypercubic lattice [25,105]. In this model, the data distribution consists either of discrete clusters with a power-law size distribution or a single dominant cluster, leading to a prediction of regime-dependent power-law neural scaling laws. Its self-similar fractal geometry suggests that data distributions can be efficiently mapped by learning sparse context features that hierarchically identify the clusters and subclusters to which inputs belong. [106] applies this data model to study scaling laws for a capacity-constrained predictor that balances between task-specific or general-purpose capabilities, finding that general capabilities emerge abruptly, before declining in relative importance. 30 B.2.3 A Mixed Perspective: Hierarchical Structure in Target Functions Finally, target functions associated with natural learning tasks may be constrained to have hierarchical structure. Functions with sparse compositional structure, consisting of a hierarchy of constituent functions that each depend only on a small number of input variables, can be expressed efficiently by deep but not shallow neural networks [107]. Because any function that is efficiently Turing-computable is compositionally sparse, [108] argue that the ability to exploit compositional sparsity is essential to deep learning’s success. One possible reason that natural target functions may be compositionally sparse is the hierarchical structure of physical generative processes that are characterized by local interactions between levels in a sequence of scales [109]. However, while deep networks can efficiently represent compositionally sparse functions, it does not guarantee that they can efficiently learn them. One function class that has been proposed as a model of functions that deep neural networks can efficiently learn hierarchically are staircase functions, which have a compositional structure in which high-order Fourier coefficients are built up from lower-order coefficients step by step [110, 111]. B.2.4 A Physics Perspective: Real-Space RG and Information-Theoretic Coarse-Graining One of the first works to relate deep learning and renormalisation was by Mehta and Schwab [9]. The authors argued that restricted Boltzmann machines (RBM) perform implicit renormalisation that is analogous to Kadanoff’s variational renormalisation scheme from condensed matter physics [112]. RBMs are unsupervised energy-based models that can learn high-dimensional probability distributions, and they are closely related to equilibrium spin models. However, later work suggested that, without additional assumptions, representations uncovered by RBMs do not always recover the types of large-scale features that are usually sought in renormalisation methods [33,109,113]. We also note that RBMs were found to be difficult to scale, and so they are not frequently used in modern ML systems. A different line of work has employed ideas from machine learning and information theory to develop explicit renormalisation methods. In the physics literature, renormalisation methods often require system-specific knowledge (such as the order parameter or the symmetries of the Hamiltonian) and hand-coded renormalisation rules. [33] developed a more general approach which discovers such information and renormalisations from data, rather than requiring that it be provided a priori. Their so-called “real-space mutual information” (RSMI) method is well-suited to studying homogeneous spatial systems. In such systems, it considers a local regionVthat is separated by a buffer regionBfrom a distant environment regionE. RSMI then identifies a lower-dimensional hidden variableHthat depends only on local regionVwhile also having large mutual informationI(H;E)withE. In simple terms,His the renormalized version ofVthat encodes relevant information about large-scale features. More recently, the RSMI method was improved by using powerful ML techniques for quantifying mutual information, and it was demonstrated on various non-trivial physical systems [114–116]. It has also been shown that, by considering graph-based distances, RSMI can be extended beyond homogeneous spatial systems to inhomogeneous systems on a graph [117]. RSMI has two intrinsic notions of scale. The first notion is the size of the buffer regionB, the non-predicted region that filters out short-range correlations, and thus sets the spatial scale of interest. The second notion is the dimensionality of the hidden variablesH, which sets the degree of information compression during coarse-graining. Recent work has shown that information compression can also be controlled by minimizing the mutual informationI(H;V ), rather than limiting the dimensionality ofH[114], revealing an important formal connection to the “information bottleneck” method [118–120]. At a high level, RSMI seeks features that capture long-range spatial correlations in statistical ensembles. Conceptually, it is related to methods for coarse-graining of dynamical systems that seek features that capture long-range temporal correlations [121–125]. In fact, recent work has developed a coarse-graining approach to dynamical systems that is similar to RSMI and information bottleneck [121], and demonstrated that it can recover important temporal features in complex physical and biological systems. Another formal approach to identify useful coarse-grainings has been introduced in [35]. Here, coarse-grainings are ‘useful’ to the degree they capture levels of description that are self-contained from an informational, causal, or computational perspective. Informational self-containment is operationalised in terms of information-theoretic terms: it corresponds to when a macro scale can optimally predict itself without considering information from finer scales. In contrast, causal and computational self-containement is formalized utilizing computational mechanics [126] – a framework that combines principles from statistical physics and theoretical computer science to investigate pattern formation in time series data. Results have shown that these properties are pervasive in paradigmatic models 31 of statistical physics and computational mechanics. Moreover, it has been found that symmetry (more specifically, dynamical equivariance) is a sufficient condition for the existence of such self-contained macro levels [127]. This work has also shown that familiar macroscopic quantities (such as magnetisation or particle count) correspond to informationally closed macro processes arising from underlying symmetries. B.2.5 A Physics Persepective: Statistical Inference as renormalisation Statistical Inference itself may be framed in terms of renormalisation. In [68] the link between traditional statistical inference using Bayesian methods and the exact renormalisation group flow was demonstrated as follows. First, consider a small incremental flow of data into your model. As the relatively infinitesimal data comes into your model one piece at a time you can use Bayes’s theorem to update your model parameters. Of course, since the amount of new data is small as compared with the amount of previous data on which the prior is based the parameter updates will be similarly infinitesimal. This infinitesimal Bayesian update equation can then be recast in terms of a first order differential equation for the change in the model parameters as a function of the amount of data. This first order differential equation can be reinterpreted as an exact RG flow equation, describing how physical parameters change with “scale”. In fact, we can read off two relevant scales that appear as a ratio describing how model parameters flow. First there is a scale associated with the model itself. This is the Fisher information metric for the model parameters. The Fisher information in some sense measures the distance in model space and so its appearance as a scale associated with the model is perfectly natural. Second, there is the scale simply given by the amount of data. Again this is natural and intuitive, the main scale on which all learning depends is data quantity. What is useful about the quantitative relationship that is described from the continuous Bayesian updating is that it provides a detailed description of the interplay between scales in the model given by the Fisher information and scales in data. Beyond the mathematics of the RG flow equation and its equivalence to the Bayesian updating equation, the basic intuition that this work demonstrates is that in traditional RG flow, one is throwing away information through “integrating out”; in statistical inference one is doing the opposite, using Bayes’s theorem to add information to your model from new data. The fact that one is sort of the inverse of the other is reflected in the fact that when mapping the theories the physical scale in the RG equation becomes mapped to the inverse of the data quantity in the Bayesian update equation. Thus, large data corresponds to the ultraviolet and small data the infrared. This is of course natural from the physics perspective, as we observe more about the universe we can distinguish between different possible UV theories. Data provides the information that allows the inverse of traditional RG to flow from IR to UV. Well observed phenomena are reproduced such as the variance in the statistical model as a function of data quantity (for small amounts of data, model variance is high, for large amounts of data the model asymptotes according to the central limit theorem). This description also allows a principled approach to model parameter pruning. One can implement a Bayesian renormalisation flow: progressively discarding parameter directions that the data cannot resolve (i.e. ‘sloppy’ modes) while retaining the informative, resolvable ones. Fisher information, in this context, measures how sensitively a probability distribution responds to changes in a parameter — directions with high Fisher information covary strongly with the output of the model, while those with low Fisher information blur into statistical noise. In settings with an underlying physical cutoff, this recovers conventional renormalisation [128], but in general it provides a principled compression/inference scheme that respects what data actually supports. In the context of neural networks, this perspective suggests that multi-scale properties of data are reflected in parameter directions with differing scales of Fisher information, so that coarse-graining by informational relevance naturally defines scales intrinsic to the model. It was further shown that this aligns with the information bottleneck and can be realized in toy neural-network pruning experiments, where parameters are hierarchically organized by their informational significance. The technique has been demonstrated with a toy autoencoder, to perform systematic model pruning. This can be thought of as a coarse-graining scheme over model parameters at each time-step of training in a way that preserves downstream performance metrics, via an information-geometric UV cutoff in parameter space. Features separated by distances smaller than a fixed resolution scale no longer contribute individually to the model’s learning and inference processes; accordingly, the model log-likelihood is oblivious to explicit short-range data (feature) interactions. As a concrete example, information-theoretic RG over UV field modes in a model initialized at NNGP, trained via SGD atL 2 loss and leading to a different NNGP, simplifies to “mass renormalisation” of a dual free statistical field theory, leading to rescaling of parameter means and covariances. This framework does not require large hyperparameter or parameter assumptions that are traditionally required by methods such as random matrix theory or dynamical mean field theory. 32 Alternatively, encoding data itself may be framed through the lens of field-theoretic renormalisation. It was found [129] that one may design an encoder-decoder architecture whose latent space is explicitly composed of n-point correlation functions (or cumulants) of the input data. The NCoder treats images (or other high-dimensional data) as draws from a lattice field theory, then reconstructs them via a perturbative expansion in correlation functions—analogous to building an effective action order by order in quantum field theory. In doing so it establishes a correspondence between perturbative renormalizability and model sufficiency: if only a finite number of correlation functions are needed, the data generating model is heuristically ‘renormalizable’. The method was demonstrated on MNIST, with benchmarks showing that generated images can be classified correctly using only up to the 3-point functions of the latent summary statistics. This suggests, empirically, that in some sense MNIST data samples are perturbatively renormalizable. Thus one might view NCoder as implementing a renormalisation-like compression in statistical observable space (moments and cumulants) while Fisher renormalisation works in parameter space; comparing them could shed light on how models choose which scales of structure to maintain or discard. B.2.6 Compressibility and Interpretability Similar work questions how different training dynamics encode features [130]; it is known that when comparing grokking (i.e. sudden generalization after memorization) with steady training regimes, the same underlying features may be learned but are represented with markedly different efficiency and compressibility. In particular, a ‘compressive regime’ emerges in steady training where there is a linear trade-off between model loss and compressibility, a phenomenon absent in grokking. While not immediately framed directly in terms of Fisher information, this compressibility picture resonates with the Fisher perspective: parameter directions that carry little information are effectively suppressed in the compressive regime, whereas grokking does not enforce such selectivity. One may consider this akin to implicit renormalisation. In the present framework, this supports the idea that reliable interpretations of compressed representations should be limited to scales where features remain distinguishable in a Fisher-information sense, providing a natural boundary for abstraction within the data structure. The work also shows that models which learn through grokking are not necessarily more compressible in terms of Bayesian Renormalisation than those which learn through steady learning. A related picture of compressibility finds direct support in recent work on sparse autoencoders. Matching pursuit SAEs [131] replace the standard shallow encoder with a greedy iterative process: at each step, the algorithm selects whichever dictionary feature best explains the current residual, subtracts its contribution, and repeats. Features selected earlier in this process tend to be coarser-grained. Practically, this suggests neural networks may be able to perform a form of lazy evaluation over their feature hierarchy, first resolving coarse-grained concepts before committing computational resources to finer distinctions. Compressibility, in this sense, is the freedom to operate at coarse or fine resolution as needed. Matryoshka SAEs [23] arrive at a similar conclusion via a different route. By training nested SAEs at progressively increasing widths, they find that capacity-constrained (narrow) models preferentially learn coarse features, with finer structure emerging only as width increases. That both methods recover the same coarse-to-fine structure strengthens the case that this organization is intrinsic to learned representations. C Glossary C.1 Physics / RG Terminology Coarse-graining: Any map from a fine-grained description (microstates, parameters, features) to a coarser one, typically by aggregating or integrating out degrees of freedom. Correlation length:A scale (ξ) characterizing how quickly correlations decay in space or time. Finiteξtypically implies exponential decay; (ξ →∞) at critical points implies scale-free, power-law correlations. Critical exponent / scaling dimension: An exponent characterizing how a quantity scales near a fixed point (e.g. with correlation length, temperature, or system size). At critical points, these are the usual critical exponents; more generally, they are eigenvalues/eigenvectors of the RG linearization. 33 Critical point:A point controlling a continuous phase transition, where correlation length becomes very large and many observables exhibit non-analytic, power-law behavior over many scales. Critical points are where universality is most dramatic. Effective field theory (EFT) / effective theory:A theory valid only up to some scale, written in terms of degrees of freedom and interactions relevant at that scale. In this paper, “effective theory” is used broadly for any coarser description of a model that preserves specified observables up to a tolerance. Fixed point (of the RG flow): A point in theory space that is invariant under RG flow (up to rescaling). Near a fixed point, the theory often exhibits simple scaling structure. Hierarchical conditional independence (as used in this paper):A Markov-like property in scale. For a good RG scheme, coarse variables at a given scale act as sufficient statistics for the finer variables directly beneath them, for the observables we care about. Conditioned on coarse variables, fine-scale degrees of freedom in one region are approximately independent of distant regions and have bounded influence on macroscopic observables (See Section B.1.5 for related formalizations using mutual information and causal states). Infrared (IR) / Ultraviolet (UV):“IR” refers to large-distance / low-energy / coarse scales; “UV” refers to short- distance / high-energy / fine scales. In this paper, IR typically corresponds to scales where safety-relevant observables live; UV corresponds to microscopic parameters or very fine features. Locality (and quasi-locality): The property that interactions in an effective theory involve only nearby degrees of freedom (or decay quickly with distance). Observable:A measurable quantity or behavior of interest used to connect theory with experiment. In a N, this could be loss on a task, performance on a benchmark, specific probe scores, statistics of internal activations, or response to interventions. Phase transition (continuous):A non-analytic change in macroscopic behavior as parameters are varied, controlled by a critical fixed point. Often accompanied by divergent correlation length and scale-invariant fluctuations. Relevant / irrelevant / marginal (directions or operators):A characterization of how linear perturbations around a point in theory space behave under an RG flow. Relevant directions grow under coarse-graining and strongly affect long-distance behavior. In an N context, these are effective features whose perturbations significantly change chosen observables. Irrelevant directions shrink and become negligible in the IR. In an N context, these form a tail with bounded impact. Marginal directions neither clearly grow nor shrink at linear order and require higher-order analysis. renormalisation (RG):A procedure that relates microscopic descriptions of a system to effective descriptions at coarser scales, by integrating out fine-grained degrees of freedom and tracking how couplings change. RG flow:The trajectory a theory traces in coupling space as we change the scale (under repeated coarse-graining and rescaling). Encodes how the effective description evolves from UV to IR. Scaling law:A relationship showing how an observable changes under rescaling of length, energy, or other parame- ters—for example, power-law dependence near criticality. In this paper, we also use “scaling law” for empirical relations in NNs (e.g. loss vs. model size / data / compute). Separation of scales:A property of a renormalisation scheme that allows fine-grained details to be integrated out without appreciably influencing coarse observables. Practically: a small set of effective degrees of freedom dominates chosen observables, and the contribution of discarded modes can be bounded. Theory space (or coupling space):An abstract space whose coordinates are the couplings (parameters) of all opera- tors allowed in a theory. RG flows are curves in this space; points represent effective theories. Universality:The phenomenon where many microscopically different models flow under RG to the same IR fixed point and share the same long-distance behavior for a wide range of observables. Universality class:A family of models whose RG flows end at the same fixed point and therefore share the same asymptotic large-scale behavior, often characterized by common scaling laws and exponents. 34 C.2 Interpretability Terminology Causal mechanistic decomposition: A way of representing a system as a hierarchy of interacting causal mechanisms such that the system’s behavior factorizes into these modules while preserving key causal counterfactuals across levels of abstraction. We treat this as a target-like structure that a successful RG scheme would approximate across scales. Circuit:A structured collection of internal components (neurons, heads, features, weights) whose combined activity implements a recognizable sub-computation (e.g., an induction head, an IOI circuit). Critical-like regime (in NNs):A region of parameter or scale space where small changes in architecture, training, or data lead to sharp or qualitative changes in effective descriptions or capabilities (e.g., capability jumps, phase-like transitions in behavior). Potentially analogous to critical regions in statistical mechanics and useful as diagnostics. Effective feature / effective degree of freedom:A coarse, possibly composite feature that emerges under a renormal- isation scheme and captures most of the contribution to some observable at a given scale. The units in which we want to express the effective theory of a model. Explicit RG (tool): A post-hoc procedure that takes a trained model or representation and constructs a coarser description (e.g., spectral truncation, clustering features, layerwise abstraction), aiming to mimic an RG step in a defined feature space. Feature (in this paper):Any component that meaningfully represents model internals in a way that can be understood by a human. Examples include a direction in activation space, an eigenvector of a kernel or covariance operator, a sparse autoencoder (SAE) feature, or a component of a generative data model. Implicit RG (theory):A theoretical description of how N training or inference implicitly implements an RG-like organization of features. Mechanistic interpretability:The process of understanding models in terms of circuits, features, and algorithms—identifying internal structures (e.g., attention heads, MLP neurons, feature combinations) that implement particular computa- tions. Representation / internal representation:The vector of activations (or a function thereof) at some layer or part of the network, taken as an internal state encoding information about inputs and context. Safety guarantee / bound:A formal statement bounding the impact of a safety-critical behavior. In this draft, we consider these to be of the form: “conditional on this effective description, perturbations confined to the irrelevant subspace cannot change observableXby more thanε.” RG-inspired separation of scales is the structural property that would make such statements possible in restricted settings. Sparse autoencoder (SAE) feature:A learned sparse basis function over activations, used to define interpretable features by reconstructing activations in a sparse code. Susceptibility (in NNs):A measure of how sensitive a model, circuit, or feature is to a structured perturbation (e.g., feature ablation, parameter change, input intervention). Inspired by linear-response/susceptibility in physics; used here as a way to quantify relevance and bound influence. Universality-like behavior (in NNs): Any regime where families of models (e.g., similar architectures and training recipes) share stable effective descriptions and scaling relationships for certain observables, even if not in the strong physics sense of universality classes at fixed points. 35