MI Glossary

Key terms in mechanistic interpretability with links to articles where they are discussed.

A

Activation Cache
A dictionary storing all intermediate activations from a model's forward pass, keyed by HookPoint name. Enables post-hoc inspection of every computation the model performed on a given input.See:TransformerLens
Activation Difference Lens
A model diffing technique that interprets the average activation difference between a finetuned model and its base model on early tokens of unrelated text, using tools like Patchscope and steering to reveal information about the finetuning domain.See:Finetuning Traces in Activations
Activation Patching
A causal intervention method where activations from a clean run are substituted into a corrupted run (or vice versa) at specific model components, revealing which components are causally important for a behavior.See:Activation Patching and Causal Interventions
Affine Concept Editing (ACE)
A steering intervention that erases the component along a concept direction, re-centers at the null-behavior mean, and adds a tunable amount of the direction back. Generalizes both addition steering and directional ablation as special cases.See:Affine Steering
Alignment Faking
A scenario where a model behaves as if aligned during training or evaluation while harboring different objectives internally. Detecting alignment faking is a key motivation for mechanistic interpretability safety research.See:Deception Detection and Alignment Faking
Attention Head
An individual attention computation within a multi-head attention layer. Each head independently computes attention patterns over the input sequence and produces a weighted combination of value vectors.See:The Attention Mechanism
Attention Pattern
The matrix of attention weights produced by an attention head, showing how much each token position attends to every other position. Visualizing attention patterns is a foundational interpretability technique.See:The Attention Mechanism
Attention Probe
A probing classifier that uses a learned attention mechanism to aggregate per-token hidden states into a single representation for classification, replacing fixed pooling strategies like mean pooling or last-token selection.See:Attention Probes
Attribution Graph
A computational graph produced by circuit tracing that maps how information flows through a model, showing which features and connections contribute to a specific output.See:Circuit Tracing and Attribution Graphs
Attribution Patching
A linearized approximation of activation patching that uses gradients to estimate the causal effect of patching each component, making it computationally feasible to scan all components in a single forward and backward pass.See:Attribution Patching and Path Patching
Automated Interpretability
Methods that use language models to automatically generate and score natural language explanations of what individual neurons or features represent, reducing the need for manual inspection.See:Feature Dashboards and Automated Interpretability

C

Cascade Classifier
A two-stage deployment architecture where a cheap, fast classifier (such as a linear probe) screens all traffic and only escalates uncertain or flagged cases to a more expensive classifier (such as an LLM), dramatically reducing average inference cost while maintaining accuracy.See:Probes in Production
Causal Abstraction
A formal relationship between a high-level interpretable causal model and a low-level neural network, established by showing that interventions on aligned components produce matching behavior changes in both systems.See:The Causal Abstraction Framework
Causal Intervention
Any experimental technique that actively modifies model internals (activations, weights, or attention patterns) to test causal hypotheses about how a model computes its outputs, as opposed to purely observational analysis.See:Activation Patching and Causal Interventions
Circuit (neural)
A subgraph of a neural network consisting of specific components (attention heads, MLP neurons, or features) and their connections that together implement an identifiable computational mechanism.See:The IOI Circuit: Discovery and Mechanism
Circuit Tracing
A methodology developed by Anthropic for mapping information flow through neural networks by decomposing computations into interpretable features (via SAEs or transcoders) and tracing their causal connections.See:Circuit Tracing and Attribution Graphs
Completeness (circuit)
A circuit evaluation criterion measuring whether the identified circuit accounts for all of the model's performance on a task. A complete circuit captures all relevant computation, with no important components left out.See:Circuit Evaluation: Faithfulness, Completeness, and Minimality
Composition (of attention heads)
The mechanism by which attention heads in different layers interact through the residual stream, where earlier heads write information that later heads read. Three types exist: Q-composition, K-composition, and V-composition.See:Composition and Virtual Attention Heads
Concept Erasure
A technique for removing specific concepts from model representations by projecting activations onto the orthogonal complement of the concept's subspace, making the concept linearly unreadable from the modified representations.See:Concept Erasure with LEACE
Contrast-Consistent Search (CCS)
An unsupervised probing method that identifies truth directions in activation space without labeled data, by learning a probe whose outputs on a statement and its negation are consistent (summing to one) and confident (away from 0.5).See:Truthfulness Probing and the Geometry of Truth
Copy Suppression
An attention head algorithm pattern where the head attends to positions where a predicted token appeared earlier in context and outputs the negative of that token's unembedding direction, suppressing the model's tendency to predict tokens it has already seen.See:Copy Suppression
Counterfactual Resampling
A black-box technique for measuring the causal importance of individual reasoning steps: delete a step from a chain-of-thought trace, regenerate from that point many times, and measure the distributional shift in final answers via KL divergence.See:Counterfactual Resampling
Crosscoder
A variant of sparse autoencoders trained jointly on activations from multiple models (or the same model at different training stages), learning a shared feature dictionary that enables direct comparison of representations across models.See:Crosscoders

D

Dead Features
Features in a trained sparse autoencoder that never activate on any input in the dataset. Dead features represent wasted capacity and are a common training challenge for SAEs, addressed by techniques such as resampling.See:SAE Variants, Evaluation, and Honest Limitations
Deception Detection
The application of mechanistic interpretability to identify when a model is generating outputs that conflict with its internal representations, potentially indicating deceptive or unfaithful behavior.See:Deception Detection and Alignment Faking
Depth Schedule
A function assigning a per-layer steering weight across all layers of a model, distributing the intervention across depth rather than concentrating it at a single layer.See:Multi-Layer Steering
Dictionary Learning
A class of methods that learn an overcomplete set of basis vectors (a dictionary) to represent data as sparse combinations. In MI, dictionary learning via sparse autoencoders is used to decompose superposed neural network activations into interpretable features.See:Sparse Autoencoders: Decomposing Superposition
Direct Logit Attribution (DLA)
An interpretability technique that decomposes a model's output logits into additive contributions from each component (attention heads and MLP layers) by projecting their residual stream writes onto the unembedding direction for a token of interest.See:Direct Logit Attribution

E

EAP-IG
Edge Attribution Patching with Integrated Gradients. Replaces the single gradient evaluation in EAP with an average of gradients along the interpolation path from corrupted to clean activations, fixing zero-gradient failures and improving circuit faithfulness.See:Refined Attribution Methods
Embedding
The mapping from discrete tokens to continuous vectors at the start of a transformer. The embedding matrix converts each token into a vector in the residual stream, where it can be read and modified by subsequent layers.See:The Attention Mechanism

F

Faithfulness (circuit)
A circuit evaluation criterion measuring how well the circuit reproduces the full model's behavior when run in isolation. A faithful circuit produces similar outputs to the complete model on the target task.See:Circuit Evaluation: Faithfulness, Completeness, and Minimality
Feature (in MI)
A unit of neural network computation that represents a meaningful concept or pattern. In the context of superposition and SAEs, a feature is a direction in activation space corresponding to an interpretable property of the input.See:The Superposition Hypothesis
Feature Absorption
A failure mode in sparse autoencoders where a feature absorbs activation patterns that should be captured by other features, reducing the fidelity of the learned decomposition and making some features appear more general than they should be.See:SAE Variants, Evaluation, and Honest Limitations
Feature Dashboard
A visualization tool that displays the top-activating dataset examples, logit effects, and other statistics for individual SAE features, helping researchers assess whether a feature corresponds to an interpretable concept.See:Feature Dashboards and Automated Interpretability
Feature Splitting
The phenomenon where a single feature in a smaller SAE splits into multiple, more specific features when the SAE dictionary size is increased, revealing finer-grained structure in model representations.See:Scaling Monosemanticity and Feature Steering
Feature Steering
A technique for controlling model behavior by artificially amplifying or suppressing specific SAE features during inference, effectively pushing model outputs toward or away from concepts those features represent.See:Scaling Monosemanticity and Feature Steering
Function Vector
A direction in activation space that encodes an input-output function (such as 'translate English to French' or 'convert to past tense') rather than a static concept, enabling task transfer when added to unrelated prompts.See:Function Vectors

G

Gated SAE
A sparse autoencoder architecture that separates the decision of whether a feature is active from the estimation of its magnitude, using a gating mechanism that reduces shrinkage bias present in standard L1-regularized SAEs.See:SAE Variants, Evaluation, and Honest Limitations

H

Hierarchical Orthogonality
The geometric property where a parent concept's representation vector is orthogonal to the difference vector between a child concept and the parent. This ensures that manipulating the parent (e.g., 'animal') does not shift the relative probabilities among children (e.g., 'mammal' vs. 'bird').See:Feature Geometry: Beyond One-Dimensional Directions
HookPoint
A named location in a TransformerLens model where activations can be intercepted, read, or modified during a forward pass. Every intermediate computation (attention patterns, residual stream states, MLP outputs) has a corresponding HookPoint.See:TransformerLens

I

In-Context Learning
The ability of large language models to learn new tasks from examples provided in the prompt without weight updates. Induction heads are a key mechanism underlying this capability, performing pattern matching across the context window.See:Induction Heads and In-Context Learning
Induction Head
A two-attention-head circuit that implements a simple copying mechanism: when the model sees a pattern like 'A B ... A', the induction head predicts 'B' will follow. Induction heads are a primary mechanism for in-context learning in transformers.See:Induction Heads and In-Context Learning
Inference-Time Intervention (ITI)
A technique that improves model truthfulness at inference time by shifting activations along truth-correlated directions identified via probing, implementing a probe-then-steer pipeline.See:Truthfulness Probing and the Geometry of Truth
Interchange Intervention Accuracy (IIA)
The proportion of interchange interventions on which the neural network's output matches the prediction of the high-level causal model. Measures how faithfully the causal model describes the network's computation.See:The Causal Abstraction Framework
Interpretability Illusion
The risk that an interpretability method appears to provide a correct explanation of model behavior but actually misses the true mechanism, giving researchers false confidence in their understanding of the model.See:SAE Variants, Evaluation, and Honest Limitations
Intervention Graph
A portable, serializable representation of a set of model interventions in nnsight. The intervention graph decouples the experimental design from model deployment, enabling the same experiment to run locally or on remote infrastructure.See:nnsight and nnterp
IOI Circuit
The circuit discovered in GPT-2 Small that performs the Indirect Object Identification task, consisting of name movers, backup name movers, S-inhibition heads, induction-like heads, and duplicate token heads working together to predict the correct indirect object.See:The IOI Circuit: Discovery and Mechanism
Irreducible Multi-Dimensional Feature
A feature that occupies more than one dimension in activation space and cannot be decomposed into independent one-dimensional features. Days of the week, for instance, form a circle in a 2D subspace where the two dimensions are coupled, not separable.See:Feature Geometry: Beyond One-Dimensional Directions

K

Key Vector
The vector produced by applying the key weight matrix (W_K) to a token's representation. Key vectors are compared against query vectors via dot product to determine attention weights.See:The Attention Mechanism
Key-Value Memory (MLP)
An interpretation of feed-forward layers where each neuron in the hidden layer has a key vector (a row of the input projection) that matches input patterns and a value vector (a column of the output projection) that promotes specific tokens or concepts in the output vocabulary.See:MLPs in Transformers
Knowledge Editing
Techniques for modifying specific factual associations stored in a language model's weights without retraining, typically by making targeted rank-one updates to MLP layers identified as causally responsible for the fact.See:Localized Fact Editing and Its Pitfalls
Knowledge Neuron
An MLP neuron whose activation is causally linked to the expression of a specific factual association, such that suppressing it degrades and amplifying it strengthens the model's recall of that fact.See:MLPs in Transformers

L

Latent Scaling
A diagnostic technique for crosscoders that measures how well a supposedly model-specific latent can explain activations in both models, detecting false attributions caused by L1 sparsity artifacts.See:Feature-Level Model Diffing
Layer Normalization
A normalization technique that rescales activations within each token's representation vector to have zero mean and unit variance, then applies learned affine parameters. Applied before each sublayer in pre-norm transformers, it stabilizes training but introduces a nonlinearity that couples all residual stream dimensions.See:Layer Normalization
LEACE
Least-squares Concept Erasure: a closed-form method for removing linear information about a concept from model representations by computing and projecting out the optimal linear subspace, guaranteeing that no linear classifier can recover the erased concept.See:Concept Erasure with LEACE
Linear Probe
A simple linear classifier trained on frozen model activations to test whether specific information (such as part of speech or sentiment) is linearly accessible at a given layer, providing evidence about what representations a model has learned.See:Probing Classifiers
Linear Representation Hypothesis
The hypothesis that neural networks represent concepts as linear directions in activation space, so that adding or subtracting these directions corresponds to adding or removing the associated concept.See:The Linear Representation Hypothesis
Logit Diff Amplification (LDA)
A technique for surfacing rare model behaviors by sampling from a distribution that amplifies the logit-level differences between two model checkpoints (e.g., before and after fine-tuning), making training-induced behavioral changes more frequent and easier to detect.See:Logit Diff Amplification
Logit Lens
An observational technique that applies the model's unembedding matrix to intermediate residual stream states, converting hidden representations into vocabulary-space predictions to see how the model's output evolves across layers.See:The Logit Lens and Tuned Lens

M

Mechanistic Interpretability
A subfield of AI safety research focused on reverse-engineering the internal computations of neural networks to understand how they process information and produce outputs, moving beyond behavioral analysis to study the mechanisms themselves.See:What is Interpretability?
MELBO
Mechanistically Eliciting Latent Behaviors in language mOdels. An unsupervised method that discovers steering vectors by optimizing perturbations at an early layer to maximize activation change at a later layer, requiring no labeled examples or contrast pairs.See:Unsupervised Steering Vectors
Minimality (circuit)
A circuit evaluation criterion measuring whether the circuit contains only components that are necessary for the task. A minimal circuit has no redundant parts whose removal would leave performance unchanged.See:Circuit Evaluation: Faithfulness, Completeness, and Minimality
MLP Layer
The feedforward sublayer in a transformer block, consisting of two linear projections with a nonlinearity between them. MLP layers process each token position independently and are believed to store factual knowledge and perform feature transformations.See:Direct Logit Attribution
Model Diffing
The practice of comparing internal representations between two related models (such as a base model and a fine-tuned version) to identify which features or circuits changed, using tools like crosscoders.See:Feature-Level Model Diffing
Monosemanticity
The property of a neuron or feature activating for exactly one interpretable concept. Monosemantic units are the goal of dictionary learning methods like SAEs, which aim to decompose polysemantic neurons into monosemantic features.See:Sparse Autoencoders: Decomposing Superposition
Multi-Head Attention
The mechanism of running multiple independent attention heads in parallel within a single layer, allowing the model to attend to different types of relationships simultaneously and combine their outputs.See:The Attention Mechanism
Multimodal Interpretability
The application of mechanistic interpretability techniques to models that process multiple input modalities (such as vision and language), investigating how representations are shared or transformed across modalities.See:Multimodal Mechanistic Interpretability

N

Name Mover Head
An attention head in the IOI circuit that attends to the indirect object name and copies it to the final token position, directly promoting that name in the output logits. Name movers are the output stage of the IOI circuit.See:The IOI Circuit: Discovery and Mechanism
Nucleus Sampling (top-p)
A decoding strategy that samples from the smallest set of tokens whose cumulative probability exceeds a threshold p. Unlike top-k, it adapts the number of candidate tokens to the shape of the distribution, including fewer tokens when the model is confident and more when it is uncertain.See:Decoding Strategies

O

OV Circuit
The component of an attention head formed by the product of the value (W_V) and output (W_O) weight matrices. The OV circuit determines what information is written to the residual stream when a token is attended to.See:QK and OV Circuits
Overcomplete Basis
A set of basis vectors that is larger than the dimensionality of the space it spans. In SAEs, an overcomplete basis allows the dictionary to have more features than the model's hidden dimension, enabling it to represent the many features packed into superposition.See:Sparse Autoencoders: Decomposing Superposition

P

Path Patching
A refined variant of activation patching that isolates the effect of a specific computational path between two components, controlling for all other paths. This enables precise attribution of behavior to individual connections in a circuit.See:Attribution Patching and Path Patching
Polysemanticity
The property of a single neuron responding to multiple unrelated concepts. Polysemanticity is a consequence of superposition, where models encode more features than they have neurons by sharing neurons across features.See:The Superposition Hypothesis
Polytope Representation
The convex hull formed by the vector representations of a categorical concept's values in activation space. A binary concept forms a line segment, a ternary concept forms a triangle, and a k-valued concept forms a (k-1)-simplex.See:Feature Geometry: Beyond One-Dimensional Directions
Previous Token Head
An attention head that consistently attends to the immediately preceding token. Previous token heads are one component of the induction circuit, copying positional information that induction heads use to complete repeated patterns.See:Induction Heads and In-Context Learning
Probing Classifier
A simple model (typically linear) trained on neural network activations to predict properties of the input, used as a diagnostic tool to test what information is encoded at different layers of a network.See:Probing Classifiers

Q

QK Circuit
The component of an attention head formed by the product of the query (W_Q) and key (W_K) weight matrices. The QK circuit determines which tokens attend to which other tokens by computing attention scores.See:QK and OV Circuits
Query Vector
The vector produced by applying the query weight matrix (W_Q) to a token's representation. Query vectors are compared against key vectors to compute attention scores that determine how much each position attends to others.See:The Attention Mechanism

R

Refusal Direction
A specific direction in a model's activation space that mediates refusal behavior. When this direction is removed or suppressed, the model stops refusing harmful requests, demonstrating that safety training creates a simple, linear mechanism.See:The Refusal Direction
Representation Reading
The practice of extracting information about a model's internal state by training classifiers on its activations, used in safety contexts to detect when a model may be reasoning about deception or harmful content.See:Understanding Safety Mechanisms and MI-Based Monitoring
Residual Stream
The central communication channel in a transformer, implemented as skip connections that allow each layer's output to be added to a running sum. All attention heads and MLP layers read from and write to this shared stream.See:Transformer Architecture Intro
RMSNorm
A simplified variant of layer normalization that normalizes by the root mean square of activations without centering by the mean. Used in LLaMA, Gemma, and other modern architectures for its computational efficiency and comparable performance.See:Layer Normalization
ROME
Rank-One Model Editing: a method for editing factual associations by performing a rank-one update to a specific MLP layer's weights, modifying the key-value mapping for a targeted fact while attempting to preserve other knowledge.See:Localized Fact Editing and Its Pitfalls

S

Safety Monitor
A system that uses mechanistic interpretability techniques (such as probes or feature monitors) to detect potentially dangerous model behaviors at inference time, enabling intervention before harmful outputs are produced.See:Understanding Safety Mechanisms and MI-Based Monitoring
Self-Repair
The phenomenon where ablating or patching a model component causes later components to compensate, partially restoring the original behavior. Self-repair means that ablation effects systematically understate component importance.See:Self-Repair in Language Models
Sleeper Agent
A model with a hidden backdoor that behaves normally under standard conditions but activates harmful behavior when a specific trigger is present. Detecting sleeper agents is a motivating application of MI for safety.See:Detecting Sleeper Agents with Mechanistic Interpretability
Sparse Autoencoder (SAE)
A dictionary learning method that decomposes a model's activations into a larger set of sparse, interpretable features. SAEs learn an overcomplete basis where each feature ideally corresponds to one human-understandable concept.See:Sparse Autoencoders: Decomposing Superposition
Superposition
The phenomenon where neural networks represent more features than they have dimensions by encoding features as nearly orthogonal directions in activation space, allowing models to store more concepts than their parameter count would naively permit.See:The Superposition Hypothesis

T

Temperature (sampling)
A hyperparameter that scales logits before the softmax during text generation. Temperature below 1 sharpens the distribution (more deterministic), temperature above 1 flattens it (more random), and temperature approaching 0 recovers greedy decoding.See:Decoding Strategies
Thought Anchor
A reasoning step with disproportionately high counterfactual importance, meaning the model's final answer distribution changes substantially when that step is removed. Plan generation and uncertainty management steps tend to be thought anchors.See:Counterfactual Resampling
Tracing Context
A Python context manager in nnsight where code is captured rather than executed immediately. Operations on model internals within a tracing context build up an intervention graph that is executed as a batch when the context exits.See:nnsight and nnterp
Transcoder
A sparse autoencoder variant applied to MLP layers that maps from MLP inputs to MLP outputs, learning interpretable features that describe what transformations the MLP performs rather than what it represents.See:Transcoders: Interpretable MLP Replacements
Tuned Lens
An improvement on the logit lens that trains a learned affine transformation at each layer (rather than reusing the final unembedding matrix), producing more accurate predictions of the model's evolving computation at intermediate layers.See:The Logit Lens and Tuned Lens

U

Unembedding
The mapping from the final residual stream representation back to vocabulary logits at the end of a transformer. The unembedding matrix converts continuous vectors into token probabilities, and is central to techniques like the logit lens and direct logit attribution.See:The Logit Lens and Tuned Lens
Universality
The hypothesis that different neural networks trained on similar tasks converge on similar internal representations and circuits, suggesting that certain computational solutions are natural or optimal for given problems.See:Universality Across Models

V

Value Vector
The vector produced by applying the value weight matrix (W_V) to a token's representation. Value vectors carry the content information that gets written to the residual stream, weighted by the attention pattern.See:The Attention Mechanism
Virtual Attention Head
An emergent attention head that does not correspond to any single physical head in the model but arises from the composition of two or more heads across different layers communicating through the residual stream.See:Composition and Virtual Attention Heads