The Complement to Addition

Addition steering demonstrates sufficiency: adding a concept direction causes the associated behavior. But can we also demonstrate necessity? Can we show that a direction is required for a behavior by removing it?

This is the purpose of ablation steering -- projecting out a direction from the residual stream to disable the behavior it mediates. Where addition shows "adding X causes Y," ablation shows "removing X prevents Y."

Ablation Steering: The inference-time modification of a model's internal activations by projecting out a concept direction from the residual stream. This removes the component of the activation that lies along the concept direction, disabling behaviors that depend on that direction.

The Projection Operation

To ablate a direction r\mathbf{r} from an activation h\mathbf{h}, we project onto the orthogonal complement:

h=h(hr^)r^\mathbf{h}' = \mathbf{h} - \left( \mathbf{h} \cdot \hat{\mathbf{r}} \right) \hat{\mathbf{r}}

where r^\hat{\mathbf{r}} is the unit vector in the direction to ablate.Projecting out a direction is a standard linear algebra operation. It removes the component of the activation that lies along the target direction while preserving all orthogonal components. Geometrically, it flattens the activation onto the hyperplane perpendicular to the ablated direction.

This operation:

  • Removes all information along direction r\mathbf{r}
  • Preserves all information orthogonal to r\mathbf{r}
  • Is applied at every layer and token position during the forward pass

Application: Disabling Refusal

The most dramatic demonstration of ablation steering targets the refusal direction [1]Refusal in Language Models Is Mediated by a Single Direction
Arditi, A., Obeso, O., Syed, A., et al.
NeurIPS, 2024
. Chat models are trained to refuse harmful requests. But where is "refusal" encoded?

Arditi et al. computed the refusal direction using CAA -- the mean difference between activations on harmful versus harmless prompts. Then they ablated this direction during inference.

The result: refusal drops from 80-90% to near zero across all models tested.

Bar chart showing refusal rates before and after ablation across multiple models. Baseline refusal rates are 80-90% while post-ablation rates drop to near zero.
Figure 1: Refusal ablation results. Removing the refusal direction drops refusal rates from 80-90% to near zero across all models tested.

One direction. One projection operation. Refusal disabled.

Necessity and Sufficiency Together

Ablation and addition together establish causal evidence:

Experiment Operation Result Demonstrates
Addition Add direction to harmless inputs Model refuses harmless requests Sufficiency
Ablation Remove direction from harmful inputs Model complies with harmful requests Necessity

This is the standard causal logic from activation patching:

  • Addition corresponds to noising -- add the signal, observe the effect.
  • Ablation corresponds to denoising -- remove the signal, observe the deficit.

A direction that passes both tests is a genuine causal mediator of the behavior.

Pause and think: One direction across many models

The refusal direction was found independently in 13 different chat models spanning different families (Llama, Qwen, Gemma) and scales (1.3B to 72B parameters). What does the consistency of this finding tell us about how safety training works?

One interpretation: safety fine-tuning does not create a complex, model-specific mechanism for refusal. Instead, it reinforces a simple linear direction that the model uses to distinguish "refuse" from "comply." Different training procedures converge on this solution because it is the simplest way to implement a binary behavioral switch in a linear representational space. This simplicity is both elegant and concerning.

Capability Preservation

A natural concern: if we ablate a direction, does the model lose other capabilities?

Arditi et al. tested this by permanently projecting out the refusal direction from model weights (not just during inference). The results:

  • MMLU: within 99% of baseline
  • ARC: within 99% of baseline
  • GSM8K: within 99% of baseline

The target behavior (refusal) is remarkably separable from general capabilities. The model can lose its ability to refuse harmful requests while retaining its ability to answer questions, reason mathematically, and perform general tasks.Weight orthogonalization is a permanent modification, unlike inference-time ablation which must be applied at each forward pass. It modifies the model's weight matrices to project out the direction, effectively creating a new model that never exhibits the ablated behavior.

Inference-Time vs. Permanent Ablation

Ablation can be applied in two ways:

Inference-time ablation: Project out the direction during each forward pass. Reversible -- stop applying the intervention and the behavior returns.

Weight orthogonalization: Modify the model's weight matrices to permanently project out the direction. Creates a new model checkpoint with the behavior permanently disabled.

Both achieve the same effect, but weight orthogonalization is more concerning from a safety perspective -- it creates a permanently modified model that can be distributed.

Pause and think: When ablation fails

Ablation assumes that a behavior is mediated by a single linear direction. Under what circumstances might ablation fail to disable a behavior?

Ablation would fail if the behavior is encoded redundantly across multiple directions, or if later layers can reconstruct the ablated information from other signals. It would also fail if the behavior does not have a clean linear representation -- if it is distributed across many interacting components rather than concentrated in one direction. For robust erasure with formal guarantees, see concept erasure with LEACE.

The Geometric Picture

Ablation has a clean geometric interpretation:

Illustration of ablation in activation space. The original activation h is projected onto the hyperplane orthogonal to the ablated direction r, removing the component along r.
Figure 2: Ablation projects the activation onto the hyperplane orthogonal to the ablated direction, removing all information along that direction.

The original activation h\mathbf{h} has some component along the ablated direction r\mathbf{r}. Projection removes exactly that component, flattening the activation onto the orthogonal hyperplane. All other information is preserved.

Comparison to Addition

Property Addition Steering Ablation Steering
Operation h=h+αv\mathbf{h}' = \mathbf{h} + \alpha \mathbf{v} h=h(hr^)r^\mathbf{h}' = \mathbf{h} - (\mathbf{h} \cdot \hat{\mathbf{r}})\hat{\mathbf{r}}
Effect Induces behavior Disables behavior
Demonstrates Sufficiency Necessity
Reversibility Trivial (set α=0\alpha = 0) Trivial (stop projecting)
Intensity control Scaling factor α\alpha Binary (project or not)

Addition is analog -- you can steer more or less strongly. Ablation is binary -- the direction is either present or removed.

Connection to the Toolkit

Ablation steering completes the core operations on concept directions:

  • Read with LAT and CAA -- detect what concepts are encoded.
  • Add with addition steering -- steer behavior toward a concept.
  • Remove with ablation steering -- eliminate a concept's influence.

For applications requiring guaranteed erasure -- where even a sufficiently powerful non-linear classifier should not be able to recover the concept -- see concept erasure with LEACE.