Why Steering Works: Toward a Unified View of Language Model Parameter Dynamics

📅 2026-02-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing control methods for large language models—such as fine-tuning, LoRA, and activation interventions—lack a unified theoretical framework, hindering systematic comparison and mechanistic understanding. This work proposes a preference–utility analysis framework that unifies diverse interventions as dynamic weight updates driven by control signals. By leveraging polarity-contrastive examples on the log-odds scale, the framework jointly quantifies a model’s preference (bias toward a target concept) and utility (coherence of generated text). Drawing on activation manifold theory, we reveal a pervasive trade-off wherein increased preference typically degrades utility, stemming from representational drift away from the effective generative manifold. Guided by this insight, we introduce the SPLIT steering algorithm, which enhances target preferences while more effectively preserving generative utility, thereby demonstrating the universality and controllability of this trade-off.

Technology Category

Application Category

📝 Abstract
Methods for controlling large language models (LLMs), including local weight fine-tuning, LoRA-based adaptation, and activation-based interventions, are often studied in isolation, obscuring their connections and making comparison difficult. In this work, we present a unified view that frames these interventions as dynamic weight updates induced by a control signal, placing them within a single conceptual framework. Building on this view, we propose a unified preference-utility analysis that separates control effects into preference, defined as the tendency toward a target concept, and utility, defined as coherent and task-valid generation, and measures both on a shared log-odds scale using polarity-paired contrastive examples. Across methods, we observe a consistent trade-off between preference and utility: stronger control increases preference while predictably reducing utility. We further explain this behavior through an activation manifold perspective, in which control shifts representations along target-concept directions to enhance preference, while utility declines primarily when interventions push representations off the model's valid-generation manifold. Finally, we introduce a new steering approach SPLIT guided by this analysis that improves preference while better preserving utility. Code is available at https://github.com/zjunlp/EasyEdit/blob/main/examples/SPLIT.md.
Problem

Research questions and friction points this paper is trying to address.

language model steering
parameter dynamics
preference-utility trade-off
activation manifold
control interventions
Innovation

Methods, ideas, or system contributions that make the work stand out.

unified framework
preference-utility trade-off
activation manifold
dynamic weight updates
SPLIT steering
🔎 Similar Papers
No similar papers found.