π€ AI Summary
Existing post-training methods for language models suffer from low efficiency, high computational cost, or insufficient control precision: weight-based fine-tuning is expensive, while prompt tuning relies on manual trial-and-error. Although activation steering (AS) shows promise, it typically requires handcrafted prompt pairs or labor-intensive feature annotation, lacking plug-and-play applicability. This paper introduces Painless Activation Steering (PAS)βthe first fully automated, lightweight AS framework that generates targeted steering vectors solely from labeled data, without manual prompt engineering or feature labeling. We further propose iPAS, an enhanced variant that significantly improves causal steering capability and synergizes effectively with in-context learning (ICL) and supervised fine-tuning (SFT). Evaluated across three open-source models and 18 diverse tasks, PAS consistently improves behavioral task performance and yields gains of 10.1%, 5.2%, and 34.8% on bias mitigation, moral reasoning, and alignment tasks, respectively.
π Abstract
Language models (LMs) are typically post-trained for desired capabilities and behaviors via weight-based or prompt-based steering, but the former is time-consuming and expensive, and the latter is not precisely controllable and often requires manual trial-and-error. While activation steering (AS) promises a cheap, fast, and controllable alternative to the two existing post-training methods, current AS techniques require hand-crafted prompt pairs or labor-intensive feature annotation, making them more inconvenient than the plug-and-play methods such as Reinforcement Learning (RL) and Supervised Fine-Tuning (SFT). We introduce Painless Activation Steering (PAS), a family of fully automated methods that make AS readily usable with any given labeled dataset, with no need for prompt construction, feature labeling, or human intervention. We evaluate PAS on three open-weight models (Llama3.1-8B-Instruct, DeepSeek-R1-Distill-8B, and Nous-Hermes-2) and 18 tasks; we find that PAS reliably improves performance for behavior tasks, but not for intelligence-oriented tasks. The introspective variant (iPAS) delivers the strongest causal steering effects (10.1% on Bias, 5.2% on Morality, and 34.8% on Alignment). We also show PAS delivers additional gains on top of In-Context Learning (ICL) and SFT. PAS constructs a fast, lightweight activation vector that can be cheaply trained, easily stored, and activated at will. Our results provide a characterization of where AS helps, where it fails, and how to deploy it as a practical, automated LM post-training option.