🤖 AI Summary
Existing large language models (LLMs) rely on heuristic activation steering for behavioral control, lacking theoretical grounding for selecting intervention locations and intensities—hindering reliable alignment with trustworthiness attributes such as factual consistency.
Method: We propose PIXEL, a position-aware activation steering framework that introduces the novel concept of attribute-aligned subspaces. Leveraging closed-form geometric optimization and orthogonal residual calibration, PIXEL enables token-level sensitivity-adaptive intervention without global hyperparameter tuning.
Contribution: Through dual-perspective representation learning and lightweight positional scanning, PIXEL significantly improves factual alignment across diverse LLMs and evaluation benchmarks—while preserving general-purpose capabilities. The approach achieves efficient, controllable, and trustworthy text generation, offering a principled alternative to heuristic steering methods.
📝 Abstract
Reliable behavior control is central to deploying large language models (LLMs) on the web. Activation steering offers a tuning-free route to align attributes (e.g., truthfulness) that ensure trustworthy generation. Prevailing approaches rely on coarse heuristics and lack a principled account of where to steer and how strongly to intervene. To this end, we propose Position-wise Injection with eXact Estimated Levels (PIXEL), a position-wise activation steering framework that, in contrast to prior work, learns a property-aligned subspace from dual views (tail-averaged and end-token) and selects intervention strength via a constrained geometric objective with a closed-form solution, thereby adapting to token-level sensitivity without global hyperparameter tuning. PIXEL further performs sample-level orthogonal residual calibration to refine the global attribute direction and employs a lightweight position-scanning routine to identify receptive injection sites. We additionally provide representation-level guarantees for the minimal-intervention rule, supporting reliable alignment. Across diverse models and evaluation paradigms, PIXEL consistently improves attribute alignment while preserving model general capabilities, offering a practical and principled method for LLMs' controllable generation. Our code is available at https://github.com/V1centNevwake/PIXEL-Adaptive-Steering