🤖 AI Summary
Current large language models (LLMs) suffer from an “alignment–reality gap”: static alignment strategies fail to adapt to dynamically evolving societal norms and policies, resulting in value misalignment, poor robustness, and high maintenance overhead. To address this, we propose TRACE—a novel framework that formalizes realignment as a programmable policy-application problem. TRACE introduces an alignment impact score to quantitatively assess preference conflicts and enables selective preference reversal, discarding, or retention—balancing correction accuracy with model performance. Leveraging a hybrid optimization pipeline—integrating conflict evaluation, preference-data categorization and filtering, and selective retraining—TRACE achieves fine-grained, low-regret updates across diverse architectures (Qwen, Gemma, Llama). Experiments demonstrate that TRACE significantly improves compliance with complex, evolving policy requirements while preserving pre-existing general-purpose capabilities.
📝 Abstract
The alignment of Large Language Models (LLMs) with human values is central to their safe deployment, yet current practice produces static, brittle, and costly-to-maintain models that fail to keep pace with evolving norms and policies. This misalignment, which we term the Alignment-Reality Gap, poses a growing challenge for reliable long-term use. Existing remedies are inadequate: large-scale re-annotation is economically prohibitive, and standard unlearning methods act as blunt instruments that erode utility rather than enable precise policy updates. We introduce TRACE (Triage and Re-align by Alignment Conflict Evaluation), a framework for principled unlearning that reconceives re-alignment as a programmatic policy application problem. TRACE programmatically triages existing preference data against a new policy, identifies high-impact conflicts via a alignment impact score, and applies a hybrid optimization that cleanly inverts, discards, or preserves preferences while safeguarding model performance. Empirical results show that TRACE achieves robust re-alignment across diverse model families (Qwen2.5-7B, Gemma-2-9B, Llama-3.1-8B). On both synthetic benchmarks and the PKU-SafeRLHF dataset under complex policy shift, TRACE enforces new principles without degrading general capabilities. Our work establishes a scalable, dynamic, and cost-effective paradigm for maintaining LLM alignment, providing a foundation for sustainable and responsible AI deployment.