🤖 AI Summary
Existing language models struggle to simultaneously ensure safety and empathy in high-stakes or emotionally charged scenarios—strict refusal risks alienating users, while unconditional compliance exacerbates harm.
Method: We propose a test-time parameter-efficient alignment framework that requires no fine-tuning of the base model. It employs hierarchical constraints during generation to produce responses that are safe, empathetic, and value-aligned. Safety is formally modeled as a lexicographic constraint optimization problem, integrating hard filtering, directional “harm vector” suppression, and a preference-aware autoregressive reward model trained jointly on multiple attributes. The framework supports user-controllable decoding.
Contribution/Results: Our method achieves state-of-the-art performance across five safety benchmarks, significantly reducing unsafe outputs while substantially improving alignment with human values—without modifying the base model’s parameters.
📝 Abstract
Current language model safety paradigms often fall short in emotionally charged or high-stakes settings, where refusal-only approaches may alienate users and naive compliance can amplify risk. We propose ProSocialAlign, a test-time, parameter-efficient framework that steers generation toward safe, empathetic, and value-aligned responses without retraining the base model. We formalize five human-centered objectives and cast safety as lexicographic constrained generation: first, applying hard constraints to eliminate harmful continuations; then optimizing for prosocial quality within the safe set. Our method combines (i) directional regulation, a harm-mitigation mechanism that subtracts a learned "harm vector" in parameter space, and (ii) preference-aware autoregressive reward modeling trained jointly across attributes with gradient conflict resolution, enabling fine-grained, user-controllable decoding. Empirical evaluations across five safety benchmarks demonstrate state-of-the-art performance, reducing unsafe leakage and boosting alignment to human values, with strong gains across multiple evaluation metrics. ProSocialAlign offers a robust and modular foundation for generating context-sensitive, safe, and human-aligned responses at inference time.