🤖 AI Summary
This work addresses the robustness of preference modeling in value alignment for large language models, focusing on how strong human preferences—those with probabilities approaching 0 or 1—induce severe prediction instability in Bradley–Terry and Plackett–Luce models. Using theoretical sensitivity analysis and formal verification, we characterize, for the first time, the critical conditions under which preference perturbations propagate across these two dominant preference models. We prove that strong preferences trigger nonlinear amplification of errors across candidate options, revealing an inherent fragility in standard preference modeling. Our results deliver a key theoretical warning for value alignment: existing preference-based methods may catastrophically destabilize under the heavy-tailed, skewed distributions prevalent in real-world human preference data—thereby compromising AI safety and trustworthiness. The study establishes necessary theoretical constraints and principled directions for designing robust alignment algorithms resilient to preference extremity.
📝 Abstract
Value alignment, which aims to ensure that large language models (LLMs) and other AI agents behave in accordance with human values, is critical for ensuring safety and trustworthiness of these systems. A key component of value alignment is the modeling of human preferences as a representation of human values. In this paper, we investigate the robustness of value alignment by examining the sensitivity of preference models. Specifically, we ask: how do changes in the probabilities of some preferences affect the predictions of these models for other preferences? To answer this question, we theoretically analyze the robustness of widely used preference models by examining their sensitivities to minor changes in preferences they model. Our findings reveal that, in the Bradley-Terry and the Placket-Luce model, the probability of a preference can change significantly as other preferences change, especially when these preferences are dominant (i.e., with probabilities near 0 or 1). We identify specific conditions where this sensitivity becomes significant for these models and discuss the practical implications for the robustness and safety of value alignment in AI systems.