🤖 AI Summary
Traditional preference learning often exhibits bias toward majority groups when aggregating multi-annotator judgments, leading to policy representational bias. This paper proposes a demographically proportional preference learning framework: first, it infers annotator demographic distributions from pairwise comparison data; second, it designs an aggregation strategy satisfying four axioms—monotonicity, Pareto efficiency, demographic proportionality, and bounded robustness (a novel axiom introduced herein); third, it employs a soft-max relaxation optimization method that preserves Condorcet winner stability while ensuring proportional representation. The approach integrates axiomatic social choice modeling, data-driven distribution inference, and relaxed optimization. Empirical evaluation on tabular recommendation and large language model alignment tasks demonstrates significant improvements in minority-group preference representation, achieving a balanced trade-off between fairness and consensus robustness.
📝 Abstract
Conventional preference learning methods often prioritize opinions held more widely when aggregating preferences from multiple evaluators. This may result in policies that are biased in favor of some types of opinions or groups. The objective of this paper is to develop a novel preference learning framework capable of aligning aggregate opinions and policies proportionally with the true population distribution of evaluator preferences. Our approach infers the feasible set of evaluator population distributions directly from pairwise comparison data. Using these estimates, the algorithm constructs a policy that satisfies foundational axioms from social choice theory, namely monotonicity and Pareto efficiency, as well as our newly-introduced axioms of population-proportional representation and population-bounded robustness. We propose a soft-max relaxation method that smoothly trade-offs population-proportional representation with the selection of the Condorcet winner (which beats all other options in pairwise comparisons). Finally, we validate the effectiveness and scalability of our approach through experiments on both tabular recommendation tasks and large-scale language model alignment.