🤖 AI Summary
To address the high cost and low efficiency of human preference annotation in large language model (LLM) alignment, this paper proposes an active learning framework that leverages the LLM itself as a parameterized, differentiable, nonlinear reward model for sample selection—departing from conventional linear reward assumptions. Our approach establishes a theoretically grounded, gradient-driven uncertainty estimation criterion and unifies implicit reward modeling, Direct Preference Optimization (DPO), and active sampling into an end-to-end optimization pipeline. Experiments across multiple LLMs and benchmark datasets demonstrate that our method achieves comparable or superior alignment performance using 30–50% fewer annotated preference pairs, significantly improving the efficiency of preference data utilization.
📝 Abstract
The recent success of using human preferences to align large language models (LLMs) has significantly improved their performance in various downstream tasks like question answering, mathematical reasoning, and code generation. However,3 achieving effective LLM alignment depends on high-quality human preference datasets. Collecting these datasets requires human preference annotation, which is costly and resource-intensive, necessitating efficient active data selection methods. Existing methods either lack a strong theoretical foundation or depend on restrictive reward function assumptions (e.g., linearity). To this end, we propose an algorithm, ActiveDPO, that uses a theoretically grounded data selection criterion for non-linear reward functions while directly leveraging the LLM itself to parameterize the reward model that is used for active data selection. As a result, ActiveDPO explicitly accounts for the influence of LLM on data selection, unlike methods that select the data without considering the LLM that is being aligned, thereby leading to more effective and efficient data collection. Extensive experiments show that ActiveDPO outperforms existing methods across various models and datasets.