🤖 AI Summary
This study systematically evaluates the effectiveness of Direct Preference Optimization (DPO) and its variants for aligning large language models (LLMs) with human preferences. We conduct a quantitative analysis across 13 multidimensional benchmarks—including MT-Bench, Big Bench, and the Open LLM Leaderboard—to assess DPO’s performance, dependence on supervised fine-tuning (SFT), the role of instruction tuning, and sensitivity to dataset scale. Key contributions: (1) DPO achieves 95% of full-dataset alignment performance using only 20% of preference data, demonstrating strong few-shot efficiency; (2) instruction tuning substantially improves factual consistency (+12.5%) and mathematical reasoning (+8.2%), though gains diminish on complex reasoning tasks; (3) we provide the first empirical evidence that DPO consistently enhances dialogue and question-answering performance (+3–7 percentage points). These findings establish theoretical foundations and practical guidelines for resource-efficient, high-fidelity LLM alignment.
📝 Abstract
This study evaluates Direct Preference Optimization (DPO) and its variants for aligning Large Language Models (LLMs) with human preferences, testing three configurations: (1) with Supervised Fine Tuning (SFT), (2) without SFT, and (3) without SFT but using an instruction tuned model. We further investigate how training set size influences model performance. Our evaluation spans 13 benchmarks covering dialogue, reasoning, mathematical problem-solving, question answering, truthfulness, MT-Bench, Big Bench, and the Open LLM Leaderboard. We find that: (1) alignment methods often achieve near optimal performance even with smaller subsets of training data; (2) although they offer limited improvements on complex reasoning tasks, they enhance mathematical problem-solving; and (3) using an instruction tuned model improves truthfulness. These insights highlight the conditions under which alignment methods excel, as well as their limitations.