🤖 AI Summary
Small language models suffer from limited performance, poor robustness, and insufficient adaptability, while existing knowledge distillation methods—such as static distillation, RLHF, and introspection—struggle to enable sustained improvement. To address this, we propose a multi-agent debate-driven lightweight knowledge distillation framework. Our approach introduces two key innovations: (1) a student–teacher multi-round collaborative debate mechanism that generates actionable error analyses and strategy-correction feedback; and (2) tree-structured direct preference optimization (T-DPO), which models debate logs as hierarchical decision paths to enhance feedback utilization efficiency and generalization. Evaluated across multiple NLP benchmarks, our method significantly improves accuracy, robustness, and generalization of small models—consistently outperforming static distillation, RLHF, and introspection baselines. These results demonstrate the effectiveness of efficient knowledge transfer under resource-constrained settings.
📝 Abstract
Large Language Models (LLMs) continue to set new standards in knowledge-intensive and complex reasoning tasks, yet their high computational demands limit widespread adoption. While distilling large models into smaller ones offers a sustainable solution, current techniques--such as static knowledge distillation, resource-intensive reinforcement learning from human feedback, or limited self-reflection--struggle to yield substantial and lasting performance gains. In this paper, we present a novel Debate and Reflect (D&R) framework that orchestrates multi-turn debates between smaller models and stronger teacher models, eliciting actionable feedback (e.g., error analysis, corrective strategies) to guide student models. Further, we introduce Tree-structured Direct Preference Optimization (T-DPO) to efficiently leverage these debate logs, organizing interactions into a hierarchical format for effective training. Empirical evaluations across diverse NLP benchmarks demonstrate that our approach significantly improves smaller-model accuracy, robustness, and generalization, outperforming conventional baselines by a large margin.