š¤ AI Summary
Despite its computational efficiency, knowledge distillationās superiority over zero-shot reinforcement learning (zero-RL) in enhancing large language modelsā (LLMs) reasoning capabilities remains poorly understood. Method: The authors conduct token-level frequency analysis and cognitive behavioral pattern identification on models trained via supervised knowledge distillation (using only 920 high-quality examples) versus zero-RL. Contribution/Results: They demonstrate that distillationāunlike zero-RLāspecifically amplifies two higher-order cognitive behaviors critical for flexible reasoning: multi-perspective thinking and metacognitive awareness. Distilled models outperform zero-RL counterparts on complex reasoning tasks and exhibit significantly higher usage of logical connectives and anthropomorphic expressionsāempirically validating a strong association between these cognitive behaviors and reasoning ability. This work uncovers the intrinsic cognitive mechanisms underlying distillation-driven reasoning improvement and establishes a novel, resource-efficient paradigm for LLM alignment.
š Abstract
Reinforcement learning (RL) has played an important role in improving the reasoning ability of large language models (LLMs). Some studies apply RL directly to extit{smaller} base models (known as zero-RL) and also achieve notable progress. However, in this paper, we show that using only 920 examples, a simple distillation method based on the base model can clearly outperform zero-RL, which typically requires much more data and computational cost. By analyzing the token frequency in model outputs, we find that the distilled model shows more flexible reasoning. It uses anthropomorphic tokens and logical connectors much more often than the zero-RL model. Further analysis reveals that distillation enhances the presence of two advanced cognitive behaviors: Multi-Perspective Thinking or Attempting and Metacognitive Awareness. Frequent occurrences of these two advanced cognitive behaviors give rise to flexible reasoning, which is essential for solving complex reasoning problems, while zero-RL fails to significantly boost the frequency of these behaviors.