🤖 AI Summary
Current hybrid reasoning large language models suffer from “reasoning leakage”: they unnecessarily generate lengthy reasoning traces even in direct-answer (no-think) mode, undermining inference efficiency and mode controllability. This work is the first to systematically identify four key factors affecting mode separability and proposes a novel training paradigm grounded in cross-problem sample construction, balanced data mixing, and two-stage fine-tuning. Crucially, the method preserves high accuracy in both think and no-think modes while substantially improving mode isolation. On MATH500, it reduces no-think output length from 1,085 to 585 tokens and slashes reasoning-support token frequency from 5,917 to 522 occurrences. The study establishes a reproducible methodology and empirical benchmark for controllable reasoning-mode modeling, advancing principled design of dual-mode LLMs.
📝 Abstract
Hybrid thinking enables LLMs to switch between reasoning and direct answering, offering a balance between efficiency and reasoning capability. Yet our experiments reveal that current hybrid thinking LLMs only achieve partial mode separation: reasoning behaviors often leak into the no-think mode. To understand and mitigate this, we analyze the factors influencing controllability and identify four that matter most: (1) larger data scale, (2) using think and no-think answers from different questions rather than the same question, (3) a moderate increase in no-think data number, and (4) a two-phase strategy that first trains reasoning ability and then applies hybrid think training. Building on these findings, we propose a practical recipe that, compared to standard training, can maintain accuracy in both modes while significantly reducing no-think output length (from $1085$ to $585$ on MATH500) and occurrences of reasoning-supportive tokens such as `` exttt{wait}''(from $5917$ to $522$ on MATH500). Our findings highlight the limitations of current hybrid thinking and offer directions for strengthening its controllability.