๐ค AI Summary
This study addresses the lack of systematic analysis regarding how discourse markersโsuch as the token โwaitโโin large language models (LLMs) are influenced by training strategies and model scale, a gap that hinders deeper understanding of LLM reasoning mechanisms. For the first time, this work establishes a direct link between token-level signals and LLM reasoning capabilities through token probability analysis, cross-model comparisons, and fine-grained behavioral tracing. The findings reveal that discourse markers exhibit significant variation across different training strategies but remain stable across model scales. While fine-tuned models can learn to utilize such tokens, their deployment remains suboptimal. These results demonstrate that specific tokens can serve as effective probes for uncovering the dynamic reasoning processes within LLMs.
๐ Abstract
The emergence of discourse-like tokens such as"wait"and"therefore"in large language models (LLMs) has offered a unique window into their reasoning processes. However, systematic analyses of how such signals vary across training strategies and model scales remain lacking. In this paper, we analyze token-level signals through token probabilities across various models. We find that specific tokens strongly correlate with reasoning correctness, varying with training strategies while remaining stable across model scales. A closer look at the"wait"token in relation to answer probability demonstrates that models fine-tuned on small-scale datasets acquire reasoning ability through such signals but exploit them only partially. This work provides a systematic lens to observe and understand the dynamics of LLM reasoning.