🤖 AI Summary
This work addresses the limited understanding of token-level mechanisms underlying reasoning improvements when fine-tuning large language models via Reinforcement Learning with Verifiable Rewards (RLVR). Through token-level distribution shift analysis, cross-sampling interventions, and probability mass reallocation, the study systematically demonstrates that RLVR operates as a highly sparse and precise optimization process, wherein performance gains are driven predominantly by a small subset of critical tokens. The authors introduce a divergence-weighted advantage signal as a diagnostic tool and show that replacing only a few RL-generated tokens suffices to recover model performance, while inserting a small number of baseline tokens causes significant degradation—highlighting the decisive role of these key tokens in shaping overall behavior.
📝 Abstract
Reinforcement learning with verifiable rewards (RLVR) has significantly improved reasoning in large language models (LLMs), yet the token-level mechanisms underlying these improvements remain unclear. We present a systematic empirical study of RLVR's distributional effects organized around three main analyses: (1) token-level characterization of distributional shifts between base and RL models, (2) the impact of token-level distributional shifts on sequence-level reasoning performance through cross-sampling interventions, and (3) fine-grained mechanics of these shifts at the token level. We find that RL fine-tuning induces highly sparse and targeted changes, with only a small fraction of token distributions exhibiting meaningful divergence between the base and RL policies. We further characterize the structure and evolution of these shifts through analyses of token entropy, positional concentration, and reallocation of probability mass. To assess the functional importance of these sparse changes, we conduct cross-sampling experiments that selectively swap token choices between the base and RL models with varying intervention budgets. We show that inserting only a small fraction of RL-sampled tokens into base generations progressively recovers RL performance gains, while injecting a similarly small number of base token choices into otherwise RL-generated sequences collapses performance to base levels, isolating a small set of token-level decisions directly responsible for RLVR's performance gains. Finally, we explore divergence-weighted variants of the advantage signal as a diagnostic intervention, finding that they can yield improvements over baselines. Together, our results shed light on the distributional changes induced by RLVR and provide a fine-grained, token-level lens for understanding RLVR fine-tuning as a targeted refinement process.