🤖 AI Summary
Autoregressive decoding constitutes the primary latency bottleneck in large language model (LLM) inference. Existing speculative decoding methods rely on offline training or auxiliary modules, incurring high computational costs and exhibiting sensitivity to distributional shift. This paper proposes a training-aware self-speculation framework: the first to integrate online learning into speculative decoding, dynamically optimizing the draft model via verifier feedback and employing a progressive scheduling strategy that bridges KL divergence minimization with reinforcement learning. Leveraging an LLM block-wise architecture, it unifies online knowledge distillation, reward-masked cross-entropy loss, and on-policy policy gradients—enabling continual learning and inference acceleration within a single, lossless deployment. Evaluated on Spec-Bench, our method achieves 2.16× end-to-end speedup, matching state-of-the-art methods like EAGLE-2, while reducing training data requirements by several orders of magnitude and significantly outperforming pure KL-based distillation baselines.
📝 Abstract
Autoregressive (AR) decoding is a major latency bottleneck for large language models. Speculative decoding (SD) accelerates AR by letting a drafter propose multi-token blocks that a verifier accepts or rejects. However, many SD systems require heavy offline training or extra components. These choices raise data/compute cost and can yield brittle drafters under distribution drift. We introduce emph{Draft, Verify, & Improve (DVI)}, a training-aware self-speculative framework that combines inference with continual online learning. We partition an LLM into a drafter and a verifier, and during generation, verifier accept/reject decisions are converted into supervision signals and used to update the drafter head. A simple emph{KL$
ightarrow$RL} schedule bootstraps calibration via online distillation and then adds reward-masked cross-entropy with a on-policy policy-gradient term, preserving lossless, single model deployment. On Spec-Bench, DVI achieves a $2.16 imes$ wall-time speedup, on par with SoTA approaches like EAGLE-2, while orders of magnitude less data for training, and ablations show that DVI outperforms KL-only online distillation. DVI demonstrates that emph{training-aware} self-speculation can deliver state-of-the-art, lossless speedups with minimal training overhead.