🤖 AI Summary
Large language models (LLMs) suffer from high inference latency due to autoregressive decoding; existing speculative decoding verification methods overemphasize distributional consistency while neglecting semantic correctness and exhibit poor generalization. Method: We propose a semantic-aware, reflective verification method that requires no additional training or model modifications. Leveraging prompt-guided self-reflection, it enables an LLM to concurrently generate both the original and self-reflective token distributions in a single forward pass, thereby performing semantic-level validation of draft tokens. Contribution/Results: This is the first work to integrate LLMs’ intrinsic reflective capability into speculative decoding verification—jointly ensuring distributional fidelity and semantic plausibility—and supports orthogonal integration with statistical validation. Experiments demonstrate significantly increased draft acceptance length, 5–15% faster decoding speed, and preserved generation quality across diverse benchmarks and model scales.
📝 Abstract
Large language models (LLMs) suffer from high inference latency due to the auto-regressive decoding process. Speculative decoding accelerates inference by generating multiple draft tokens using a lightweight model and verifying them in parallel. However, existing verification methods rely heavily on distributional consistency while overlooking semantic correctness, thereby limiting the potential speedup of speculative decoding. While some methods employ additional models for relaxed verification of draft tokens, they often fail to generalize effectively to more diverse or open-domain settings. In this work, we propose Reflective Verification, a training-free and semantics-aware approach that achieves a better trade-off between correctness and efficiency. Specifically, we leverage the inherent reflective capacity of LLMs to semantically assess the correctness of draft tokens in parallel during verification. Using prompt-based probing, we obtain both the original and reflective distributions of draft tokens in a single forward pass. The fusion of these distributions enables semantic-level verification of draft tokens that incorporates both consistency and correctness. Experiments across multiple domain benchmarks and model scales demonstrate that our method significantly increases the acceptance length of draft tokens without compromising model performance. Furthermore, we find that the proposed Reflective Verification is orthogonal to existing statistical verification methods, and their combination yields additional 5$sim$15% improvements in decoding speed.