🤖 AI Summary
The verification stage of speculative decoding becomes a significant computational bottleneck—especially under long-context and Mixture-of-Experts (MoE) settings—where existing sparsification methods fail to accommodate its multi-draft parallelism and cross-layer dependencies.
Method: We systematically identify structural redundancy in the verification stage across three dimensions: attention, feed-forward networks (FFNs), and MoE expert selection. We propose a joint sparsification framework tailored to verification, integrating sparse attention, sparse FFNs, and dynamic MoE expert pruning, augmented by a retrieval reuse mechanism across draft tokens and layers—requiring no additional training.
Contribution/Results: Evaluated on summarization, question answering, and mathematical reasoning tasks, our method reduces verification-stage computation by up to 58%, maintains stable token acceptance rates, and achieves superior efficiency–accuracy trade-offs without compromising model performance.
📝 Abstract
Speculative decoding accelerates autoregressive language model inference by verifying multiple draft tokens in parallel. However, the verification stage often becomes the dominant computational bottleneck, especially for long-context inputs and mixture-of-experts (MoE) models. Existing sparsification methods are designed primarily for standard token-by-token autoregressive decoding to remove substantial computational redundancy in LLMs. This work systematically adopts different sparse methods on the verification stage of the speculative decoding and identifies structured redundancy across multiple dimensions. Based on these observations, we propose a sparse verification framework that jointly sparsifies attention, FFN, and MoE components during the verification stage to reduce the dominant computation cost. The framework further incorporates an inter-draft token and inter-layer retrieval reuse strategy to further reduce redundant computation without introducing additional training. Extensive experiments across summarization, question answering, and mathematical reasoning datasets demonstrate that the proposed methods achieve favorable efficiency-accuracy trade-offs, while maintaining stable acceptance length.