WavefrontDiffusion: Dynamic Decoding Schedule or Improved Reasoning

📅 2025-11-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing diffusion language models (DLMs) suffer from inherent limitations in their denoising strategies: standard sequential diffusion is prone to premature sequence termination due to context truncation, while block-wise diffusion disrupts semantic coherence and logical reasoning through fixed-length token segmentation. To address this, we propose WavefrontDiffusion—a dynamic wavefront decoding framework that initiates denoising from already-determined tokens and adaptively expands the active decoding region outward, thereby aligning the denoising process with linguistic semantic structure. Our method preserves the parallel efficiency and computational budget of block-based decoding while enabling fine-grained, structure-aware adaptive denoising. Evaluated on four reasoning and code-generation benchmarks, WavefrontDiffusion significantly outperforms existing DLMs, achieving state-of-the-art performance in semantic fidelity, logical consistency, and task-specific accuracy.

Technology Category

Application Category

📝 Abstract
Diffusion Language Models (DLMs) have shown strong potential for text generation and are becoming a competitive alternative to autoregressive models. The denoising strategy plays an important role in determining the quality of their outputs. Mainstream denoising strategies include Standard Diffusion and BlockDiffusion. Standard Diffusion performs global denoising without restricting the update range, often finalizing incomplete context and causing premature end-of-sequence predictions. BlockDiffusion updates fixed-size blocks in a preset order, but its rigid structure can break apart coherent semantic units and disrupt reasoning. We present WavefrontDiffusion, a dynamic decoding approach that expands a wavefront of active tokens outward from finalized positions. This adaptive process follows the natural flow of semantic structure while keeping computational cost equal to block-based methods. Across four benchmarks in reasoning and code generation, WavefrontDiffusion achieves state-of-the-art performance while producing outputs with higher semantic fidelity, showing the value of adaptive scheduling for more coherent and efficient generation.
Problem

Research questions and friction points this paper is trying to address.

Dynamic decoding schedule improves reasoning in diffusion language models
Adaptive token expansion maintains semantic coherence during generation
Overcoming premature termination and rigid block structure limitations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic wavefront expansion from finalized tokens
Adaptive scheduling following natural semantic flow
Computational cost equivalent to block-based methods
🔎 Similar Papers
No similar papers found.
Haojin Yang
Haojin Yang
Hasso Plattner Institute
Efficient Deep LearningEdge AISustainable AI
R
Rui Hu
Software Institute, Nanjing University
Zequn Sun
Zequn Sun
Nanjing University
Knowledge GraphLarge Language Model
Yujun Cai
Yujun Cai
NTU → Meta → Lecturer(Assistant Professor) @UQ
Multi-Modal PerceptionVision-Language Models
R
Rui Zhou
School of Software and Microelectronics, Peking University
Y
Yiwei Wang
Department of Computer Science, University of California at Merced