π€ AI Summary
This work addresses the challenge that existing test-time scaling methods, which rely on autoregressive decoding, are ill-suited for discrete diffusion language models due to their parallel generation nature. To bridge this gap, the authors propose Prism, the first efficient test-time scaling framework tailored for such models. Prism dynamically allocates computational resources through Hierarchical Trajectory Search (HTS), enhances output diversity via local branch re-masking, and replaces external verifiers with an endogenous Self-Verification Feedback (SVF) mechanism. Evaluated on four benchmarks spanning mathematical reasoning and code generation, Prism matches or approaches best-of-N performance while using significantly fewer function evaluations, demonstrating both its efficiency and broad applicability.
π Abstract
Inference-time compute has re-emerged as a practical way to improve LLM reasoning. Most test-time scaling (TTS) algorithms rely on autoregressive decoding, which is ill-suited to discrete diffusion language models (dLLMs) due to their parallel decoding over the entire sequence. As a result, developing effective and efficient TTS methods to unlock dLLMs'full generative potential remains an underexplored challenge. To address this, we propose Prism (Pruning, Remasking, and Integrated Self-verification Method), an efficient TTS framework for dLLMs that (i) performs Hierarchical Trajectory Search (HTS) which dynamically prunes and reallocates compute in an early-to-mid denoising window, (ii) introduces Local branching with partial remasking to explore diverse implementations while preserving high-confidence tokens, and (iii) replaces external verifiers with Self-Verified Feedback (SVF) obtained via self-evaluation prompts on intermediate completions. Across four mathematical reasoning and code generation benchmarks on three dLLMs, including LLaDA 8B Instruct, Dream 7B Instruct, and LLaDA 2.0-mini, our Prism achieves a favorable performance-efficiency trade-off, matching best-of-N performance with substantially fewer function evaluations (NFE). The code is released at https://github.com/viiika/Prism.