Region-Adaptive Sampling for Diffusion Transformers

πŸ“… 2025-02-14
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Diffusion models (DMs), particularly DiT-based architectures, suffer from slow inference due to fixed-step and global sampling schemes. This work proposes a training-free, plug-and-play region-adaptive sampling strategy that exploits DiT’s token-level spatial flexibility: leveraging cross-step attention heatmaps to model temporal continuity, it dynamically identifies high-attention semantic regions and updates only their corresponding noise tokens, while reusing cached noise for the rest. To our knowledge, this is the first method enabling dynamic, spatially granular sampling in DiT without fine-tuning or additional parameters. Evaluated on Stable Diffusion 3 and Lumina-Next-T2I, it achieves 2.36Γ— and 2.51Γ— speedup, respectively, with negligible degradation in FID and CLIP-Score. A user study confirms no statistically significant difference in visual quality between the accelerated and baseline outputs at 1.6Γ— speedup.

Technology Category

Application Category

πŸ“ Abstract
Diffusion models (DMs) have become the leading choice for generative tasks across diverse domains. However, their reliance on multiple sequential forward passes significantly limits real-time performance. Previous acceleration methods have primarily focused on reducing the number of sampling steps or reusing intermediate results, failing to leverage variations across spatial regions within the image due to the constraints of convolutional U-Net structures. By harnessing the flexibility of Diffusion Transformers (DiTs) in handling variable number of tokens, we introduce RAS, a novel, training-free sampling strategy that dynamically assigns different sampling ratios to regions within an image based on the focus of the DiT model. Our key observation is that during each sampling step, the model concentrates on semantically meaningful regions, and these areas of focus exhibit strong continuity across consecutive steps. Leveraging this insight, RAS updates only the regions currently in focus, while other regions are updated using cached noise from the previous step. The model's focus is determined based on the output from the preceding step, capitalizing on the temporal consistency we observed. We evaluate RAS on Stable Diffusion 3 and Lumina-Next-T2I, achieving speedups up to 2.36x and 2.51x, respectively, with minimal degradation in generation quality. Additionally, a user study reveals that RAS delivers comparable qualities under human evaluation while achieving a 1.6x speedup. Our approach makes a significant step towards more efficient diffusion transformers, enhancing their potential for real-time applications.
Problem

Research questions and friction points this paper is trying to address.

Enhance real-time performance of diffusion models.
Reduce sequential forward pass limitations in DMs.
Optimize sampling strategy in Diffusion Transformers.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic region sampling ratios
Focus-based noise caching
Temporal consistency utilization
πŸ”Ž Similar Papers
No similar papers found.