AdaCorrection: Adaptive Offset Cache Correction for Accurate Diffusion Transformers

📅 2026-02-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Diffusion Transformers suffer from high computational overhead during inference due to iterative denoising, and existing caching-based acceleration methods often employ static reuse strategies that are prone to temporal drift and cache misalignment, degrading generation quality. To address this, this work proposes AdaCorrection, a novel framework that introduces, for the first time, a lightweight spatiotemporal signal-driven adaptive cache correction mechanism. Without requiring additional supervision or retraining, AdaCorrection dynamically evaluates cache validity and adaptively fuses newly computed activations. Extensive experiments demonstrate that the method significantly improves inference efficiency in both image and video generation tasks while maintaining FID scores close to those of the original model, outperforming existing caching acceleration approaches in generation quality.

Technology Category

Application Category

📝 Abstract
Diffusion Transformers (DiTs) achieve state-of-the-art performance in high-fidelity image and video generation but suffer from expensive inference due to their iterative denoising structure. While prior methods accelerate sampling by caching intermediate features, they rely on static reuse schedules or coarse-grained heuristics, which often lead to temporal drift and cache misalignment that significantly degrade generation quality. We introduce \textbf{AdaCorrection}, an adaptive offset cache correction framework that maintains high generation fidelity while enabling efficient cache reuse across Transformer layers during diffusion inference. At each timestep, AdaCorrection estimates cache validity with lightweight spatio-temporal signals and adaptively blends cached and fresh activations. This correction is computed on-the-fly without additional supervision or retraining. Our approach achieves strong generation quality with minimal computational overhead, maintaining near-original FID while providing moderate acceleration. Experiments on image and video diffusion benchmarks show that AdaCorrection consistently improves generation performance.
Problem

Research questions and friction points this paper is trying to address.

Diffusion Transformers
inference acceleration
cache reuse
temporal drift
generation quality
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive Cache Correction
Diffusion Transformers
Efficient Inference
Temporal Drift Mitigation
On-the-fly Activation Blending
🔎 Similar Papers
No similar papers found.