🤖 AI Summary
Video diffusion Transformers incur prohibitive computational and memory overhead, hindering practical deployment. To address this, we propose a synergistic compression framework integrating quantization and attention sparsification. Our method innovatively combines multi-scale saliency-guided attention distillation with second-order sparse attention reparameterization, dynamically balancing quantization noise mitigation and sparse information preservation. We further enhance robustness and fidelity of the compressed attention mechanism via temporal-stability-driven residual reparameterization, global structural guidance, and local saliency supervision. Evaluated on HunyuanVideo-13B, our approach achieves 20.88 PSNR—surpassing the best quantization baseline by +4.03—while reducing model storage by 3.68× and accelerating end-to-end inference by 1.88×. These results significantly outperform existing single-dimension compression methods.
📝 Abstract
Diffusion transformers exhibit remarkable video generation capability, yet their prohibitive computational and memory costs hinder practical deployment. Model quantization and attention sparsification are two promising directions for compression, but each alone suffers severe performance degradation under aggressive compression. Combining them promises compounded efficiency gains, but naive integration is ineffective. The sparsity-induced information loss exacerbates quantization noise, leading to amplified attention shifts. To address this, we propose extbf{QuantSparse}, a unified framework that integrates model quantization with attention sparsification. Specifically, we introduce extit{Multi-Scale Salient Attention Distillation}, which leverages both global structural guidance and local salient supervision to mitigate quantization-induced bias. In addition, we develop extit{Second-Order Sparse Attention Reparameterization}, which exploits the temporal stability of second-order residuals to efficiently recover information lost under sparsity. Experiments on HunyuanVideo-13B demonstrate that QuantSparse achieves 20.88 PSNR, substantially outperforming the state-of-the-art quantization baseline Q-VDiT (16.85 PSNR), while simultaneously delivering a extbf{3.68$ imes$} reduction in storage and extbf{1.88$ imes$} acceleration in end-to-end inference. Our code will be released in https://github.com/wlfeng0509/QuantSparse.