π€ AI Summary
Existing diffusion-based acceleration methods for long portrait animation suffer from severe identity (ID) consistency degradation. To address this, we propose an end-to-end framework for infinite-length video generation. Our method builds upon a video diffusion Transformer augmented with a pre-trained facial feature extractor. Key contributions include: (1) a normalized facial expression module that explicitly disentangles and stabilizes ID and expression; (2) an adaptive higher-order latent derivative prediction mechanism to enhance temporal modeling fidelity; and (3) a dynamic sliding window strategy with overlapping-region weighted fusion for efficient local-global collaborative inference. Evaluated on standard benchmarks, our framework achieves 6Γ inference speedup over baseline diffusion models while significantly outperforming state-of-the-art methods in both ID preservation and motion coherence. Quantitative metrics and qualitative assessments consistently demonstrate superior performance.
π Abstract
Current diffusion-based acceleration methods for long-portrait animation struggle to ensure identity (ID) consistency. This paper presents FlashPortrait, an end-to-end video diffusion transformer capable of synthesizing ID-preserving, infinite-length videos while achieving up to 6x acceleration in inference speed. In particular, FlashPortrait begins by computing the identity-agnostic facial expression features with an off-the-shelf extractor. It then introduces a Normalized Facial Expression Block to align facial features with diffusion latents by normalizing them with their respective means and variances, thereby improving identity stability in facial modeling. During inference, FlashPortrait adopts a dynamic sliding-window scheme with weighted blending in overlapping areas, ensuring smooth transitions and ID consistency in long animations. In each context window, based on the latent variation rate at particular timesteps and the derivative magnitude ratio among diffusion layers, FlashPortrait utilizes higher-order latent derivatives at the current timestep to directly predict latents at future timesteps, thereby skipping several denoising steps and achieving 6x speed acceleration. Experiments on benchmarks show the effectiveness of FlashPortrait both qualitatively and quantitatively.