π€ AI Summary
Existing text-to-video (T2V) methods struggle with fine-grained facial dynamics modeling and cross-frame identity consistency. To address this, we propose an identity-preserving T2V generation framework. Our method introduces: (1) a Mixture of Cross-Attention mechanism that explicitly models both text-video cross-modal alignment and inter-frame temporal dependencies; (2) hierarchical temporal pooling coupled with a latent-space video-aware perceptual loss to enhance temporal stability of identity features and fidelity of facial details; and (3) an end-to-end DiT-based architecture trained on CelebIPVidβa newly curated high-quality human face video dataset. Quantitative and qualitative evaluations demonstrate that our approach outperforms state-of-the-art T2V models by over 5% on facial similarity metrics, achieving significant improvements in identity consistency and visual quality.
π Abstract
Achieving ID-preserving text-to-video (T2V) generation remains challenging despite recent advances in diffusion-based models. Existing approaches often fail to capture fine-grained facial dynamics or maintain temporal identity coherence. To address these limitations, we propose MoCA, a novel Video Diffusion Model built on a Diffusion Transformer (DiT) backbone, incorporating a Mixture of Cross-Attention mechanism inspired by the Mixture-of-Experts paradigm. Our framework improves inter-frame identity consistency by embedding MoCA layers into each DiT block, where Hierarchical Temporal Pooling captures identity features over varying timescales, and Temporal-Aware Cross-Attention Experts dynamically model spatiotemporal relationships. We further incorporate a Latent Video Perceptual Loss to enhance identity coherence and fine-grained details across video frames. To train this model, we collect CelebIPVid, a dataset of 10,000 high-resolution videos from 1,000 diverse individuals, promoting cross-ethnicity generalization. Extensive experiments on CelebIPVid show that MoCA outperforms existing T2V methods by over 5% across Face similarity.