🤖 AI Summary
Hair dynamics simulation faces significant generalization challenges across diverse hairstyles, body morphologies, and motion types. This paper introduces the first Transformer-based two-stage neural framework: a static network employs a Transformer to predict initial, penetration-free hair coverage geometry; a dynamic network incorporates cross-attention to jointly encode static geometry and motion inputs, generating high-fidelity secondary motion sequences. By pioneering the integration of Transformers into hair simulation—augmented with physics-aware loss functions—the method substantially improves generalization and stability on unseen long-hair configurations and abrupt motions. The approach enables real-time static inference and dynamic sequence generation, preserving fine hair strand details, eliminating body penetration, and delivering high-quality, multi-hairstyle dynamic simulation under arbitrary poses.
📝 Abstract
Simulating hair dynamics that generalize across arbitrary hairstyles, body shapes, and motions is a critical challenge. Our novel two-stage neural solution is the first to leverage Transformer-based architectures for such a broad generalization. We propose a Transformer-powered static network that predicts static draped shapes for any hairstyle, effectively resolving hair-body penetrations and preserving hair fidelity. Subsequently, a dynamic network with a novel cross-attention mechanism fuses static hair features with kinematic input to generate expressive dynamics and complex secondary motions. This dynamic network also allows for efficient fine-tuning of challenging motion sequences, such as abrupt head movements. Our method offers real-time inference for both static single-frame drapes and dynamic drapes over pose sequences. Our method demonstrates high-fidelity and generalizable dynamic hair across various styles, guided by physics-informed losses, and can resolve penetrations even for complex, unseen long hairstyles, highlighting its broad generalization.