🤖 AI Summary
Existing digital human hair modeling methods suffer from reliance on multi-view capture setups, computationally expensive volumetric estimation, and low synthesis fidelity. To address these limitations, this paper proposes a single-image, high-fidelity hair geometry synthesis method leveraging knowledge distillation from pretrained generative models. We represent head-hair geometry using signed distance fields (SDFs) and distill geometric priors directly from the latent space of a pre-trained diffusion model via single-image inversion and latent-space manipulation. Our framework jointly predicts semantic segmentation masks and 3D directional fields to guide geometry reconstruction. Crucially, it operates without specialized hardware and generates topologically clean, manifold hair meshes—along with corresponding semantic and directional maps—in under five seconds. Experiments demonstrate substantial improvements over state-of-the-art approaches in both synthesis speed and visual fidelity. This work establishes a new paradigm for real-time, lightweight, and high-fidelity digital human hair modeling.
📝 Abstract
Achieving realistic hair strand synthesis is essential for creating lifelike digital humans, but producing high-fidelity hair strand geometry remains a significant challenge. Existing methods require a complex setup for data acquisition, involving multi-view images captured in constrained studio environments. Additionally, these methods have longer hair volume estimation and strand synthesis times, which hinder efficiency. We introduce PanoHair, a model that estimates head geometry as signed distance fields using knowledge distillation from a pre-trained generative teacher model for head synthesis. Our approach enables the prediction of semantic segmentation masks and 3D orientations specifically for the hair region of the estimated geometry. Our method is generative and can generate diverse hairstyles with latent space manipulations. For real images, our approach involves an inversion process to infer latent codes and produces visually appealing hair strands, offering a streamlined alternative to complex multi-view data acquisition setups. Given the latent code, PanoHair generates a clean manifold mesh for the hair region in under 5 seconds, along with semantic and orientation maps, marking a significant improvement over existing methods, as demonstrated in our experiments.