3D Gaussian Head Avatars with Expressive Dynamic Appearances by Compact Tensorial Representations

📅 2025-04-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing methods struggle to simultaneously achieve high dynamic texture fidelity, real-time rendering performance, and storage efficiency in 3D head modeling. This paper proposes a tensorized compact representation framework: neutral geometry and appearance are encoded in static tri-planes, while expressive dynamic details are modeled via lightweight 1D feature curves. To enhance training stability and reconstruction quality, we introduce an adaptive truncation opacity regularization and a class-balanced sampling strategy. Our method integrates 3D Gaussian splatting, 3DMM-based deformation priors, tri-plane encoding, and 1D decoding for high-fidelity, expression-driven dynamic appearance reconstruction. Experiments demonstrate that our approach maintains real-time rendering capability (≥30 FPS) while reducing model size by over 60%. It exhibits strong generalization across diverse input configurations—including single-image and multi-image settings—and enables cross-identity transfer.

Technology Category

Application Category

📝 Abstract
Recent studies have combined 3D Gaussian and 3D Morphable Models (3DMM) to construct high-quality 3D head avatars. In this line of research, existing methods either fail to capture the dynamic textures or incur significant overhead in terms of runtime speed or storage space. To this end, we propose a novel method that addresses all the aforementioned demands. In specific, we introduce an expressive and compact representation that encodes texture-related attributes of the 3D Gaussians in the tensorial format. We store appearance of neutral expression in static tri-planes, and represents dynamic texture details for different expressions using lightweight 1D feature lines, which are then decoded into opacity offset relative to the neutral face. We further propose adaptive truncated opacity penalty and class-balanced sampling to improve generalization across different expressions. Experiments show this design enables accurate face dynamic details capturing while maintains real-time rendering and significantly reduces storage costs, thus broadening the applicability to more scenarios.
Problem

Research questions and friction points this paper is trying to address.

Capturing dynamic textures in 3D head avatars efficiently
Reducing runtime and storage overhead in avatar rendering
Improving generalization across facial expressions in avatars
Innovation

Methods, ideas, or system contributions that make the work stand out.

Compact tensorial representation for dynamic textures
Static tri-planes and lightweight 1D feature lines
Adaptive opacity penalty and balanced sampling
🔎 Similar Papers
No similar papers found.
Y
Yating Wang
Shanghai Jiao Tong University
X
Xuan Wang
AntGroup Research
Ran Yi
Ran Yi
Associate Professor, Shanghai Jiao Tong University
Computer VisionComputer Graphics
Yanbo Fan
Yanbo Fan
Nanjing University
AI SafetyAIGC
J
Jichen Hu
Shanghai Jiao Tong University
J
Jingcheng Zhu
Shanghai Jiao Tong University
L
Lizhuang Ma
Shanghai Jiao Tong University