Consistent Zero-shot 3D Texture Synthesis Using Geometry-aware Diffusion and Temporal Video Models

📅 2025-06-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing 3D texture synthesis methods suffer from single-view input constraints and inadequate geometric understanding, leading to inconsistent inter-patch seams, semantic distortion in occluded regions, and temporal flickering under dynamic scenes. To address these limitations, we propose the first video-generation-based framework for 3D texture synthesis, innovatively integrating geometry-aware diffusion with structural UV-space modeling. Our method employs mesh-structured conditioning to enforce geometric awareness during generation and introduces a spatiotemporal diffusion strategy explicitly defined over the UV parameterization—thereby modeling both inter-patch topological relationships and inter-frame temporal dependencies. Experiments demonstrate significant improvements over state-of-the-art approaches in three key metrics: texture fidelity, seam coherence, and temporal stability. The framework enables high-fidelity, dynamically consistent, and real-time 3D content generation.

Technology Category

Application Category

📝 Abstract
Current texture synthesis methods, which generate textures from fixed viewpoints, suffer from inconsistencies due to the lack of global context and geometric understanding. Meanwhile, recent advancements in video generation models have demonstrated remarkable success in achieving temporally consistent videos. In this paper, we introduce VideoTex, a novel framework for seamless texture synthesis that leverages video generation models to address both spatial and temporal inconsistencies in 3D textures. Our approach incorporates geometry-aware conditions, enabling precise utilization of 3D mesh structures. Additionally, we propose a structure-wise UV diffusion strategy, which enhances the generation of occluded areas by preserving semantic information, resulting in smoother and more coherent textures. VideoTex not only achieves smoother transitions across UV boundaries but also ensures high-quality, temporally stable textures across video frames. Extensive experiments demonstrate that VideoTex outperforms existing methods in texture fidelity, seam blending, and stability, paving the way for dynamic real-time applications that demand both visual quality and temporal coherence.
Problem

Research questions and friction points this paper is trying to address.

Address inconsistent 3D textures from fixed viewpoints
Leverage video models for spatial-temporal texture coherence
Enhance occluded area generation with geometry-aware conditions
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages video models for consistent textures
Uses geometry-aware 3D mesh conditions
Proposes structure-wise UV diffusion strategy
🔎 Similar Papers
No similar papers found.
Donggoo Kang
Donggoo Kang
Chung-Ang University, Department of Image
Computer visionGraphics AIVision-Language Model
Jangyeong Kim
Jangyeong Kim
Researcher, NCSOFT inc.
D
Dasol Jeong
Department of Image, Chung-Ang University
J
Junyoung Choi
Graphics AI Lab, NCSOFT
J
Jeonga Wi
Graphics AI Lab, NCSOFT
H
Hyunmin Lee
Department of Image, Chung-Ang University
J
Joonho Gwon
Department of Computer Science, University of Seoul
Joonki Paik
Joonki Paik
Chung-Ang University
Image ProcessingVideo Analysis