RoomPainter: View-Integrated Diffusion for Consistent Indoor Scene Texturing

📅 2024-12-21
🏛️ arXiv.org
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
Indoor scene texture synthesis faces challenges including inter-view inconsistency, visible seams, and high computational overhead. This paper introduces RoomPainter, the first zero-shot 2D diffusion-based framework for indoor texture synthesis. It achieves efficient, high-fidelity, and geometrically consistent 3D textures via a global–local two-stage strategy. In the first stage, multi-view integrated sampling (MVIS) combined with neighborhood-enhanced attention generates holistic textures. In the second stage, a repaint-based multi-view refinement strategy (MVRS) performs occlusion-aware, instance-level fine-tuning. Crucially, RoomPainter requires no fine-tuning or iterative optimization. It significantly reduces computational cost while achieving state-of-the-art visual quality, inter-view consistency, and seamless texture synthesis.

Technology Category

Application Category

📝 Abstract
Indoor scene texture synthesis has garnered significant interest due to its important potential applications in virtual reality, digital media and creative arts. Existing diffusion-model-based researches either rely on per-view inpainting techniques, which are plagued by severe cross-view inconsistencies and conspicuous seams, or adopt optimization-based approaches that involve substantial computational overhead. In this work, we present RoomPainter, a framework that seamlessly integrates efficiency and consistency to achieve high-fidelity texturing of indoor scenes. The core of RoomPainter features a zero-shot technique that effectively adapts a 2D diffusion model for 3D-consistent texture synthesis, along with a two-stage generation strategy that ensures both global and local consistency. Specifically, we introduce Attention-Guided Multi-View Integrated Sampling (MVIS) combined with a neighbor-integrated attention mechanism for zero-shot texture map generation. Using the MVIS, we firstly generate texture map for the entire room to ensure global consistency, then adopt its variant, namely Attention-Guided Multi-View Integrated Repaint Sampling (MVRS) to repaint individual instances within the room, thereby further enhancing local consistency and addressing the occlusion problem. Experiments demonstrate that RoomPainter achieves superior performance for indoor scene texture synthesis in visual quality, global consistency and generation efficiency.
Problem

Research questions and friction points this paper is trying to address.

Achieving cross-view consistency in indoor scene texturing
Reducing computational overhead in texture synthesis
Enhancing global and local texture consistency efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

Zero-shot 2D diffusion for 3D textures
Two-stage global-local consistency strategy
Attention-guided multi-view integrated sampling
🔎 Similar Papers
No similar papers found.