Personalize Your Gaussian: Consistent 3D Scene Personalization from a Single Image

📅 2025-05-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the core challenges of multi-view inconsistency and poor reference consistency in single-reference-image-driven 3D scene personalization, this paper proposes a progressive single-view appearance propagation framework. Our method integrates the 3D generation priors of pre-trained diffusion models with geometry-guided LoRA-based efficient fine-tuning, enabling cross-view appearance consistency modeling directly on 3D Gaussian Splatting (3DGS) representations and effectively mitigating viewpoint bias induced by single-view input. Crucially, we decouple appearance propagation into coarse-grained geometric constraint enforcement and fine-grained texture refinement via iterative optimization, significantly improving both multi-view consistency and image fidelity. Extensive experiments demonstrate that our approach consistently outperforms existing single-image 3D personalization methods on real-world scenes. This work establishes a new paradigm for few-shot 3D content creation.

Technology Category

Application Category

📝 Abstract
Personalizing 3D scenes from a single reference image enables intuitive user-guided editing, which requires achieving both multi-view consistency across perspectives and referential consistency with the input image. However, these goals are particularly challenging due to the viewpoint bias caused by the limited perspective provided in a single image. Lacking the mechanisms to effectively expand reference information beyond the original view, existing methods of image-conditioned 3DGS personalization often suffer from this viewpoint bias and struggle to produce consistent results. Therefore, in this paper, we present Consistent Personalization for 3D Gaussian Splatting (CP-GS), a framework that progressively propagates the single-view reference appearance to novel perspectives. In particular, CP-GS integrates pre-trained image-to-3D generation and iterative LoRA fine-tuning to extract and extend the reference appearance, and finally produces faithful multi-view guidance images and the personalized 3DGS outputs through a view-consistent generation process guided by geometric cues. Extensive experiments on real-world scenes show that our CP-GS effectively mitigates the viewpoint bias, achieving high-quality personalization that significantly outperforms existing methods. The code will be released at https://github.com/Yuxuan-W/CP-GS.
Problem

Research questions and friction points this paper is trying to address.

Achieving multi-view consistency from single image
Overcoming viewpoint bias in 3D scene personalization
Propagating reference appearance to novel perspectives
Innovation

Methods, ideas, or system contributions that make the work stand out.

Propagates single-view reference appearance progressively
Integrates pre-trained image-to-3D generation and LoRA fine-tuning
Uses view-consistent generation guided by geometric cues
🔎 Similar Papers
No similar papers found.
Y
Yuxuan Wang
Nanyang Technological University
Xuanyu Yi
Xuanyu Yi
ByteDance Seed
3D VisionGenerative Model
Qingshan Xu
Qingshan Xu
Nanyang Technological University
Computer Vision3D Reconstruction
Y
Yuan Zhou
Nanyang Technological University
L
Long Chen
Hong Kong University of Science and Technology
H
Hanwang Zhang
Nanyang Technological University