Improving Novel view synthesis of 360$^circ$ Scenes in Extremely Sparse Views by Jointly Training Hemisphere Sampled Synthetic Images

📅 2025-05-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses two key challenges in 360° novel view synthesis from only four input views: (1) failure of pose estimation under extreme sparsity, and (2) insufficient 3D scene coverage. To tackle these, we propose a hemispherical dense sampling and joint rendering strategy to strengthen geometric priors, and introduce a point-cloud-rendered image augmentation fine-tuning paradigm to mitigate overfitting and artifact generation. Our method integrates DUSt3R for robust pose initialization, 3D Gaussian Splatting for efficient scene representation, hemispherical-view rendering, and diffusion-based post-processing. Evaluated under the challenging 4-view setting, our approach significantly improves both completeness and visual fidelity of synthesized novel views: 3D spatial coverage increases by 27.3%, and artifact incidence decreases by 41.6%, outperforming existing state-of-the-art methods.

Technology Category

Application Category

📝 Abstract
Novel view synthesis in 360$^circ$ scenes from extremely sparse input views is essential for applications like virtual reality and augmented reality. This paper presents a novel framework for novel view synthesis in extremely sparse-view cases. As typical structure-from-motion methods are unable to estimate camera poses in extremely sparse-view cases, we apply DUSt3R to estimate camera poses and generate a dense point cloud. Using the poses of estimated cameras, we densely sample additional views from the upper hemisphere space of the scenes, from which we render synthetic images together with the point cloud. Training 3D Gaussian Splatting model on a combination of reference images from sparse views and densely sampled synthetic images allows a larger scene coverage in 3D space, addressing the overfitting challenge due to the limited input in sparse-view cases. Retraining a diffusion-based image enhancement model on our created dataset, we further improve the quality of the point-cloud-rendered images by removing artifacts. We compare our framework with benchmark methods in cases of only four input views, demonstrating significant improvement in novel view synthesis under extremely sparse-view conditions for 360$^circ$ scenes.
Problem

Research questions and friction points this paper is trying to address.

Novel view synthesis in 360° scenes from extremely sparse views
Estimating camera poses and generating dense point clouds in sparse-view cases
Improving 3D scene coverage and reducing overfitting with synthetic images
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses DUSt3R for camera pose estimation
Trains 3D Gaussian Splatting with synthetic images
Enhances images using diffusion-based model
🔎 Similar Papers
No similar papers found.
G
Guangan Chen
Image Processing and Interpretation (IPI), IMEC research group at Ghent University, Belgium
A
A. Truong
Image Processing and Interpretation (IPI), IMEC research group at Ghent University, Belgium
Hanhe Lin
Hanhe Lin
SSEN@University of Dundee
visual quality assessmentmedical image analysiscomputer vision
M
M. Vlaminck
Image Processing and Interpretation (IPI), IMEC research group at Ghent University, Belgium
W
Wilfried Philips
Image Processing and Interpretation (IPI), IMEC research group at Ghent University, Belgium
H
H. Luong
Image Processing and Interpretation (IPI), IMEC research group at Ghent University, Belgium