Segmentation-Guided Neural Radiance Fields for Novel Street View Synthesis

📅 2025-03-18
🏛️ Proceedings of the 20th International Joint Conference on Computer Vision, Imaging and Computer Graphics Theory and Applications
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address distortions in large-scale outdoor street-scene NeRF reconstruction caused by dynamic objects, sparse camera coverage, illumination variations, and texture scarcity, this paper proposes a semantic-guided robust NeRF method. Methodologically, it integrates Grounded SAM–generated segmentation masks—enabling dynamic object removal, sky modeling, and ground-plane geometric regularization—for the first time; introduces learnable appearance embeddings to adaptively correct inter-view illumination inconsistencies; and unifies semantic guidance, multi-scale volumetric rendering, and implicit geometric constraints within the ZipNeRF framework. Evaluated on real-world street-scene datasets, the method achieves significant improvements over baselines: synthesized images exhibit fewer artifacts and sharper edges, with PSNR increased by 2.1 dB and SSIM improved by 0.032.

Technology Category

Application Category

📝 Abstract
Recent advances in Neural Radiance Fields (NeRF) have shown great potential in 3D reconstruction and novel view synthesis, particularly for indoor and small-scale scenes. However, extending NeRF to large-scale outdoor environments presents challenges such as transient objects, sparse cameras and textures, and varying lighting conditions. In this paper, we propose a segmentation-guided enhancement to NeRF for outdoor street scenes, focusing on complex urban environments. Our approach extends ZipNeRF and utilizes Grounded SAM for segmentation mask generation, enabling effective handling of transient objects, modeling of the sky, and regularization of the ground. We also introduce appearance embeddings to adapt to inconsistent lighting across view sequences. Experimental results demonstrate that our method outperforms the baseline ZipNeRF, improving novel view synthesis quality with fewer artifacts and sharper details.
Problem

Research questions and friction points this paper is trying to address.

Extends NeRF for large-scale outdoor street scenes
Handles transient objects and varying lighting conditions
Improves novel view synthesis quality with fewer artifacts
Innovation

Methods, ideas, or system contributions that make the work stand out.

Segmentation-guided NeRF for outdoor scenes
Uses Grounded SAM for transient object handling
Appearance embeddings adapt to lighting variations
🔎 Similar Papers
No similar papers found.
Y
Yizhou Li
Institute of Science Tokyo, Tokyo, Japan
Yusuke Monno
Yusuke Monno
Institute of Science Tokyo
Image ProcessingImage SensorsComputer VisionBiomedical Engineering
Masatoshi Okutomi
Masatoshi Okutomi
Tokyo Institute of Technology
computer visionimage processingartificial intelligence
Y
Yuuichi Tanaka
Micware Mobility Co., Ltd., Hyogo, Japan
S
Seiichi Kataoka
Micware Automotive Co., Ltd., Hyogo, Japan
T
Teruaki Kosiba
Micware Navigations Co., Ltd, Hyogo, Japan