🤖 AI Summary
This work addresses the degraded geometric and appearance reconstruction in novel view synthesis under sparse-view settings, particularly in regions with insufficient co-visibility. To this end, we propose a co-visibility map-guided 3D Gaussian splatting framework. Our method introduces a co-visibility-aware adaptive weighting scheme for the rendering loss, an uncertainty-driven proximity classifier, and a COLMAP point cloud–enhanced Gaussian initialization strategy. Crucially, it is the first to explicitly model co-visibility as a geometric prior, enabling joint optimization across high- and low-uncertainty regions. Extensive experiments demonstrate state-of-the-art performance on the Mip-NeRF 360 and LLFF benchmarks. Moreover, our approach exhibits strong robustness and generalization across varying degrees of input sparsity, outperforming existing methods consistently in both quantitative metrics and visual quality.
📝 Abstract
We propose Covisibility Map-based Gaussian Splatting (CoMapGS), designed to recover underrepresented sparse regions in sparse novel view synthesis. CoMapGS addresses both high- and low-uncertainty regions by constructing covisibility maps, enhancing initial point clouds, and applying uncertainty-aware weighted supervision using a proximity classifier. Our contributions are threefold: (1) CoMapGS reframes novel view synthesis by leveraging covisibility maps as a core component to address region-specific uncertainty; (2) Enhanced initial point clouds for both low- and high-uncertainty regions compensate for sparse COLMAP-derived point clouds, improving reconstruction quality and benefiting few-shot 3DGS methods; (3) Adaptive supervision with covisibility-score-based weighting and proximity classification achieves consistent performance gains across scenes with varying sparsity scores derived from covisibility maps. Experimental results demonstrate that CoMapGS outperforms state-of-the-art methods on datasets including Mip-NeRF 360 and LLFF.