🤖 AI Summary
To address the challenges of registration for large-scale point clouds under extremely low overlap (e.g., <5%), including poor robustness, high computational cost, and registration failure, this paper proposes the first modality-transformative, BEV-driven registration framework. It projects 3D point clouds onto bird’s-eye-view (BEV) images, leverages SuperPoint and SuperGlue for 2D keypoint detection and matching, and recovers high-accuracy 3D correspondences via inverse geometric mapping. By decoupling registration from direct 3D geometric constraints, the method exploits maximal BEV overlap and complementary image-space features, thereby alleviating reliance on local 3D overlap. Evaluated on the GrAco dataset—a large-scale (8 km²), multi-source LiDAR benchmark with ultra-low overlap—the approach significantly outperforms state-of-the-art methods including ICP, FPFH, and RPMNet, achieving a 27.6% improvement in registration accuracy and reducing the failure rate to just 1.2%.
📝 Abstract
Large-scale scene point cloud registration with limited overlap is a challenging task due to computational load and constrained data acquisition. To tackle these issues, we propose a point cloud registration method, MT-PCR, based on Modality Transformation. MT-PCR leverages a BEV capturing the maximal overlap information to improve the accuracy and utilizes images to provide complementary spatial features. Specifically, MT-PCR converts 3D point clouds to BEV images and eastimates correspondence by 2D image keypoints extraction and matching. Subsequently, the 2D correspondence estimates are then transformed back to 3D point clouds using inverse mapping. We have applied MT-PCR to Terrestrial Laser Scanning and Aerial Laser Scanning point cloud registration on the GrAco dataset, involving 8 low-overlap, square-kilometer scale registration scenarios. Experiments and comparisons with commonly used methods demonstrate that MT-PCR can achieve superior accuracy and robustness in large-scale scenes with limited overlap.