🤖 AI Summary
This work addresses the insufficient translation/rotation invariance of deep learning–Gaussian mixture model (GMM) hybrid methods (e.g., DeepGMR) for partial-to-partial point cloud registration. We propose an Attention-guided Reference Point Self-adaptation (ARPS) mechanism that eliminates reliance on overlapping-region correspondence. Instead, an attention module dynamically identifies common reference points across two point clouds and integrates them into the GMM framework to enforce feature consistency. The resulting ARPS layer is plug-and-play, enhancing existing GMM-based registrars (e.g., DeepGMR, UGMMReg) with improved robustness and accuracy under rigid transformations. Extensive evaluation on multiple benchmarks demonstrates that our method surpasses state-of-the-art approaches leveraging Transformers or attention blocks for common-region extraction, validating the effectiveness and generalizability of transformation-invariant feature modeling.
📝 Abstract
This study investigates the impact of the invariance of feature vectors for partial-to-partial point set registration under translation and rotation of input point sets, particularly in the realm of techniques based on deep learning and Gaussian mixture models (GMMs). We reveal both theoretical and practical problems associated with such deep-learning-based registration methods using GMMs, with a particular focus on the limitations of DeepGMR, a pioneering study in this line, to the partial-to-partial point set registration. Our primary goal is to uncover the causes behind such methods and propose a comprehensible solution for that. To address this, we introduce an attention-based reference point shifting (ARPS) layer, which robustly identifies a common reference point of two partial point sets, thereby acquiring transformation-invariant features. The ARPS layer employs a well-studied attention module to find a common reference point rather than the overlap region. Owing to this, it significantly enhances the performance of DeepGMR and its recent variant, UGMMReg. Furthermore, these extension models outperform even prior deep learning methods using attention blocks and Transformer to extract the overlap region or common reference points. We believe these findings provide deeper insights into registration methods using deep learning and GMMs.