π€ AI Summary
Visual place recognition (VPR) fails in long-term underwater ecological monitoring due to severe illumination variations, high turbidity, and texture scarcity. Method: This paper proposes a VPR-matching-segmentation co-design framework: (1) we introduce SQUIDLE+ VPRβthe first large-scale, cross-temporal, multi-platform, multi-environment underwater VPR benchmark; (2) we integrate SuperPoint/SuperGlue-based feature matching, semantic segmentation, and video frame sequence alignment to achieve robust scene re-identification, rigid registration, and interpretable ecological change analysis. Contribution/Results: Experiments demonstrate substantial improvements in relocalization accuracy and generalization across long-term (day-to-year), low-texture, and highly dynamic underwater environments. The framework has been validated on real autonomous underwater vehicles, confirming its efficacy for quantitative ecological change assessment.
π Abstract
Effective monitoring of underwater ecosystems is crucial for tracking environmental changes, guiding conservation efforts, and ensuring long-term ecosystem health. However, automating underwater ecosystem management with robotic platforms remains challenging due to the complexities of underwater imagery, which pose significant difficulties for traditional visual localization methods. We propose an integrated pipeline that combines Visual Place Recognition (VPR), feature matching, and image segmentation on video-derived images. This method enables robust identification of revisited areas, estimation of rigid transformations, and downstream analysis of ecosystem changes. Furthermore, we introduce the SQUIDLE+ VPR Benchmark-the first large-scale underwater VPR benchmark designed to leverage an extensive collection of unstructured data from multiple robotic platforms, spanning time intervals from days to years. The dataset encompasses diverse trajectories, arbitrary overlap and diverse seafloor types captured under varying environmental conditions, including differences in depth, lighting, and turbidity. Our code is available at: https://github.com/bev-gorry/underloc