AquaticVision: Benchmarking Visual SLAM in Underwater Environment with Events and Frames

📅 2025-05-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing underwater visual SLAM datasets suffer from a lack of high-precision ground-truth trajectories, hindering quantitative algorithm evaluation; moreover, challenging conditions—such as low illumination and high turbidity—severely degrade localization and mapping performance. To address these limitations, we introduce AquaticVision, the first underwater dual-modal SLAM benchmark dataset equipped with optical motion-capture ground-truth trajectories. It is the first to synchronously capture event-based (DVS) and RGB frame-based data, accompanied by rigorous spatiotemporal calibration. The dataset encompasses diverse representative underwater scenarios, thereby filling a critical gap in objective, reproducible SLAM evaluation for aquatic environments. Upon open release, AquaticVision significantly improves pose estimation robustness and reproducibility under low-light and turbid conditions. It provides a standardized, reliable benchmark for algorithm development, ablation studies, and fair performance comparison in underwater visual SLAM.

Technology Category

Application Category

📝 Abstract
Many underwater applications, such as offshore asset inspections, rely on visual inspection and detailed 3D reconstruction. Recent advancements in underwater visual SLAM systems for aquatic environments have garnered significant attention in marine robotics research. However, existing underwater visual SLAM datasets often lack groundtruth trajectory data, making it difficult to objectively compare the performance of different SLAM algorithms based solely on qualitative results or COLMAP reconstruction. In this paper, we present a novel underwater dataset that includes ground truth trajectory data obtained using a motion capture system. Additionally, for the first time, we release visual data that includes both events and frames for benchmarking underwater visual positioning. By providing event camera data, we aim to facilitate the development of more robust and advanced underwater visual SLAM algorithms. The use of event cameras can help mitigate challenges posed by extremely low light or hazy underwater conditions. The webpage of our dataset is https://sites.google.com/view/aquaticvision-lias.
Problem

Research questions and friction points this paper is trying to address.

Lack of groundtruth data for underwater SLAM evaluation
Need for robust SLAM in low-light/hazy underwater conditions
Absence of combined event-frame datasets for aquatic benchmarking
Innovation

Methods, ideas, or system contributions that make the work stand out.

Novel underwater dataset with ground truth trajectory
Includes both event and frame visual data
Event cameras mitigate low light challenges
🔎 Similar Papers
No similar papers found.
Y
Yifan Peng
School of Data Science, The Chinese University of Hong Kong, Shenzhen, P. R. China
Y
Yuze Hong
School of Data Science, The Chinese University of Hong Kong, Shenzhen, P. R. China
Ziyang Hong
Ziyang Hong
The Chinese University of Hong Kong, Shenzhen
Mobile RoboticsRadarUnderwater RoboticsRobot LearningSLAM
A
Apple P. Y. Chui
School of Life Sciences, The Chinese University of Hong Kong, Hongkong
Junfeng Wu
Junfeng Wu
Huazhong University of Science and Technology
Computer Vision