ClimbingCap: Multi-Modal Dataset and Method for Rock Climbing in World Coordinate

📅 2025-03-27
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing human motion reconstruction (HMR) methods are primarily designed for ground-level locomotion and struggle to accurately recover global 3D pose during off-ground climbing activities—largely due to the absence of large-scale, climbing-specific datasets with 3D global coordinate annotations and HMR frameworks tailored to aerial, non-planar environments. To address this, we introduce AscendMotion, the first large-scale multimodal (RGB/LiDAR/IMU) climbing motion dataset, comprising 412k frames captured from 22 climbers traversing challenging routes on 12 distinct climbing walls. We further propose ClimbingCap, a novel end-to-end neural motion reconstruction framework featuring joint optimization across RGB and LiDAR coordinate systems, incorporating LiDAR-derived global geometric constraints to enforce spatial consistency. Extensive experiments demonstrate that ClimbingCap significantly improves global position and orientation accuracy over state-of-the-art HMR methods, establishing a new benchmark for off-ground human motion analysis.

Technology Category

Application Category

📝 Abstract
Human Motion Recovery (HMR) research mainly focuses on ground-based motions such as running. The study on capturing climbing motion, an off-ground motion, is sparse. This is partly due to the limited availability of climbing motion datasets, especially large-scale and challenging 3D labeled datasets. To address the insufficiency of climbing motion datasets, we collect AscendMotion, a large-scale well-annotated, and challenging climbing motion dataset. It consists of 412k RGB, LiDAR frames, and IMU measurements, including the challenging climbing motions of 22 skilled climbing coaches across 12 different rock walls. Capturing the climbing motions is challenging as it requires precise recovery of not only the complex pose but also the global position of climbers. Although multiple global HMR methods have been proposed, they cannot faithfully capture climbing motions. To address the limitations of HMR methods for climbing, we propose ClimbingCap, a motion recovery method that reconstructs continuous 3D human climbing motion in a global coordinate system. One key insight is to use the RGB and LiDAR modalities to separately reconstruct motions in camera coordinates and global coordinates and to optimize them jointly. We demonstrate the quality of the AscendMotion dataset and present promising results from ClimbingCap. The AscendMotion dataset and source code release publicly at href{this link}{http://www.lidarhumanmotion.net/climbingcap/}
Problem

Research questions and friction points this paper is trying to address.

Lack of large-scale 3D labeled climbing motion datasets
Inability of existing HMR methods to capture climbing motions accurately
Need for precise recovery of climber pose and global position
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses RGB and LiDAR for motion reconstruction
Joint optimization in camera and global coordinates
Large-scale annotated climbing dataset AscendMotion
🔎 Similar Papers
No similar papers found.
M
Ming Yan
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; National Institute for Data Science in Health and Medicine, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
X
Xincheng Lin
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
Y
Yuhua Luo
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
S
Shuqi Fan
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
Yudi Dai
Yudi Dai
ETH Zurich
Computer VisionPoint CloudMotion captureHuman-Scene interactionMulti-modal
Q
Qixin Zhong
China National Climbing Team
L
Lincai Zhong
Ningbo Sports Work Training Team
Yuexin Ma
Yuexin Ma
Assistant Professor, School of Information Science and Technology, ShanghaiTech University
computer visionembodied AIautonomous driving
L
Lan Xu
ShanghaiTech University
Chenglu Wen
Chenglu Wen
Professor of Xiamen University
3D visionpoint cloudsmobile mappingrobotics
S
Siqi Shen
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University
C
Cheng Wang
Fujian Key Laboratory of Sensing and Computing for Smart Cities, Xiamen University; Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, School of Informatics, Xiamen University