🤖 AI Summary
To address bottlenecks in soft robot 3D shape reconstruction—namely, reliance on artificial markers, large-scale annotated training data, or complex multi-camera setups—this paper proposes a marker-free, training-free vision-based method that achieves real-time, robust 3D reconstruction solely from the robot’s natural surface texture. The approach employs a hierarchical matching strategy: at the lower level, appearance-based feature tracking and implicit visual landmark matching enable multi-view local alignment; at the upper level, kinematic constraints guide global optimization, decoupling deformation modeling from rigid-body motion. The framework requires no specialized background, labeled datasets, or custom hardware, enabling real-time continuum robot tracking in dynamic environments (average end-effector localization error: 2.6% of total length) and demonstrating stability and practicality in closed-loop control. Its core contribution is the first end-to-end, unsupervised 3D shape reconstruction paradigm fully grounded in natural appearance cues.
📝 Abstract
Accurate shape reconstruction is essential for precise control and reliable operation of soft robots. Compared to sensor-based approaches, vision-based methods offer advantages in cost, simplicity, and ease of deployment. However, existing vision-based methods often rely on complex camera setups, specific backgrounds, or large-scale training datasets, limiting their practicality in real-world scenarios. In this work, we propose a vision-based, markerless, and training-free framework for soft robot shape reconstruction that directly leverages the robot's natural surface appearance. These surface features act as implicit visual markers, enabling a hierarchical matching strategy that decouples local partition alignment from global kinematic optimization. Requiring only an initial 3D reconstruction and kinematic alignment, our method achieves real-time shape tracking across diverse environments while maintaining robustness to occlusions and variations in camera viewpoints. Experimental validation on a continuum soft robot demonstrates an average tip error of 2.6% during real-time operation, as well as stable performance in practical closed-loop control tasks. These results highlight the potential of the proposed approach for reliable, low-cost deployment in dynamic real-world settings.