🤖 AI Summary
To address the challenge of autonomous locomotion-manipulation for the COBRA snake robot in unstructured environments, this work introduces the first end-to-end vision-guided locomanipulation system. Methodologically, it integrates YOLOv8 for real-time object detection with binocular stereo vision, proposes a lightweight 6-DOF pose estimation algorithm, and designs a closed-loop feedback control strategy based on kinematic modeling. This enables, for the first time, full 6-DOF closed-loop locomotion-manipulation coordination on a snake robot. Experiments in realistic settings demonstrate robust performance across object detection, 6-DOF pose estimation, grasping, transport, and millimeter-precision placement—achieving >15 FPS while significantly enhancing task robustness and real-time responsiveness. This work establishes a scalable, perception-planning-control integrated paradigm for autonomous operation of soft and hyper-redundant robots.
📝 Abstract
This paper presents the development and integration of a vision-guided loco-manipulation pipeline for Northeastern University's snake robot, COBRA. The system leverages a YOLOv8-based object detection model and depth data from an onboard stereo camera to estimate the 6-DOF pose of target objects in real time. We introduce a framework for autonomous detection and control, enabling closed-loop loco-manipulation for transporting objects to specified goal locations. Additionally, we demonstrate open-loop experiments in which COBRA successfully performs real-time object detection and loco-manipulation tasks.