π€ AI Summary
Legged robots struggle to simultaneously achieve autonomous exploration and dexterous physical interaction in unstructured environments. To address this, we propose a hierarchical loco-manipulation framework: a high-level, semantic-goal-conditioned autonomous exploration strategy for target localization and navigation; and a low-level, unified whole-body torque control policy that coordinates leg and manipulator motions for fine-grained force interaction. The framework decouples high-level semantic decision-making from low-level motion execution, enabling, for the first time on embodied legged robots, unified real-time closed-loop modeling of semantic navigation, dynamic target search, and multi-joint coordinated manipulation. Experiments demonstrate the navigation moduleβs ability to accurately localize specified semantic targets in dynamic scenes, establishing a systematic foundation for end-to-end mobile manipulation.
π Abstract
Seamless loco-manipulation in unstructured environments requires robots to leverage autonomous exploration alongside whole-body control for physical interaction. In this work, we introduce HANDO (Hierarchical Autonomous Navigation and Dexterous Omni-loco-manipulation), a two-layer framework designed for legged robots equipped with manipulators to perform human-centered mobile manipulation tasks. The first layer utilizes a goal-conditioned autonomous exploration policy to guide the robot to semantically specified targets, such as a black office chair in a dynamic environment. The second layer employs a unified whole-body loco-manipulation policy to coordinate the arm and legs for precise interaction tasks-for example, handing a drink to a person seated on the chair. We have conducted an initial deployment of the navigation module, and will continue to pursue finer-grained deployment of whole-body loco-manipulation.