MobileH2R: Learning Generalizable Human to Mobile Robot Handover Exclusively from Scalable and Diverse Synthetic Data

📅 2025-01-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of enabling mobile robots to reliably receive objects from humans in large-scale workspaces, this paper proposes a vision-driven human–robot handover learning framework trained exclusively on synthetic data. Methodologically, we introduce (i) a full-body pose-controllable synthetic data generation pipeline; (ii) an automated safe demonstration construction mechanism; and (iii) a novel 4D (3D spatial + temporal) imitation learning paradigm for end-to-end base–manipulator coordinated closed-loop control. Our key contribution is the first demonstration that high-quality, large-scale, and diverse synthetic data alone—without any real human demonstrations—can train a handover policy with strong generalization capability. Both simulation and real-robot experiments show over 15% improvement in handover success rate, validating that synthetic data can effectively replace real annotated data, significantly enhancing model generalization and deployment efficiency.

Technology Category

Application Category

📝 Abstract
This paper introduces MobileH2R, a framework for learning generalizable vision-based human-to-mobile-robot (H2MR) handover skills. Unlike traditional fixed-base handovers, this task requires a mobile robot to reliably receive objects in a large workspace enabled by its mobility. Our key insight is that generalizable handover skills can be developed in simulators using high-quality synthetic data, without the need for real-world demonstrations. To achieve this, we propose a scalable pipeline for generating diverse synthetic full-body human motion data, an automated method for creating safe and imitation-friendly demonstrations, and an efficient 4D imitation learning method for distilling large-scale demonstrations into closed-loop policies with base-arm coordination. Experimental evaluations in both simulators and the real world show significant improvements (at least +15% success rate) over baseline methods in all cases. Experiments also validate that large-scale and diverse synthetic data greatly enhances robot learning, highlighting our scalable framework.
Problem

Research questions and friction points this paper is trying to address.

Mobile Robotics
Human-Robot Interaction
Object Handover
Innovation

Methods, ideas, or system contributions that make the work stand out.

MobileH2R
Virtual Data Training
Human-Robot Interaction
🔎 Similar Papers
No similar papers found.
Z
Zifan Wang
Tsinghua University, Galbot
Ziqing Chen
Ziqing Chen
Tsinghua University
robotics
J
Junyu Chen
Tsinghua University
J
Jilong Wang
Galbot, Peking University
Y
Yuxin Yang
Galbot
Yunze Liu
Yunze Liu
IIIS, Tsinghua University; Memories.ai Research
AI Memories3D Computer VisionEmbodied AIEgocentric Video
Xueyi Liu
Xueyi Liu
Institute of Automation, Chinese Academy of Sciences
H
He Wang
Galbot, Peking University
L
Li Yi
Tsinghua University, Shanghai Artificial Intelligence Laboratory, Shanghai Qi Zhi Institute