🤖 AI Summary
This work addresses the sim-to-real transfer challenge for RGB-only vision-driven humanoid robots performing robust manipulation of diverse articulated objects (e.g., various door types) in real-world environments. We propose a two-stage reset-exploration strategy and GRPO-based policy fine-tuning in simulation. Leveraging GPU-accelerated photorealistic simulation, multi-level physical and visual randomization, and a teacher–student–bootstrapping learning framework, we achieve end-to-end, long-horizon pixel-to-full-body control policy transfer. To our knowledge, this is the first demonstration of zero-shot operation of unseen door types by a humanoid robot using only RGB input. Our method reduces task completion time by 31.7% compared to human teleoperation, validating the effectiveness and scalability of full-stack simulation training for complex, dexterous human–robot interaction tasks.
📝 Abstract
Recent progress in GPU-accelerated, photorealistic simulation has opened a scalable data-generation path for robot learning, where massive physics and visual randomization allow policies to generalize beyond curated environments. Building on these advances, we develop a teacher-student-bootstrap learning framework for vision-based humanoid loco-manipulation, using articulated-object interaction as a representative high-difficulty benchmark. Our approach introduces a staged-reset exploration strategy that stabilizes long-horizon privileged-policy training, and a GRPO-based fine-tuning procedure that mitigates partial observability and improves closed-loop consistency in sim-to-real RL. Trained entirely on simulation data, the resulting policy achieves robust zero-shot performance across diverse door types and outperforms human teleoperators by up to 31.7% in task completion time under the same whole-body control stack. This represents the first humanoid sim-to-real policy capable of diverse articulated loco-manipulation using pure RGB perception.