I-CTRL: Imitation to Control Humanoid Robots Through Constrained Reinforcement Learning

📅 2024-05-14
🏛️ arXiv.org
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the core challenge in humanoid robot motion imitation: high visual fidelity versus physical feasibility. We propose a constrained reinforcement learning (CRL) framework that leverages kinematically feasible but dynamically infeasible motion-retargeted trajectories as initial references, then refines them under full physical constraints to achieve high-fidelity, dynamically stable gait transfer. Our method integrates motion retargeting, trajectory-following optimization, and high-fidelity physics simulation (PyBullet/MuJoCo), employing a compact, general-purpose reward function. Crucially, we formulate motion imitation for the first time as constrained fine-tuning of non-physical reference trajectories. Experiments demonstrate significant improvements in tracking accuracy and execution stability across multiple large-scale human motion datasets. Moreover, the framework successfully generalizes to four distinct bipedal robot morphologies, validating its robustness, versatility, and practical applicability.

Technology Category

Application Category

📝 Abstract
This paper addresses the critical need for refining robot motions that, despite achieving a high visual similarity through human-to-humanoid retargeting methods, fall short of practical execution in the physical realm. Existing techniques in the graphics community often prioritize visual fidelity over physics-based feasibility, posing a significant challenge for deploying bipedal systems in practical applications. Our research introduces a constrained reinforcement learning algorithm to produce physics-based high-quality motion imitation onto legged humanoid robots that enhance motion resemblance while successfully following the reference human trajectory. We name our framework: I-CTRL. By reformulating the motion imitation problem as a constrained refinement over non-physics-based retargeted motions, our framework excels in motion imitation with simple and unique rewards that generalize across four robots. Moreover, our framework can follow large-scale motion datasets with a unique RL agent. The proposed approach signifies a crucial step forward in advancing the control of bipedal robots, emphasizing the importance of aligning visual and physical realism for successful motion imitation.
Problem

Research questions and friction points this paper is trying to address.

Enhancing humanoid robot motion imitation
Bridging visual and physical motion realism
Managing large-scale motion datasets efficiently
Innovation

Methods, ideas, or system contributions that make the work stand out.

Constrained reinforcement learning
Automatic priority scheduler
Physics-based motion imitation
🔎 Similar Papers
No similar papers found.
Yashuai Yan
Yashuai Yan
Ph.D. candidate, Vienna University of Technology
Machine LearningReinforcement LearningRobotics
E
Esteve Valls Mascaro
Autonomous Systems Lab, Technische Universität Wien (TU Wien), Vienna, Austria
T
Tobias Egle
Robotics Systems Lab, Technische Universität Wien (TU Wien), Vienna, Austria
Dongheui Lee
Dongheui Lee
Professor, Technische Universität Wien (TU Wien) // German Aerospace Center (DLR)
RoboticsMachine LearningHuman Robot InteractionHumanoid robots