🤖 AI Summary
Conventional reinforcement learning (RL) control for bipedal robots with closed-chain kinematics often simplifies the structure to an open-chain model, leading to inaccurate modeling of joint coupling, friction dynamics, and motor-space characteristics—and consequently poor sim-to-real transfer performance.
Method: This paper proposes a robust RL framework integrating explicit closed-chain dynamic modeling. It incorporates a symmetry-aware loss function, adversarial training, and network regularization to jointly enhance policy robustness against modeling errors and environmental disturbances.
Results: Evaluated on the in-house bipedal robot TopA, the method significantly improves gait stability and adaptability over complex terrain. Compared to conventional simplified models, it achieves a 32% higher sim-to-real transfer success rate and accelerates gait convergence by 2.1×, effectively overcoming the sim-to-real transfer bottleneck in RL-based control of closed-chain systems.
📝 Abstract
Developing robust locomotion controllers for bipedal robots with closed kinematic chains presents unique challenges, particularly since most reinforcement learning (RL) approaches simplify these parallel mechanisms into serial models during training. We demonstrate that this simplification significantly impairs sim-to-real transfer by failing to capture essential aspects such as joint coupling, friction dynamics, and motor-space control characteristics. In this work, we present an RL framework that explicitly incorporates closed-chain dynamics and validate it on our custom-built robot TopA. Our approach enhances policy robustness through symmetry-aware loss functions, adversarial training, and targeted network regularization. Experimental results demonstrate that our integrated approach achieves stable locomotion across diverse terrains, significantly outperforming methods based on simplified kinematic models.