Coordinated Humanoid Robot Locomotion with Symmetry Equivariant Reinforcement Learning Policy

📅 2025-08-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing deep reinforcement learning methods neglect the morphological symmetry of humanoid robots, leading to uncoordinated locomotion and poor balance control. To address this, we propose SE-Policy—a novel policy optimization framework that rigorously enforces spatial symmetry: it embeds *equivariance* into the policy network to ensure coordinated bilateral actuation, and imposes *invariance* in the critic network to enhance training stability—both without introducing additional hyperparameters. Built upon symmetric equivariant neural networks, SE-Policy enables spatiotemporally consistent motion control across simulation and real-world deployment. Evaluated on velocity-tracking tasks in MuJoCo and on the Unitree G1 robot, SE-Policy achieves up to 40% higher tracking accuracy over state-of-the-art methods, while significantly improving motion coordination, robustness to disturbances, and generalization across diverse terrains and gaits.

Technology Category

Application Category

📝 Abstract
The human nervous system exhibits bilateral symmetry, enabling coordinated and balanced movements. However, existing Deep Reinforcement Learning (DRL) methods for humanoid robots neglect morphological symmetry of the robot, leading to uncoordinated and suboptimal behaviors. Inspired by human motor control, we propose Symmetry Equivariant Policy (SE-Policy), a new DRL framework that embeds strict symmetry equivariance in the actor and symmetry invariance in the critic without additional hyperparameters. SE-Policy enforces consistent behaviors across symmetric observations, producing temporally and spatially coordinated motions with higher task performance. Extensive experiments on velocity tracking tasks, conducted in both simulation and real-world deployment with the Unitree G1 humanoid robot, demonstrate that SE-Policy improves tracking accuracy by up to 40% compared to state-of-the-art baselines, while achieving superior spatial-temporal coordination. These results demonstrate the effectiveness of SE-Policy and its broad applicability to humanoid robots.
Problem

Research questions and friction points this paper is trying to address.

Improving humanoid robot locomotion coordination using symmetry equivariant reinforcement learning
Addressing suboptimal behaviors in DRL by embedding morphological symmetry in policies
Enhancing tracking accuracy and spatial-temporal coordination in humanoid robot movements
Innovation

Methods, ideas, or system contributions that make the work stand out.

Symmetry Equivariant Policy for coordinated locomotion
Embeds symmetry equivariance without extra hyperparameters
Improves tracking accuracy by 40% in experiments
🔎 Similar Papers
No similar papers found.
Buqing Nie
Buqing Nie
Shanghai Jiao Tong University
Reinforcement LearningRobot Learning
Y
Yang Zhang
MoE Key Lab of Artificial Intelligence and AI Institute, Shanghai Jiao Tong University, Shanghai, China
R
Rongjun Jin
MoE Key Lab of Artificial Intelligence and AI Institute, Shanghai Jiao Tong University, Shanghai, China
Zhanxiang Cao
Zhanxiang Cao
上海交通大学
RoboticsReinforcement LearningLegged Robot
Huangxuan Lin
Huangxuan Lin
Shanghai Jiao Tong University
X
Xiaokang Yang
MoE Key Lab of Artificial Intelligence and AI Institute, Shanghai Jiao Tong University, Shanghai, China
Y
Yue Gao
MoE Key Lab of Artificial Intelligence and AI Institute, Shanghai Jiao Tong University, Shanghai, China