๐ค AI Summary
To address the low mobility efficiency of conventional rovers in low-gravity environments (e.g., Moon and Mars), this work proposes Olympus, a hopping quadruped robot designed to tackle autonomous in-flight attitude regulation. Methodologically, we integrate physics-based multi-objective structural optimization, Proximal Policy Optimization (PPO)-based reinforcement learning (RL) for flight control policy design, and a real-time embedded attitude control framework. Notably, this is the first successful deployment of RL for in-flight attitude control on a planetary hopping robot, coupled with high-fidelity sim-to-real transfer. Experimental validation on physical hardware demonstrates robust multimodal aerial tumbling and stable landing: vertical hops reach 1.8 m, forward hops exceed 3.2 m, and attitude tracking error remains below 3ยฐ. These results significantly enhance maneuverability and mission adaptability under low-gravity conditions.
๐ Abstract
Exploring planetary bodies with lower gravity, such as the moon and Mars, allows legged robots to utilize jumping as an efficient form of locomotion thus giving them a valuable advantage over traditional rovers for exploration. Motivated by this fact, this paper presents the design, simulation, and learning-based"in-flight"attitude control of Olympus, a jumping legged robot tailored to the gravity of Mars. First, the design requirements are outlined followed by detailing how simulation enabled optimizing the robot's design - from its legs to the overall configuration - towards high vertical jumping, forward jumping distance, and in-flight attitude reorientation. Subsequently, the reinforcement learning policy used to track desired in-flight attitude maneuvers is presented. Successfully crossing the sim2real gap, extensive experimental studies of attitude reorientation tests are demonstrated.