Reinforcement Learning for Robust Athletic Intelligence: Lessons from the 2nd 'AI Olympics with RealAIGym' Competition

📅 2025-03-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses robust swing-up control of an underactuated, chaotic double-pendulum system on real hardware, focusing on sim-to-real transfer of reinforcement learning (RL) controllers. We propose an integrated RL deployment framework combining domain randomization, online adaptation, and embedded real-time control, and systematically evaluate four state-of-the-art algorithms—PPO, SAC, TD3, and DreamerV3. To our knowledge, this is the first benchmark study at IROS comparing algorithmic robustness and transferability on a chaotic physical platform, establishing a new hardware-centric evaluation paradigm for “movement intelligence.” All methods successfully achieve swing-up control on the physical double pendulum. DreamerV3 demonstrates superior sim-to-real transfer performance, while PPO exhibits the highest disturbance robustness—achieving a 42% improvement in success rate under external perturbations.

Technology Category

Application Category

📝 Abstract
In the field of robotics many different approaches ranging from classical planning over optimal control to reinforcement learning (RL) are developed and borrowed from other fields to achieve reliable control in diverse tasks. In order to get a clear understanding of their individual strengths and weaknesses and their applicability in real world robotic scenarios is it important to benchmark and compare their performances not only in a simulation but also on real hardware. The '2nd AI Olympics with RealAIGym' competition was held at the IROS 2024 conference to contribute to this cause and evaluate different controllers according to their ability to solve a dynamic control problem on an underactuated double pendulum system with chaotic dynamics. This paper describes the four different RL methods submitted by the participating teams, presents their performance in the swing-up task on a real double pendulum, measured against various criteria, and discusses their transferability from simulation to real hardware and their robustness to external disturbances.
Problem

Research questions and friction points this paper is trying to address.

Benchmark RL methods for robotic control tasks
Evaluate RL performance on real hardware systems
Assess robustness and simulation-to-reality transferability
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reinforcement Learning for dynamic control tasks
Benchmarking RL methods on real hardware
Evaluating robustness and transferability from simulation
🔎 Similar Papers
No similar papers found.
F
Felix Wiebe
Robotics Innovation Center, German Research Center for Artificial Intelligence (DFKI), Germany
N
Niccolo Turcato
Department of Information Engineering, University of Padova, Italy
Alberto Dalla Libera
Alberto Dalla Libera
Research Fellow, University of Padova
J
Jean Seong Bjorn Choe
Korea University, Seoul, South Korea
B
Bumkyu Choi
Korea University, Seoul, South Korea
T
Tim Lukas Faust
Department of Information Engineering, University of Padova, Italy
H
Habib Maraqten
Department of Information Engineering, University of Padova, Italy
E
Erfan Aghadavoodi
Robotics Innovation Center, German Research Center for Artificial Intelligence (DFKI), Germany
M
Marco Cali
Department of Information Engineering, University of Padova, Italy
Alberto Sinigaglia
Alberto Sinigaglia
PhD student
Deep Reinforcement LearningDeep Learning
Giulio Giacomuzzo
Giulio Giacomuzzo
PhD student, University of Padova
Learning for controlHuman Robot Interaction
Diego Romeres
Diego Romeres
Senior Principal Research Scientist & Team Leader at Mitsubishi Electric Research Laboratories
RoboticsMachine LearningBayesian EstimationOptimization
J
Jong-kook Kim
Korea University, Seoul, South Korea
G
Gian Antonio Susto
Department of Information Engineering, University of Padova, Italy
S
Shubham Vyas
Korea University, Seoul, South Korea
Dennis Mronga
Dennis Mronga
Postdoc, DFKI Robotics Innovation Center
Humanoid Robot Motion Planning and Control
Boris Belousov
Boris Belousov
Senior Researcher at German Research Centre for Artificial Intelligence (DFKI GmbH)
Robot LearningReinforcement LearningMachine LearningRobotics
J
Jan Peters
Technical University of Darmstadt, Germany; Center for Cognitive Science, Germany; Hessian.AI, Germany
Frank Kirchner
Frank Kirchner
Professor für Robotik, Universität Bremen, DFKI
artificial intelligenceroboticsmachine learningHuman-Machine-Interfacewalking robots
S
Shivesh Kumar
Robotics Innovation Center, German Research Center for Artificial Intelligence (DFKI), Germany; Chalmers University of Technology, Sweden