Reinforcement Learning Within the Classical Robotics Stack: A Case Study in Robot Soccer

📅 2024-12-12
🏛️ arXiv.org
📈 Citations: 1
✨ Influential: 0
📄 PDF
🤖 AI Summary
Addressing real-time, partially observable multi-robot decision-making under dynamic environments. Method: We propose a hybrid hierarchical architecture that tightly integrates model-free reinforcement learning (PPO/SAC) into the classical robotics stack via sub-behavior decomposition and heuristic scheduling, enabling end-to-end decision-making; augmented by multi-fidelity sim2real transfer (Gazebo → physical platform) and co-optimization of motion planning and state estimation modules. Contribution/Results: Our work introduces the first tightly coupled integration mechanism between RL modules and conventional robot software architectures, alongside a lightweight generalization strategy—“sub-behavior learning + heuristic selection”—that ensures millisecond-level latency while significantly improving robustness and environmental adaptability. The system secured first place in the Shield Challenge of the RoboCup Standard Platform League 2024. Real-robot evaluations demonstrate high task success rates, low end-to-end latency, and deployment stability.

Technology Category

Application Category

📝 Abstract
Robot decision-making in partially observable, real-time, dynamic, and multi-agent environments remains a difficult and unsolved challenge. Model-free reinforcement learning (RL) is a promising approach to learning decision-making in such domains, however, end-to-end RL in complex environments is often intractable. To address this challenge in the RoboCup Standard Platform League (SPL) domain, we developed a novel architecture integrating RL within a classical robotics stack, while employing a multi-fidelity sim2real approach and decomposing behavior into learned sub-behaviors with heuristic selection. Our architecture led to victory in the 2024 RoboCup SPL Challenge Shield Division. In this work, we fully describe our system's architecture and empirically analyze key design decisions that contributed to its success. Our approach demonstrates how RL-based behaviors can be integrated into complete robot behavior architectures.
Problem

Research questions and friction points this paper is trying to address.

Robot decision-making in complex, dynamic environments
Integration of RL within classical robotics stack
Sim2real approach for multi-agent robot soccer
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates RL within classical robotics stack
Uses multi-fidelity sim2real approach
Decomposes behavior into learned sub-behaviors
🔎 Similar Papers
No similar papers found.
A
Adam Labiosa
University of Wisconsin–Madison
Z
Zhihan Wang
The University of Texas at Austin
Siddhant Agarwal
Siddhant Agarwal
The University of Texas at Austin
Reinforcement LearningAdvesarial AttacksExplainable AIRobotics
W
William Cong
University of Wisconsin–Madison
G
Geethika Hemkumar
The University of Texas at Austin
A
A. Harish
University of Wisconsin–Madison
B
Benjamin Hong
University of Wisconsin–Madison
J
Josh Kelle
The University of Texas at Austin
C
Chen Li
University of Wisconsin–Madison
Y
Yuhao Li
University of Wisconsin–Madison
Z
Zisen Shao
University of Wisconsin–Madison
P
Peter Stone
The University of Texas at Austin, Sony AI
J
Josiah P. Hanna
University of Wisconsin–Madison