On learning racing policies with reinforcement learning

πŸ“… 2025-04-03
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the insufficient reliability of fully autonomous driving under extreme operating conditions, this paper proposes an end-to-end reinforcement learning (RL) framework designed for zero-shot deployment on real vehicles, using autonomous racing as a representative task. Methodologically, it integrates domain randomization, high-fidelity actuator dynamics modeling, and a lightweight CNN-LSTM policy network, trained via PPO or SAC algorithms to yield high-performance driving policies. The key contributions are threefold: (1) the first demonstration of an RL policy achieving zero-shot deployment on the F1TENTH platform that outperforms human expert drivers; (2) superior performance over state-of-the-art model predictive control (MPC) approaches in extreme dynamic maneuvers; and (3) empirical validation of deep RL’s feasibility and advantages forζžι™ dynamic vehicle control, establishing a transferable technical pathway toward highly reliable autonomous driving.

Technology Category

Application Category

πŸ“ Abstract
Fully autonomous vehicles promise enhanced safety and efficiency. However, ensuring reliable operation in challenging corner cases requires control algorithms capable of performing at the vehicle limits. We address this requirement by considering the task of autonomous racing and propose solving it by learning a racing policy using Reinforcement Learning (RL). Our approach leverages domain randomization, actuator dynamics modeling, and policy architecture design to enable reliable and safe zero-shot deployment on a real platform. Evaluated on the F1TENTH race car, our RL policy not only surpasses a state-of-the-art Model Predictive Control (MPC), but, to the best of our knowledge, also represents the first instance of an RL policy outperforming expert human drivers in RC racing. This work identifies the key factors driving this performance improvement, providing critical insights for the design of robust RL-based control strategies for autonomous vehicles.
Problem

Research questions and friction points this paper is trying to address.

Develops RL policy for autonomous racing vehicles
Enhances performance beyond MPC and human drivers
Ensures safe zero-shot deployment on real platforms
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Reinforcement Learning for racing policy
Leverages domain randomization and actuator modeling
First RL policy outperforming human drivers
πŸ”Ž Similar Papers
No similar papers found.
G
Grzegorz Czechmanowski
IDEAS NCBR, Warsaw, Poland and with Institute of Robotics and Machine Intelligence, Poznan University of Technology, Poznan, Poland
J
Jan Wkegrzynowski
IDEAS NCBR, Warsaw, Poland and with Institute of Robotics and Machine Intelligence, Poznan University of Technology, Poznan, Poland
Piotr Kicki
Piotr Kicki
Poznan University of Technology
RoboticsDeep Neural NetworksMotion PlanningImage ProcessingMachine Learning
Krzysztof Walas
Krzysztof Walas
Assistant Professor, Institute of Robotics and Machine Intelligence, Poznan University of Technology
walking robotsenvironment perceptionphysical parameters perception