JuggleRL: Mastering Ball Juggling with a Quadrotor via Deep Reinforcement Learning

📅 2025-09-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the aerial juggling task of a quadrotor using a racket to hit and catch a ball—a challenge demanding millisecond-precision timing, high-dynamic contact stability, and strong environmental adaptability. We present the first application of deep reinforcement learning (DRL) to this nonlinear, underactuated, multimodal contact control problem. Our method integrates system dynamics modeling, hierarchical reward shaping, and domain randomization within a simulation-based training framework, while co-designing a low-level controller and a lightweight communication protocol. Real-world zero-shot deployment achieves an average of 311 consecutive hits (peak: 462), substantially outperforming conventional model-based approaches (3.1 hits). Moreover, without fine-tuning, the policy generalizes robustly to ultra-light 5g balls, sustaining an average of 145.9 hits—demonstrating exceptional robustness and cross-environment generalization capability.

Technology Category

Application Category

📝 Abstract
Aerial robots interacting with objects must perform precise, contact-rich maneuvers under uncertainty. In this paper, we study the problem of aerial ball juggling using a quadrotor equipped with a racket, a task that demands accurate timing, stable control, and continuous adaptation. We propose JuggleRL, the first reinforcement learning-based system for aerial juggling. It learns closed-loop policies in large-scale simulation using systematic calibration of quadrotor and ball dynamics to reduce the sim-to-real gap. The training incorporates reward shaping to encourage racket-centered hits and sustained juggling, as well as domain randomization over ball position and coefficient of restitution to enhance robustness and transferability. The learned policy outputs mid-level commands executed by a low-level controller and is deployed zero-shot on real hardware, where an enhanced perception module with a lightweight communication protocol reduces delays in high-frequency state estimation and ensures real-time control. Experiments show that JuggleRL achieves an average of $311$ hits over $10$ consecutive trials in the real world, with a maximum of $462$ hits observed, far exceeding a model-based baseline that reaches at most $14$ hits with an average of $3.1$. Moreover, the policy generalizes to unseen conditions, successfully juggling a lighter $5$ g ball with an average of $145.9$ hits. This work demonstrates that reinforcement learning can empower aerial robots with robust and stable control in dynamic interaction tasks.
Problem

Research questions and friction points this paper is trying to address.

Develops reinforcement learning for quadrotor ball juggling under uncertainty
Reduces sim-to-real gap via calibrated dynamics and domain randomization
Enables robust real-world deployment with zero-shot transfer and generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Deep reinforcement learning for aerial juggling
Systematic simulation calibration and domain randomization
Real-time control with enhanced perception module
🔎 Similar Papers
No similar papers found.
Shilong Ji
Shilong Ji
Tsinghua University
roboticsRL
Y
Yinuo Chen
Tsinghua University, Beijing, 100084, China.
C
Chuqi Wang
Tsinghua University, Beijing, 100084, China.
J
Jiayu Chen
Tsinghua University, Beijing, 100084, China.
Ruize Zhang
Ruize Zhang
Tsinghua University
Reinforcement LearningMulti-Agent SystemsEmbodied AIGame Theory
F
Feng Gao
Tsinghua University, Beijing, 100084, China.
W
Wenhao Tang
Tsinghua University, Beijing, 100084, China.
S
Shu'ang Yu
Tsinghua University, Beijing, 100084, China.
S
Sirui Xiang
Tsinghua University, Beijing, 100084, China.
X
Xinlei Chen
Tsinghua University, Beijing, 100084, China.
C
Chao Yu
Tsinghua University, Beijing, 100084, China.; Beijing Zhongguancun Academy, Beijing 100094, China.
Y
Yu Wang
Tsinghua University, Beijing, 100084, China.