Achieving Human Level Competitive Robot Table Tennis

πŸ“… 2024-08-07
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 13
✨ Influential: 1
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of achieving human-level real-time performance for robots in realistic adversarial tasks, focusing on table tennisβ€”a highly dynamic, interactive domain. Methodologically, we propose the first end-to-end learning-based robotic system capable of sustaining stable amateur-human competitive proficiency. Our approach features a hierarchical reinforcement learning architecture comprising a physics-aware low-level skill controller and a high-level skill selector; introduces the first automatic curriculum learning framework grounded in real-task distribution, coupled with zero-shot sim-to-real transfer; and incorporates online opponent modeling and real-time skill switching to enable closed-loop visuomotor control. In 29 human-robot matches, the system achieved a 45% win rate overall (100% against beginners, 55% against intermediate players), demonstrating breakthrough capabilities in adapting to unknown opponent behaviors, dynamically scaling task difficulty, and seamless simulation-to-reality transfer.

Technology Category

Application Category

πŸ“ Abstract
Achieving human-level speed and performance on real world tasks is a north star for the robotics research community. This work takes a step towards that goal and presents the first learned robot agent that reaches amateur human-level performance in competitive table tennis. Table tennis is a physically demanding sport which requires human players to undergo years of training to achieve an advanced level of proficiency. In this paper, we contribute (1) a hierarchical and modular policy architecture consisting of (i) low level controllers with their detailed skill descriptors which model the agent's capabilities and help to bridge the sim-to-real gap and (ii) a high level controller that chooses the low level skills, (2) techniques for enabling zero-shot sim-to-real including an iterative approach to defining the task distribution that is grounded in the real-world and defines an automatic curriculum, and (3) real time adaptation to unseen opponents. Policy performance was assessed through 29 robot vs. human matches of which the robot won 45% (13/29). All humans were unseen players and their skill level varied from beginner to tournament level. Whilst the robot lost all matches vs. the most advanced players it won 100% matches vs. beginners and 55% matches vs. intermediate players, demonstrating solidly amateur human-level performance. Videos of the matches can be viewed at https://sites.google.com/view/competitive-robot-table-tennis
Problem

Research questions and friction points this paper is trying to address.

Achieving human-level robot performance in competitive table tennis
Bridging sim-to-real gap with hierarchical policy architecture
Enabling real-time adaptation to unseen opponents
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hierarchical modular policy architecture for robot control
Zero-shot sim-to-real with automatic task curriculum
Real-time adaptation to unseen opponents
πŸ”Ž Similar Papers
No similar papers found.
David B. D'Ambrosio
David B. D'Ambrosio
Lila Sciences
Artificial IntelligenceNeural NetworksNeuroevolutionRoboticsMultiagent Learning
Saminda Abeyruwan
Saminda Abeyruwan
University of Miami
Artificial IntelligenceAutonomous LearningMachine LearningReinforcement LearningSemantic Web
Laura Graesser
Laura Graesser
Robotics at Google DeepMind
roboticsreinforcement learningevolutionary strategiesemergent communication
Atil Iscen
Atil Iscen
Google
RoboticsReinforcement LearningEvolutionary AlgorithmsMulti-agent Learning
Heni Ben Amor
Heni Ben Amor
Associate Professor, Arizona State University
Human-Robot InteractionRoboticsMotor Skill LearningArtificial Intelligence
Alex Bewley
Alex Bewley
Google DeepMind
RoboticsMachine LearningComputer VisionVision Language Models
B
Barney J. Reed
Core contributors (Alphabetized)
K
Krista Reymann
Core contributors (Alphabetized)
Leila Takayama
Leila Takayama
Core contributors (Alphabetized)
Yuval Tassa
Yuval Tassa
Senior Research Scientist, Google DeepMind
Optimal ControlMotor ControlRoboticsPhysical Simulation
Krzysztof Choromanski
Krzysztof Choromanski
Google DeepMind Robotics & Columbia University
roboticsreinforcement learningefficient Transformersquasi Monte Carlo methods
E
Erwin Coumans
Google DeepMind
Deepali Jain
Deepali Jain
Google Deepmind
Artificial IntelligenceRoboticsReinforcement Learning
Navdeep Jaitly
Navdeep Jaitly
Apple
Machine LearningLanguage ModelingSpeech ModelingComputational BiologyRobotics
Natasha Jaques
Natasha Jaques
University of Washington, Google Research
Social reinforcement learningMachine learningdeep learningmulti-agenthuman-AI interaction
S
Satoshi Kataoka
Google DeepMind
Y
Yuheng Kuang
Google DeepMind
N
Nevena Lazic
Google DeepMind
R
Reza Mahjourian
Google DeepMind
S
Sherry Moore
Google DeepMind
K
Kenneth Oslund
Google DeepMind
A
Anish Shankar
Google DeepMind
Vikas Sindhwani
Vikas Sindhwani
Google DeepMind Robotics
AIRoboticsAI SafetyMachine LearningOptimization
Vincent Vanhoucke
Vincent Vanhoucke
Distinguished Engineer, Waymo
Robot LearningRoboticsMachine LearningComputer VisionArtificial Intelligence
Grace Vesom
Grace Vesom
Google DeepMind
Computer VisionMachine LearningShape Representation
P
Peng Xu
Google DeepMind
P
Pannag R. Sanketi
Google DeepMind