ZEST: Zero-shot Embodied Skill Transfer for Athletic Robot Control

📅 2026-01-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of deploying agile multi-contact locomotion on humanoid robots, which typically requires extensive skill-specific customization and parameter tuning. The authors propose ZEST, a framework that leverages reinforcement learning to train whole-body control policies end-to-end from diverse motion data—including motion capture, monocular video, and animation—without relying on contact labels, reference windows, state estimators, or intricate reward shaping. By integrating adaptive sampling with a model-based auxiliary torque curriculum, ZEST achieves zero-shot generalization across behaviors and platforms under moderate domain randomization in simulation. Experiments demonstrate successful reproduction of complex multi-contact skills such as crawling and breakdancing on the Atlas robot, direct transfer of dance and box-jumping motions from video to both Atlas and Unitree G1, and even consecutive backflips on the quadrupedal Spot robot, highlighting its strong cross-modal and cross-morphology generalization capabilities.

Technology Category

Application Category

📝 Abstract
Achieving robust, human-like whole-body control on humanoid robots for agile, contact-rich behaviors remains a central challenge, demanding heavy per-skill engineering and a brittle process of tuning controllers. We introduce ZEST (Zero-shot Embodied Skill Transfer), a streamlined motion-imitation framework that trains policies via reinforcement learning from diverse sources -- high-fidelity motion capture, noisy monocular video, and non-physics-constrained animation -- and deploys them to hardware zero-shot. ZEST generalizes across behaviors and platforms while avoiding contact labels, reference or observation windows, state estimators, and extensive reward shaping. Its training pipeline combines adaptive sampling, which focuses training on difficult motion segments, and an automatic curriculum using a model-based assistive wrench, together enabling dynamic, long-horizon maneuvers. We further provide a procedure for selecting joint-level gains from approximate analytical armature values for closed-chain actuators, along with a refined model of actuators. Trained entirely in simulation with moderate domain randomization, ZEST demonstrates remarkable generality. On Boston Dynamics'Atlas humanoid, ZEST learns dynamic, multi-contact skills (e.g., army crawl, breakdancing) from motion capture. It transfers expressive dance and scene-interaction skills, such as box-climbing, directly from videos to Atlas and the Unitree G1. Furthermore, it extends across morphologies to the Spot quadruped, enabling acrobatics, such as a continuous backflip, through animation. Together, these results demonstrate robust zero-shot deployment across heterogeneous data sources and embodiments, establishing ZEST as a scalable interface between biological movements and their robotic counterparts.
Problem

Research questions and friction points this paper is trying to address.

humanoid robot control
zero-shot transfer
whole-body control
multi-contact skills
embodied skill transfer
Innovation

Methods, ideas, or system contributions that make the work stand out.

Zero-shot transfer
Motion imitation
Reinforcement learning
Embodied intelligence
Domain randomization
🔎 Similar Papers
No similar papers found.
J
Jean Pierre Sleiman
RAI Institute, USA
H
He Li
RAI Institute, USA
Alphonsus Adu-Bredu
Alphonsus Adu-Bredu
PhD Student, University of Michigan
RoboticsMotion PlanningControlsMachine learning
Robin Deits
Robin Deits
Graduate Student, Computer Science and Artificial Intelligence Laboratory, MIT
RoboticsControlsLanguageArtificial IntelligenceManipulation
A
Arun Kumar
Boston Dynamics, USA
K
Kevin Bergamin
Boston Dynamics, USA
Mohak Bhardwaj
Mohak Bhardwaj
University of Washington
RoboticsMachine LearningArtificial IntelligencePlanning and Control
S
Scott Biddlestone
RAI Institute, USA
N
Nicola Burger
RAI Institute, USA
M
Matthew A. Estrada
RAI Institute, USA
F
Francesco Iacobelli
RAI Institute, USA
Twan Koolen
Twan Koolen
Robotics Engineer, Boston Dynamics
RoboticsControl TheoryHumanoidsBalance
Alexander Lambert
Alexander Lambert
Research Scientist, Boston Dynamics
RoboticsOptimal ControlMachine LearningReinforcement LearningArtificial Intelligence
E
Erica Lin
RAI Institute, USA
M. Eva Mungai
M. Eva Mungai
University of Michigan, Ann Arbor
bipedal locomotionroboticsnonlinear controloptimizationsafety and verification
Z
Zach Nobles
RAI Institute, USA
S
Shane Rozen-Levy
Boston Dynamics, USA
Y
Yuyao Shi
RAI Institute, USA
Jiashun Wang
Jiashun Wang
Carnegie Mellon University
Computer VisionComputer GraphicsRobotics
J
Jakob Welner
Boston Dynamics, USA
F
Fangzhou Yu
RAI Institute, USA
Mike Zhang
Mike Zhang
Aalborg University (Copenhagen)
Artificial IntelligenceNatural Language ProcessingInformation ExtractionNLP Applications
A
Alfred Rizzi
RAI Institute, USA
Jessica Hodgins
Jessica Hodgins
Professor of Computer Science and Robotics, Carnegie Mellon University
Computer GraphicsAnimationHumanoid Robotics
Sylvain Bertrand
Sylvain Bertrand
ONERA
control systemsMPCUAVroboticsmulti-agent systems