Generalized Animal Imitator: Agile Locomotion with Versatile Motion Prior

📅 2023-10-02
🏛️ Conference on Robot Learning
📈 Citations: 15
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenges of multi-behavior coordinated learning, smooth mode transitions, and cross-task generalization in quadrupedal robots, this paper proposes the Versatile Instructable Motion prior (VIM) framework. Methodologically, VIM introduces a novel dual-objective reward design—comprising functional and stylistic rewards—and unifies motion imitation, sparse instruction guidance, kinematic constraints, and sim-to-real transfer within a single end-to-end policy. This enables concurrent learning and execution of over ten high-dynamic, low-level locomotion skills—including running, turning, jumping, and backflips. Our key contributions are: (i) the first demonstration of real-time learning and seamless switching among diverse agile locomotion behaviors using a single controller; (ii) substantial improvements in task coverage and motion naturalness; and (iii) simultaneous validation in both simulation and on physical hardware, with demonstrated cross-domain policy transfer capability.
📝 Abstract
The agility of animals, particularly in complex activities such as running, turning, jumping, and backflipping, stands as an exemplar for robotic system design. Transferring this suite of behaviors to legged robotic systems introduces essential inquiries: How can a robot learn multiple locomotion behaviors simultaneously? How can the robot execute these tasks with a smooth transition? How to integrate these skills for wide-range applications? This paper introduces the Versatile Instructable Motion prior (VIM) - a Reinforcement Learning framework designed to incorporate a range of agile locomotion tasks suitable for advanced robotic applications. Our framework enables legged robots to learn diverse agile low-level skills by imitating animal motions and manually designed motions. Our Functionality reward guides the robot's ability to adopt varied skills, and our Stylization reward ensures that robot motions align with reference motions. Our evaluations of the VIM framework span both simulation and the real world. Our framework allows a robot to concurrently learn diverse agile locomotion skills using a single learning-based controller in the real world. Videos can be found on our website: https://rchalyang.github.io/VIM/
Problem

Research questions and friction points this paper is trying to address.

How robots learn multiple locomotion behaviors simultaneously
How robots achieve smooth transitions between diverse agile motions
How to integrate versatile skills for wide-range robotic applications
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reinforcement Learning framework for agile locomotion
Functionality reward for diverse skill adoption
Stylization reward aligning with reference motions
🔎 Similar Papers
R
Ruihan Yang
UC San Diego
Zhuoqun Chen
Zhuoqun Chen
Duke University
RoboticsReinforcement Learning
J
Jianhan Ma
UC San Diego
Chongyi Zheng
Chongyi Zheng
Princeton University
Reinforcement LearningMachine Learning
Y
Yiyu Chen
University of Southern California
Q
Quan Nguyen
University of Southern California
X
Xiaolong Wang
UC San Diego