DreamControl-v2: Simpler and Scalable Autonomous Humanoid Skills via Trainable Guided Diffusion Priors

📅 2026-03-31
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of insufficient robustness in motion planning and manipulation skills for humanoid robots performing long-horizon, highly interactive tasks. The authors propose an end-to-end trainable guided diffusion prior that integrates multimodal human and robot motion data within a unified embodied representation space to directly generate high-quality reference trajectories. This approach enables automated synthesis of diverse motor skills without manual intervention and effectively supports downstream reinforcement learning policies. Experimental results on both simulation and the real-world Unitree G1 humanoid robot demonstrate that the proposed method significantly enhances policy robustness and broadens the repertoire of executable skills.
📝 Abstract
Developing robust autonomous loco-manipulation skills for humanoids remains an open problem in robotics. While RL has been applied successfully to legged locomotion, applying it to complex, interaction-rich manipulation tasks is harder given long-horizon planning challenges for manipulation. A recent approach along these lines is DreamControl, which addresses these issues by leveraging off-the-shelf human motion diffusion models as a generative prior to guide RL policies during training. In this paper, we investigate the impact of DreamControl's motion prior and propose an improved framework that trains a guided diffusion model directly in the humanoid robot's motion space, aggregating diverse human and robot datasets into a unified embodiment space. We demonstrate that our approach captures a wider range of skills due to the larger training data mixture and establishes a more automated pipeline by removing the need for manual filtering interventions. Furthermore, we show that scaling the generation of reference trajectories is important for achieving robust downstream RL policies. We validate our approach through extensive experiments in simulation and on a real Unitree-G1.
Problem

Research questions and friction points this paper is trying to address.

humanoid robotics
loco-manipulation
reinforcement learning
long-horizon planning
autonomous skills
Innovation

Methods, ideas, or system contributions that make the work stand out.

guided diffusion
humanoid robotics
loco-manipulation
embodiment space
scalable trajectory generation
🔎 Similar Papers
No similar papers found.
S
Sudarshan Harithas
General Robotics, USA; Brown University, USA
S
Sangkyung Kwak
General Robotics, USA
P
Pushkal Katara
General Robotics, USA
S
Srujan Deolasee
General Robotics, USA
Dvij Kalaria
Dvij Kalaria
PhD in AI Robotics, UC Berkeley
RoboticsAritificial IntelligenceGame TheoryMachine Learning
Srinath Sridhar
Srinath Sridhar
John E. Savage Assistant Professor of Computer Science, Brown University
Computer VisionArtificial IntelligenceRoboticsHuman-Computer Interaction
S
Sai Vemprala
General Robotics, USA
Ashish Kapoor
Ashish Kapoor
CEO and Chief Scientist, General Robotics
Machine LearningRoboticsQuantum ComputingComputer Vision
J
Jonathan Chung-Kuan Huang
General Robotics, USA