đ¤ AI Summary
Robot manipulation policy learning in dynamic environments heavily relies on large-scale human demonstrations, incurring high data collection costs. Method: This paper proposes a novel method that generates high-quality training data from minimal human demonstrations. Its core innovation is the first integration of Dynamic Movement Primitives (DMPs) with a subtask segmentation mechanism, enabling real-time, adaptive generalization to object pose, robot state, and scene geometry variations. The approach unifies behavior cloning, trajectory generation, and hierarchical task decomposition to support data-efficient cross-scene generalization. Contribution/Results: Evaluated on long-horizon, high-contact dynamic tasksâincluding cube stacking and cup insertion into drawersâthe method significantly reduces dependence on human demonstrations while improving policy robustness and generalization capability across diverse scenarios.
đ Abstract
Learning robust manipulation policies typically requires large and diverse datasets, the collection of which is time-consuming, labor-intensive, and often impractical for dynamic environments. In this work, we introduce DynaMimicGen (D-MG), a scalable dataset generation framework that enables policy training from minimal human supervision while uniquely supporting dynamic task settings. Given only a few human demonstrations, D-MG first segments the demonstrations into meaningful sub-tasks, then leverages Dynamic Movement Primitives (DMPs) to adapt and generalize the demonstrated behaviors to novel and dynamically changing environments. Improving prior methods that rely on static assumptions or simplistic trajectory interpolation, D-MG produces smooth, realistic, and task-consistent Cartesian trajectories that adapt in real time to changes in object poses, robot states, or scene geometry during task execution. Our method supports different scenarios - including scene layouts, object instances, and robot configurations - making it suitable for both static and highly dynamic manipulation tasks. We show that robot agents trained via imitation learning on D-MG-generated data achieve strong performance across long-horizon and contact-rich benchmarks, including tasks like cube stacking and placing mugs in drawers, even under unpredictable environment changes. By eliminating the need for extensive human demonstrations and enabling generalization in dynamic settings, D-MG offers a powerful and efficient alternative to manual data collection, paving the way toward scalable, autonomous robot learning.