🤖 AI Summary
To address challenges in multi-arm robotic collaboration—including collision avoidance, latency in real-time trajectory generation, and poor adaptability to dynamic environments—this paper proposes an RL-DMP hierarchical planning framework. At the high level, a Proximal Policy Optimization (PPO) reinforcement learning policy network enables task-level decision-making and zero-shot transfer across tasks; at the low level, Dynamic Movement Primitives (DMPs) ensure motion smoothness and millisecond-scale trajectory execution. The framework introduces, for the first time, a tightly coupled coordination mechanism between the policy and execution layers. Evaluated in PyBullet simulations with multiple UR5e manipulators performing complex collaborative tasks, it achieves an average trajectory generation latency of <50 ms. It significantly outperforms pure RL or pure DMP baselines in collision-free success rate, while demonstrating strong task generalization and real-time capability.
📝 Abstract
Robotic tasks often require multiple manipulators to enhance task efficiency and speed, but this increases complexity in terms of collaboration, collision avoidance, and the expanded state-action space. To address these challenges, we propose a multi-level approach combining Reinforcement Learning (RL) and Dynamic Movement Primitives (DMP) to generate adaptive, real-time trajectories for new tasks in dynamic environments using a demonstration library. This method ensures collision-free trajectory generation and efficient collaborative motion planning. We validate the approach through experiments in the PyBullet simulation environment with UR5e robotic manipulators.