A Task-Efficient Reinforcement Learning Task-Motion Planner for Safe Human-Robot Cooperation

πŸ“… 2025-10-14
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
In human–robot collaboration (HRC), safety mechanisms often compromise task efficiency, leading to frequent human interventions, robot fallback motions, task failures, and repeated motion replanning. To address this, we propose a two-tiered collaborative framework integrating reinforcement learning (RL) and interactive motion planning: an RL-based task-level policy adaptively selects high-level actions, while a motion-level planner generates safe, dynamically feasible trajectories conditioned on real-time human pose estimation; bidirectional feedback between layers jointly optimizes safety and efficiency. Evaluated in both simulation and on a physical collaborative robot platform, our framework significantly reduces target instruction repetition (βˆ’42%) and replanning frequency (βˆ’58%) compared to conventional hard-coded approaches, while maintaining collision avoidance and high task success rates. This enables a principled trade-off between safety and efficiency in dynamic, unstructured environments.

Technology Category

Application Category

πŸ“ Abstract
In a Human-Robot Cooperation (HRC) environment, safety and efficiency are the two core properties to evaluate robot performance. However, safety mechanisms usually hinder task efficiency since human intervention will cause backup motions and goal failures of the robot. Frequent motion replanning will increase the computational load and the chance of failure. In this paper, we present a hybrid Reinforcement Learning (RL) planning framework which is comprised of an interactive motion planner and a RL task planner. The RL task planner attempts to choose statistically safe and efficient task sequences based on the feedback from the motion planner, while the motion planner keeps the task execution process collision-free by detecting human arm motions and deploying new paths when the previous path is not valid anymore. Intuitively, the RL agent will learn to avoid dangerous tasks, while the motion planner ensures that the chosen tasks are safe. The proposed framework is validated on the cobot in both simulation and the real world, we compare the planner with hard-coded task motion planning methods. The results show that our planning framework can 1) react to uncertain human motions at both joint and task levels; 2) reduce the times of repeating failed goal commands; 3) reduce the total number of replanning requests.
Problem

Research questions and friction points this paper is trying to address.

Ensuring robot safety while maintaining task efficiency in human-robot cooperation
Reducing motion replanning frequency caused by unpredictable human interventions
Developing hybrid reinforcement learning framework for safe task-motion coordination
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hybrid RL framework combines task and motion planning
RL task planner selects safe and efficient sequences
Motion planner ensures collision-free execution with human detection
πŸ”Ž Similar Papers
No similar papers found.
Gaoyuan Liu
Gaoyuan Liu
Department of Mechanical Engineering, Vrije Universiteit Brussel, Brussels, Belgium
J
Joris de Winter
Department of Mechanical Engineering, Vrije Universiteit Brussel, Brussels, Belgium
K
Kelly Merckaert
Department of Mechanical Engineering, Vrije Universiteit Brussel, Brussels, Belgium
D
Denis Steckelmacher
Artificial Intelligence (AI) Lab, Vrije Universiteit Brussel, Brussels, Belgium
A
Ann Nowe
Artificial Intelligence (AI) Lab, Vrije Universiteit Brussel, Brussels, Belgium
Bram Vanderborght
Bram Vanderborght
Vrije Universiteit Brussel and imec
human robot interaction for health and manufacturing