Re4MPC: Reactive Nonlinear MPC for Multi-model Motion Planning via Deep Reinforcement Learning

📅 2025-06-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high computational cost of conventional nonlinear model predictive control (NMPC) in real-time motion planning for high-degree-of-freedom robots (e.g., mobile manipulators), this paper proposes a deep reinforcement learning (DRL)-driven reactive NMPC framework. The method reformulates NMPC as a learnable reactive decision-making problem, establishing a mathematically rigorous co-optimization framework that jointly adapts the model structure, cost function, and constraint set in real time. Leveraging multi-model dynamics representation and online trajectory optimization, the approach achieves a 32% improvement in end-effector goal attainment success rate over a full-body NMPC baseline in simulation, while reducing average planning latency by 5.8×. This demonstrates a significant advance in balancing real-time performance with motion accuracy for complex robotic systems.

Technology Category

Application Category

📝 Abstract
Traditional motion planning methods for robots with many degrees-of-freedom, such as mobile manipulators, are often computationally prohibitive for real-world settings. In this paper, we propose a novel multi-model motion planning pipeline, termed Re4MPC, which computes trajectories using Nonlinear Model Predictive Control (NMPC). Re4MPC generates trajectories in a computationally efficient manner by reactively selecting the model, cost, and constraints of the NMPC problem depending on the complexity of the task and robot state. The policy for this reactive decision-making is learned via a Deep Reinforcement Learning (DRL) framework. We introduce a mathematical formulation to integrate NMPC into this DRL framework. To validate our methodology and design choices, we evaluate DRL training and test outcomes in a physics-based simulation involving a mobile manipulator. Experimental results demonstrate that Re4MPC is more computationally efficient and achieves higher success rates in reaching end-effector goals than the NMPC baseline, which computes whole-body trajectories without our learning mechanism.
Problem

Research questions and friction points this paper is trying to address.

Efficient motion planning for high-DOF robots
Reactive model selection for NMPC via DRL
Improving computational efficiency and success rates
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reactive NMPC for efficient multi-model planning
DRL-based policy for adaptive model selection
Integrated NMPC-DRL framework for robotics
🔎 Similar Papers
2024-10-07arXiv.orgCitations: 9
N
Nesğet Ünver Akmandor
Motional AD Inc., Boston, MA, USA
S
Sarvesh Prajapati
Institute for Experiential Robotics, Northeastern University, Boston, Massachusetts, USA
Mark Zolotas
Mark Zolotas
Research Scientist, Toyota Research Institute
Shared ControlHuman-Robot InteractionExtended RealityRepresentation Learning
T
Tacşkin Padir
Professor of Electrical and Computer Engineering at Northeastern University and Amazon Scholar