Receding-Horizon Control via Drifting Models

📅 2026-04-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work proposes Drifting MPC, a novel framework for offline reinforcement learning in settings where the system dynamics are unknown and trajectory simulation is infeasible. Drifting MPC uniquely integrates a drift generative model with model predictive control to learn a conditional trajectory distribution from offline data that balances data support and cost optimality. The method explicitly optimizes task-specific costs while maintaining fidelity to the empirical data distribution, and it is theoretically shown that the resulting distribution constitutes the unique solution that optimally trades off optimality against consistency with the data prior. Empirical results demonstrate that Drifting MPC efficiently generates near-optimal trajectories with low per-step computational overhead, significantly reducing trajectory generation time compared to diffusion-model baselines.
📝 Abstract
We study the problem of trajectory optimization in settings where the system dynamics are unknown and it is not possible to simulate trajectories through a surrogate model. When an offline dataset of trajectories is available, an agent could directly learn a trajectory generator by distribution matching. However, this approach only recovers the behavior distribution in the dataset, and does not in general produce a model that minimizes a desired cost criterion. In this work, we propose Drifting MPC, an offline trajectory optimization framework that combines drifting generative models with receding-horizon planning under unknown dynamics. The goal of Drifting MPC is to learn, from an offline dataset of trajectories, a conditional distribution over trajectories that is both supported by the data and biased toward optimal plans. We show that the resulting distribution learned by Drifting MPC is the unique solution of an objective that trades off optimality with closeness to the offline prior. Empirically, we show that Drifting MPC can generate near-optimal trajectories while retaining the one-step inference efficiency of drifting models and substantially reducing generation time relative to diffusion-based baselines.
Problem

Research questions and friction points this paper is trying to address.

trajectory optimization
unknown dynamics
offline dataset
receding-horizon control
generative models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Drifting MPC
receding-horizon control
offline trajectory optimization
drifting generative models
unknown dynamics
🔎 Similar Papers
No similar papers found.