🤖 AI Summary
Randomized optimal control often converges to local optima in complex non-convex environments and lacks the capacity to learn from historical trajectories or adapt online. To address this, we propose a memory-augmented adaptive control framework that, for the first time, embeds historical trajectory experience into potential field theory—yielding a dynamic memory potential field capable of online modeling of state-space topology and adaptive policy refinement. Our approach requires no domain-specific prior knowledge or offline training, while guaranteeing both non-convex escape capability and asymptotic convergence. Integrated with Model Predictive Path Integral (MPPI) control for real-time optimization, it significantly improves control performance on challenging non-convex tasks. Extensive experiments demonstrate its effectiveness, robustness, and computational efficiency on high-dimensional, nonlinear systems—including robotic dynamics—under realistic constraints.
📝 Abstract
Stochastic optimal control methods often struggle in complex non-convex landscapes, frequently becoming trapped in local optima due to their inability to learn from historical trajectory data. This paper introduces Memory-Augmented Potential Field Theory, a unified mathematical framework that integrates historical experience into stochastic optimal control. Our approach dynamically constructs memory-based potential fields that identify and encode key topological features of the state space, enabling controllers to automatically learn from past experiences and adapt their optimization strategy. We provide a theoretical analysis showing that memory-augmented potential fields possess non-convex escape properties, asymptotic convergence characteristics, and computational efficiency. We implement this theoretical framework in a Memory-Augmented Model Predictive Path Integral (MPPI) controller that demonstrates significantly improved performance in challenging non-convex environments. The framework represents a generalizable approach to experience-based learning within control systems (especially robotic dynamics), enhancing their ability to navigate complex state spaces without requiring specialized domain knowledge or extensive offline training.