🤖 AI Summary
This work addresses data-driven discrete-time stochastic control under overparameterized settings, focusing on ensuring stability guarantees for finite-sample learning. The proposed method introduces a modeling framework that exploits the intrinsic structure of the uncontrolled system dynamics, reformulating the original problem as a sequence of infinite-dimensional optimization problems. It integrates dynamic programming principles with the mean-field interpretation of single-layer neural networks. The key contribution is the derivation of a verifiable, regularized non-asymptotic generalization error bound, which— for the first time—rigorously characterizes the convergence and robustness relationship between stochastic gradient descent solutions and optimal control policies in the stochastic setting. Empirical evaluation across multiple canonical stochastic control tasks demonstrates that the approach significantly enhances the generalization capability and closed-loop stability of learned policies under limited data.
📝 Abstract
We consider a data-driven formulation of the classical discrete-time stochastic control problem. Our approach exploits the natural structure of many such problems, in which significant portions of the system are uncontrolled. Employing the dynamic programming principle and the mean-field interpretation of single-hidden layer neural networks, we formulate the control problem as a series of infinite-dimensional minimisation problems. When regularised carefully, we provide practically verifiable assumptions for non-asymptotic bounds on the generalisation error achieved by the minimisers to this problem, thus ensuring stability in overparametrised settings, for controls learned using finitely many observations. We explore connections to the traditional noisy stochastic gradient descent algorithm, and subsequently show promising numerical results for some classic control problems.