Convergence of Stochastic Gradient Methods for Wide Two-Layer Physics-Informed Neural Networks

πŸ“… 2025-08-29
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the convergence of wide two-layer physics-informed neural networks (PINNs) under stochastic gradient descent (SGD) and stochastic gradient flow (SGF), a longstanding challenge unaddressed by prior deterministic analyses. Method: Leveraging overparameterization, we establish high-probability linear convergence by jointly ensuring Gram matrix positive definiteness throughout training and rigorously characterizing the stochastic dynamics induced by SGD/SGFβ€”thereby controlling random perturbations across diverse common activation functions. Contribution/Results: Our analysis extends PINN convergence theory to stochastic optimization for the first time, revealing an intrinsic linear convergence mechanism of SGD/SGF in solving partial differential equations. It provides the first rigorous theoretical foundation for training PINNs via stochastic algorithms, bridging a critical gap between practical stochastic training and theoretical guarantees in scientific machine learning.

Technology Category

Application Category

πŸ“ Abstract
Physics informed neural networks (PINNs) represent a very popular class of neural solvers for partial differential equations. In practice, one often employs stochastic gradient descent type algorithms to train the neural network. Therefore, the convergence guarantee of stochastic gradient descent is of fundamental importance. In this work, we establish the linear convergence of stochastic gradient descent / flow in training over-parameterized two layer PINNs for a general class of activation functions in the sense of high probability. These results extend the existing result [18] in which gradient descent was analyzed. The challenge of the analysis lies in handling the dynamic randomness introduced by stochastic optimization methods. The key of the analysis lies in ensuring the positive definiteness of suitable Gram matrices during the training. The analysis sheds insight into the dynamics of the optimization process, and provides guarantees on the neural networks trained by stochastic algorithms.
Problem

Research questions and friction points this paper is trying to address.

Analyzing convergence of stochastic gradient methods
Training over-parameterized physics-informed neural networks
Ensuring positive definiteness of Gram matrices
Innovation

Methods, ideas, or system contributions that make the work stand out.

Stochastic gradient descent for PINNs
Linear convergence in over-parameterized networks
Positive definiteness of Gram matrices
πŸ”Ž Similar Papers
No similar papers found.
Bangti Jin
Bangti Jin
Department of Mathematics, The Chinese University of Hong Kong
inverse problemsnumerical analysis
L
Longjun Wu
Department of Mathematics, The Chinese University of Hong Kong, Shatin, N.T., Hong Kong, P.R. China