🤖 AI Summary
Physics-informed neural networks (PINNs) face challenges—including ill-conditioned optimization, slow convergence, and poor generalization—when solving parametric partial differential equations (PDEs). This paper proposes a data-driven neural solver that parameterizes adaptive gradient descent as a neural network, jointly modeling distributions of PDE coefficients and initial/boundary conditions under physical constraints, while dynamically conditioning the optimizer to alleviate loss function ill-conditioning. To our knowledge, this is the first work to introduce neural solvers into parametric PDE settings, enabling end-to-end training via implicit differentiation and backpropagation. Experiments demonstrate a 2–5× speedup in training with enhanced convergence stability. At inference, the solver generalizes robustly to unseen parameter combinations, significantly reducing required iterations while maintaining high accuracy.
📝 Abstract
Physics-informed deep learning often faces optimization challenges due to the complexity of solving partial differential equations (PDEs), which involve exploring large solution spaces, require numerous iterations, and can lead to unstable training. These challenges arise particularly from the ill-conditioning of the optimization problem, caused by the differential terms in the loss function. To address these issues, we propose learning a solver, i.e., solving PDEs using a physics-informed iterative algorithm trained on data. Our method learns to condition a gradient descent algorithm that automatically adapts to each PDE instance, significantly accelerating and stabilizing the optimization process and enabling faster convergence of physics-aware models. Furthermore, while traditional physics-informed methods solve for a single PDE instance, our approach addresses parametric PDEs. Specifically, our method integrates the physical loss gradient with the PDE parameters to solve over a distribution of PDE parameters, including coefficients, initial conditions, or boundary conditions. We demonstrate the effectiveness of our method through empirical experiments on multiple datasets, comparing training and test-time optimization performance.