Self-adaptive weights based on balanced residual decay rate for physics-informed neural networks and deep operator networks

๐Ÿ“… 2024-06-28
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 8
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the slow convergence and low accuracy of Physics-Informed Neural Networks (PINNs) and Deep Operator Networks (DeepONets) in solving partial differential equations (PDEs)โ€”caused by highly heterogeneous residual decay rates across training pointsโ€”this work identifies imbalanced residual decay as the primary cause of training failure. We propose a point-wise residual decay rate balancing mechanism with adaptive weighting. Our method introduces a bounded, hyperparameter-free, and low-uncertainty residual equalization criterion, integrating dynamic residual monitoring, gradient-aware weight updates, and a unified PINN/DeepONet co-optimization framework. Evaluated on diverse PDE benchmarks, our approach achieves significantly higher accuracy and faster convergence compared to state-of-the-art adaptive methods, reduces computational cost by over 30%, and cuts training uncertainty by 50%.

Technology Category

Application Category

๐Ÿ“ Abstract
Physics-informed deep learning has emerged as a promising alternative for solving partial differential equations. However, for complex problems, training these networks can still be challenging, often resulting in unsatisfactory accuracy and efficiency. In this work, we demonstrate that the failure of plain physics-informed neural networks arises from the significant discrepancy in the convergence speed of residuals at different training points, where the slowest convergence speed dominates the overall solution convergence. Based on these observations, we propose a point-wise adaptive weighting method that balances the residual decay rate across different training points. The performance of our proposed adaptive weighting method is compared with current state-of-the-art adaptive weighting methods on benchmark problems for both physics-informed neural networks and physics-informed deep operator networks. Through extensive numerical results we demonstrate that our proposed approach of balanced residual decay rates offers several advantages, including bounded weights, high prediction accuracy, fast convergence speed, low training uncertainty, low computational cost and ease of hyperparameter tuning.
Problem

Research questions and friction points this paper is trying to address.

Balancing residual decay rates across training points
Improving accuracy and efficiency in physics-informed networks
Addressing slow convergence in complex differential equations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Pointwise adaptive weighting balances residual decay
Method ensures bounded weights and fast convergence
Low computational cost with easy hyperparameter tuning
๐Ÿ”Ž Similar Papers
No similar papers found.
W
Wenqian Chen
Advanced Computing, Mathematics and Data Division, Pacific Northwest National Laboratory, Richland, WA 99354, USA
A
Amanda A. Howard
Advanced Computing, Mathematics and Data Division, Pacific Northwest National Laboratory, Richland, WA 99354, USA
P
P. Stinis
Advanced Computing, Mathematics and Data Division, Pacific Northwest National Laboratory, Richland, WA 99354, USA