🤖 AI Summary
This work addresses the lack of rigorous error theory for neural-network-based solvers of partial differential equations (PDEs). We systematically investigate the approximation capacity of deep fully connected networks for weak solutions of PDEs in the Sobolev space $W^{n,infty}$. Under general activation functions—including common non-polynomial ones—we derive high-order approximation error estimates in the $W^{m,p}$ norm for $m < n$ and $1 leq p leq infty$. Crucially, we identify and rigorously prove a “superconvergence” phenomenon: deep networks can achieve convergence rates exceeding the intrinsic limits of classical finite element and spectral methods. This result fills a fundamental theoretical gap in the error analysis of deep learning for scientific computing. By establishing a unified, scalable Sobolev framework, our work provides a solid theoretical foundation for both the reliability and efficiency of neural PDE solvers.
📝 Abstract
This paper establishes a comprehensive approximation result for deep fully-connected neural networks with commonly-used and general activation functions in Sobolev spaces $W^{n,infty}$, with errors measured in the $W^{m,p}$-norm for $m < n$ and $1le p le infty$. The derived rates surpass those of classical numerical approximation techniques, such as finite element and spectral methods, exhibiting a phenomenon we refer to as emph{super-convergence}. Our analysis shows that deep networks with general activations can approximate weak solutions of partial differential equations (PDEs) with superior accuracy compared to traditional numerical methods at the approximation level. Furthermore, this work closes a significant gap in the error-estimation theory for neural-network-based approaches to PDEs, offering a unified theoretical foundation for their use in scientific computing.