Physics-Informed Laplace Neural Operator for Solving Partial Differential Equations

📅 2026-02-13
📈 Citations: 0
Influential: 0
📄 PDF

Technology Category

Application Category

📝 Abstract
Neural operators have emerged as fast surrogate solvers for parametric partial differential equations (PDEs). However, purely data-driven models often require extensive training data and can generalize poorly, especially in small-data regimes and under unseen (out-of-distribution) input functions that are not represented in the training data. To address these limitations, we propose the Physics-Informed Laplace Neural Operator (PILNO), which enhances the Laplace Neural Operator (LNO) by embedding governing physics into training through PDE, boundary condition, and initial condition residuals. To improve expressivity, we first introduce an Advanced LNO (ALNO) backbone that retains a pole-residue transient representation while replacing the steady-state branch with an FNO-style Fourier multiplier. To make physics-informed training both data-efficient and robust, PILNO further leverages (i) virtual inputs: an unlabeled ensemble of input functions spanning a broad spectral range that provides abundant physics-only supervision and explicitly targets out-of-distribution (OOD) regimes; and (ii) temporal-causality weighting: a time-decaying reweighting of the physics residual that prioritizes early-time dynamics and stabilizes optimization for time-dependent PDEs. Across four representative benchmarks -- Burgers'equation, Darcy flow, a reaction-diffusion system, and a forced KdV equation -- PILNO consistently improves accuracy in small-data settings (e.g., N_train<= 27), reduces run-to-run variability across random seeds, and achieves stronger OOD generalization than purely data-driven baselines.
Problem

Research questions and friction points this paper is trying to address.

neural operators
partial differential equations
data efficiency
out-of-distribution generalization
physics-informed learning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Physics-Informed Neural Operator
Laplace Neural Operator
Virtual Inputs
Temporal-Causality Weighting
Out-of-Distribution Generalization
H
Heechang Kim
Department of Mathematics, Pohang University of Science and Technology (POSTECH), Pohang, 37673, Republic of Korea
Q
Qianying Cao
Division of Applied Mathematics, Brown University, Providence, 02906, Rhode Island, United States
H
Hyomin Shin
Department of Mathematics, Pohang University of Science and Technology (POSTECH), Pohang, 37673, Republic of Korea
Seungchul Lee
Seungchul Lee
Invited Professor, School of Computing, KAIST
Mobile computingIoTSocial computing
George Em Karniadakis
George Em Karniadakis
The Charles Pitts Robinson and John Palmer Barstow Professor of Applied Mathematics and Engineering
Math+Machine LearningProbabilistic Scientific ComputingStochastic Multiscale Modeling
Minseok Choi
Minseok Choi
Kyung Hee University
Wireless caching networkFederated learningStochastic network optimizationReinforcement learning