🤖 AI Summary
This work addresses the limited data efficiency and poor out-of-distribution (OOD) generalization of existing neural operator methods, which often neglect underlying physical principles—particularly when facing parameter variations or simulation-to-reality transfer. To overcome these limitations, we propose a multi-physics joint training framework that explicitly integrates the original partial differential equations (PDEs) with their simplified canonical forms directly into the neural operator training process. This architecture-agnostic approach is compatible with diverse neural operator designs and consistently enhances model robustness under parameter shifts and cross-domain scenarios. Extensive experiments across multiple 1D, 2D, and 3D PDE tasks demonstrate significant reductions in normalized root mean square error (nRMSE), confirming improved data efficiency and superior OOD generalization performance.
📝 Abstract
Recent advances in scientific machine learning (SciML) have enabled neural operators (NOs) to serve as powerful surrogates for modeling the dynamic evolution of physical systems governed by partial differential equations (PDEs). While existing approaches focus primarily on learning simulations from the target PDE, they often overlook more fundamental physical principles underlying these equations. Inspired by how numerical solvers are compatible with simulations of different settings of PDEs, we propose a multiphysics training framework that jointly learns from both the original PDEs and their simplified basic forms. Our framework enhances data efficiency, reduces predictive errors, and improves out-of-distribution (OOD) generalization, particularly in scenarios involving shifts of physical parameters and synthetic-to-real transfer. Our method is architecture-agnostic and demonstrates consistent improvements in normalized root mean square error (nRMSE) across a wide range of 1D/2D/3D PDE problems. Through extensive experiments, we show that explicit incorporation of fundamental physics knowledge significantly strengthens the generalization ability of neural operators. We will release models and codes at https://sites.google.com/view/sciml-fundemental-pde.