π€ AI Summary
This work addresses the challenges of gradient imbalance and physical inconsistency in physics-informed neural networks (PINNs) when reconstructing physical fields from sparse, irregular observations. To overcome these limitations, the authors propose a novel physics-guided Transformer architecture that explicitly integrates physical priors into the self-attention mechanism. Specifically, additive biases derived from the heat kernel encode diffusion dynamics and temporal causality, while FiLM-modulated sinusoidal implicit neural networks adaptively regulate spectral responses. Experiments demonstrate that the model significantly enhances stability and generalization under data scarcity: it achieves a relative LΒ² error of 5.9Γ10β»Β³ in one-dimensional heat equation reconstruction with only 100 observation points, and simultaneously attains low PDE residual (8.3Γ10β»β΄) and low relative error (0.034) in two-dimensional flow past a cylinder, outperforming single-objective optimization approaches.
π Abstract
Reconstructing continuous physical fields from sparse, irregular observations is a central challenge in scientific machine learning, particularly for systems governed by partial differential equations (PDEs). Existing physics-informed methods typically enforce governing equations as soft penalty terms during optimization, often leading to gradient imbalance, instability, and degraded physical consistency under limited data. We introduce the Physics-Guided Transformer (PGT), a neural architecture that embeds physical structure directly into the self-attention mechanism. Specifically, PGT incorporates a heat-kernel-derived additive bias into attention logits, encoding diffusion dynamics and temporal causality within the representation. Query coordinates attend to these physics-conditioned context tokens, and the resulting features are decoded using a FiLM-modulated sinusoidal implicit network that adaptively controls spectral response. We evaluate PGT on the one-dimensional heat equation and two-dimensional incompressible Navier-Stokes systems. In sparse 1D reconstruction with 100 observations, PGT achieves a relative L2 error of 5.9e-3, significantly outperforming both PINNs and sinusoidal representations. In the 2D cylinder wake problem, PGT uniquely achieves both low PDE residual (8.3e-4) and competitive relative error (0.034), outperforming methods that optimize only one objective. These results demonstrate that embedding physics within attention improves stability, generalization, and physical fidelity under data-scarce conditions.