Efficient Transformer-Inspired Variants of Physics-Informed Deep Operator Networks

📅 2025-09-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the trade-off between accuracy and efficiency in DeepONets for solving partial differential equations (PDEs), this paper proposes a novel Transformer-inspired DeepONet variant. The core innovation is a non-intrusive bidirectional cross-attention mechanism that enables dynamic, symmetric fusion of query-point and input-function information between the branch and trunk networks. This design preserves the original DeepONet’s architectural simplicity and training efficiency while substantially enhancing representational capacity and naturally aligning the optimal variant with the underlying PDE physics. Evaluated on four benchmark PDEs—advection, diffusion-reaction, Burgers, and KdV equations—the method matches or exceeds the accuracy of state-of-the-art DeepONet variants while significantly accelerating training. Robustness and generalizability are rigorously validated via Wilcoxon signed-rank tests, Glass’s Delta effect sizes, and Spearman rank correlation.

Technology Category

Application Category

📝 Abstract
Operator learning has emerged as a promising tool for accelerating the solution of partial differential equations (PDEs). The Deep Operator Networks (DeepONets) represent a pioneering framework in this area: the "vanilla" DeepONet is valued for its simplicity and efficiency, while the modified DeepONet achieves higher accuracy at the cost of increased training time. In this work, we propose a series of Transformer-inspired DeepONet variants that introduce bidirectional cross-conditioning between the branch and trunk networks in DeepONet. Query-point information is injected into the branch network and input-function information into the trunk network, enabling dynamic dependencies while preserving the simplicity and efficiency of the "vanilla" DeepONet in a non-intrusive manner. Experiments on four PDE benchmarks -- advection, diffusion-reaction, Burgers', and Korteweg-de Vries equations -- show that for each case, there exists a variant that matches or surpasses the accuracy of the modified DeepONet while offering improved training efficiency. Moreover, the best-performing variant for each equation aligns naturally with the equation's underlying characteristics, suggesting that the effectiveness of cross-conditioning depends on the characteristics of the equation and its underlying physics. To ensure robustness, we validate the effectiveness of our variants through a range of rigorous statistical analyses, among them the Wilcoxon Two One-Sided Test, Glass's Delta, and Spearman's rank correlation.
Problem

Research questions and friction points this paper is trying to address.

Accelerating PDE solutions with efficient operator learning
Improving DeepONet accuracy without sacrificing training efficiency
Introducing cross-conditioning for dynamic dependencies in networks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transformer-inspired bidirectional cross-conditioning networks
Dynamic dependencies with preserved simplicity and efficiency
Equation-specific variant selection based on physics characteristics
🔎 Similar Papers
No similar papers found.
Z
Zhi-Feng Wei
Advanced Computing, Mathematics, and Data Division, Pacific Northwest National Laboratory, Richland, WA 99354, USA
W
Wenqian Chen
Advanced Computing, Mathematics, and Data Division, Pacific Northwest National Laboratory, Richland, WA 99354, USA
Panos Stinis
Panos Stinis
Pacific Northwest National Laboratory
Scientific computing