🤖 AI Summary
To address the low data efficiency and inaccurate gradients that undermine optimization stability in PDE-constrained optimization (PDECO), this work introduces three key innovations: (1) an optimization-driven active sampling and training paradigm that prioritizes gradient-sensitive regions to enhance data utilization; (2) a Virtual-Fourier Layer that explicitly models and corrects high-order derivative errors via spectral-domain regularization; and (3) a hybrid optimization framework integrating neural operators with numerical solvers to balance learning speed and numerical robustness. Experiments demonstrate substantial improvements: average relative error in derivative prediction is reduced by 37–58%; gradient-based optimization exhibits enhanced stability; convergence speed accelerates by 2.1–3.4× across diverse PDECO tasks; and overall optimization success rate increases by 22–41% compared to purely data-driven or purely numerical approaches.
📝 Abstract
PDE-Constrained Optimization (PDECO) problems can be accelerated significantly by employing gradient-based methods with surrogate models like neural operators compared to traditional numerical solvers. However, this approach faces two key challenges: (1) **Data inefficiency**: Lack of efficient data sampling and effective training for neural operators, particularly for optimization purpose. (2) **Instability**: High risk of optimization derailment due to inaccurate neural operator predictions and gradients. To address these challenges, we propose a novel framework: (1) **Optimization-oriented training**: we leverage data from full steps of traditional optimization algorithms and employ a specialized training method for neural operators. (2) **Enhanced derivative learning**: We introduce a *Virtual-Fourier* layer to enhance derivative learning within the neural operator, a crucial aspect for gradient-based optimization. (3) **Hybrid optimization**: We implement a hybrid approach that integrates neural operators with numerical solvers, providing robust regularization for the optimization process. Our extensive experimental results demonstrate the effectiveness of our model in accurately learning operators and their derivatives. Furthermore, our hybrid optimization approach exhibits robust convergence.