🤖 AI Summary
To address critical limitations of machine learning (ML) models in computational fluid dynamics (CFD)—including insufficient accuracy, physical inconsistency, poor generalizability, and computational inefficiency—this work organizes the first ML surrogate modeling competition dedicated to 2D airfoil aerodynamic simulation. We propose a multidimensional scientific evaluation framework assessing accuracy, physical consistency, generalization, and computational efficiency; curate a high-fidelity aerodynamic dataset based on OpenFOAM; and design a hybrid modeling paradigm integrating physics-informed constraints, graph neural networks, attention mechanisms, and uncertainty calibration, enhanced by distributionally robust training and multi-objective joint evaluation. Among 240+ participating teams, the top-performing model achieves a composite score surpassing the OpenFOAM baseline for the first time and demonstrates engineering viability on key scientific metrics, thereby validating the efficacy of physics-guided surrogate modeling.
📝 Abstract
The integration of machine learning (ML) into the physical sciences is reshaping computational paradigms, offering the potential to accelerate demanding simulations such as computational fluid dynamics (CFD). Yet, persistent challenges in accuracy, generalization, and physical consistency hinder the practical deployment of ML models in scientific domains. To address these limitations and systematically benchmark progress, we organized the ML4CFD competition, centered on surrogate modeling for aerodynamic simulations over two-dimensional airfoils. The competition attracted over 240 teams, who were provided with a curated dataset generated via OpenFOAM and evaluated through a multi-criteria framework encompassing predictive accuracy, physical fidelity, computational efficiency, and out-of-distribution generalization. This retrospective analysis reviews the competition outcomes, highlighting several approaches that outperformed baselines under our global evaluation score. Notably, the top entry exceeded the performance of the original OpenFOAM solver on aggregate metrics, illustrating the promise of ML-based surrogates to outperform traditional solvers under tailored criteria. Drawing from these results, we analyze the key design principles of top submissions, assess the robustness of our evaluation framework, and offer guidance for future scientific ML challenges.