Interpreting CFD Surrogates through Sparse Autoencoders

📅 2025-07-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Graph neural network (GNN)-driven CFD surrogate models suffer from limited interpretability, hindering their deployment in safety-critical applications. Method: We propose a post-hoc interpretability framework based on sparse autoencoders (SAEs), which constructs an overcomplete dictionary in the pre-trained GNN’s node embedding space to learn physically meaningful implicit primitives—e.g., vorticity structures and shear layers—marking the first application of SAEs to interpret graph-based CFD surrogates. Contribution/Results: The learned dictionary exhibits unambiguity and model-agnosticism, enabling cross-architecture interpretation and visualization-based validation. Experiments demonstrate precise identification and disentanglement of key flow physics without modifying the original surrogate model, significantly enhancing decision transparency and trustworthiness. This approach provides a novel pathway toward regulatory compliance and engineering adoption of high-fidelity CFD surrogates.

Technology Category

Application Category

📝 Abstract
Learning-based surrogate models have become a practical alternative to high-fidelity CFD solvers, but their latent representations remain opaque and hinder adoption in safety-critical or regulation-bound settings. This work introduces a posthoc interpretability framework for graph-based surrogate models used in computational fluid dynamics (CFD) by leveraging sparse autoencoders (SAEs). By obtaining an overcomplete basis in the node embedding space of a pretrained surrogate, the method extracts a dictionary of interpretable latent features. The approach enables the identification of monosemantic concepts aligned with physical phenomena such as vorticity or flow structures, offering a model-agnostic pathway to enhance explainability and trustworthiness in CFD applications.
Problem

Research questions and friction points this paper is trying to address.

Interpret latent representations of CFD surrogate models
Enhance explainability of graph-based CFD surrogates
Align learned features with physical phenomena
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leveraging sparse autoencoders for interpretability
Extracting interpretable latent features dictionary
Identifying monosemantic concepts with physical phenomena
🔎 Similar Papers
No similar papers found.