D-Flow SGLD: Source-Space Posterior Sampling for Scientific Inverse Problems with Flow Matching

📅 2026-02-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses posterior reconstruction of high-dimensional physical states in scientific inverse problems under sparse, noisy observations, where fidelity to physical laws and uncertainty quantification must be jointly preserved. Existing flow matching models lack training-free conditional generation and efficient posterior sampling mechanisms. We propose D-Flow SGLD, a differentiable posterior inference framework that integrates preconditioned stochastic gradient Langevin dynamics (SGLD) in the source space, enabling flexible adaptation to arbitrary observation operators without retraining the prior or altering the flow dynamics. We systematically categorize two inference strategies for injecting measurement information in flow matching and establish a scalable, training-free posterior sampling framework. Experiments on 2D toy distributions, Kuramoto–Sivashinsky chaotic trajectories, and wall turbulence reconstruction demonstrate superior balance among measurement consistency, posterior diversity, and physical/statistical fidelity.

Technology Category

Application Category

📝 Abstract
Data assimilation and scientific inverse problems require reconstructing high-dimensional physical states from sparse and noisy observations, ideally with uncertainty-aware posterior samples that remain faithful to learned priors and governing physics. While training-free conditional generation is well developed for diffusion models, corresponding conditioning and posterior sampling strategies for Flow Matching (FM) priors remain comparatively under-explored, especially on scientific benchmarks where fidelity must be assessed beyond measurement misfit. In this work, we study training-free conditional generation for scientific inverse problems under FM priors and organize existing inference-time strategies by where measurement information is injected: (i) guided transport dynamics that perturb sampling trajectories using likelihood information, and (ii) source-distribution inference that performs posterior inference over the source variable while keeping the learned transport fixed. Building on the latter, we propose D-Flow SGLD, a source-space posterior sampling method that augments differentiable source inference with preconditioned stochastic gradient Langevin dynamics, enabling scalable exploration of the source posterior induced by new measurement operators without retraining the prior or modifying the learned FM dynamics. We benchmark representative methods from both families on a hierarchy of problems: 2D toy posteriors, chaotic Kuramoto-Sivashinsky trajectories, and wall-bounded turbulence reconstruction. Across these settings, we quantify trade-offs among measurement assimilation, posterior diversity, and physics/statistics fidelity, and establish D-Flow SGLD as a practical FM-compatible posterior sampler for scientific inverse problems.
Problem

Research questions and friction points this paper is trying to address.

scientific inverse problems
data assimilation
posterior sampling
flow matching
uncertainty quantification
Innovation

Methods, ideas, or system contributions that make the work stand out.

Flow Matching
Posterior Sampling
Source-Space Inference
Stochastic Gradient Langevin Dynamics
Scientific Inverse Problems
M
Meet Hemant Parikh
Sibley School of Mechanical and Aerospace Engineering, Cornell University, Ithaca, NY, USA
Y
Yaqin Chen
Department of Aerospace and Mechanical Engineering, University of Notre Dame, Notre Dame, IN, USA
Jian-Xun Wang
Jian-Xun Wang
Associate Professor, Cornell University
Scientific Machine LearningAI for ScienceCFDData AssimilationComputational Physics