🤖 AI Summary
Bayesian optimal experimental design (OED) suffers from reliance on explicit likelihoods, computationally expensive expected information gain (EIG) estimation, and difficulty in achieving high-fidelity posterior approximation. Method: This work introduces normalizing flows—specifically RealNVP and FFJORD—into the variational OED framework for the first time, enabling differentiable, flexible, and high-fidelity posterior modeling. We propose an end-to-end trainable utility maximization objective that avoids gradient bias induced by nested Monte Carlo estimation and non-differentiable utilities. Experiment policies and posterior approximations are jointly optimized via reparameterization and ELBO maximization. Results: On nonlinear dynamical systems and medical imaging simulation tasks, our approach improves experimental efficiency by 37% and reduces posterior uncertainty calibration error by 52% compared to conventional methods.