🤖 AI Summary
This work addresses the challenge of efficiently simulating conditional diffusion processes—diffusion bridges—in Euclidean space, where conventional score-based or MCMC-based methods suffer from high computational cost and poor robustness in rare-event and multimodal settings. We propose a neural variational path measure approximation framework that directly parameterizes diffusion bridges without requiring backward-process solving or MCMC sampling—a first in the literature. By jointly optimizing a variational objective over the path space, the model end-to-end learns the conditional transition dynamics. Experiments demonstrate that our method achieves sampling efficiency comparable to unconditional forward diffusion, while significantly improving stability and accuracy in rare-event and multimodal regimes.
📝 Abstract
We propose a novel method for simulating conditioned diffusion processes (diffusion bridges) in Euclidean spaces. By training a neural network to approximate bridge dynamics, our approach eliminates the need for computationally intensive Markov Chain Monte Carlo (MCMC) methods or reverse-process modeling. Compared to existing methods, it offers greater robustness across various diffusion specifications and conditioning scenarios. This applies in particular to rare events and multimodal distributions, which pose challenges for score-learning- and MCMC-based approaches. We propose a flexible variational family for approximating the diffusion bridge path measure which is partially specified by a neural network. Once trained, it enables efficient independent sampling at a cost comparable to sampling the unconditioned (forward) process.