đ€ AI Summary
This work addresses challenging linear and nonlinear inverse problems in vision and scientific imaging through a novel Bayesian inference framework based on conditional diffusion posterior modeling. To overcome inherent bias in existing training-free diffusion posterior methodsâparticularly during latent-space inferenceâwe introduce, for the first time, the Relative Trajectory Balancing (RTB) objective from off-policy reinforcement learning into diffusion model training. This enables efficient adaptation from a pre-trained unconditional prior to a task-driven conditional posterior without fine-tuning or additional supervision. Guided by RTB-based off-policy retrospective exploration, our method significantly improves posterior fidelity and calibration. Extensive experiments demonstrate consistent superiority over state-of-the-art training-free approaches across diverse inverse problemsâincluding deblurring, super-resolution, and tomographic reconstructionâestablishing a new paradigm for Bayesian inverse problem solving.
đ Abstract
This paper presents a practical application of Relative Trajectory Balance (RTB), a recently introduced off-policy reinforcement learning (RL) objective that can asymptotically solve Bayesian inverse problems optimally. We extend the original work by using RTB to train conditional diffusion model posteriors from pretrained unconditional priors for challenging linear and non-linear inverse problems in vision, and science. We use the objective alongside techniques such as off-policy backtracking exploration to improve training. Importantly, our results show that existing training-free diffusion posterior methods struggle to perform effective posterior inference in latent space due to inherent biases.