Solving Bayesian inverse problems with diffusion priors and off-policy RL

📅 2025-03-12
📈 Citations: 0
✹ Influential: 0
📄 PDF
đŸ€– AI Summary
This work addresses challenging linear and nonlinear inverse problems in vision and scientific imaging through a novel Bayesian inference framework based on conditional diffusion posterior modeling. To overcome inherent bias in existing training-free diffusion posterior methods—particularly during latent-space inference—we introduce, for the first time, the Relative Trajectory Balancing (RTB) objective from off-policy reinforcement learning into diffusion model training. This enables efficient adaptation from a pre-trained unconditional prior to a task-driven conditional posterior without fine-tuning or additional supervision. Guided by RTB-based off-policy retrospective exploration, our method significantly improves posterior fidelity and calibration. Extensive experiments demonstrate consistent superiority over state-of-the-art training-free approaches across diverse inverse problems—including deblurring, super-resolution, and tomographic reconstruction—establishing a new paradigm for Bayesian inverse problem solving.

Technology Category

Application Category

📝 Abstract
This paper presents a practical application of Relative Trajectory Balance (RTB), a recently introduced off-policy reinforcement learning (RL) objective that can asymptotically solve Bayesian inverse problems optimally. We extend the original work by using RTB to train conditional diffusion model posteriors from pretrained unconditional priors for challenging linear and non-linear inverse problems in vision, and science. We use the objective alongside techniques such as off-policy backtracking exploration to improve training. Importantly, our results show that existing training-free diffusion posterior methods struggle to perform effective posterior inference in latent space due to inherent biases.
Problem

Research questions and friction points this paper is trying to address.

Solving Bayesian inverse problems using diffusion priors and off-policy RL.
Training conditional diffusion models for challenging inverse problems in vision and science.
Addressing biases in existing diffusion posterior methods for latent space inference.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Relative Trajectory Balance for RL
Trains conditional diffusion model posteriors
Employs off-policy backtracking exploration
🔎 Similar Papers
2024-05-22Neural Information Processing SystemsCitations: 33
Luca Scimeca
Luca Scimeca
Postdoctoral Research Fellow at Mila AI Institute
Deep LearningComputer VisionProbabilistic InferenceScientific DiscoveryRobotics
Siddarth Venkatraman
Siddarth Venkatraman
Mila, University of Montreal
Artificial IntelligenceRobotics
Moksh Jain
Moksh Jain
Mila, Université de Montréal
probabilistic machine learning
M
Minsu Kim
Mila, Université de Montréal
Marcin Sendera
Marcin Sendera
PhD Student, Jagiellonian University, Research Intern at Mila - Quebec AI Institute,
deep learningmeta-learningfew-shot learninggenerative modelsnormalizing flows
M
Mohsin Hasan
Mila, Université de Montréal
Luke Rowe
Luke Rowe
Ph.D. Student, Mila / University of Montreal
Autonomous DrivingMachine LearningComputer Vision
Sarthak Mittal
Sarthak Mittal
Mila, Université de Montréal
Generative ModelsProbabilistic ModelsBayesian InferenceMachine Learning
Pablo Lemos
Pablo Lemos
Mila - Université de Montreal
CosmologyMachine LearningBayesian InferenceGenerative models
Emmanuel Bengio
Emmanuel Bengio
McGill University; Recursion/Valence Labs
Machine LearningDeep LearningReinforcement learning
Alexandre Adam
Alexandre Adam
University of Montreal
AstrophysicsCosmologyDeep Learning
Jarrid Rector-Brooks
Jarrid Rector-Brooks
Université de Montréal, Mila, Caltech
Generative modelingMachine learningComputer science
Yashar Hezaveh
Yashar Hezaveh
Stanford University
Laurence Perreault-Levasseur
Laurence Perreault-Levasseur
Associate Professor, Université de Montréal
CosmologyData ScienceMachine Learning
Yoshua Bengio
Yoshua Bengio
Professor of computer science, University of Montreal, Mila, IVADO, CIFAR
Machine learningdeep learningartificial intelligence
Glen Berseth
Glen Berseth
Assitant Professor - Université de Montréal
Reinforcement LearningRoboticsDeep LearningMachine Learning
Nikolay Malkin
Nikolay Malkin
University of Edinburgh