Rethinking Losses for Diffusion Bridge Samplers

๐Ÿ“… 2025-06-12
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This paper addresses the theoretical deficiency of the widely adopted Log Variance (LV) loss in diffusion bridge samplersโ€”namely, its lack of rigorous derivation and misalignment between optimization objective and sampling performance. We propose rKL-LD, a novel loss function grounded in the reverse KL divergence and leveraging the log-derivative trick, which combines theoretical rigor with practical superiority. We first demonstrate that LV cannot be justified via the data processing inequality within the diffusion bridge framework, whereas rKL-LD emerges naturally and admits a clear variational interpretation. Experiments across multiple challenging benchmarks show that rKL-LD significantly improves sample quality (FID reduced by 12โ€“28%), enhances training stability, and exhibits greater robustness to hyperparameter choices. Our work establishes a new theoretical standard and practical paradigm for diffusion bridge modeling.

Technology Category

Application Category

๐Ÿ“ Abstract
Diffusion bridges are a promising class of deep-learning methods for sampling from unnormalized distributions. Recent works show that the Log Variance (LV) loss consistently outperforms the reverse Kullback-Leibler (rKL) loss when using the reparametrization trick to compute rKL-gradients. While the on-policy LV loss yields identical gradients to the rKL loss when combined with the log-derivative trick for diffusion samplers with non-learnable forward processes, this equivalence does not hold for diffusion bridges or when diffusion coefficients are learned. Based on this insight we argue that for diffusion bridges the LV loss does not represent an optimization objective that can be motivated like the rKL loss via the data processing inequality. Our analysis shows that employing the rKL loss with the log-derivative trick (rKL-LD) does not only avoid these conceptual problems but also consistently outperforms the LV loss. Experimental results with different types of diffusion bridges on challenging benchmarks show that samplers trained with the rKL-LD loss achieve better performance. From a practical perspective we find that rKL-LD requires significantly less hyperparameter optimization and yields more stable training behavior.
Problem

Research questions and friction points this paper is trying to address.

Comparing Log Variance and rKL losses for diffusion bridges
Analyzing gradient equivalence in diffusion bridge samplers
Evaluating rKL-LD loss performance in diffusion bridge training
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses Log Variance loss for diffusion bridges
Employs rKL loss with log-derivative trick
Reduces hyperparameter optimization needs significantly
๐Ÿ”Ž Similar Papers
S
Sebastian Sanokowski
ELLIS Unit Linz, LIT AI Lab, Johannes Kepler University Linz, Austria
L
Lukas Gruber
ELLIS Unit Linz, LIT AI Lab, Johannes Kepler University Linz, Austria
C
Christoph Bartmann
ELLIS Unit Linz, LIT AI Lab, Johannes Kepler University Linz, Austria
Sepp Hochreiter
Sepp Hochreiter
Institute for Machine Learning, Johannes Kepler University Linz
Machine LearningDeep LearningArtificial IntelligenceNeural NetworksBioinformatics
Sebastian Lehner
Sebastian Lehner
JKU Linz, Institute for Machine Learning