đ€ AI Summary
Standard discriminator guidance employs cross-entropy loss, yet theoretical analysis reveals it increases the KL divergence between the model and data distributionsâespecially under discriminator overfittingâdegrading generation quality. Method: We propose the first theoretically guaranteed discriminator guidance framework converging to the true data distribution: we reformulate the discriminator objective to directly minimize KL divergence, ensuring that Score-Matching diffusion models are guided toward the target distribution with strict theoretical guarantees. Contribution/Results: Through rigorous convergence analysis and extensive experiments across multiple datasets, our method consistently suppresses distributional shift, improving FID by 12â28% across all benchmarks. It significantly enhances sample fidelity and training robustness, establishing a principled foundation for reliable, high-quality diffusion-based generation.
đ Abstract
Discriminator Guidance has become a popular method for efficiently refining pre-trained Score-Matching Diffusion models. However, in this paper, we demonstrate that the standard implementation of this technique does not necessarily lead to a distribution closer to the real data distribution. Specifically, we show that training the discriminator using Cross-Entropy loss, as commonly done, can in fact increase the Kullback-Leibler divergence between the model and target distributions, particularly when the discriminator overfits. To address this, we propose a theoretically sound training objective for discriminator guidance that properly minimizes the KL divergence. We analyze its properties and demonstrate empirically across multiple datasets that our proposed method consistently improves over the conventional method by producing samples of higher quality.