🤖 AI Summary
This work reveals the dimensionality dependence of Classifier-Free Guidance (CFG): in low dimensions, CFG tends to distort the target distribution and attenuate sample variance, whereas in high- or infinite-dimensional settings, it converges rigorously to the true conditional distribution—exhibiting a “blessing of dimensionality.” To address this, we propose a nonlinear CFG generalization framework that enhances guidance flexibility without additional computational overhead. We establish, for the first time, a rigorous convergence theory for CFG in high-dimensional regimes. Our analysis combines high-dimensional probability theory, Gaussian mixture model (GMM) simulations, and text-to-image generation experiments. Theoretical results guarantee asymptotic convergence under mild conditions; empirically, nonlinear CFG significantly outperforms standard CFG in CLIPScore (+2.1) and FID (−14.3), while preserving sampling efficiency.
📝 Abstract
Recent studies have raised concerns about the effectiveness of Classifier-Free Guidance (CFG), indicating that in low-dimensional settings, it can lead to overshooting the target distribution and reducing sample diversity. In this work, we demonstrate that in infinite and sufficiently high-dimensional contexts CFG effectively reproduces the target distribution, revealing a blessing-of-dimensionality result. Additionally, we explore finite-dimensional effects, precisely characterizing overshoot and variance reduction. Based on our analysis, we introduce non-linear generalizations of CFG. Through numerical simulations on Gaussian mixtures and experiments on class-conditional and text-to-image diffusion models, we validate our analysis and show that our non-linear CFG offers improved flexibility and generation quality without additional computation cost.