🤖 AI Summary
This work addresses the first-order bias in score function estimation under Gaussian noise, which compromises the validity of Langevin sampling. To resolve this, we propose the Noise-Corrected Langevin Algorithm (NCLA), the first method to systematically eliminate noise-induced first-order bias within the Langevin framework using only a single noise-level score function, thereby enabling unbiased sampling. NCLA introduces a “semi-denoising” sampling paradigm that integrates score matching, Taylor-expansion-based error correction, and iterative noise addition/subtraction—achieving both theoretical rigor and interpretability. Experiments demonstrate that NCLA significantly improves sample quality and accelerates convergence on both synthetic distributions and image generation tasks, without requiring multi-scale noise schedules or reverse diffusion processes.
📝 Abstract
The Langevin algorithm is a classic method for sampling from a given pdf in a real space. In its basic version, it only requires knowledge of the gradient of the log-density, also called the score function. However, in deep learning, it is often easier to learn the so-called"noisy-data score function", i.e. the gradient of the log-density of noisy data, more precisely when Gaussian noise is added to the data. Such an estimate is biased and complicates the use of the Langevin method. Here, we propose a noise-corrected version of the Langevin algorithm, where the bias due to noisy data is removed, at least regarding first-order terms. Unlike diffusion models, our algorithm needs to know the noisy score function for one single noise level only. We further propose a simple special case which has an interesting intuitive interpretation of iteratively adding noise the data and then attempting to remove half of that noise.