🤖 AI Summary
This work proposes a nonparametric sampling method based on Wasserstein gradient descent to address the limitations of traditional Markov chain Monte Carlo (MCMC) and variational inference in high-dimensional or multimodal distributions. By directly optimizing the Kullback–Leibler (KL) functional over the space of probability measures and leveraging particle approximations together with score matching, the approach enables efficient sampling. Theoretically, this study establishes, for the first time, convergence guarantees for Wasserstein gradient descent within two important subclasses of probability measures, thereby providing a novel theoretical foundation for optimization-driven sampling. Empirical results demonstrate that the proposed method significantly outperforms standard MCMC and parametric variational Bayesian approaches in both high-dimensional and multimodal settings.
📝 Abstract
This paper studies the optimization of the KL functional on the Wasserstein space of probability measures, and develops a sampling framework based on Wasserstein gradient descent (WGD). We identify two important subclasses of the Wasserstein space for which the WGD scheme is guaranteed to converge, thereby providing new theoretical foundations for optimization-based sampling methods on measure spaces. For practical implementation, we construct a particle-based WGD algorithm in which the score function is estimated via score matching. Through a series of numerical experiments, we demonstrate that WGD can provide good approximation to a variety of complex target distributions, including those that pose substantial challenges for standard MCMC and parametric variational Bayes methods. These results suggest that WGD offers a promising and flexible alternative for scalable Bayesian inference in high-dimensional or multimodal settings.