🤖 AI Summary
This work studies the high-dimensional convergence of black-box variational inference (BBVI) under the mean-field location-scale variational family. For strongly log-concave and log-smooth target distributions, we establish, for the first time, a nearly dimension-free iteration complexity of $O(log d)$, substantially mitigating the “curse of dimensionality.” Our theoretical analysis reveals that the gradient variance bound cannot be improved via Hessian spectral bounds; achieves $O(log d)$ convergence under sub-Gaussian approximating families; eliminates explicit dimension dependence entirely in the constant-Hessian case; and yields an $O(d^{2/k})$ bound for heavy-tailed distribution families. The analysis integrates reparameterization gradients, stochastic optimization theory, high-dimensional probability inequalities, and log-concave distribution theory. This provides the first tight, dimension-adaptive convergence guarantee for BBVI.
📝 Abstract
We prove that, given a mean-field location-scale variational family, black-box variational inference (BBVI) with the reparametrization gradient converges at an almost dimension-independent rate. Specifically, for strongly log-concave and log-smooth targets, the number of iterations for BBVI with a sub-Gaussian family to achieve an objective $epsilon$-close to the global optimum is $mathrm{O}(log d)$, which improves over the $mathrm{O}(d)$ dependence of full-rank location-scale families. For heavy-tailed families, we provide a weaker $mathrm{O}(d^{2/k})$ dimension dependence, where $k$ is the number of finite moments. Additionally, if the Hessian of the target log-density is constant, the complexity is free of any explicit dimension dependence. We also prove that our bound on the gradient variance, which is key to our result, cannot be improved using only spectral bounds on the Hessian of the target log-density.