Optimization Guarantees for Square-Root Natural-Gradient Variational Inference

📅 2025-07-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Natural-gradient variational inference (NGVI) exhibits fast empirical convergence but lacks rigorous theoretical convergence guarantees—even under idealized settings such as concave log-likelihoods and Gaussian approximations. Method: We introduce a square-root parameterization of the covariance matrix, embedding the natural gradient flow into a Euclidean geometric framework. This circumvents analytical challenges arising from singularity of the Fisher information matrix and complex Riemannian curvature inherent in standard parameterizations. Contribution/Results: We establish, for the first time, global convergence of the continuous-time natural gradient flow for Gaussian variational inference and derive explicit convergence rate bounds for its discrete-time counterpart. Our analysis unifies Euclidean and Wasserstein geometric perspectives. Experiments demonstrate that the proposed method significantly outperforms both standard Euclidean-gradient descent and Wasserstein natural-gradient methods in terms of convergence speed and numerical stability.

Technology Category

Application Category

📝 Abstract
Variational inference with natural-gradient descent often shows fast convergence in practice, but its theoretical convergence guarantees have been challenging to establish. This is true even for the simplest cases that involve concave log-likelihoods and use a Gaussian approximation. We show that the challenge can be circumvented for such cases using a square-root parameterization for the Gaussian covariance. This approach establishes novel convergence guarantees for natural-gradient variational-Gaussian inference and its continuous-time gradient flow. Our experiments demonstrate the effectiveness of natural gradient methods and highlight their advantages over algorithms that use Euclidean or Wasserstein geometries.
Problem

Research questions and friction points this paper is trying to address.

Theoretical convergence guarantees for natural-gradient variational inference
Square-root parameterization for Gaussian covariance in optimization
Advantages of natural gradient over Euclidean or Wasserstein methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Square-root parameterization for Gaussian covariance
Novel convergence guarantees for variational-Gaussian inference
Natural gradient methods outperform Euclidean alternatives
🔎 Similar Papers
No similar papers found.