Walking on the Fiber: A Simple Geometric Approximation for Bayesian Neural Networks

📅 2025-12-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Exact posterior inference in Bayesian neural networks (BNNs) is computationally intractable, while mainstream approximations—e.g., Laplace approximation—suffer from poor scalability and inaccurate posterior estimation in deep networks. To address this, we propose a geometric parameter-space deformation framework for efficient posterior approximation. Leveraging the low-dimensional manifold structure of loss minima in overparameterized networks, our method learns an explicit, invertible coordinate transformation that maps the complex posterior to a standard distribution—enabling direct sampling without iterative optimization. Theoretically grounded and empirically scalable, our approach achieves uncertainty calibration and predictive accuracy on par with or superior to state-of-the-art methods (e.g., Laplace, SWAG, K-FAC), at significantly lower computational cost across multiple benchmarks. This work introduces a novel paradigm for scalable deep Bayesian inference.

Technology Category

Application Category

📝 Abstract
Bayesian Neural Networks provide a principled framework for uncertainty quantification by modeling the posterior distribution of network parameters. However, exact posterior inference is computationally intractable, and widely used approximations like the Laplace method struggle with scalability and posterior accuracy in modern deep networks. In this work, we revisit sampling techniques for posterior exploration, proposing a simple variation tailored to efficiently sample from the posterior in over-parameterized networks by leveraging the low-dimensional structure of loss minima. Building on this, we introduce a model that learns a deformation of the parameter space, enabling rapid posterior sampling without requiring iterative methods. Empirical results demonstrate that our approach achieves competitive posterior approximations with improved scalability compared to recent refinement techniques. These contributions provide a practical alternative for Bayesian inference in deep learning.
Problem

Research questions and friction points this paper is trying to address.

Efficiently sample Bayesian posterior in over-parameterized networks
Leverage low-dimensional structure of loss minima for scalability
Enable rapid posterior sampling without iterative methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

Sampling posterior via low-dimensional loss minima
Deforming parameter space for rapid sampling
Scalable Bayesian inference without iterative methods
🔎 Similar Papers
No similar papers found.