🤖 AI Summary
In Bayesian neural networks (BNNs), selecting weight priors faces a fundamental trade-off: overly strong shrinkage degrades discriminative feature representation, while insufficient shrinkage fails to suppress noise. To address this, we propose the R²-induced Dirichlet Decomposition (R2D2) prior—the first to integrate the R² statistic into the Dirichlet decomposition framework—enabling dynamic balancing between sparsity and preservation of task-critical features. We design a variational Gibbs inference algorithm that ensures convergence stability and posterior consistency under non-convex shrinkage objectives. Theoretically, we characterize both the evidence lower bound (ELBO) and the posterior contraction rate. Experiments across natural and medical image classification, as well as uncertainty estimation tasks, demonstrate that R2D2 significantly improves predictive accuracy, calibration, and robustness, while effectively distinguishing noise-corrupted weights from discriminative signal components.
📝 Abstract
Bayesian neural networks (BNNs) treat neural network weights as random variables, which aim to provide posterior uncertainty estimates and avoid overfitting by performing inference on the posterior weights. However, selection of appropriate prior distributions remains a challenging task, and BNNs may suffer from catastrophic inflated variance or poor predictive performance when poor choices are made for the priors. Existing BNN designs apply different priors to weights, while the behaviours of these priors make it difficult to sufficiently shrink noisy signals or they are prone to overshrinking important signals in the weights. To alleviate this problem, we propose a novel R2D2-Net, which imposes the $R^{2}$-induced Dirichlet Decomposition (R2D2) prior to the BNN weights. The R2D2-Net can effectively shrink irrelevant coefficients towards zero, while preventing key features from over-shrinkage. To approximate the posterior distribution of weights more accurately, we further propose a variational Gibbs inference algorithm that combines the Gibbs updating procedure and gradient-based optimization. This strategy enhances stability and consistency in estimation when the variational objective involving the shrinkage parameters is non-convex. We also analyze the evidence lower bound (ELBO) and the posterior concentration rates from a theoretical perspective. Experiments on both natural and medical image classification and uncertainty estimation tasks demonstrate satisfactory performances of our method.