🤖 AI Summary
This work addresses the challenge of encoding Gaussian process (GP) function-space priors in Bayesian neural networks (BNNs). We propose a prior-matching method based on learnable, hyperparameter-conditioned activation functions. Our approach is the first to jointly incorporate activation function trainability and explicit hyperparameter dependence, systematically resolving key issues in function-space priors—including identifiability, symmetry, and principled loss design—while enabling evidence-driven model selection via marginal likelihood maximization. Under wide single-hidden-layer BNN architectures, the method achieves high-fidelity approximation of GP priors. Empirical evaluation demonstrates substantial improvements in uncertainty calibration, generalization robustness, and risk-sensitive decision-making. The effectiveness is validated across both standard regression and risk-aware prediction tasks.
📝 Abstract
Function-space priors in Bayesian Neural Networks provide a more intuitive approach to embedding beliefs directly into the model's output, thereby enhancing regularization, uncertainty quantification, and risk-aware decision-making. However, imposing function-space priors on BNNs is challenging. We address this task through optimization techniques that explore how trainable activations can accommodate complex priors and match intricate target function distributions. We discuss critical learning challenges, including identifiability, loss construction, and symmetries that arise in this context. Furthermore, we enable evidence maximization to facilitate model selection by conditioning the functional priors on additional hyperparameters. Our empirical findings demonstrate that even BNNs with a single wide hidden layer, when equipped with these adaptive trainable activations and conditioning strategies, can effectively achieve high-fidelity function-space priors, providing a robust and flexible framework for enhancing Bayesian neural network performance.