Activation-Space Uncertainty Quantification for Pretrained Networks

📅 2026-02-16
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of efficiently and reliably estimating epistemic uncertainty in frozen pre-trained models. The authors propose GAPA, a method that constructs a Gaussian process model in the activation space, leveraging a replaceable nonlinear activation function to enable closed-form quantification of epistemic uncertainty. By shifting Bayesian modeling from weight space to activation space, GAPA requires neither retraining, sampling, nor backpropagation, and yields uncertainty estimates with a single forward pass. Integrating sparse variational inducing points, local k-nearest-neighbor conditioning, and cached training activations, GAPA achieves competitive or superior performance compared to existing post-hoc baselines in out-of-distribution detection and calibration across diverse tasks—including regression, classification, image segmentation, and language modeling—while maintaining high efficiency at test time.

Technology Category

Application Category

📝 Abstract
Reliable uncertainty estimates are crucial for deploying pretrained models; yet, many strong methods for quantifying uncertainty require retraining, Monte Carlo sampling, or expensive second-order computations and may alter a frozen backbone's predictions. To address this, we introduce Gaussian Process Activations (GAPA), a post-hoc method that shifts Bayesian modeling from weights to activations. GAPA replaces standard nonlinearities with Gaussian-process activations whose posterior mean exactly matches the original activation, preserving the backbone's point predictions by construction while providing closed-form epistemic variances in activation space. To scale to modern architectures, we use a sparse variational inducing-point approximation over cached training activations, combined with local k-nearest-neighbor subset conditioning, enabling deterministic single-pass uncertainty propagation without sampling, backpropagation, or second-order information. Across regression, classification, image segmentation, and language modeling, GAPA matches or outperforms strong post-hoc baselines in calibration and out-of-distribution detection while remaining efficient at test time.
Problem

Research questions and friction points this paper is trying to address.

uncertainty quantification
pretrained networks
post-hoc calibration
epistemic uncertainty
frozen backbone
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gaussian Process Activations
activation-space uncertainty
post-hoc uncertainty quantification
epistemic uncertainty
sparse variational approximation
🔎 Similar Papers
No similar papers found.
Richard Bergna
Richard Bergna
University of Cambridge
Probabilistic Machine LearningGaussian ProcessGNNBayesian Methods
Stefan Depeweg
Stefan Depeweg
Siemens AG
Machine LearningNeural NetworksReinforcement LearningBayesian Inference
S
Sergio Calvo-Ordoñez
Mathematical Institute and Oxford-Man Institute, University of Oxford, Oxford, UK
Jonathan Plenk
Jonathan Plenk
University of Oxford
Deep Learning TheoryUncertainty QuantificationMathematical Finance
A
Alvaro Cartea
Mathematical Institute and Oxford-Man Institute, University of Oxford, Oxford, UK
J
Jose Miguel Hernández-Lobato
Department of Engineering, University of Cambridge, Cambridge, UK