PCS-UQ: Uncertainty Quantification via the Predictability-Computability-Stability Framework

๐Ÿ“… 2025-05-13
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing uncertainty quantification (UQ) methods for machine learning in high-stakes settings face critical limitations: traditional approaches rely on correctly specified generative models and are sensitive to model misspecification, while conformal prediction ignores model selection and yields overly conservative prediction intervals. Method: We propose a novel UQ framework grounded in the Predictability-Computability-Stability (PCS) paradigmโ€”first systematically integrating PCS principles into UQ. Our approach selects optimal models via predictive checking, evaluates algorithmic and sample variability using multi-Bootstrap, and introduces a local adaptive calibration mechanism. Contribution/Results: Theoretically, our method guarantees valid coverage probability. Empirically, across 17 regression and 6 classification datasets, plus 3 vision benchmarks, it reduces average prediction interval/set width by 20% compared to split conformal prediction, while substantially improving subgroup robustness and compactness.

Technology Category

Application Category

๐Ÿ“ Abstract
As machine learning (ML) models are increasingly deployed in high-stakes domains, trustworthy uncertainty quantification (UQ) is critical for ensuring the safety and reliability of these models. Traditional UQ methods rely on specifying a true generative model and are not robust to misspecification. On the other hand, conformal inference allows for arbitrary ML models but does not consider model selection, which leads to large interval sizes. We tackle these drawbacks by proposing a UQ method based on the predictability, computability, and stability (PCS) framework for veridical data science proposed by Yu and Kumbier. Specifically, PCS-UQ addresses model selection by using a prediction check to screen out unsuitable models. PCS-UQ then fits these screened algorithms across multiple bootstraps to assess inter-sample variability and algorithmic instability, enabling more reliable uncertainty estimates. Further, we propose a novel calibration scheme that improves local adaptivity of our prediction sets. Experiments across $17$ regression and $6$ classification datasets show that PCS-UQ achieves the desired coverage and reduces width over conformal approaches by $approx 20%$. Further, our local analysis shows PCS-UQ often achieves target coverage across subgroups while conformal methods fail to do so. For large deep-learning models, we propose computationally efficient approximation schemes that avoid the expensive multiple bootstrap trainings of PCS-UQ. Across three computer vision benchmarks, PCS-UQ reduces prediction set size over conformal methods by $20%$. Theoretically, we show a modified PCS-UQ algorithm is a form of split conformal inference and achieves the desired coverage with exchangeable data.
Problem

Research questions and friction points this paper is trying to address.

Addresses unreliable uncertainty quantification in ML models
Improves model selection and stability via PCS framework
Reduces prediction interval size compared to conformal methods
Innovation

Methods, ideas, or system contributions that make the work stand out.

UQ method based on PCS framework
Prediction check screens unsuitable models
Novel calibration improves local adaptivity
๐Ÿ”Ž Similar Papers
No similar papers found.
Abhineet Agarwal
Abhineet Agarwal
Statistics PhD, University of California, Berkeley
Large Language ModelsAI ExplainabilityCausal InferenceBandits
M
Michael Xiao
Department of Statistics, University of California, Berkeley
R
Rebecca Barter
Department of Epidemiology, University of Utah
Omer Ronen
Omer Ronen
UC Berkeley
B
Boyu Fan
Department of Statistics, University of California, Berkeley
B
Bin Yu
Department of Statistics, University of California, Berkeley; Department of Electrical Engineering and Computer Science, University of California, Berkeley