๐ค AI Summary
This paper addresses three key challenges in constructing dynamic confidence sets for black-box models in online conformal prediction: (1) fragile reliance on i.i.d. or exchangeability assumptions; (2) first-order online optimization under adversarial environments requiring pre-specified quantiles and suffering from monotonicity distortion due to loss linearization; and (3) the difficulty of simultaneously achieving low regret and statistical fidelity under multiple confidence-level queries. We propose the first Bayesian-regularized empirical distribution framework, integrating nonlinear Follow-the-Regularized-Leader (FTRL), online quantile forecasting, and a multi-head shared-belief update architecture. Our method imposes no distributional assumptions, inherently supports U-calibration, and guarantees low-regret coverage across multiple confidence levels for arbitrary input sequencesโwhile automatically attaining exact coverage under i.i.d. settings. Empirically, it significantly outperforms existing baselines, especially under non-stationary data, demonstrating strong robustness.
๐ Abstract
Based on the framework of Conformal Prediction (CP), we study the online construction of confidence sets given a black-box machine learning model. By converting the target confidence levels into quantile levels, the problem can be reduced to predicting the quantiles (in hindsight) of a sequentially revealed data sequence. Two very different approaches have been studied previously: (i) Assuming the data sequence is iid or exchangeable, one could maintain the empirical distribution of the observed data as an algorithmic belief, and directly predict its quantiles. (ii) Due to the fragility of statistical assumptions, a recent trend is to consider the non-distributional, adversarial setting and apply first-order online optimization algorithms to moving quantile losses. However, it requires the oracle knowledge of the target quantile level, and suffers from a previously overlooked monotonicity issue due to the associated loss linearization. This paper presents an adaptive CP algorithm that combines their strengths. Without any statistical assumption, it is able to answer multiple arbitrary confidence level queries with low regret, while also overcoming the monotonicity issue suffered by first-order optimization baselines. Furthermore, if the data sequence is actually iid, then the same algorithm is automatically equipped with the"correct"coverage probability guarantee. To achieve such strengths, our key technical innovation is to regularize the aforementioned algorithmic belief (the empirical distribution) by a Bayesian prior, which robustifies it by simulating a non-linearized Follow the Regularized Leader (FTRL) algorithm on the output. Such a belief update backbone is shared by prediction heads targeting different confidence levels, bringing practical benefits analogous to the recently proposed concept of U-calibration (Kleinberg et al., 2023).