๐ค AI Summary
This work addresses the challenge of dynamically constructing statistically valid prediction sets under semi-bandit feedback in online learningโwhere the true label is observable only if it falls within the predicted set. We introduce conformal prediction to the stochastic online setting for the first time. Our method proposes the first adaptive algorithm with a sublinear regret bound, eliminating reliance on an i.i.d. calibration set. By integrating stochastic optimization and online learning theory, we design a confidence-set-based dynamic threshold update mechanism that simultaneously ensures coverage guarantees and minimizes regret. Empirical evaluation across document retrieval, image classification, and auction pricing demonstrates significant improvements over baselines, validating both statistical validity and practical utility.
๐ Abstract
Conformal prediction has emerged as an effective strategy for uncertainty quantification by modifying a model to output sets of labels instead of a single label. These prediction sets come with the guarantee that they contain the true label with high probability. However, conformal prediction typically requires a large calibration dataset of i.i.d. examples. We consider the online learning setting, where examples arrive over time, and the goal is to construct prediction sets dynamically. Departing from existing work, we assume semi-bandit feedback, where we only observe the true label if it is contained in the prediction set. For instance, consider calibrating a document retrieval model to a new domain; in this setting, a user would only be able to provide the true label if the target document is in the prediction set of retrieved documents. We propose a novel conformal prediction algorithm targeted at this setting, and prove that it obtains sublinear regret compared to the optimal conformal predictor. We evaluate our algorithm on a retrieval task, an image classification task, and an auction price-setting task, and demonstrate that it empirically achieves good performance compared to several baselines.