Attention-Based Feature Online Conformal Prediction for Time Series

📅 2025-11-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Standard online conformal prediction (OCP) suffers from two key limitations in time-series forecasting: (1) reliance on simplistic nonconformity scores defined solely in the output space, and (2) uniform weighting of historical data, rendering it ill-suited for non-stationary processes and distributional shifts. To address these, we propose the first neural-feature-space extension of OCP, incorporating an attention mechanism to adaptively weight past samples based on both task relevance and evolving distribution dynamics. Our method integrates pretrained feature representations, feature-space nonconformity scoring, and dynamic quantile estimation—guaranteeing long-term marginal coverage theoretically. Experiments across diverse synthetic and real-world time-series benchmarks demonstrate that our approach reduces average prediction interval width by 88% while strictly maintaining the target coverage level. These results validate the effectiveness and robustness of feature-space calibration coupled with attention-based adaptive weighting.

Technology Category

Application Category

📝 Abstract
Online conformal prediction (OCP) wraps around any pre-trained predictor to produce prediction sets with coverage guarantees that hold irrespective of temporal dependencies or distribution shifts. However, standard OCP faces two key limitations: it operates in the output space using simple nonconformity (NC) scores, and it treats all historical observations uniformly when estimating quantiles. This paper introduces attention-based feature OCP (AFOCP), which addresses both limitations through two key innovations. First, AFOCP operates in the feature space of pre-trained neural networks, leveraging learned representations to construct more compact prediction sets by concentrating on task-relevant information while suppressing nuisance variation. Second, AFOCP incorporates an attention mechanism that adaptively weights historical observations based on their relevance to the current test point, effectively handling non-stationarity and distribution shifts. We provide theoretical guarantees showing that AFOCP maintains long-term coverage while provably achieving smaller prediction intervals than standard OCP under mild regularity conditions. Extensive experiments on synthetic and real-world time series datasets demonstrate that AFOCP consistently reduces the size of prediction intervals by as much as $88%$ as compared to OCP, while maintaining target coverage levels, validating the benefits of both feature-space calibration and attention-based adaptive weighting.
Problem

Research questions and friction points this paper is trying to address.

Improves prediction set compactness using feature-space representations
Adapts to distribution shifts via attention-based historical weighting
Maintains coverage guarantees while reducing interval sizes
Innovation

Methods, ideas, or system contributions that make the work stand out.

Operates in feature space of neural networks
Uses attention mechanism for adaptive weighting
Maintains coverage with smaller prediction intervals
🔎 Similar Papers
No similar papers found.
M
Meiyi Zhu
King’s Communications, Learning & Information Processing (KCLIP) lab, Department of Engineering, King’s College London, London WC2R 2LS, U.K.
Caili Guo
Caili Guo
Beijing University of Posts and Telecommunications
wireless communicationcognitive radiostatistical signal processingsocial multimedia computingbig data processing,vehic
C
Chunyan Feng
Beijing Key Laboratory of Network System Architecture and Convergence, School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
Osvaldo Simeone
Osvaldo Simeone
King's College London
Information theorymachine learningquantum information processingwireless systems