Online Estimation with Rolling Validation: Adaptive Nonparametric Estimation with Streaming Data

๐Ÿ“… 2023-10-18
๐Ÿ“ˆ Citations: 1
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Existing methods for online nonparametric estimation on streaming data lack efficient, adaptive hyperparameter selection mechanisms. Method: We propose Weighted Rolling Validation (WRV), a low-overhead, fully online model selection framework that generalizes leave-one-out cross-validation to the streaming setting via temporal weighting of historical validation samples. Grounded in statistical stability assumptions, WRV dynamically assigns time-decaying weights without requiring additional storage or retraining, and is compatible with stochastic gradientโ€“based nonparametric estimators. Contribution/Results: We establish theoretical guarantees showing WRV achieves adaptive convergence rates. Empirically, WRV exhibits high sensitivity to subtle performance differences among candidate estimators, incurs negligible computational overhead, and significantly improves prediction accuracy and robustness. To our knowledge, WRV is the first lightweight, theoretically grounded hyperparameter adaptation mechanism for online nonparametric learning.
๐Ÿ“ Abstract
Online nonparametric estimators are gaining popularity due to their efficient computation and competitive generalization abilities. An important example includes variants of stochastic gradient descent. These algorithms often take one sample point at a time and incrementally update the parameter estimate of interest. In this work, we consider model selection/hyperparameter tuning for such online algorithms. We propose a weighted rolling validation procedure, an online variant of leave-one-out cross-validation, that costs minimal extra computation for many typical stochastic gradient descent estimators and maintains their online nature. Similar to batch cross-validation, it can boost base estimators to achieve better heuristic performance and adaptive convergence rate. Our analysis is straightforward, relying mainly on some general statistical stability assumptions. The simulation study underscores the significance of diverging weights in practice and demonstrates its favorable sensitivity even when there is only a slim difference between candidate estimators.
Problem

Research questions and friction points this paper is trying to address.

Online model selection for streaming data algorithms
Hyperparameter tuning in online nonparametric estimation
Weighted rolling validation for adaptive convergence rates
Innovation

Methods, ideas, or system contributions that make the work stand out.

Weighted rolling validation for online algorithms
Minimal extra computation for SGD estimators
Boosts performance with adaptive convergence rates
๐Ÿ”Ž Similar Papers
No similar papers found.
T
Tianyu Zhang
Department of Statistics and Applied Probability, University of California, Santa Barbara
Jing Lei
Jing Lei
Carnegie Mellon University
Probability and Statistics