🤖 AI Summary
This paper addresses the scalability bottleneck of kernel ridge regression (KRR) under covariate shift for large-scale datasets. We propose an efficient algorithm based on random subspace projection in a reproducing kernel Hilbert space (RKHS). Our key contribution is the first theoretical guarantee—under covariate shift—that random projection achieves near-optimal statistical error rate $O(n^{-1/2})$ while reducing time and memory complexity from $O(n^3)$ and $O(n^2)$ to $O(mn)$ and $O(m)$, respectively, where $m ll n$. Crucially, the method avoids explicit estimation of covariate shift weights, ensuring both theoretical rigor and practical deployability. Empirical results demonstrate significant speedups (several-fold acceleration) and substantial memory reduction, without sacrificing predictive accuracy.
📝 Abstract
This paper addresses the covariate shift problem in the context of nonparametric regression within reproducing kernel Hilbert spaces (RKHSs). Covariate shift arises in supervised learning when the input distributions of the training and test data differ, presenting additional challenges for learning. Although kernel methods have optimal statistical properties, their high computational demands in terms of time and, particularly, memory, limit their scalability to large datasets. To address this limitation, the main focus of this paper is to explore the trade-off between computational efficiency and statistical accuracy under covariate shift. We investigate the use of random projections where the hypothesis space consists of a random subspace within a given RKHS. Our results show that, even in the presence of covariate shift, significant computational savings can be achieved without compromising learning performance.