π€ AI Summary
This paper investigates the learning performance of regularized least squares (i.e., kernel ridge regression) in reproducing kernel Hilbert spaces under heavy-tailed noiseβwhere only finite higher-order moments exist, violating the sub-exponential assumption. Methodologically, it introduces an integral operator framework combined with a Hilbert-space extension of the Fuk-Nagaev inequality. The analysis yields an excess risk bound comprising a sub-Gaussian leading term and a polynomial remainder term. Crucially, this is the first result establishing sub-Gaussian convergence rates under heavy-tailed noise, rigorously proving the asymptotic robustness of kernel ridge regression. Under standard eigenvalue decay conditions, the method achieves minimax-optimal convergence rates. These advances break the two-decade reliance on sub-exponential noise assumptions in kernel learning theory, substantially broadening both the applicability and theoretical foundation of kernel ridge regression.
π Abstract
This paper examines the performance of ridge regression in reproducing kernel Hilbert spaces in the presence of noise that exhibits a finite number of higher moments. We establish excess risk bounds consisting of subgaussian and polynomial terms based on the well known integral operator framework. The dominant subgaussian component allows to achieve convergence rates that have previously only been derived under subexponential noise - a prevalent assumption in related work from the last two decades. These rates are optimal under standard eigenvalue decay conditions, demonstrating the asymptotic robustness of regularized least squares against heavy-tailed noise. Our derivations are based on a Fuk-Nagaev inequality for Hilbert-space valued random variables.