A Regularization-Sharpness Tradeoff for Linear Interpolators

📅 2026-02-13
📈 Citations: 0
Influential: 0
📄 PDF

Technology Category

Application Category

📝 Abstract
The rule of thumb regarding the relationship between the bias-variance tradeoff and model size plays a key role in classical machine learning, but is now well-known to break down in the overparameterized setting as per the double descent curve. In particular, minimum-norm interpolating estimators can perform well, suggesting the need for new tradeoff in these settings. Accordingly, we propose a regularization-sharpness tradeoff for overparameterized linear regression with an $\ell^p$ penalty. Inspired by the interpolating information criterion, our framework decomposes the selection penalty into a regularization term (quantifying the alignment of the regularizer and the interpolator) and a geometric sharpness term on the interpolating manifold (quantifying the effect of local perturbations), yielding a tradeoff analogous to bias-variance. Building on prior analyses that established this information criterion for ridge regularizers, this work first provides a general expression of the interpolating information criterion for $\ell^p$ regularizers where $p \ge 2$. Subsequently, we extend this to the LASSO interpolator with $\ell^1$ regularizer, which induces stronger sparsity. Empirical results on real-world datasets with random Fourier features and polynomials validate our theory, demonstrating how the tradeoff terms can distinguish performant linear interpolators from weaker ones.
Problem

Research questions and friction points this paper is trying to address.

overparameterization
linear interpolation
regularization
sharpness
generalization
Innovation

Methods, ideas, or system contributions that make the work stand out.

regularization-sharpness tradeoff
interpolating information criterion
overparameterized linear regression
ℓ^p regularization
double descent
🔎 Similar Papers
2023-06-19Journal of Artificial Intelligence ResearchCitations: 1
Q
Qingyi Hu
School of Mathematics and Statistics, University of Melbourne, Australia
Liam Hodgkinson
Liam Hodgkinson
University of Melbourne
probabilistic machine learningdeep learning theory