A Compositional Kernel Model for Feature Learning

πŸ“… 2025-09-17
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenges of identifying relevant variables and suppressing noise in nonlinear feature learning. We propose a coordinate-wise reweighted composite kernel ridge regression framework. Methodologically, we integrate variational analysis with variable selection theory to systematically characterize how different kernels affect feature recoverability: we theoretically prove that ℓ₁-type kernels (e.g., Laplacian) enable exact recovery of nonlinearly relevant features at stationary points, whereas Gaussian kernels only guarantee linear feature recovery. Furthermore, under Gaussian noise, both the global optimum and stationary points exhibit consistent variable screening consistencyβ€”i.e., asymptotically eliminating irrelevant coordinates while correctly identifying truly relevant ones. This work establishes the first unified analytical framework for kernel-based nonlinear feature selection that simultaneously provides rigorous theoretical guarantees and mechanistic interpretation.

Technology Category

Application Category

πŸ“ Abstract
We study a compositional variant of kernel ridge regression in which the predictor is applied to a coordinate-wise reweighting of the inputs. Formulated as a variational problem, this model provides a simple testbed for feature learning in compositional architectures. From the perspective of variable selection, we show how relevant variables are recovered while noise variables are eliminated. We establish guarantees showing that both global minimizers and stationary points discard noise coordinates when the noise variables are Gaussian distributed. A central finding is that $ell_1$-type kernels, such as the Laplace kernel, succeed in recovering features contributing to nonlinear effects at stationary points, whereas Gaussian kernels recover only linear ones.
Problem

Research questions and friction points this paper is trying to address.

Studying compositional kernel ridge regression for feature learning
Analyzing variable selection to recover relevant and eliminate noise variables
Comparing Laplace and Gaussian kernels for nonlinear feature recovery
Innovation

Methods, ideas, or system contributions that make the work stand out.

Compositional kernel ridge regression model
Variational formulation for feature learning
Laplace kernel recovers nonlinear features
πŸ”Ž Similar Papers
No similar papers found.
Feng Ruan
Feng Ruan
Department of EECS, University of California, Berkeley
Machine LearningStatistics
K
Keli Liu
Company E
M
Michael Jordan
Department of Statistics and Computer Science, University of California, Berkeley