Joint Learning in the Gaussian Single Index Model

πŸ“… 2025-05-27
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper studies joint learning of a one-dimensional projection direction $w^*$ and a univariate function $varphi^*$ in high-dimensional Gaussian settings, i.e., estimating $f(x) = varphi^*(langle w^*, x angle)$β€”a problem exhibiting both nonconvexity and the interplay between representation learning and nonlinear regression. We propose a gradient-flow-based alternating algorithm and establish, for the first time, its global convergence guarantee, revealing the counterintuitive phenomenon that negatively correlated initializations still converge. To characterize convergence behavior, we introduce the β€œinformation index,” which quantifies how Gaussian regularity governs convergence rates. Theoretically, we prove that the convergence rate is determined by the Gaussian regularity of the target function within an appropriately adapted reproducing kernel Hilbert space (RKHS). Empirically, our RKHS-based implementation demonstrates high efficiency, flexibility, and strong robustness, achieving superior generalization performance in high-dimensional sparse structure learning.

Technology Category

Application Category

πŸ“ Abstract
We consider the problem of jointly learning a one-dimensional projection and a univariate function in high-dimensional Gaussian models. Specifically, we study predictors of the form $f(x)=varphi^star(langle w^star, x angle)$, where both the direction $w^star in mathcal{S}_{d-1}$, the sphere of $mathbb{R}^d$, and the function $varphi^star: mathbb{R} o mathbb{R}$ are learned from Gaussian data. This setting captures a fundamental non-convex problem at the intersection of representation learning and nonlinear regression. We analyze the gradient flow dynamics of a natural alternating scheme and prove convergence, with a rate controlled by the information exponent reflecting the extit{Gaussian regularity} of the function $varphi^star$. Strikingly, our analysis shows that convergence still occurs even when the initial direction is negatively correlated with the target. On the practical side, we demonstrate that such joint learning can be effectively implemented using a Reproducing Kernel Hilbert Space (RKHS) adapted to the structure of the problem, enabling efficient and flexible estimation of the univariate function. Our results offer both theoretical insight and practical methodology for learning low-dimensional structure in high-dimensional settings.
Problem

Research questions and friction points this paper is trying to address.

Jointly learning projection and function in high-dimensional Gaussian models
Analyzing gradient flow dynamics for non-convex representation learning
Implementing efficient estimation using RKHS adapted to problem structure
Innovation

Methods, ideas, or system contributions that make the work stand out.

Joint learning of projection and univariate function
Gradient flow dynamics with convergence analysis
RKHS adapted for efficient function estimation
πŸ”Ž Similar Papers
No similar papers found.
Loucas Pillaud-Vivien
Loucas Pillaud-Vivien
Ecole Nationale Des Ponts
Machine LearningOptimization
A
Adrien Schertzer
Goethe University, Frankfurt am main, Germany