đ¤ AI Summary
This work investigates the mechanism underlying the spontaneous emergence of low-rank structure in neural network weights during training. Under significantly weaker assumptionsâarbitrary depth and width, full-parameter training, smooth loss, infinitesimal regularization, and convergence only to a second-order stationary pointâwe establish the universality of this phenomenon. We introduce a key âderandomization lemmaâ that, combined with expectation-function analysis and perturbed gradient descent, rigorously proves that the first-layer weight matrix converges to low-rank structure throughout training. This structural bias substantially reduces the sample complexity required for generalization. Moreover, it enables end-to-end, provably correct neural solvers for combinatorial optimization problemsâincluding MAX-CUT approximation and JohnsonâLindenstrauss embeddingâby leveraging the emergent low-rank geometry. Crucially, our theory dispenses with strong initialization requirements, architecture-specific constraints, or stringent convergence assumptions (e.g., global optimality), thereby providing, for the first time, a rigorous foundation for structural discovery under broad, practically relevant settings.
đ Abstract
Understanding the dynamics of feature learning in neural networks (NNs) remains a significant challenge. The work of (Mousavi-Hosseini et al., 2023) analyzes a multiple index teacher-student setting and shows that a two-layer student attains a low-rank structure in its first-layer weights when trained with stochastic gradient descent (SGD) and a strong regularizer. This structural property is known to reduce sample complexity of generalization. Indeed, in a second step, the same authors establish algorithm-specific learning guarantees under additional assumptions. In this paper, we focus exclusively on the structure discovery aspect and study it under weaker assumptions, more specifically: we allow (a) NNs of arbitrary size and depth, (b) with all parameters trainable, (c) under any smooth loss function, (d) tiny regularization, and (e) trained by any method that attains a second-order stationary point (SOSP), e.g. perturbed gradient descent (PGD). At the core of our approach is a key $ extit{derandomization}$ lemma, which states that optimizing the function $mathbb{E}_{mathbf{x}} left[g_θ(mathbf{W}mathbf{x} + mathbf{b})
ight]$ converges to a point where $mathbf{W} = mathbf{0}$, under mild conditions. The fundamental nature of this lemma directly explains structure discovery and has immediate applications in other domains including an end-to-end approximation for MAXCUT, and computing Johnson-Lindenstrauss embeddings.