🤖 AI Summary
This work investigates the feature learning mechanism in deep neural networks, aiming to validate and generalize the Neural Feature Hypothesis (NFA)—the conjecture that the Gram matrix of first-layer weights asymptotically scales with the outer product of input gradients.
Method: We extend NFA systematically to L-layer networks, employing gradient flow dynamics and mean-field theory within a novel Average Gradient Outer Product (AGOP) analytical framework; we conduct comprehensive numerical experiments across optimizers, initialization schemes, and regularization settings.
Contribution/Results: We prove that NFA holds asymptotically under balanced initialization with exponent α = 1/L; remarkably, even under unbalanced initialization, weight decay restores its asymptotic validity. Our analysis not only confirms the depth-dependent scaling law but also constructs explicit counterexamples where NFA fails in nonlinear networks, thereby precisely characterizing its regime of validity.
📝 Abstract
Understanding feature learning is an important open question in establishing a mathematical foundation for deep neural networks. The Neural Feature Ansatz (NFA) states that after training, the Gram matrix of the first-layer weights of a deep neural network is proportional to some power $α>0$ of the average gradient outer product (AGOP) of this network with respect to its inputs. Assuming gradient flow dynamics with balanced weight initialization, the NFA was proven to hold throughout training for two-layer linear networks with exponent $α= 1/2$ (Radhakrishnan et al., 2024). We extend this result to networks with $L geq 2$ layers, showing that the NFA holds with exponent $α= 1/L$, thus demonstrating a depth dependency of the NFA. Furthermore, we prove that for unbalanced initialization, the NFA holds asymptotically through training if weight decay is applied. We also provide counterexamples showing that the NFA does not hold for some network architectures with nonlinear activations, even when these networks fit arbitrarily well the training data. We thoroughly validate our theoretical results through numerical experiments across a variety of optimization algorithms, weight decay rates and initialization schemes.