🤖 AI Summary
This work investigates the condition number stability of mass and stiffness matrices induced by shallow ReLU$^k$ neural networks on the unit sphere $mathbb{S}^d$, focusing on antipodally quasi-uniform node sets. It establishes sharp, quantitative estimates of the condition number and its interplay with approximation power and numerical stability. Methodologically, the paper introduces a novel analytical framework unifying harmonic analysis and spherical approximation theory to derive complete asymptotic characterizations of the eigenvalue spectrum: the smallest eigenvalues correspond to low-degree spherical harmonics, while the largest eigenvalues dominate high-degree components. This yields tight, full-spectrum estimates for both extremal eigenvalues. The resulting optimal condition number bound precisely quantifies the fundamental trade-off between expressive capacity and solver stability. These results provide a rigorous theoretical foundation and practical criteria for analyzing and training spherical neural networks.
📝 Abstract
We present an estimation of the condition numbers of the emph{mass} and emph{stiffness} matrices arising from shallow ReLU$^k$ neural networks defined on the unit sphere~$mathbb{S}^d$. In particular, when ${ heta_j^*}_{j=1}^n subset mathbb{S}^d$ is emph{antipodally quasi-uniform}, the condition number is sharp. Indeed, in this case, we obtain sharp asymptotic estimates for the full spectrum of eigenvalues and characterize the structure of the corresponding eigenspaces, showing that the smallest eigenvalues are associated with an eigenbasis of low-degree polynomials while the largest eigenvalues are linked to high-degree polynomials. This spectral analysis establishes a precise correspondence between the approximation power of the network and its numerical stability.