🤖 AI Summary
This work establishes the asymptotic normality of rank-$k$ matrix estimators in generalized low-rank matrix sensing. The core challenge arises from the degeneracy of the Hessian matrix in Euclidean space, induced by rotational symmetry of the loss function. To address this, the authors model the parameter space as the quotient manifold $mathbb{R}^{d imes k} / O(k)$ and reformulate estimation within the horizontal space of this Riemannian quotient manifold. This constitutes the first asymptotic normality analysis for such problems under non-Euclidean geometric structure. Theoretically, they prove that $sqrt{n}(hat{varphi} - varphi^*)$ converges in distribution to $mathcal{N}(0, (H^*)^{-1})$, where $H^*$ denotes the effective Hessian restricted to the horizontal space. This result overcomes fundamental limitations of classical Euclidean analysis and provides the first rigorous large-sample statistical inference guarantee for low-rank matrix sensing under generalized convex losses.
📝 Abstract
We prove an asymptotic normality guarantee for generalized low-rank matrix sensing -- i.e., matrix sensing under a general convex loss $arell(langle X,M
angle,y^*)$, where $Minmathbb{R}^{d imes d}$ is the unknown rank-$k$ matrix, $X$ is a measurement matrix, and $y^*$ is the corresponding measurement. Our analysis relies on tools from Riemannian geometry to handle degeneracy of the Hessian of the loss due to rotational symmetry in the parameter space. In particular, we parameterize the manifold of low-rank matrices by $ar hetaar heta^ op$, where $ar hetainmathbb{R}^{d imes k}$. Then, assuming the minimizer of the empirical loss $ar heta^0inmathbb{R}^{d imes k}$ is in a constant size ball around the true parameters $ar heta^*$, we prove $sqrt{n}(phi^0-phi^*)xrightarrow{D}N(0,(H^*)^{-1})$ as $n oinfty$, where $phi^0$ and $phi^*$ are representations of $ar heta^*$ and $ar heta^0$ in the horizontal space of the Riemannian quotient manifold $mathbb{R}^{d imes k}/ ext{O}(k)$, and $H^*$ is the Hessian of the true loss in the same representation.