🤖 AI Summary
This work investigates the convergence properties and implicit bias of matrix-form stochastic mirror descent (SMD) in over-parameterized, multi-output high-dimensional problems such as multiclass classification and matrix completion. By extending the implicit bias theory of vector-valued SMD to the matrix setting, it reveals for the first time how the Bregman divergence induced by the mirror map governs the uniqueness of interpolating solutions and the associated inductive bias. The theoretical analysis demonstrates that, under over-parameterization, matrix SMD converges exponentially to the unique solution that both interpolates the data and minimizes the Bregman divergence from the initialization. This result elucidates the pivotal role of the mirror map in shaping the generalization behavior of the learned model.
📝 Abstract
We investigate Stochastic Mirror Descent (SMD) with matrix parameters and vector-valued predictions, a framework relevant to multi-class classification and matrix completion problems. Focusing on the overparameterized regime, where the total number of parameters exceeds the number of training samples, we prove that SMD with matrix mirror functions $ψ(\cdot)$ converges exponentially to a global interpolator. Furthermore, we generalize classical implicit bias results of vector SMD by demonstrating that the matrix SMD algorithm converges to the unique solution minimizing the Bregman divergence induced by $ψ(\cdot)$ from initialization subject to interpolating the data. These findings reveal how matrix mirror maps dictate inductive bias in high-dimensional, multi-output problems.