🤖 AI Summary
This work addresses the lack of generalization theory for Randomized Coordinate Descent (RCD) in pairwise learning tasks (e.g., ranking, metric learning). Methodologically, it introduces parameter stability analysis—previously unexplored in pairwise learning—for RCD, integrates early stopping to balance optimization and estimation errors, and derives expected generalization upper bounds for both convex and strongly convex objectives. The theoretical contributions are twofold: (i) it establishes tight generalization bounds under standard pairwise loss functions; and (ii) under a low-noise condition, it achieves the optimal $O(1/n)$ convergence rate for excess risk. To the best of our knowledge, this is the first systematic theoretical guarantee for RCD in pairwise learning and the first stability-based generalization analysis in this setting, thereby filling a critical gap in the literature.
📝 Abstract
Pairwise learning includes various machine learning tasks, with ranking and metric learning serving as the primary representatives. While randomized coordinate descent (RCD) is popular in various learning problems, there is much less theoretical analysis on the generalization behavior of models trained by RCD, especially under the pairwise learning framework. In this paper, we consider the generalization of RCD for pairwise learning. We measure the on-average argument stability for both convex and strongly convex objective functions, based on which we develop generalization bounds in expectation. The early-stopping strategy is adopted to quantify the balance between estimation and optimization. Our analysis further incorporates the low-noise setting into the excess risk bound to achieve the optimistic bound as $O(1/n)$, where $n$ is the sample size.