π€ AI Summary
This work investigates the optimal universal convergence rates for binary classification in the agnostic setting, dispensing with the classical realizability assumption. By integrating tools from statistical learning theory, combinatorial complexity, and concentration inequalities, it establishes the first βtetrachotomyβ theorem for agnostic learning, demonstrating that the optimal excess risk convergence rate for any concept class must fall into one of four categories: $e^{-n}$, $e^{-o(n)}$, $o(n^{-1/2})$, or arbitrarily slow. This framework fully characterizes the spectrum of achievable rates in agnostic learning and provides a decision mechanism grounded in a succinct combinatorial structure, thereby laying the theoretical foundation for universal convergence rates in this setting.
π Abstract
We provide a complete theory of optimal universal rates for binary classification in the agnostic setting. This extends the realizable-case theory of Bousquet, Hanneke, Moran, van Handel, and Yehudayoff (2021) by removing the realizability assumption on the distribution. We identify a fundamental tetrachotomy of optimal rates: for every concept class, the optimal universal rate of convergence of the excess error rate is one of $e^{-n}$, $e^{-o(n)}$, $o(n^{-1/2})$, or arbitrarily slow. We further identify simple combinatorial structures which determine which of these categories any given concept class falls into.