Last-Iterate Convergence Properties of Regret-Matching Algorithms in Games

📅 2023-11-01
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Regret Matching⁺ (RM⁺) and its variants—despite widespread use in two-player zero-sum games—lack last-iterate convergence guarantees, and their limit-point behavior remains poorly understood. This work first characterizes the geometric structure of RM⁺ limit points, revealing the fundamental non-convergence of standard, predictive, and alternating RM⁺ variants even in simple 3×3 matrix games. Method: We propose two novel algorithms—Smooth Predictive RM⁺ and Extragradient RM⁺—that integrate smoothing, predictive updates, extragradient steps, and adaptive restarts. Contribution/Results: We establish asymptotic last-iterate convergence for both algorithms, achieving the optimal sublinear rate O(1/√t); with restarts, linear convergence is attained. Leveraging non-monotone operator analysis, our convergence proofs require no restrictive rate assumptions. Numerical experiments confirm the failure of existing RM⁺ variants and demonstrate the efficacy and robustness of the proposed methods.
📝 Abstract
We study last-iterate convergence properties of algorithms for solving two-player zero-sum games based on Regret Matching$^+$ (RM$^+$). Despite their widespread use for solving real games, virtually nothing is known about their last-iterate convergence. A major obstacle to analyzing RM-type dynamics is that their regret operators lack Lipschitzness and (pseudo)monotonicity. We start by showing numerically that several variants used in practice, such as RM$^+$, predictive RM$^+$ and alternating RM$^+$, all lack last-iterate convergence guarantees even on a simple $3 imes 3$ matrix game. We then prove that recent variants of these algorithms based on a smoothing technique, extragradient RM$^{+}$ and smooth Predictive RM$^+$, enjoy asymptotic last-iterate convergence (without a rate), $1/sqrt{t}$ best-iterate convergence, and when combined with restarting, linear-rate last-iterate convergence. Our analysis builds on a new characterization of the geometric structure of the limit points of our algorithms, marking a significant departure from most of the literature on last-iterate convergence. We believe that our analysis may be of independent interest and offers a fresh perspective for studying last-iterate convergence in algorithms based on non-monotone operators.
Problem

Research questions and friction points this paper is trying to address.

Analyzing last-iterate convergence in Regret-Matching algorithms.
Overcoming lack of Lipschitzness and monotonicity in RM dynamics.
Proving convergence for smoothed variants of RM algorithms.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Smoothing technique enhances RM+ convergence.
Extragradient RM+ ensures asymptotic last-iterate convergence.
Restarting achieves linear-rate last-iterate convergence.
🔎 Similar Papers
No similar papers found.