🤖 AI Summary
This work studies constrained minimax optimization problems where the objective is nonconvex in the primal variable and strongly concave in the dual variable. We propose a first-order augmented Lagrangian framework whose subproblems are unconstrained nonconvex–strongly concave minimax problems, and design a specialized first-order inner solver that exploits strong concavity to accelerate convergence. Under standard assumptions, our method achieves— for the first time— a gradient complexity of $O(varepsilon^{-3.5}logvarepsilon^{-1})$, improving upon the previous best rate by a factor of $varepsilon^{-0.5}$. Moreover, it efficiently computes an $varepsilon$-KKT point. This work provides a new algorithmic tool for applications such as nonconvex adversarial training and constrained games, offering both theoretical guarantees and computational tractability.
📝 Abstract
In this paper we study a nonconvex-strongly-concave constrained minimax problem. Specifically, we propose a first-order augmented Lagrangian method for solving it, whose subproblems are nonconvex-strongly-concave unconstrained minimax problems and suitably solved by a first-order method developed in this paper that leverages the strong concavity structure. Under suitable assumptions, the proposed method achieves an emph{operation complexity} of $O(varepsilon^{-3.5}logvarepsilon^{-1})$, measured in terms of its fundamental operations, for finding an $varepsilon$-KKT solution of the constrained minimax problem, which improves the previous best-known operation complexity by a factor of $varepsilon^{-0.5}$.