Finding a Multiple Follower Stackelberg Equilibrium: A Fully First-Order Method

πŸ“… 2025-09-09
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This paper addresses the efficient computation of Ξ΅-stationary equilibria in multi-follower Stackelberg games, proposing the first fully first-order algorithm with theoretical convergence guarantees. Methodologically, it reformulates the bilevel game as a single-level constrained optimization problem, approximating the leader’s gradient via the Lagrangian function and relying solely on first-order derivative information; follower best responses are approximated via gradient descent, while leader and follower strategies are updated jointly to handle bilevel coupling. Key contributions include: (1) the first convergence guarantee for a fully first-order method in this setting; (2) elimination of second-order derivatives or implicit differentiation, drastically reducing computational complexity; and (3) strong scalability and numerical stability in multi-follower scenarios. Experiments demonstrate superior performance in accuracy, efficiency, and robustness compared to existing approaches.

Technology Category

Application Category

πŸ“ Abstract
In this work, we propose the first fully first-order method to compute an epsilon stationary Stackelberg equilibrium with convergence guarantees. To achieve this, we first reframe the leader follower interaction as single level constrained optimization. Second, we define the Lagrangian and show that it can approximate the leaders gradient in response to the equilibrium reached by followers with only first-order gradient evaluations. These findings suggest a fully first order algorithm that alternates between (i) approximating followers best responses through gradient descent and (ii) updating the leaders strategy via approximating the gradient using Lagrangian.
Problem

Research questions and friction points this paper is trying to address.

Finding a multiple follower Stackelberg equilibrium solution
Developing a fully first-order method with convergence guarantees
Approximating leader's gradient using Lagrangian without second-order information
Innovation

Methods, ideas, or system contributions that make the work stand out.

Reframe leader-follower as constrained optimization
Define Lagrangian to approximate leader's gradient
Alternate follower response and leader update
πŸ”Ž Similar Papers
No similar papers found.
A
April Niu
School of Industrial and Systems Engineering, Georgia Tech
K
Kai Wang
School of Computational Science and Engineering, Georgia Tech
Juba Ziani
Juba Ziani
Georgia Institute of Technology (ISyE)
Mechanism designdata marketsfairnessprivacylearning theory