π€ AI Summary
Existing penalty-based methods for bilevel optimization (BLO) with coupling constraints suffer from high computational overhead due to inner-loop iterations and necessitate small outer-loop step sizes owing to stringent smoothness requirements, leading to poor convergence complexity.
Method: This paper proposes a novel penalty reformulation that decouples upper- and lower-level variables, substantially reducing reliance on objective function smoothness. Building upon this, we design PBGD-Freeβa single-loop algorithm that eliminates inner-loop optimization and replaces the standard Lipschitz gradient assumption with a curvature condition, enabling smaller penalty coefficients and tighter gradient control.
Contribution/Results: We establish theoretical convergence guarantees under mild assumptions. Empirical evaluation on SVM hyperparameter tuning and large-model fine-tuning demonstrates significant reductions in iteration complexity and substantial improvements in training efficiency.
π Abstract
Penalty-based methods have become popular for solving bilevel optimization (BLO) problems, thanks to their effective first-order nature. However, they often require inner-loop iterations to solve the lower-level (LL) problem and small outer-loop step sizes to handle the increased smoothness induced by large penalty terms, leading to suboptimal complexity. This work considers the general BLO problems with coupled constraints (CCs) and leverages a novel penalty reformulation that decouples the upper- and lower-level variables. This yields an improved analysis of the smoothness constant, enabling larger step sizes and reduced iteration complexity for Penalty-Based Gradient Descent algorithms in ALTernating fashion (ALT-PBGD). Building on the insight of reduced smoothness, we propose PBGD-Free, a novel fully single-loop algorithm that avoids inner loops for the uncoupled constraint BLO. For BLO with CCs, PBGD-Free employs an efficient inner-loop with substantially reduced iteration complexity. Furthermore, we propose a novel curvature condition describing the "flatness" of the upper-level objective with respect to the LL variable. This condition relaxes the traditional upper-level Lipschitz requirement, enables smaller penalty constant choices, and results in a negligible penalty gradient term during upper-level variable updates. We provide rigorous convergence analysis and validate the method's efficacy through hyperparameter optimization for support vector machines and fine-tuning of large language models.