A Provably Convergent Plug-and-Play Framework for Stochastic Bilevel Optimization

📅 2025-05-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the long-standing open problem of convergence and sample complexity in stochastic bilevel optimization: Can the optimal $O(1/varepsilon^2)$ sample complexity—matching that of single-level stochastic optimization—be achieved? To this end, we propose PnPBO, a plug-and-play single-loop framework that flexibly incorporates diverse stochastic gradient estimators (e.g., PAGE, ZeroSARAH) and introduces upper-level variable moving averaging for enhanced stability. For the first time, we establish a unified convergence analysis accommodating both biased and unbiased gradient estimators, rigorously proving convergence to an $varepsilon$-stationary point with optimal $O(1/varepsilon^2)$ sample complexity. Empirical evaluations on hyperparameter optimization and data distillation demonstrate that PnPBO significantly outperforms existing methods, thereby resolving the fundamental question of whether bilevel optimization can achieve sample efficiency on par with its single-level counterpart.

Technology Category

Application Category

📝 Abstract
Bilevel optimization has recently attracted significant attention in machine learning due to its wide range of applications and advanced hierarchical optimization capabilities. In this paper, we propose a plug-and-play framework, named PnPBO, for developing and analyzing stochastic bilevel optimization methods. This framework integrates both modern unbiased and biased stochastic estimators into the single-loop bilevel optimization framework introduced in [9], with several improvements. In the implementation of PnPBO, all stochastic estimators for different variables can be independently incorporated, and an additional moving average technique is applied when using an unbiased estimator for the upper-level variable. In the theoretical analysis, we provide a unified convergence and complexity analysis for PnPBO, demonstrating that the adaptation of various stochastic estimators (including PAGE, ZeroSARAH, and mixed strategies) within the PnPBO framework achieves optimal sample complexity, comparable to that of single-level optimization. This resolves the open question of whether the optimal complexity bounds for solving bilevel optimization are identical to those for single-level optimization. Finally, we empirically validate our framework, demonstrating its effectiveness on several benchmark problems and confirming our theoretical findings.
Problem

Research questions and friction points this paper is trying to address.

Develops a plug-and-play framework for stochastic bilevel optimization
Integrates unbiased and biased estimators for optimal sample complexity
Resolves if bilevel optimization complexity matches single-level optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Plug-and-play framework integrates unbiased and biased estimators
Moving average technique enhances upper-level variable estimation
Unified convergence analysis achieves optimal sample complexity
🔎 Similar Papers
Tianshu Chu
Tianshu Chu
Stanford University
reinforcement learningintelligent transportation system
D
Dachuan Xu
Institute of Operations Research and Information Engineering, Beijing University of Technology
W
Wei Yao
National Center for Applied Mathematics Shenzhen, and Department of Mathematics, Southern University of Science and Technology
Chengming Yu
Chengming Yu
School of Science, Beijing University of Posts and Telecommunications; National Center for Applied Mathematics Shenzhen
J
Jin Zhang
Department of Mathematics, and National Center for Applied Mathematics Shenzhen, Southern University of Science and Technology