Achieving ${O}(epsilon^{-1.5})$ Complexity in Hessian/Jacobian-free Stochastic Bilevel Optimization

📅 2023-12-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
本文解决了无需二阶导数计算的随机双层优化问题,提出了一种名为FdeHBO的优化器,通过单循环结构和动量更新,实现了O(ε^{-1.5})的样本复杂度。
📝 Abstract
In this paper, we revisit the bilevel optimization problem, in which the upper-level objective function is generally nonconvex and the lower-level objective function is strongly convex. Although this type of problem has been studied extensively, it still remains an open question how to achieve an ${O}(epsilon^{-1.5})$ sample complexity in Hessian/Jacobian-free stochastic bilevel optimization without any second-order derivative computation. To fill this gap, we propose a novel Hessian/Jacobian-free bilevel optimizer named FdeHBO, which features a simple fully single-loop structure, a projection-aided finite-difference Hessian/Jacobian-vector approximation, and momentum-based updates. Theoretically, we show that FdeHBO requires ${O}(epsilon^{-1.5})$ iterations (each using ${O}(1)$ samples and only first-order gradient information) to find an $epsilon$-accurate stationary point. As far as we know, this is the first Hessian/Jacobian-free method with an ${O}(epsilon^{-1.5})$ sample complexity for nonconvex-strongly-convex stochastic bilevel optimization.
Problem

Research questions and friction points this paper is trying to address.

Achieving O(ε⁻¹·⁵) complexity in Hessian-free bilevel optimization
Solving nonconvex-strongly-convex stochastic bilevel optimization efficiently
Eliminating second-order derivative computation in bilevel optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Fully single-loop structure optimizer
Projection-aided finite-difference approximation
Momentum-based updates for efficiency