🤖 AI Summary
This work investigates the convergence rate of nonlinear two-timescale stochastic approximation under state- and time-dependent (nonstationary) noise. Addressing the limitation of existing theory in handling such nonstationary noise, we propose a Lyapunov-based analytical framework that integrates the two-timescale update mechanism with Polyak–Ruppert averaging. We establish, for the first time, a polynomial upper bound on the convergence rate. Moreover, we uncover a novel phenomenon: as the noise parameters approach stationarity, the algorithm achieves exponential convergence—a property previously unobserved. Our theoretical results uniformly apply to canonical settings including stochastic gradient descent and stochastic bilevel optimization. Numerical experiments corroborate both the tightness of the derived convergence order and the practical efficacy of the proposed framework.
📝 Abstract
The nonlinear two-time-scale stochastic approximation is widely studied under conditions of bounded variances in noise. Motivated by recent advances that allow for variability linked to the current state or time, we consider state- and time-dependent noises. We show that the Lyapunov function exhibits polynomial convergence rates in both cases, with the rate of polynomial delay depending on the parameters of state- or time-dependent noises. Notably, if the state noise parameters fully approach their limiting value, the Lyapunov function achieves an exponential convergence rate. We provide two numerical examples to illustrate our theoretical findings in the context of stochastic gradient descent with Polyak-Ruppert averaging and stochastic bilevel optimization.