🤖 AI Summary
This work addresses the bias and higher-order error bounds of constant-step-size linear stochastic approximation (LSA) under Markovian noise. We show that Polyak–Ruppert (PR) averaging fails to eliminate the dominant $O(alpha)$ bias term. To resolve this, we introduce a novel analytical framework based on linearization decomposition and, for the first time, incorporate Richardson–Romberg (RR) extrapolation into LSA to systematically cancel the first-order bias. Through rigorous higher-order moment analysis, we derive tight moment bounds for the RR-iterated estimator and prove that its asymptotic variance matches the optimal covariance matrix of the original LSA, while its bias is reduced to $O(alpha^2)$. This constitutes the first higher-order bias correction for LSA under non-i.i.d. (Markovian) noise, significantly improving both convergence accuracy and statistical efficiency.
📝 Abstract
In this paper, we study the bias and high-order error bounds of the Linear Stochastic Approximation (LSA) algorithm with Polyak-Ruppert (PR) averaging under Markovian noise. We focus on the version of the algorithm with constant step size $α$ and propose a novel decomposition of the bias via a linearization technique. We analyze the structure of the bias and show that the leading-order term is linear in $α$ and cannot be eliminated by PR averaging. To address this, we apply the Richardson-Romberg (RR) extrapolation procedure, which effectively cancels the leading bias term. We derive high-order moment bounds for the RR iterates and show that the leading error term aligns with the asymptotically optimal covariance matrix of the vanilla averaged LSA iterates.