🤖 AI Summary
We address online policy optimization for infinite-horizon average-reward Markov decision processes (MDPs). We propose two novel policy gradient algorithms: one employs implicit gradient transport (IGT) for variance reduction, and the other incorporates Hessian-assisted updates to accelerate convergence. Theoretically, we establish, for the first time without assuming $L$-smoothness of the policy parameterization, that the average-reward objective satisfies an approximate $L$-smoothness property—circumventing a key limitation of prior analyses. This enables derivation of two expected regret bounds: $ ilde{mathcal{O}}(T^{2/3})$ and $ ilde{mathcal{O}}(sqrt{T})$, with the latter matching the information-theoretic lower bound and strictly improving upon the previous best-known rate of $ ilde{mathcal{O}}(T^{3/4})$. Empirical evaluations confirm the efficacy of our methods. Our work provides a more rigorous theoretical foundation for policy gradient methods in average-reward MDPs.
📝 Abstract
We present two Policy Gradient-based algorithms with general parametrization in the context of infinite-horizon average reward Markov Decision Process (MDP). The first one employs Implicit Gradient Transport for variance reduction, ensuring an expected regret of the order $ ilde{mathcal{O}}(T^{2/3})$. The second approach, rooted in Hessian-based techniques, ensures an expected regret of the order $ ilde{mathcal{O}}(sqrt{T})$. These results significantly improve the state-of-the-art $ ilde{mathcal{O}}(T^{3/4})$ regret and achieve the theoretical lower bound. We also show that the average-reward function is approximately $L$-smooth, a result that was previously assumed in earlier works.