Towards Parameter-Free Temporal Difference Learning

๐Ÿ“… 2026-03-02
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the gap between theory and practice in temporal difference (TD) learning, where traditional theoretical analyses rely on hard-to-estimate problem-specific parametersโ€”such as the minimum eigenvalue of the feature covariance matrix or the mixing time of the Markov chain. To overcome this limitation, the paper proposes a parameter-free TD(0) algorithm with exponentially decaying step sizes that requires no projection, iterate averaging, or prior knowledge of the environment. The method is proven to converge under both i.i.d. and Markov sampling settings without depending on any instance-specific constants. Notably, it achieves convergence rates comparable to existing approaches while guaranteeing an optimal bias-variance trade-off for the final iterate in the i.i.d. setting, thereby significantly narrowing the divide between theoretical guarantees and practical performance.

Technology Category

Application Category

๐Ÿ“ Abstract
Temporal difference (TD) learning is a fundamental algorithm for estimating value functions in reinforcement learning. Recent finite-time analyses of TD with linear function approximation quantify its theoretical convergence rate. However, they often require setting the algorithm parameters using problem-dependent quantities that are difficult to estimate in practice -- such as the minimum eigenvalue of the feature covariance (\(ฯ‰\)) or the mixing time of the underlying Markov chain (\(ฯ„_{\text{mix}}\)). In addition, some analyses rely on nonstandard and impractical modifications, exacerbating the gap between theory and practice. To address these limitations, we use an exponential step-size schedule with the standard TD(0) algorithm. We analyze the resulting method under two sampling regimes: independent and identically distributed (i.i.d.) sampling from the stationary distribution, and the more practical Markovian sampling along a single trajectory. In the i.i.d.\ setting, the proposed algorithm does not require knowledge of problem-dependent quantities such as \(ฯ‰\), and attains the optimal bias-variance trade-off for the last iterate. In the Markovian setting, we propose a regularized TD(0) algorithm with an exponential step-size schedule. The resulting algorithm achieves a comparable convergence rate to prior works, without requiring projections, iterate averaging, or knowledge of \(ฯ„_{\text{mix}}\) or \(ฯ‰\).
Problem

Research questions and friction points this paper is trying to address.

Temporal Difference Learning
Parameter-Free
Linear Function Approximation
Markovian Sampling
Convergence Analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

parameter-free
temporal difference learning
exponential step-size
Markovian sampling
linear function approximation
๐Ÿ”Ž Similar Papers
No similar papers found.