🤖 AI Summary
In offline reinforcement learning, distributional shift—particularly under data scarcity or a high proportion of out-of-distribution samples—induces harmful cross-covariance bias in temporal difference (TD) updates, which can mislead policy optimization. To address this issue, this work proposes Clustered Cross-Covariance Control (C⁴), a method that partitions the replay buffer via clustering to enable localized experience sampling and introduces an explicit gradient penalty term to correct the bias. C⁴ mitigates excessive conservatism while preserving the policy constraint framework and the lower-bound property of the objective function. Experimental results demonstrate that C⁴ achieves up to a 30% improvement in returns over existing methods on small-scale and highly out-of-distribution datasets, with notably enhanced training stability.
📝 Abstract
A fundamental challenge in offline reinforcement learning is distributional shift. Scarce data or datasets dominated by out-of-distribution (OOD) areas exacerbate this issue. Our theoretical analysis and experiments show that the standard squared error objective induces a harmful TD cross covariance. This effect amplifies in OOD areas, biasing optimization and degrading policy learning. To counteract this mechanism, we develop two complementary strategies: partitioned buffer sampling that restricts updates to localized replay partitions, attenuates irregular covariance effects, and aligns update directions, yielding a scheme that is easy to integrate with existing implementations, namely Clustered Cross-Covariance Control for TD (C^4). We also introduce an explicit gradient-based corrective penalty that cancels the covariance induced bias within each update. We prove that buffer partitioning preserves the lower bound property of the maximization objective, and that these constraints mitigate excessive conservatism in extreme OOD areas without altering the core behavior of policy constrained offline reinforcement learning. Empirically, our method showcases higher stability and up to 30% improvement in returns over prior methods, especially with small datasets and splits that emphasize OOD areas.