🤖 AI Summary
Offline reinforcement learning suffers from degraded policy performance due to insufficient target-domain samples.
Method: We propose the first theoretically grounded cross-domain adaptation framework leveraging abundant source-domain data (e.g., simulations) to enhance learning. We formally characterize how source/target data weighting affects policy performance, derive a closed-form optimal weighting solution and a generalization error bound, and prove convergence to a neighborhood of the optimal policy. Our approach integrates bias-variance trade-off analysis, weighted empirical risk minimization, and conservative policy constraints.
Results: On the Procgen benchmark under low-data regimes, our method achieves an average 23.6% improvement over baselines, empirically validating both the tightness of our theoretical bound and the efficacy of the learned weighting strategy.
📝 Abstract
Offline reinforcement learning (RL) learns effective policies from a static target dataset. The performance of state-of-the-art offline RL algorithms notwithstanding, it relies on the quality and size of the target dataset and it degrades if limited samples in the target dataset are available, which is often the case in real-world applications. To address this issue, domain adaptation that leverages auxiliary samples from related source datasets (such as simulators) can be beneficial. However, establishing the optimal way to trade off the source and target datasets while ensuring provably theoretical guarantees remains an open challenge. To the best of our knowledge, this paper proposes the first framework that theoretically explores the impact of the weights assigned to each dataset on the performance of offline RL. In particular, we establish performance bounds and the existence of an optimal weight, which can be computed in closed form under simplifying assumptions. We also provide algorithmic guarantees in terms of convergence to a neighborhood of the optimum. Notably, these results depend on the quality of the source dataset and the number of samples from the target dataset. Our empirical results on the well-known Procgen benchmark substantiate our theoretical contributions.