π€ AI Summary
To address the low sample efficiency of goal-conditioned reinforcement learning (GCRL) under sparse rewards, this paper proposes Hindsight Goal-conditioned Regularization (HGR), a novel hindsight experience utilization mechanism integrating goal generation and action regularization. Unlike conventional Hindsight Experience Replay (HER), which only relabels achieved goals, HGR generates semantically plausible goals and enforces action consistency constraints to improve the reasonableness and generalizability of goal-action pairs. Additionally, it introduces hindsight self-imitation regularization to enhance policy reuse of successful trajectories. By embedding these components within the HER framework, HGR achieves more efficient experience replay and policy optimization. Experimental results demonstrate that HGR significantly improves sample efficiency across diverse navigation and manipulation tasks, consistently outperforming state-of-the-art GCRL baselines.
π Abstract
Goal-conditioned reinforcement learning (GCRL) with sparse rewards remains a fundamental challenge in reinforcement learning. While hindsight experience replay (HER) has shown promise by relabeling collected trajectories with achieved goals, we argue that trajectory relabeling alone does not fully exploit the available experiences in off-policy GCRL methods, resulting in limited sample efficiency. In this paper, we propose Hindsight Goal-conditioned Regularization (HGR), a technique that generates action regularization priors based on hindsight goals. When combined with hindsight self-imitation regularization (HSR), our approach enables off-policy RL algorithms to maximize experience utilization. Compared to existing GCRL methods that employ HER and self-imitation techniques, our hindsight regularizations achieve substantially more efficient sample reuse and the best performances, which we empirically demonstrate on a suite of navigation and manipulation tasks.