🤖 AI Summary
This study investigates the intrinsic mechanisms underlying “emergent exploration” in unsupervised reinforcement learning (RL), focusing on Single-Goal Contrastive RL (SGCRL)—a reward-free paradigm. We reveal how SGCRL implicitly shapes exploration-driven reward signals via self-supervised representation learning. Through theoretical analysis and controlled ablation experiments, we demonstrate that SGCRL autonomously learns low-rank state representations, dynamically reshapes the reward landscape, enables adaptive switching between exploration and exploitation, and inherently supports safety-aware policies. Crucially, unlike conventional function-approximation–based explanations, our work provides the first geometric characterization of exploration dynamics rooted in representation structure. Empirically, SGCRL significantly improves exploration efficiency in long-horizon, reward-free tasks and generalizes effectively to safety-constrained environments. Our findings establish a novel, interpretable, and scalable mechanistic framework for unsupervised RL grounded in representation geometry.
📝 Abstract
In this work, we take a first step toward elucidating the mechanisms behind emergent exploration in unsupervised reinforcement learning. We study Single-Goal Contrastive Reinforcement Learning (SGCRL), a self-supervised algorithm capable of solving challenging long-horizon goal-reaching tasks without external rewards or curricula. We combine theoretical analysis of the algorithm's objective function with controlled experiments to understand what drives its exploration. We show that SGCRL maximizes implicit rewards shaped by its learned representations. These representations automatically modify the reward landscape to promote exploration before reaching the goal and exploitation thereafter. Our experiments also demonstrate that these exploration dynamics arise from learning low-rank representations of the state space rather than from neural network function approximation. Our improved understanding enables us to adapt SGCRL to perform safety-aware exploration.