🤖 AI Summary
To address the low sample efficiency of Actor-Critic agents in deep reinforcement learning, this paper proposes Simplex Embedding—a novel architectural component that constrains both policy and value function representations to the probability simplex geometry. By introducing sparse, discretized features, it injects a strong geometric inductive bias into representation learning. This lightweight, plug-and-play embedding layer integrates seamlessly into mainstream algorithms—including FastTD3, FastSAC, and PPO—without increasing training overhead. It significantly improves gradient quality and value function stability. Empirical evaluation across continuous control (MuJoCo) and discrete control (Atari) benchmarks demonstrates substantial reductions in environment interactions required to reach target performance, alongside improved final policy performance. These results validate the effectiveness of simplex-geometric priors in enhancing representation learning and policy optimization.
📝 Abstract
Recent works have proposed accelerating the wall-clock training time of actor-critic methods via the use of large-scale environment parallelization; unfortunately, these can sometimes still require large number of environment interactions to achieve a desired level of performance. Noting that well-structured representations can improve the generalization and sample efficiency of deep reinforcement learning (RL) agents, we propose the use of simplicial embeddings: lightweight representation layers that constrain embeddings to simplicial structures. This geometric inductive bias results in sparse and discrete features that stabilize critic bootstrapping and strengthen policy gradients. When applied to FastTD3, FastSAC, and PPO, simplicial embeddings consistently improve sample efficiency and final performance across a variety of continuous- and discrete-control environments, without any loss in runtime speed.