Surrogate-Assisted Evolutionary Reinforcement Learning Based on Autoencoder and Hyperbolic Neural Network

📅 2025-05-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the high policy evaluation cost and low search efficiency in Evolutionary Reinforcement Learning (ERL), this paper proposes a surrogate-assisted framework integrating autoencoders and hyperbolic neural networks. Our method introduces the first learnable low-dimensional embedding for ERL policies and constructs a classification-based surrogate model to jointly enable efficient pre-screening and quality assessment of high-dimensional deep neural network policies. By synergistically combining the autoencoder’s nonlinear dimensionality reduction capability with hyperbolic space’s intrinsic capacity to model hierarchical policy structures, our approach significantly improves exploration quality and convergence speed. Evaluated on ten Atari and four MuJoCo benchmark tasks, it consistently outperforms state-of-the-art ERL baselines. Visualization analysis further confirms superior search trajectories, more thorough exploration, and faster convergence.

Technology Category

Application Category

📝 Abstract
Evolutionary Reinforcement Learning (ERL), training the Reinforcement Learning (RL) policies with Evolutionary Algorithms (EAs), have demonstrated enhanced exploration capabilities and greater robustness than using traditional policy gradient. However, ERL suffers from the high computational costs and low search efficiency, as EAs require evaluating numerous candidate policies with expensive simulations, many of which are ineffective and do not contribute meaningfully to the training. One intuitive way to reduce the ineffective evaluations is to adopt the surrogates. Unfortunately, existing ERL policies are often modeled as deep neural networks (DNNs) and thus naturally represented as high-dimensional vectors containing millions of weights, which makes the building of effective surrogates for ERL policies extremely challenging. This paper proposes a novel surrogate-assisted ERL that integrates Autoencoders (AE) and Hyperbolic Neural Networks (HNN). Specifically, AE compresses high-dimensional policies into low-dimensional representations while extracting key features as the inputs for the surrogate. HNN, functioning as a classification-based surrogate model, can learn complex nonlinear relationships from sampled data and enable more accurate pre-selection of the sampled policies without real evaluations. The experiments on 10 Atari and 4 Mujoco games have verified that the proposed method outperforms previous approaches significantly. The search trajectories guided by AE and HNN are also visually demonstrated to be more effective, in terms of both exploration and convergence. This paper not only presents the first learnable policy embedding and surrogate-modeling modules for high-dimensional ERL policies, but also empirically reveals when and why they can be successful.
Problem

Research questions and friction points this paper is trying to address.

Reducing computational costs in Evolutionary Reinforcement Learning
Improving search efficiency for high-dimensional policy evaluation
Enhancing policy selection accuracy using surrogate models
Innovation

Methods, ideas, or system contributions that make the work stand out.

Autoencoder compresses high-dimensional policies
Hyperbolic Neural Network classifies policies effectively
Combined method reduces computational costs significantly
🔎 Similar Papers
No similar papers found.
Bingdong Li
Bingdong Li
East China Normal University
evolutionary computationmachine learningblack-box optimization
M
Mei Jiang
Senior Member, IEEE
H
Hong Qian
P
Peng Yang
Senior Member, IEEE
Wenjing Hong
Wenjing Hong
Shenzhen University
K
Ke Tang
Fellow, IEEE