A Survey of Continual Reinforcement Learning

📅 2025-06-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Reinforcement learning (RL) suffers from high data and computational costs, poor cross-task generalization, and catastrophic forgetting—key bottlenecks hindering its continual deployment in dynamic real-world environments. To address these challenges, this work systematically surveys continual reinforcement learning (CRL), proposing the first taxonomy grounded in a “knowledge storage–transfer” perspective, categorizing CRL methods into four principled classes. It unifies evaluation metrics, benchmark tasks, and scenario constraints. Through bibliometric analysis, methodological abstraction, and cross-benchmark comparative study, the survey covers dominant paradigms—including experience replay, regularization, parameter isolation, and meta-learning. Furthermore, it constructs the first structured, holistic CRL landscape, clarifying community consensus and open debates while identifying domain-specific challenges and evolutionary trajectories. The findings provide a rigorous theoretical framework and practical guidelines for algorithm design, standardized evaluation, and real-world CRL deployment.

Technology Category

Application Category

📝 Abstract
Reinforcement Learning (RL) is an important machine learning paradigm for solving sequential decision-making problems. Recent years have witnessed remarkable progress in this field due to the rapid development of deep neural networks. However, the success of RL currently relies on extensive training data and computational resources. In addition, RL's limited ability to generalize across tasks restricts its applicability in dynamic and real-world environments. With the arisen of Continual Learning (CL), Continual Reinforcement Learning (CRL) has emerged as a promising research direction to address these limitations by enabling agents to learn continuously, adapt to new tasks, and retain previously acquired knowledge. In this survey, we provide a comprehensive examination of CRL, focusing on its core concepts, challenges, and methodologies. Firstly, we conduct a detailed review of existing works, organizing and analyzing their metrics, tasks, benchmarks, and scenario settings. Secondly, we propose a new taxonomy of CRL methods, categorizing them into four types from the perspective of knowledge storage and/or transfer. Finally, our analysis highlights the unique challenges of CRL and provides practical insights into future directions.
Problem

Research questions and friction points this paper is trying to address.

Addresses RL's reliance on extensive training data
Improves RL's generalization across diverse tasks
Enables continuous learning and knowledge retention
Innovation

Methods, ideas, or system contributions that make the work stand out.

Survey of continual reinforcement learning methods
Taxonomy of CRL knowledge storage techniques
Analysis of CRL challenges and future directions
🔎 Similar Papers
No similar papers found.
C
Chaofan Pan
School of Computing and Artificial Intelligence, Southwestern University of Finance and Economics, Chengdu, 611130, China
X
Xin Yang
School of Computing and Artificial Intelligence, Southwestern University of Finance and Economics, Chengdu, 611130, China
Yanhua Li
Yanhua Li
Associate Professor at Worcester Polytechnic Institute
Spatial-Temporal Data ScienceArtificial IntelligenceSmart Cities
W
Wei Wei
Key Laboratory of Computational Intelligence and Chinese Information Processing of Ministry of Education, School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, 030006, China
Tianrui Li
Tianrui Li
School of Computing and Artificial Intelligence, Southwest Jiaotong University
Big Data IntelligenceUrban ComputingGranular Computing
B
Bo An
College of Data Science and Computing, Nanyang Technological University, 639798, Singapore
Jiye Liang
Jiye Liang
Shanxi University