Efficient Reinforcement Learning by Reducing Forgetting with Elephant Activation Functions

📅 2025-09-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In reinforcement learning, catastrophic forgetting arises from training dynamics imbalance in neural networks. This work identifies gradient sparsity as a key mechanism for mitigating forgetting and proposes the Elephant Activation Function (EAF), a novel activation function that jointly enforces output sparsity and gradient sparsity to significantly enhance representation stability. Within value-based RL frameworks, replacing standard activations with EAF—without architectural or algorithmic modifications—substantially suppresses forgetting across multitask benchmarks including Atari and Continual Control: average forgetting rate decreases by 32.7%, sample efficiency improves by 1.8×, and memory overhead remains unchanged. To our knowledge, this is the first work to formalize gradient sparsity as a core design principle for anti-forgetting, establishing an interpretable, lightweight, and deployable paradigm for scalable and continual neural network learning.

Technology Category

Application Category

📝 Abstract
Catastrophic forgetting has remained a significant challenge for efficient reinforcement learning for decades (Ring 1994, Rivest and Precup 2003). While recent works have proposed effective methods to mitigate this issue, they mainly focus on the algorithmic side. Meanwhile, we do not fully understand what architectural properties of neural networks lead to catastrophic forgetting. This study aims to fill this gap by studying the role of activation functions in the training dynamics of neural networks and their impact on catastrophic forgetting in reinforcement learning setup. Our study reveals that, besides sparse representations, the gradient sparsity of activation functions also plays an important role in reducing forgetting. Based on this insight, we propose a new class of activation functions, elephant activation functions, that can generate both sparse outputs and sparse gradients. We show that by simply replacing classical activation functions with elephant activation functions in the neural networks of value-based algorithms, we can significantly improve the resilience of neural networks to catastrophic forgetting, thus making reinforcement learning more sample-efficient and memory-efficient.
Problem

Research questions and friction points this paper is trying to address.

Understanding neural network architectural properties causing catastrophic forgetting
Studying activation functions' role in reinforcement learning training dynamics
Developing activation functions to reduce forgetting and improve efficiency
Innovation

Methods, ideas, or system contributions that make the work stand out.

Elephant activation functions reduce forgetting
Sparse outputs and gradients improve resilience
Replacing classical activations increases sample efficiency
🔎 Similar Papers
No similar papers found.
Qingfeng Lan
Qingfeng Lan
PhD student @ University of Alberta
Reinforcement LearningLarge Language ModelContinual LearningMeta Learning
G
G. Vasan
Department of Computing Science, University of Alberta
A
A. R. Mahmood
Department of Computing Science, University of Alberta, CIFAR AI Chair, Amii