Centralized Reward Agent for Knowledge Sharing and Transfer in Multi-Task Reinforcement Learning

📅 2024-08-20
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
To address inefficient training and poor generalization caused by sparse rewards in multi-task reinforcement learning, this paper proposes the Centralized Reward Agent (CRA) framework. Methodologically, CRA integrates multi-agent reward shaping, distributed policy training, knowledge distillation, and reward signal propagation—supporting both discrete and continuous action spaces. Its core innovation lies in modeling shaped rewards as transferable knowledge carriers for the first time, enabling cross-task knowledge sharing and transfer via a centralized reward mechanism, thereby facilitating zero-shot and few-shot reward generalization. Evaluated on benchmarks including Meta-World, CRA significantly improves multi-task learning efficiency, generalization performance, and rapid adaptation to unseen tasks. Experimental results validate the effectiveness and robustness of treating rewards as a knowledge representation paradigm.

Technology Category

Application Category

📝 Abstract
Reward shaping is effective in addressing the sparse-reward challenge in reinforcement learning by providing immediate feedback through auxiliary informative rewards. Based on the reward shaping strategy, we propose a novel multi-task reinforcement learning framework that integrates a centralized reward agent (CRA) and multiple distributed policy agents. The CRA functions as a knowledge pool, which aims to distill knowledge from various tasks and distribute it to individual policy agents to improve learning efficiency. Specifically, the shaped rewards serve as a straightforward metric to encode knowledge. This framework not only enhances knowledge sharing across established tasks but also adapts to new tasks by transferring meaningful reward signals. We validate the proposed method on both discrete and continuous domains, including the representative meta world benchmark, demonstrating its robustness in multi-task sparse-reward settings and its effective transferability to unseen tasks.
Problem

Research questions and friction points this paper is trying to address.

Addresses sparse-reward challenge in multi-task reinforcement learning
Proposes centralized reward agent for knowledge sharing across tasks
Enhances learning efficiency and transferability to unseen tasks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Centralized reward agent for knowledge sharing
Distilled knowledge via shaped rewards encoding
Transferable reward signals for new tasks
🔎 Similar Papers
H
Haozhe Ma
National University of Singapore
Z
Zhengding Luo
Nanyang Technological University
T
Thanh Vinh Vo
National University of Singapore
K
Kuankuan Sima
National University of Singapore
Tze-Yun Leong
Tze-Yun Leong
National University of Singapore
Artificial intelligencebiomedical informatics