๐ค AI Summary
In safety-critical reinforcement learning under partial observability, existing methods suffer from inadequate risk identification and poor policy generalization due to insufficient exploitation of privileged information. Method: This paper proposes the ACPOMDP (Augmented Completely observable Privileged Markov Decision Process) theoretical frameworkโthe first to systematically integrate privileged information into partially observable Markov decision process modeling. We design a privileged representation alignment mechanism and an asymmetric actor-critic architecture within a world model framework, enabling efficient privileged-information-guided training and privilege-free safe inference. Contribution/Results: Our approach guarantees strict safety constraints while significantly improving task performance and training stability. It outperforms state-of-the-art safe RL and privileged learning baselines across multiple benchmark tasks, demonstrating superior convergence, generalization, and engineering practicality.
๐ Abstract
Partial observability presents a significant challenge for safe reinforcement learning, as it impedes the identification of potential risks and rewards. Leveraging specific types of privileged information during training to mitigate the effects of partial observability has yielded notable empirical successes. In this paper, we propose Asymmetric Constrained Partially Observable Markov Decision Processes (ACPOMDPs) to theoretically examine the advantages of incorporating privileged information. Building upon ACPOMDPs, we propose the Privileged Information Guided Dreamer, a model-based safe reinforcement learning approach that leverages privileged information to enhance the agent's safety and performance through privileged representation alignment and an asymmetric actor-critic structure. Our empirical results demonstrate that our approach significantly outperforms existing methods in terms of safety and task-centric performance. Meanwhile, compared to alternative privileged model-based reinforcement learning methods, our approach exhibits superior performance and ease of training.