π€ AI Summary
In real-world deployments, AI systems are vulnerable to environmental disturbances and out-of-distribution (OOD) sensor noise, leading to policy instability and safety risks. To address this, we propose a robust reinforcement learning framework grounded in βsurprisalβ awareness within world models: it dynamically detects and suppresses anomalous sensory inputs via multi- or single-representation rejection sampling, thereby enhancing world model stability under unknown perturbations. The method is architecture-agnostic, seamlessly integrating with state-of-the-art world models including DreamerV3 and Cosmos. Extensive evaluation in CARLA and Safety Gymnasium demonstrates that our approach significantly preserves policy performance across diverse noise types and intensities, validating both its effectiveness and generalizability. The implementation is publicly available.
π Abstract
AI systems deployed in the real world must contend with distractions and out-of-distribution (OOD) noise that can destabilize their policies and lead to unsafe behavior. While robust training can reduce sensitivity to some forms of noise, it is infeasible to anticipate all possible OOD conditions. To mitigate this issue, we develop an algorithm that leverages a world model's inherent measure of surprise to reduce the impact of noise in world model--based reinforcement learning agents. We introduce both multi-representation and single-representation rejection sampling, enabling robustness to settings with multiple faulty sensors or a single faulty sensor. While the introduction of noise typically degrades agent performance, we show that our techniques preserve performance relative to baselines under varying types and levels of noise across multiple environments within self-driving simulation domains (CARLA and Safety Gymnasium). Furthermore, we demonstrate that our methods enhance the stability of two state-of-the-art world models with markedly different underlying architectures: Cosmos and DreamerV3. Together, these results highlight the robustness of our approach across world modeling domains. We release our code at https://github.com/Bluefin-Tuna/WISER .