🤖 AI Summary
This work addresses the critical challenge of out-of-distribution (OOD) detection in continual learning (CL). We propose five principles for co-designing OOD detection and CL, and introduce CLOD—the first unified benchmark for continual OOD detection—encompassing multi-domain streaming OOD scenarios and a dedicated evaluation protocol quantifying the forgetting–detection trade-off. Methodologically, we integrate uncertainty-aware modeling, task-incremental feature disentanglement, energy-based OOD scoring, gradient-aware distribution shift monitoring, and a lightweight online calibration mechanism. Evaluated across 12 CL benchmarks, our approach reduces average false positive rate at 95% true positive rate (FPR95) by 37% while mitigating CL accuracy degradation by 2.1 percentage points. To our knowledge, this is the first method achieving joint optimization of OOD detection robustness and CL stability, substantially enhancing reliability and adaptability of AI systems in open, dynamic environments.
📝 Abstract
Recent years have witnessed significant progress in the development of machine learning models across a wide range of fields, fueled by increased computational resources, large-scale datasets, and the rise of deep learning architectures. From malware detection to enabling autonomous navigation, modern machine learning systems have demonstrated remarkable capabilities. However, as these models are deployed in ever-changing real-world scenarios, their ability to remain reliable and adaptive over time becomes increasingly important. For example, in the real world, new malware families are continuously developed, whereas autonomous driving cars are employed in many different cities and weather conditions. Models trained in fixed settings can not respond effectively to novel conditions encountered post-deployment. In fact, most machine learning models are still developed under the assumption that training and test data are independent and identically distributed (i.i.d.), i.e., sampled from the same underlying (unknown) distribution. While this assumption simplifies model development and evaluation, it does not hold in many real-world applications, where data changes over time and unexpected inputs frequently occur. Retraining models from scratch whenever new data appears is computationally expensive, time-consuming, and impractical in resource-constrained environments. These limitations underscore the need for Continual Learning (CL), which enables models to incrementally learn from evolving data streams without forgetting past knowledge, and Out-of-Distribution (OOD) detection, which allows systems to identify and respond to novel or anomalous inputs. Jointly addressing both challenges is critical to developing robust, efficient, and adaptive AI systems.