Out-of-Distribution Detection Methods Answer the Wrong Questions

📅 2025-07-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current mainstream out-of-distribution (OOD) detection methods suffer from a fundamental objective misalignment: supervised models trained exclusively on in-distribution (ID) data erroneously equate high predictive uncertainty or large feature-space distances with OOD samples, leading to irreducible detection errors. Method: The authors systematically analyze major paradigms—including uncertainty estimation, feature-logit hybrid models, density estimation, and generative approaches—and rigorously prove their theoretical limitations under common distribution shifts. Contribution/Results: They introduce the “objective misalignment” framework, establishing that OOD detection is inherently a distribution discrimination task—not an uncertainty or distance regression problem. Crucially, they demonstrate that prevalent mitigation strategies—such as anomaly exposure, model architecture expansion, or cognitive uncertainty modeling—cannot rectify this foundational flaw. This work provides critical theoretical grounding for paradigmatic reformulation of OOD detection.

Technology Category

Application Category

📝 Abstract
To detect distribution shifts and improve model safety, many out-of-distribution (OOD) detection methods rely on the predictive uncertainty or features of supervised models trained on in-distribution data. In this paper, we critically re-examine this popular family of OOD detection procedures, and we argue that these methods are fundamentally answering the wrong questions for OOD detection. There is no simple fix to this misalignment, since a classifier trained only on in-distribution classes cannot be expected to identify OOD points; for instance, a cat-dog classifier may confidently misclassify an airplane if it contains features that distinguish cats from dogs, despite generally appearing nothing alike. We find that uncertainty-based methods incorrectly conflate high uncertainty with being OOD, while feature-based methods incorrectly conflate far feature-space distance with being OOD. We show how these pathologies manifest as irreducible errors in OOD detection and identify common settings where these methods are ineffective. Additionally, interventions to improve OOD detection such as feature-logit hybrid methods, scaling of model and data size, epistemic uncertainty representation, and outlier exposure also fail to address this fundamental misalignment in objectives. We additionally consider unsupervised density estimation and generative models for OOD detection, which we show have their own fundamental limitations.
Problem

Research questions and friction points this paper is trying to address.

Current OOD detection methods misidentify distribution shifts
Classifier trained on in-distribution data fails to detect OOD
Uncertainty and feature-based methods conflate wrong OOD indicators
Innovation

Methods, ideas, or system contributions that make the work stand out.

Critiques predictive uncertainty for OOD detection
Highlights flaws in feature-space distance methods
Examines limitations of unsupervised density models
🔎 Similar Papers
No similar papers found.