🤖 AI Summary
Open-world semantic segmentation confronts the dual challenges of unknown-class (out-of-distribution, OOD) detection and pixel-wise segmentation. Existing entropy- or confidence-threshold-based approaches exhibit limited discriminative capability for OOD identification. This paper introduces the first integration of the Wasserstein distance into an evidential deep learning framework, leveraging the geometric structure of the probability simplex to model OOD uncertainty. We propose a novel loss comprising Wasserstein distance regularization, Dempster–Shafer evidence fusion, KL divergence–based epistemic regularization, and Dice-based structural consistency constraints. Our method preserves in-distribution (ID) segmentation accuracy while significantly improving OOD detection: it achieves an AUROC gain of over 8% compared to state-of-the-art uncertainty quantification methods and sets new records across multiple benchmarks. The core contribution is a principled, geometrically interpretable, and statistically robust uncertainty modeling paradigm for open-world segmentation.
📝 Abstract
Deep neural networks achieve superior performance in semantic segmentation, but are limited to a predefined set of classes, which leads to failures when they encounter unknown objects in open-world scenarios. Recognizing and segmenting these out-of-distribution (OOD) objects is crucial for safety-critical applications such as automated driving. In this work, we present an evidence segmentation framework using a Wasserstein loss, which captures distributional distances while respecting the probability simplex geometry. Combined with Kullback-Leibler regularization and Dice structural consistency terms, our approach leads to improved OOD segmentation performance compared to uncertainty-based approaches.