🤖 AI Summary
Deep neural networks suffer from degraded reliability under distributional shift in real-world settings; disagreement discrepancy serves as a key robustness metric, yet its 0–1 loss is non-differentiable, necessitating surrogate losses. Method: We theoretically establish, for the first time, that existing surrogates are Bayes-inconsistent—maximizing them does not guarantee maximization of the true disagreement discrepancy. To address this, we propose a novel Bayes-consistent surrogate loss, derived from tight upper and lower bounds of disagreement discrepancy and instantiated via cross-entropy, yielding a differentiable, end-to-end optimizable objective. Contribution/Results: Extensive experiments across multiple benchmark datasets and adversarial distribution shift scenarios demonstrate that our method significantly improves both the accuracy and robustness of disagreement discrepancy estimation, thereby providing a more principled theoretical foundation and practical tool for out-of-distribution generalization.
📝 Abstract
Deep neural networks often fail when deployed in real-world contexts due to distribution shift, a critical barrier to building safe and reliable systems. An emerging approach to address this problem relies on emph{disagreement discrepancy} -- a measure of how the disagreement between two models changes under a shifting distribution. The process of maximizing this measure has seen applications in bounding error under shifts, testing for harmful shifts, and training more robust models. However, this optimization involves the non-differentiable zero-one loss, necessitating the use of practical surrogate losses. We prove that existing surrogates for disagreement discrepancy are not Bayes consistent, revealing a fundamental flaw: maximizing these surrogates can fail to maximize the true disagreement discrepancy. To address this, we introduce new theoretical results providing both upper and lower bounds on the optimality gap for such surrogates. Guided by this theory, we propose a novel disagreement loss that, when paired with cross-entropy, yields a provably consistent surrogate for disagreement discrepancy. Empirical evaluations across diverse benchmarks demonstrate that our method provides more accurate and robust estimates of disagreement discrepancy than existing approaches, particularly under challenging adversarial conditions.