Beyond Calibration: Assessing the Probabilistic Fit of Neural Regressors via Conditional Congruence

📅 2024-05-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Deep neural network regressors often suffer from pointwise miscalibration—i.e., a mismatch between the predicted and true conditional distributions—hindering reliable single-sample uncertainty quantification. Existing calibration metrics (e.g., Expected Calibration Error, ECE) assess only marginal probability alignment and lack input-level interpretability. To address this, we propose *conditional congruence*, a stricter probabilistic fidelity criterion requiring alignment of predicted and ground-truth conditional distributions *at each input*. We define the Conditional Congruence Error (CCE) as an interpretable, pointwise reliability diagnostic. Methodologically, we develop a nonparametric distance estimation framework based on conditional kernel mean embeddings, enabling scalable application to high-dimensional image regression. Experiments demonstrate that CCE precisely quantifies distributional miscalibration on synthetic data, scales effectively to real-world image regression tasks, and substantially improves discrimination of reliability for out-of-distribution samples.

Technology Category

Application Category

📝 Abstract
While significant progress has been made in specifying neural networks capable of representing uncertainty, deep networks still often suffer from overconfidence and misaligned predictive distributions. Existing approaches for addressing this misalignment are primarily developed under the framework of calibration, with common metrics such as Expected Calibration Error (ECE). However, calibration can only provide a strictly marginal assessment of probabilistic alignment. Consequently, calibration metrics such as ECE are distribution-wise measures and cannot diagnose the point-wise reliability of individual inputs, which is important for real-world decision-making. We propose a stronger condition, which we term conditional congruence, for assessing probabilistic fit. We also introduce a metric, Conditional Congruence Error (CCE), that uses conditional kernel mean embeddings to estimate the distance, at any point, between the learned predictive distribution and the empirical, conditional distribution in a dataset. We show that using CCE to measure congruence 1) accurately quantifies misalignment between distributions when the data generating process is known, 2) effectively scales to real-world, high dimensional image regression tasks, and 3) can be used to gauge model reliability on unseen instances.
Problem

Research questions and friction points this paper is trying to address.

Assessing probabilistic fit of neural networks via conditional congruence
Addressing overconfidence and misaligned predictive distributions in deep networks
Providing point-wise reliability assessment beyond marginal calibration metrics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Proposes conditional congruence for probabilistic fit assessment
Introduces CCE metric using kernel mean embeddings
Estimates distance between predictive and empirical distributions
S
Spencer Young
Delicious AI
P
P. Jenkins
Brigham Young University