Revisiting Reweighted Risk for Calibration: AURC, Focal Loss, and Inverse Focal Loss

📅 2025-05-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the theoretical relationship between weighted risk functions—including focal loss, inverse focal loss, and the Area-Under-the-Reliability-Curve (AURC)—and model calibration error. To bridge selective classification and calibration objectives, we propose a differentiable AURC optimization framework. We introduce SoftRank to enable end-to-end AURC minimization and support flexible confidence score function (CSF) modeling. Theoretical analysis reveals that inverse focal loss aligns more closely with calibration objectives than focal loss. Extensive experiments across multiple datasets and model architectures demonstrate that our method significantly improves class-wise calibration performance, achieving state-of-the-art (SOTA) results—outperforming standard cross-entropy and focal loss baselines.

Technology Category

Application Category

📝 Abstract
Several variants of reweighted risk functionals, such as focal losss, inverse focal loss, and the Area Under the Risk-Coverage Curve (AURC), have been proposed in the literature and claims have been made in relation to their calibration properties. However, focal loss and inverse focal loss propose vastly different weighting schemes. In this paper, we revisit a broad class of weighted risk functions commonly used in deep learning and establish a principled connection between these reweighting schemes and calibration errors. We show that minimizing calibration error is closely linked to the selective classification paradigm and demonstrate that optimizing a regularized variant of the AURC naturally leads to improved calibration. This regularized AURC shares a similar reweighting strategy with inverse focal loss, lending support to the idea that focal loss is less principled when calibration is a desired outcome. Direct AURC optimization offers greater flexibility through the choice of confidence score functions (CSFs). To enable gradient-based optimization, we introduce a differentiable formulation of the regularized AURC using the SoftRank technique. Empirical evaluations demonstrate that our AURC-based loss achieves competitive class-wise calibration performance across a range of datasets and model architectures.
Problem

Research questions and friction points this paper is trying to address.

Analyzing reweighted risk functions for model calibration
Linking calibration errors to weighted risk schemes
Optimizing regularized AURC for improved calibration performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Regularized AURC improves calibration error
Differentiable AURC via SoftRank technique
AURC shares strategy with inverse focal loss
🔎 Similar Papers
No similar papers found.
H
Han Zhou
ESAT-PSI, KU Leuven, Belgium
S
Sebastian G.Gruber
ESAT-PSI, KU Leuven, Belgium
Teodora Popordanoska
Teodora Popordanoska
Postdoc, KU Leuven
Deep LearningMachine LearningComputer Vision
M
M. Blaschko
ESAT-PSI, KU Leuven, Belgium