🤖 AI Summary
This study addresses the challenge of probability calibration in competing risks survival analysis, where existing calibration methods are often inapplicable, leading to unreliable predicted probabilities. To tackle this issue, the authors develop a dedicated calibration framework tailored to the multi-event and time-to-event nature of competing risks. They introduce two novel calibration metrics grounded in proper scoring rules, along with corresponding algorithms for calibration assessment, statistical testing, and recalibration. The proposed approach significantly enhances the reliability of predicted probabilities while preserving the model’s original discriminative performance. This work thus provides both theoretical foundations and practical tools for generating trustworthy survival predictions in competing risks settings.
📝 Abstract
Survival analysis deals with modeling the time until an event occurs, and accurate probability estimates are crucial for decision-making, particularly in the competing-risks setting where multiple events are possible. While recent work has addressed calibration in standard survival analysis, the competing-risks setting remains under-explored as it is harder (the calibration applies to both probabilities across classes and time horizon). We show that existing calibration measures are not suited to the competing-risk setting and that recent models do not give well-behaved probabilities. To address this, we introduce a dedicated framework with two novel calibration measures that are minimized for oracle estimators (i.e., both measures are proper). We also introduce some methods to estimate, test, and correct the calibration. Our recalibration methods yield good probabilities while preserving discrimination.