Beyond validation loss: Clinically-tailored optimization metrics improve a model's clinical performance

📅 2026-01-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the mismatch between conventional machine learning practices and the specific performance requirements of clinical tasks in healthcare settings. Traditional approaches rely on differentiable validation losses for model optimization, which often fail to align with clinically meaningful outcomes. To bridge this gap, the paper proposes replacing standard loss functions with non-differentiable yet clinically interpretable custom metrics to guide critical optimization decisions—such as hyperparameter selection and training termination—thereby redefining the model validation pipeline. In two controlled experiments, models optimized using this framework demonstrated significantly superior performance on key clinical tasks compared to those guided by conventional differentiable validation losses. This approach overcomes the inherent limitation of relying solely on differentiable objectives and better aligns medical AI development with real-world clinical goals.

Technology Category

Application Category

📝 Abstract
A key task in ML is to optimize models at various stages, e.g. by choosing hyperparameters or picking a stopping point. A traditional ML approach is to use validation loss, i.e. to apply the training loss function on a validation set to guide these optimizations. However, ML for healthcare has a distinct goal from traditional ML: Models must perform well relative to specific clinical requirements, vs. relative to the loss function used for training. These clinical requirements can be captured more precisely by tailored metrics. Since many optimization tasks do not require the driving metric to be differentiable, they allow a wider range of options, including the use of metrics tailored to be clinically-relevant. In this paper we describe two controlled experiments which show how the use of clinically-tailored metrics provide superior model optimization compared to validation loss, in the sense of better performance on the clinical task. The use of clinically-relevant metrics for optimization entails some extra effort, to define the metrics and to code them into the pipeline. But it can yield models that better meet the central goal of ML for healthcare: strong performance in the clinic.
Problem

Research questions and friction points this paper is trying to address.

clinical performance
optimization metrics
validation loss
machine learning for healthcare
clinically-tailored metrics
Innovation

Methods, ideas, or system contributions that make the work stand out.

clinically-tailored metrics
validation loss
model optimization
clinical performance
healthcare AI
🔎 Similar Papers
No similar papers found.
C
Charles B. Delahunt
1 formerly Global Health Labs, Bellevue, WA.; 2 University of Washington, Seattle, WA.
C
C. Mehanian
1 University of Oregon, Eugene, OR.
Daniel E. Shea
Daniel E. Shea
Global Health Labs
machine learningdeep learningmedicinehealthengineering
M
M. Horning
1 formerly Global Health Labs, Bellevue, WA.