🤖 AI Summary
Medical machine learning often employs frequency-based weighting, causing models to prioritize common conditions while neglecting rare yet clinically critical cases—termed the “average-patient fallacy”—contravening precision medicine principles. To address this, we propose the Clinical Weighting Objective (CWO), a clinical utility-driven framework that redefines case rarity and implements an ethics-aware gradient reweighting mechanism to mitigate gradient suppression of rare cases in ensemble models. Our methodology includes quantification of rare-case performance gaps, calibration error analysis, validation via clinical vignettes, and structured weight design. Evaluated across multi-center oncology, cardiology, and ophthalmology datasets, CWO significantly improves detection of rare treatment responders, atypical acute presentations, and vision-threatening variants—reducing false-negative rates and enhancing both clinical applicability and algorithmic fairness.
📝 Abstract
Machine learning in medicine is typically optimized for population averages. This frequency weighted training privileges common presentations and marginalizes rare yet clinically critical cases, a bias we call the average patient fallacy. In mixture models, gradients from rare cases are suppressed by prevalence, creating a direct conflict with precision medicine. Clinical vignettes in oncology, cardiology, and ophthalmology show how this yields missed rare responders, delayed recognition of atypical emergencies, and underperformance on vision-threatening variants. We propose operational fixes: Rare Case Performance Gap, Rare Case Calibration Error, a prevalence utility definition of rarity, and clinically weighted objectives that surface ethical priorities. Weight selection should follow structured deliberation. AI in medicine must detect exceptional cases because of their significance.