🤖 AI Summary
Medical imaging AI models often exhibit bias with respect to demographic attributes (e.g., sex, race), undermining cross-population generalizability and clinical safety. To address this, we propose the first statistical evaluation framework grounded in counterfactual invariance—requiring no ground-truth counterfactual data—to quantify implicit model dependence on sensitive attributes. Our method innovatively integrates conditional latent diffusion models to generate controllable counterfactual samples, coupled with hypothesis testing for interpretable, statistically rigorous bias quantification. We validate the framework on multi-source chest X-ray datasets (CheXpert and MIMIC-CXR), demonstrating significant improvements over existing baselines in both bias detection and mitigation, while enhancing model robustness. The implementation is publicly available.
📝 Abstract
Machine learning (ML) systems for medical imaging have demonstrated remarkable diagnostic capabilities, but their susceptibility to biases poses significant risks, since biases may negatively impact generalization performance. In this paper, we introduce a novel statistical framework to evaluate the dependency of medical imaging ML models on sensitive attributes, such as demographics. Our method leverages the concept of counterfactual invariance, measuring the extent to which a model's predictions remain unchanged under hypothetical changes to sensitive attributes. We present a practical algorithm that combines conditional latent diffusion models with statistical hypothesis testing to identify and quantify such biases without requiring direct access to counterfactual data. Through experiments on synthetic datasets and large-scale real-world medical imaging datasets, including extsc{cheXpert} and MIMIC-CXR, we demonstrate that our approach aligns closely with counterfactual fairness principles and outperforms standard baselines. This work provides a robust tool to ensure that ML diagnostic systems generalize well, e.g., across demographic groups, offering a critical step towards AI safety in healthcare. Code: https://github.com/Neferpitou3871/AI-Alignment-Medical-Imaging.