FairREAD: Re-fusing Demographic Attributes after Disentanglement for Fair Medical Image Classification

๐Ÿ“… 2024-12-20
๐Ÿ›๏ธ arXiv.org
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
In medical image classification, removing sensitive attributes (e.g., demographic features) often erodes clinically relevant information, hindering simultaneous improvement of fairness and diagnostic performance. To address this, we propose a โ€œdecouple-then-controllably-refuseโ€ paradigm: first, orthogonal constraints and adversarial training disentangle bias-laden and clinically informative components within sensitive attributes; second, a controllable feature re-fusion mechanism selectively preserves the latter; third, subgroup-specific adaptive decision thresholds are introduced to jointly optimize fairness and diagnostic accuracy. Evaluated on a large-scale chest X-ray dataset, our method reduces equalized odds disparity by over 60% while maintaining baseline diagnostic accuracy. This work establishes the first framework for separable modeling and selective utilization of bias versus clinical signals embedded in sensitive attributes, offering a novel direction for fairness-aware medical AI.

Technology Category

Application Category

๐Ÿ“ Abstract
Recent advancements in deep learning have shown transformative potential in medical imaging, yet concerns about fairness persist due to performance disparities across demographic subgroups. Existing methods aim to address these biases by mitigating sensitive attributes in image data; however, these attributes often carry clinically relevant information, and their removal can compromise model performance-a highly undesirable outcome. To address this challenge, we propose Fair Re-fusion After Disentanglement (FairREAD), a novel, simple, and efficient framework that mitigates unfairness by re-integrating sensitive demographic attributes into fair image representations. FairREAD employs orthogonality constraints and adversarial training to disentangle demographic information while using a controlled re-fusion mechanism to preserve clinically relevant details. Additionally, subgroup-specific threshold adjustments ensure equitable performance across demographic groups. Comprehensive evaluations on a large-scale clinical X-ray dataset demonstrate that FairREAD significantly reduces unfairness metrics while maintaining diagnostic accuracy, establishing a new benchmark for fairness and performance in medical image classification.
Problem

Research questions and friction points this paper is trying to address.

Mitigates unfairness in medical image classification across demographics
Preserves clinically relevant information while removing biases
Ensures equitable performance without compromising diagnostic accuracy
Innovation

Methods, ideas, or system contributions that make the work stand out.

Disentangles demographic info via orthogonality constraints
Re-fuses attributes to preserve clinical relevance
Adjusts thresholds for equitable subgroup performance
๐Ÿ”Ž Similar Papers
No similar papers found.
Y
Yicheng Gao
Department of Radiology, Northwestern University, Chicago, IL, USA
Jinkui Hao
Jinkui Hao
Northwestern University
Medical image analysisComputer-Aided Diagnosis
B
Bo Zhou
Department of Radiology, Northwestern University, Chicago, IL, USA