Fairness in Multi-modal Medical Diagnosis with Demonstration Selection

📅 2025-11-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Multimodal large language models (MLLMs) exhibit significant demographic biases—e.g., along gender and race dimensions—in medical image reasoning, yet existing debiasing methods rely on large-scale annotated data or model fine-tuning, limiting compatibility with foundational LLMs. To address this, we propose a fine-tuning-free, annotation-free fair in-context learning framework. Its core innovation is the Fairness-Aware Demonstration Selection (FADS) mechanism, which—novelly for multimodal contexts—integrates clustering-based sampling to jointly optimize demographic balance and semantic relevance. FADS combines clustering-driven exemplar selection, cross-modal semantic matching, and demographic balancing strategies. Evaluated across multiple medical imaging benchmarks, our method reduces inter-group performance disparities by an average of 38% while preserving diagnostic accuracy. Results demonstrate strong efficacy, generalizability across tasks and datasets, and deployment friendliness—enabling fairer MLLM inference without architectural or training modifications.

Technology Category

Application Category

📝 Abstract
Multimodal large language models (MLLMs) have shown strong potential for medical image reasoning, yet fairness across demographic groups remains a major concern. Existing debiasing methods often rely on large labeled datasets or fine-tuning, which are impractical for foundation-scale models. We explore In-Context Learning (ICL) as a lightweight, tuning-free alternative for improving fairness. Through systematic analysis, we find that conventional demonstration selection (DS) strategies fail to ensure fairness due to demographic imbalance in selected exemplars. To address this, we propose Fairness-Aware Demonstration Selection (FADS), which builds demographically balanced and semantically relevant demonstrations via clustering-based sampling. Experiments on multiple medical imaging benchmarks show that FADS consistently reduces gender-, race-, and ethnicity-related disparities while maintaining strong accuracy, offering an efficient and scalable path toward fair medical image reasoning. These results highlight the potential of fairness-aware in-context learning as a scalable and data-efficient solution for equitable medical image reasoning.
Problem

Research questions and friction points this paper is trying to address.

Addressing demographic fairness in medical image diagnosis using MLLMs
Overcoming limitations of conventional demonstration selection strategies
Reducing gender, race and ethnicity disparities in medical imaging
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses clustering-based sampling for balanced demonstrations
Implements fairness-aware in-context learning without fine-tuning
Ensures demographic balance while maintaining semantic relevance
🔎 Similar Papers
No similar papers found.
D
Dawei Li
Arizona State University
Z
Zijian Gu
University of Rochester
P
Peng Wang
University of Virginia
C
Chuhan Song
UCL
Zhen Tan
Zhen Tan
Ph.D. at Arizona State University
Data MiningMachine LearningAI for ScienceUser-centric ExplanationResponsible AI
Mohan Zhang
Mohan Zhang
Ph.D., CS@UNC Chapel Hill;
LLM
Tianlong Chen
Tianlong Chen
Assistant Professor, CS@UNC Chapel Hill; Chief AI Scientist, hireEZ
Machine LearningAI4ScienceComputer VisionSparsity
Y
Yu Tian
University of Central Florida
S
Song Wang
University of Central Florida