Addressing Bias in VLMs for Glaucoma Detection Without Protected Attribute Supervision

📅 2025-08-12
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address subgroup performance disparity in vision-language models (VLMs) for glaucoma detection—caused by latent demographic biases—this paper proposes an unsupervised debiasing method that requires no protected attribute labels. First, unsupervised clustering is performed on image embeddings to infer proxy subgroups. Second, a dual-contrastive learning framework is introduced, integrating CLIP-style cross-modal contrastive learning with SimCLR-style intra-image contrastive learning; a gradient-similarity weighting mechanism dynamically adjusts the joint loss weights for the top-k hard samples. This adaptive strategy strengthens discriminative boundaries across subgroups, thereby mitigating performance gaps. Evaluated on the Harvard FairVLMed dataset, the method significantly improves both Equalized Subgroup AUC and Groupwise AUC, yielding fairer and more robust glaucoma screening without access to sensitive demographic annotations.

Technology Category

Application Category

📝 Abstract
Vision-Language Models (VLMs) have achieved remarkable success on multimodal tasks such as image-text retrieval and zero-shot classification, yet they can exhibit demographic biases even when explicit protected attributes are absent during training. In this work, we focus on automated glaucoma screening from retinal fundus images, a critical application given that glaucoma is a leading cause of irreversible blindness and disproportionately affects underserved populations. Building on a reweighting-based contrastive learning framework, we introduce an attribute-agnostic debiasing method that (i) infers proxy subgroups via unsupervised clustering of image-image embeddings, (ii) computes gradient-similarity weights between the CLIP-style multimodal loss and a SimCLR-style image-pair contrastive loss, and (iii) applies these weights in a joint, top-$k$ weighted objective to upweight underperforming clusters. This label-free approach adaptively targets the hardest examples, thereby reducing subgroup disparities. We evaluate our method on the Harvard FairVLMed glaucoma subset, reporting Equalized Odds Distance (EOD), Equalized Subgroup AUC (ES AUC), and Groupwise AUC to demonstrate equitable performance across inferred demographic subgroups.
Problem

Research questions and friction points this paper is trying to address.

Detect glaucoma without bias in VLMs
Address demographic bias without protected attributes
Improve fairness in automated glaucoma screening
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unsupervised clustering for proxy subgroups
Gradient-similarity weights for debiasing
Top-k weighted objective for underperforming clusters
🔎 Similar Papers
No similar papers found.
Ahsan Habib Akash
Ahsan Habib Akash
Graduate Reasearch Assistant | West Virginia University
Computer Vision | Artificial Intelligence | Deep Learning
G
Greg Murray
West Virginia University, Morgantown, WV, USA
A
Annahita Amireskandari
West Virginia University, Morgantown, WV, USA
Joel Palko
Joel Palko
West Virginia University
Ophthalmology
C
Carol Laxson
West Virginia University, Morgantown, WV, USA
Binod Bhattarai
Binod Bhattarai
Assistant Professor, University of Aberdeen
Machine LearningMedical Image AnalysisComputer Vision
P
Prashnna Gyawali
West Virginia University, Morgantown, WV, USA