Attention IoU: Examining Biases in CelebA using Attention Maps

๐Ÿ“… 2025-03-25
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This paper addresses implicit bias in vision models arising from dataset biases (e.g., CelebA), proposing Attention-IoUโ€”a novel internal representation bias metric grounded in attention maps. Unlike conventional external evaluations relying on subgroup accuracy, Attention-IoU quantifies spatial overlap (Intersection-over-Union) between attention regions and features correlated with sensitive attributes (e.g., gender), thereby exposing unannotated confounders and non-explicit bias pathways (e.g., โ€œgenderโ€“glassesโ€ coupling). Key contributions include: (i) the first systematic use of attention mechanisms to disentangle multidimensional attribute coupling biases; (ii) identification of latent bias sources beyond annotated labels in CelebA; and (iii) empirical validation on synthetic Waterbirds data and controlled resampling experiments, demonstrating statistically significant improvements over existing bias detection methods.

Technology Category

Application Category

๐Ÿ“ Abstract
Computer vision models have been shown to exhibit and amplify biases across a wide array of datasets and tasks. Existing methods for quantifying bias in classification models primarily focus on dataset distribution and model performance on subgroups, overlooking the internal workings of a model. We introduce the Attention-IoU (Attention Intersection over Union) metric and related scores, which use attention maps to reveal biases within a model's internal representations and identify image features potentially causing the biases. First, we validate Attention-IoU on the synthetic Waterbirds dataset, showing that the metric accurately measures model bias. We then analyze the CelebA dataset, finding that Attention-IoU uncovers correlations beyond accuracy disparities. Through an investigation of individual attributes through the protected attribute of Male, we examine the distinct ways biases are represented in CelebA. Lastly, by subsampling the training set to change attribute correlations, we demonstrate that Attention-IoU reveals potential confounding variables not present in dataset labels.
Problem

Research questions and friction points this paper is trying to address.

Measuring model bias using attention maps in computer vision
Identifying image features causing biases in classification models
Analyzing biases in CelebA beyond dataset distribution
Innovation

Methods, ideas, or system contributions that make the work stand out.

Attention-IoU metric for bias detection
Analyzes internal model representations via attention maps
Identifies confounding variables beyond dataset labels
๐Ÿ”Ž Similar Papers
No similar papers found.