Understanding challenges to the interpretation of disaggregated evaluations of algorithmic fairness

📅 2025-06-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper identifies fundamental limitations of disaggregated evaluation (i.e., subgroup performance analysis) in algorithmic fairness assessment: (i) when data are representative but reflect real-world inequities, equal subgroup performance can be misleading; and (ii) under unknown selection bias, both disaggregated evaluation and conditional independence tests fail. Crucially, evaluation stability is shown to depend critically on the underlying data-generating mechanism. Method: We propose a causally enhanced, robust evaluation paradigm integrating causal graph modeling, explicit causal assumptions, weighted performance estimation, and distributional shift analysis. Contribution/Results: Our framework predicts metric stability, delineates the validity boundaries of fairness evaluation methods across distinct bias regimes, and provides actionable, causally grounded guidance for fairness practice—marking the first systematic treatment of evaluation robustness through a causal lens.

Technology Category

Application Category

📝 Abstract
Disaggregated evaluation across subgroups is critical for assessing the fairness of machine learning models, but its uncritical use can mislead practitioners. We show that equal performance across subgroups is an unreliable measure of fairness when data are representative of the relevant populations but reflective of real-world disparities. Furthermore, when data are not representative due to selection bias, both disaggregated evaluation and alternative approaches based on conditional independence testing may be invalid without explicit assumptions regarding the bias mechanism. We use causal graphical models to predict metric stability across subgroups under different data generating processes. Our framework suggests complementing disaggregated evaluations with explicit causal assumptions and analysis to control for confounding and distribution shift, including conditional independence testing and weighted performance estimation. These findings have broad implications for how practitioners design and interpret model assessments given the ubiquity of disaggregated evaluation.
Problem

Research questions and friction points this paper is trying to address.

Assessing fairness via subgroup performance can mislead without causal analysis
Non-representative data invalidates fairness tests without bias assumptions
Causal models needed to stabilize metrics across biased subgroups
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses causal graphical models for fairness analysis
Combines disaggregated evaluations with causal assumptions
Proposes weighted performance estimation for bias control
🔎 Similar Papers
No similar papers found.
Stephen R. Pfohl
Stephen R. Pfohl
Google Research
Machine LearningHealthcareAlgorithmic Fairness
N
Natalie Harris
Google Research, Mountain View, CA, USA
Chirag Nagpal
Chirag Nagpal
AI Research Scientist, Meta
AI AlignmentML for HealthcareGraphical ModelsSurvival Analysis
David Madras
David Madras
University of Toronto
Machine Learning
V
Vishwali Mhasawade
New York University, New York, NY, USA
Olawale Salaudeen
Olawale Salaudeen
Postdoctoral Associate, MIT
Trustworthy AIAI for SocietyDistribution ShiftCausality
A
Awa Dieng
Google DeepMind, Mountain View, CA, USA
S
Shannon Sequeira
Google Research, Mountain View, CA, USA
S
S. Arciniegas
The Hopsital for Sick Children, Toronto, ON, Canada
L
Lillian Sung
The Hopsital for Sick Children, Toronto, ON, Canada
N
Nnamdi Ezeanochie
Google Research, Mountain View, CA, USA
H
Heather Cole-Lewis
Google Research, Mountain View, CA, USA
Katherine Heller
Katherine Heller
Google Research
Machine LearningHealth AIEthical AI
Sanmi Koyejo
Sanmi Koyejo
Assistant Professor, Stanford University
Machine LearningHealthcare AINeuroinformatics
Alexander D'Amour
Alexander D'Amour
Research Scientist, Google DeepMind
StatisticsMachine LearningArtificial Intelligence