A Counterfactual LLM Framework for Detecting Human Biases: A Case Study of Sex/Gender in Emergency Triage

📅 2025-11-21
📈 Citations: 0
✹ Influential: 0
📄 PDF
đŸ€– AI Summary
This study addresses potential gender bias in emergency department triage by proposing the first domain-agnostic counterfactual large language model (LLM) framework designed for human clinical decision support. Methodologically, it generates counterfactual patient pairs differing solely in gender—while rigorously controlling all other clinical features—to isolate and quantify gender’s causal effect on triage scores; it jointly models explicit and implicit gender cues from both structured variables and unstructured clinical text. Its key contribution is the first integration of counterfactual reasoning with LLMs for clinical bias detection. Evaluated on 150,000 real-world French ED cases, the framework reveals that women are systematically assigned 2.1% lower triage priority on average—projecting over 200,000 annual misclassifications in France. Critically, these findings replicate robustly across languages and healthcare systems, as validated on the English-language MIMIC-IV dataset.

Technology Category

Application Category

📝 Abstract
We present a novel, domain-agnostic counterfactual approach that uses Large Language Models (LLMs) to quantify gender disparities in human clinical decision-making. The method trains an LLM to emulate observed decisions, then evaluates counterfactual pairs in which only gender is flipped, estimating directional disparities while holding all other clinical factors constant. We study emergency triage, validating the approach on more than 150,000 admissions to the Bordeaux University Hospital (France) and replicating results on a subset of MIMIC-IV across a different language, population, and healthcare system. In the Bordeaux cohort, otherwise identical presentations were approximately 2.1% more likely to receive a lower-severity triage score when presented as female rather than male; scaled to national emergency volumes in France, this corresponds to more than 200,000 lower-severity assignments per year. Modality-specific analyses indicate that both explicit tabular gender indicators and implicit textual gender cues contribute to the disparity. Beyond emergency care, the approach supports bias audits in other settings (e.g., hiring, academic, and justice decisions), providing a scalable tool to detect and address inequities in real-world decision-making.
Problem

Research questions and friction points this paper is trying to address.

Quantifying gender disparities in human clinical decision-making using LLMs
Detecting bias through counterfactual analysis of emergency triage decisions
Developing scalable framework to audit biases across different domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uses counterfactual LLM framework to detect biases
Trains LLM to emulate decisions then flips gender
Validates method across different languages and populations
🔎 Similar Papers
No similar papers found.
A
Ariel Guerra-Adames
AHeaD Team, Bordeaux Population Health Research Center, Université de Bordeaux, Bordeaux, F-33000, France.
M
Marta Avalos-Fernandez
SISTM Team, Bordeaux Population Health Research Center, Université de Bordeaux, Bordeaux, F-33000, France.
O
Océane Dorémus
AHeaD Team, Bordeaux Population Health Research Center, Université de Bordeaux, Bordeaux, F-33000, France.
Leo Anthony Celi
Leo Anthony Celi
Massachusetts Institute of Technology
C
Cédric Gil-Jardiné
AHeaD Team, Bordeaux Population Health Research Center, Université de Bordeaux, Bordeaux, F-33000, France.
Emmanuel Lagarde
Emmanuel Lagarde
INSERM