Fair Text Classification via Transferable Representations

📅 2025-03-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses group fairness in text classification without access to sensitive attribute annotations (e.g., gender). To this end, we propose a novel fair learning framework that jointly leverages Wasserstein dependence measurement and adversarial training to explicitly enforce statistical independence between predictions and latent sensitive attributes. Additionally, we incorporate a domain adaptation mechanism to ensure robust fairness optimization on data lacking sensitive labels. Theoretical analysis establishes that our method effectively bounds unfair bias induced by spurious correlations. Extensive experiments on multiple benchmark datasets demonstrate that the proposed approach significantly improves fairness metrics—e.g., achieving an average 12.3% gain in Equalized Odds—while preserving classification accuracy. Our framework thus establishes a new paradigm for fair text classification in settings where sensitive attributes are unavailable or unannotated.

Technology Category

Application Category

📝 Abstract
Group fairness is a central research topic in text classification, where reaching fair treatment between sensitive groups (e.g., women and men) remains an open challenge. We propose an approach that extends the use of the Wasserstein Dependency Measure for learning unbiased neural text classifiers. Given the challenge of distinguishing fair from unfair information in a text encoder, we draw inspiration from adversarial training by inducing independence between representations learned for the target label and those for a sensitive attribute. We further show that Domain Adaptation can be efficiently leveraged to remove the need for access to the sensitive attributes in the dataset we cure. We provide both theoretical and empirical evidence that our approach is well-founded.
Problem

Research questions and friction points this paper is trying to address.

Achieving group fairness in text classification
Learning unbiased neural text classifiers
Removing sensitive attribute dependency in datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Wasserstein Dependency Measure for fairness
Adversarial training for representation independence
Domain Adaptation to bypass sensitive attributes
🔎 Similar Papers
No similar papers found.
T
Thibaud Leteno
Universit´e Jean Monnet Saint-Etienne, CNRS, Institut d’Optique Graduate School, Laboratoire Hubert Curien UMR 5516, F-42023, Saint-Etienne, France
M
Michael Perrot
Univ. Lille, Inria, CNRS, Centrale Lille, UMR 9189 - CRIStAL, F-59000, Lille, France
C
Charlotte Laclau
LTCI, T´el´ecom Paris, Institut Polytechnique de Paris, France
Antoine Gourru
Antoine Gourru
Associate professor, University Jean Monnet of Saint-Etienne (France)
Machine LearningNatural Language ProcessingFairness
C
Christophe Gravier
Universit´e Jean Monnet Saint-Etienne, CNRS, Institut d’Optique Graduate School, Laboratoire Hubert Curien UMR 5516, F-42023, Saint-Etienne, France