🤖 AI Summary
This work addresses the lack of verifiable probabilistic guarantees for group fairness in downstream predictions within representation learning. We propose FRG, the first representation learning framework enabling high-confidence (PAC-style) fairness control. FRG jointly optimizes representations and fairness constraints via adversarial inference, ensuring—uniformly across arbitrary downstream models—that demographic disparities (e.g., mean difference, prediction error difference) are bounded within a user-specified threshold $varepsilon$ with probability at least $1-delta$. The framework integrates high-confidence risk bounding with an adjustable fairness boundary mechanism, offering theoretical guarantees and implementation simplicity. Experiments on three real-world datasets demonstrate that FRG consistently achieves superior and more robust unfairness mitigation compared to six state-of-the-art methods, across diverse downstream models—including logistic regression, MLPs, and XGBoost—and multiple prediction tasks.
📝 Abstract
Representation learning is increasingly applied to generate representations that generalize well across multiple downstream tasks. Ensuring fairness guarantees in representation learning is crucial to prevent unfairness toward specific demographic groups in downstream tasks. In this work, we formally introduce the task of learning representations that achieve high-confidence fairness. We aim to guarantee that demographic disparity in every downstream prediction remains bounded by a *user-defined* error threshold $ε$, with *controllable* high probability. To this end, we propose the ***F**air **R**epresentation learning with high-confidence **G**uarantees (FRG)* framework, which provides these high-confidence fairness guarantees by leveraging an optimized adversarial model. We empirically evaluate FRG on three real-world datasets, comparing its performance to six state-of-the-art fair representation learning methods. Our results demonstrate that FRG consistently bounds unfairness across a range of downstream models and tasks.