Simulating Biases for Interpretable Fairness in Offline and Online Classifiers

📅 2025-07-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Predictive models often inherit and amplify systemic biases present in training data, undermining decision fairness. To address this, we propose a controllable bias injection framework for synthetic data generation, using proxy base models to emulate structural biases in loan approval and producing synthetic datasets with tunable bias intensity. Building upon this, we systematically evaluate the efficacy of preprocessing, in-processing, and post-processing debiasing methods under both offline and online learning settings. We further introduce a novel second-order Shapley value-based interpretability method to quantitatively characterize how debiasing strategies alter feature dependency structures and model reliance mechanisms. Experiments demonstrate that our framework enables precise bias replication and calibration, while revealing distinct internal decision-logic interventions induced by different debiasing techniques—thereby significantly enhancing the interpretability and controllability of fairness evaluation.

Technology Category

Application Category

📝 Abstract
Predictive models often reinforce biases which were originally embedded in their training data, through skewed decisions. In such cases, mitigation methods are critical to ensure that, regardless of the prevailing disparities, model outcomes are adjusted to be fair. To assess this, datasets could be systematically generated with specific biases, to train machine learning classifiers. Then, predictive outcomes could aid in the understanding of this bias embedding process. Hence, an agent-based model (ABM), depicting a loan application process that represents various systemic biases across two demographic groups, was developed to produce synthetic datasets. Then, by applying classifiers trained on them to predict loan outcomes, we can assess how biased data leads to unfairness. This highlights a main contribution of this work: a framework for synthetic dataset generation with controllable bias injection. We also contribute with a novel explainability technique, which shows how mitigations affect the way classifiers leverage data features, via second-order Shapley values. In experiments, both offline and online learning approaches are employed. Mitigations are applied at different stages of the modelling pipeline, such as during pre-processing and in-processing.
Problem

Research questions and friction points this paper is trying to address.

Simulating biases in training data for fair classifiers
Assessing bias impact on loan prediction fairness
Developing controllable bias injection framework for datasets
Innovation

Methods, ideas, or system contributions that make the work stand out.

Agent-based model for synthetic biased datasets
Controllable bias injection framework
Second-order Shapley explainability technique
🔎 Similar Papers
2024-08-10AAAI Conference on Artificial IntelligenceCitations: 0
R
Ricardo Inácio
Faculdade de Engenharia da Universidade do Porto, Portugal
Z
Zafeiris Kokkinogenis
Faculdade de Engenharia da Universidade do Porto, Portugal
Vitor Cerqueira
Vitor Cerqueira
University of Porto, Faculty of Engineering
Machine learningTime series
C
Carlos Soares
Faculdade de Engenharia da Universidade do Porto, Portugal