Fair Feature Importance Scores via Feature Occlusion and Permutation

πŸ“… 2026-02-09
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses a critical gap in existing feature importance methods, which predominantly focus on predictive accuracy and fail to quantify the impact of features on model fairness. The paper introduces, for the first time, two model-agnostic measures of fair feature importance: one evaluates a feature’s contribution by assessing changes in fairness before and after its permutation, while the other enhances computational efficiency by masking specific features during training and leveraging minipatch learning to compare resulting fairness disparities. The proposed approach is interpretable, scalable, and computationally efficient, accurately identifying fairness-critical features across multiple prediction tasks. By offering a practical tool for fairness-aware model analysis, this method advances the development of responsible AI systems and fills a key void in fairness attribution research.

Technology Category

Application Category

πŸ“ Abstract
As machine learning models increasingly impact society, their opaque nature poses challenges to trust and accountability, particularly in fairness contexts. Understanding how individual features influence model outcomes is crucial for building interpretable and equitable models. While feature importance metrics for accuracy are well-established, methods for assessing feature contributions to fairness remain underexplored. We propose two model-agnostic approaches to measure fair feature importance. First, we propose to compare model fairness before and after permuting feature values. This simple intervention-based approach decouples a feature and model predictions to measure its contribution to training. Second, we evaluate the fairness of models trained with and without a given feature. This occlusion-based score enjoys dramatic computational simplification via minipatch learning. Our empirical results reflect the simplicity and effectiveness of our proposed metrics for multiple predictive tasks. Both methods offer simple, scalable, and interpretable solutions to quantify the influence of features on fairness, providing new tools for responsible machine learning development.
Problem

Research questions and friction points this paper is trying to address.

fairness
feature importance
machine learning
interpretability
model accountability
Innovation

Methods, ideas, or system contributions that make the work stand out.

fair feature importance
feature occlusion
feature permutation
model-agnostic interpretability
minipatch learning
πŸ”Ž Similar Papers
No similar papers found.