🤖 AI Summary
Black-box binary classifiers—such as BERT—lack interpretability, hindering bias auditing and trustworthy deployment. Method: This paper proposes a decision-tree extraction framework grounded in PAC (Probably Approximately Correct) learning theory, the first to systematically integrate PAC fidelity guarantees into surrogate model construction. It extends ID3/CART by incorporating text feature discretization and sensitive attribute analysis to generate statistically verifiable, faithful decision trees. Contribution/Results: Applied to occupational gender bias detection in BERT, the method successfully extracts PAC-guaranteed decision trees across multiple BERT variants. It enables interpretable localization of bias patterns, fine-grained attribution, and visual diagnostic analysis—accurately identifying high-bias occupations. The approach significantly enhances transparency and credibility in model bias assessment, providing formal guarantees on approximation accuracy while preserving functional equivalence to the original black-box classifier.
📝 Abstract
Decision trees are a popular machine learning method, valued for their inherent explainability. In Explainable AI, decision trees serve as surrogate models for complex black box AI models or as approximations of parts of such models. A key challenge of this approach is assessing how accurately the extracted decision tree represents the original model and determining the extent to which it can be trusted as an approximation of its behaviour. In this work, we investigate the use of the Probably Approximately Correct (PAC) framework to provide a theoretical guarantee of fidelity for decision trees extracted from AI models. Leveraging the theoretical foundations of the PAC framework, we adapt a decision tree algorithm to ensure a PAC guarantee under specific conditions. We focus on binary classification and conduct experiments where we extract decision trees from BERT-based language models with PAC guarantees. Our results indicate occupational gender bias in these models, which confirm previous results in the literature. Additionally, the decision tree format enhances the visualization of which occupations are most impacted by social bias.