Towards AI Accountability Infrastructure: Gaps and Opportunities in AI Audit Tooling

📅 2024-02-27
🏛️ arXiv.org
📈 Citations: 19
Influential: 2
📄 PDF
🤖 AI Summary
Current AI auditing tools predominantly focus on model performance evaluation, failing to support end-to-end accountability practices—including harm identification, evidence construction, stakeholder engagement, and intervention advocacy. Method: We conducted in-depth interviews with 35 practitioners and systematically crawled, cataloged, and coded 435 auditing tools to develop a需求–capability mapping framework and an ecosystem gap diagnostic model. Contribution/Results: Our analysis reveals systematic deficiencies across four critical capabilities: traceability, multi-stakeholder participation, evidentiary chain generation, and intervention support—constituting the first empirical diagnosis of such gaps. Building on these findings, we propose the “AI Accountability Infrastructure” paradigm, shifting beyond narrow assessment-centric design toward cross-stage coordination and multi-role adaptability. The study delivers an empirically grounded, prioritized roadmap for designing next-generation, accountability-oriented AI auditing tools.

Technology Category

Application Category

📝 Abstract
Audits are critical mechanisms for identifying the risks and limitations of deployed artificial intelligence (AI) systems. However, the effective execution of AI audits remains incredibly difficult, and practitioners often need to make use of various tools to support their efforts. Drawing on interviews with 35 AI audit practitioners and a landscape analysis of 435 tools, we compare the current ecosystem of AI audit tooling to practitioner needs. While many tools are designed to help set standards and evaluate AI systems, they often fall short in supporting accountability. We outline challenges practitioners faced in their efforts to use AI audit tools and highlight areas for future tool development beyond evaluation -- from harms discovery to advocacy. We conclude that the available resources do not currently support the full scope of AI audit practitioners' needs and recommend that the field move beyond tools for just evaluation and towards more comprehensive infrastructure for AI accountability.
Problem

Research questions and friction points this paper is trying to address.

Identifies gaps in AI audit tools for accountability.
Highlights challenges in using tools for AI system evaluation.
Recommends comprehensive infrastructure for AI accountability needs.
Innovation

Methods, ideas, or system contributions that make the work stand out.

Interviews with 35 AI audit practitioners
Analysis of 435 AI audit tools
Recommend comprehensive AI accountability infrastructure
🔎 Similar Papers
No similar papers found.
Victor Ojewale
Victor Ojewale
Brown University
Algorithm AuditsMachine LearningHuman-Computer Interaction
R
Ryan Steed
Carnegie Mellon University, Pittsburgh, Pennsylvania, USA
B
Briana Vecchione
Data & Society, New York City, New York, USA
A
Abeba Birhane
Mozilla Foundation and Trinity College Dublin, Dublin, Ireland
Inioluwa Deborah Raji
Inioluwa Deborah Raji
UC Berkeley
machine learningevaluationauditingalgorithmssociety