When AI Fails, What Works? A Data-Driven Taxonomy of Real-World AI Risk Mitigation Strategies

📅 2026-03-04
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the lack of structured mechanisms for mitigating systemic risks—legal, reputational, and financial—arising from failures of high-risk AI systems. Leveraging 6,893 incident reports extracted from a corpus of 9,705 media articles, the authors employ structured prompt engineering to identify and categorize mitigation strategies. They extend MIT’s existing AI risk mitigation taxonomy by introducing four new strategic categories: “Correction and Containment,” “Legal/Regulatory and Enforcement,” “Financial and Market Controls,” and “Avoidance and Denial,” which collectively encompass 67% of newly identified subcategories. The resulting dataset comprises 32 distinct mitigation labels, supported by 23,994 annotations—including 9,629 instances of previously undocumented patterns—thereby substantially enhancing diagnostic and intervention capabilities for emerging systemic failures and improving post-deployment monitoring efficacy.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) are increasingly embedded in high-stakes workflows, where failures propagate beyond isolated model errors into systemic breakdowns that can lead to legal exposure, reputational damage, and material financial losses. Building on this shift from model-centric risks to end-to-end system vulnerabilities, we analyze real-world AI incident reporting and mitigation actions to derive an empirically grounded taxonomy that links failure dynamics to actionable interventions. Using a unified corpus of 9,705 media-reported AI incident articles, we extract explicit mitigation actions from 6,893 texts via structured prompting and then systematically classify responses to extend MIT's AI Risk Mitigation Taxonomy. Our taxonomy introduces four new mitigation categories, including 1) Corrective and Restrictive Actions, 2) Legal/Regulatory and Enforcement Actions, 3) Financial, Economic, and Market Controls, and 4) Avoidance and Denial, capturing response patterns that are becoming increasingly prevalent as AI deployment and regulation evolve. Quantitatively, we label the mitigation dataset with 32 distinct labels, producing 23,994 label assignments; 9,629 of these reflect previously unseen mitigation patterns, yielding a 67% increase of the original subcategory coverage and substantially enhancing the taxonomy's applicability to emerging systemic failure modes. By structuring incident responses, the paper strengthens "diagnosis-to-prescription" guidance and advances continuous, taxonomy-aligned post-deployment monitoring to prevent cascading incidents and downstream impact.
Problem

Research questions and friction points this paper is trying to address.

AI risk mitigation
systemic failure
large language models
incident response
taxonomy
Innovation

Methods, ideas, or system contributions that make the work stand out.

AI risk mitigation
taxonomy extension
real-world incident analysis
systemic failure
post-deployment monitoring
🔎 Similar Papers
No similar papers found.
E
Evgenija Popchanovska
Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University, 1000 Skopje, North Macedonia
A
Ana Gjorgjevikj
Faculty of Computer Science and Engineering, Ss. Cyril and Methodius University, 1000 Skopje, North Macedonia
M
Maryan Rizinski
Department of Computer Science, Metropolitan College, Boston University, Boston, MA 02215, USA
L
Lubomir T. Chitkushev
Department of Computer Science, Metropolitan College, Boston University, Boston, MA 02215, USA
Irena Vodenska
Irena Vodenska
Boston University
Environmental Social and Governance (ESG) investingDisinformation in the era of climate crisisNews Sentiments and Systemic R
Dimitar Trajanov
Dimitar Trajanov
Prof.@ Ss. Cyril and Methodius University in Skopje & Visiting Research Prof.@ Boston University, US
Data scienceAI AgentsNLPSemantic webOpen data