๐ค AI Summary
This paper addresses the lack of systematic design and evaluation frameworks for prompt-based natural language explanations (NLEs) in transparent AI governance. We propose the first taxonomy specifically tailored to prompt-driven NLEs, structured along three dimensions: context, generation & presentation, and evaluation. The taxonomy integrates task objectives, data characteristics, stakeholder needs, generation mechanisms, interaction modalities, output formats, and user-centered evaluation criteria into a multidimensional, structured classification. Its key innovation lies in adapting eXplainable AI (XAI) taxonomic principles to the prompt-based NLE paradigmโmoving beyond conventional model-internal explanations. The resulting framework provides researchers, auditors, and policymakers with actionable design guidelines and standardized evaluation benchmarks, thereby enhancing the practical utility of NLEs in transparency, auditability, and regulatory compliance. (149 words)
๐ Abstract
Effective AI governance requires structured approaches for stakeholders to access and verify AI system behavior. With the rise of large language models, Natural Language Explanations (NLEs) are now key to articulating model behavior, which necessitates a focused examination of their characteristics and governance implications. We draw on Explainable AI (XAI) literature to create an updated XAI taxonomy, adapted to prompt-based NLEs, across three dimensions: (1) Context, including task, data, audience, and goals; (2) Generation and Presentation, covering generation methods, inputs, interactivity, outputs, and forms; and (3) Evaluation, focusing on content, presentation, and user-centered properties, as well as the setting of the evaluation. This taxonomy provides a framework for researchers, auditors, and policymakers to characterize, design, and enhance NLEs for transparent AI systems.