🤖 AI Summary
This work addresses a critical limitation in existing explainable AI (XAI) methods—such as Chain-of-Thought—which follow a “reason-then-conclude” paradigm and often suffer from misalignment between reasoning and conclusions due to logical gaps or hallucinations, thereby undermining trustworthiness and verifiability. To overcome this, the authors propose a novel “conclusion-first, structured justification” paradigm, termed Result→Justify, which for the first time integrates structured communication norms from professional writing (e.g., CREAC/BLUF) into XAI. They introduce a Structured Explanation Framework (SEF) and define six quantitative metrics capturing structural coherence and factual consistency. Experiments across four tasks in three high-stakes domains demonstrate that SEF-generated explanations exhibit significant correlations between all six metrics and model correctness (r = 0.20–0.42, p < 0.001), achieving an accuracy of 83.9%—a 5.3% improvement over Chain-of-Thought.
📝 Abstract
Explainable AI (XAI) in high-stakes domains should help stakeholders trust and verify system outputs. Yet Chain-of-Thought methods reason before concluding, and logical gaps or hallucinations can yield conclusions that do not reliably align with their rationale. Thus, we propose"Result ->Justify", which constrains the output communication to present a conclusion before its structured justification. We introduce SEF (Structured Explainability Framework), operationalizing professional conventions (e.g., CREAC, BLUF) via six metrics for structure and grounding. Experiments across four tasks in three domains validate this approach: all six metrics correlate with correctness (r=0.20-0.42; p<0.001), and SEF achieves 83.9% accuracy (+5.3 over CoT). These results suggest structured justification can improve verifiability and may also improve reliability.