Robust Explanations Through Uncertainty Decomposition: A Path to Trustworthier AI

📅 2025-07-17
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address declining model interpretability and consequent trust deficits in complex machine learning systems, this paper proposes a trustworthy XAI framework grounded in uncertainty decoupling. Methodologically, it is the first to disentangle aleatoric from epistemic uncertainty and leverages epistemic uncertainty as a reliability metric for explanations: (i) as a rejection threshold for low-fidelity explanations, and (ii) as a dynamic signal to adapt explanation strategies—e.g., feature attribution or counterfactual generation—to local uncertainty conditions. The framework integrates Bayesian neural networks, rigorous uncertainty quantification, and state-of-the-art XAI techniques. Extensive experiments across diverse models—including traditional machine learning and deep neural networks—demonstrate substantial improvements in explanation stability and robustness. Crucially, it effectively filters unreliable explanations, thereby enhancing user trust in AI-driven decisions.

Technology Category

Application Category

📝 Abstract
Recent advancements in machine learning have emphasized the need for transparency in model predictions, particularly as interpretability diminishes when using increasingly complex architectures. In this paper, we propose leveraging prediction uncertainty as a complementary approach to classical explainability methods. Specifically, we distinguish between aleatoric (data-related) and epistemic (model-related) uncertainty to guide the selection of appropriate explanations. Epistemic uncertainty serves as a rejection criterion for unreliable explanations and, in itself, provides insight into insufficient training (a new form of explanation). Aleatoric uncertainty informs the choice between feature-importance explanations and counterfactual explanations. This leverages a framework of explainability methods driven by uncertainty quantification and disentanglement. Our experiments demonstrate the impact of this uncertainty-aware approach on the robustness and attainability of explanations in both traditional machine learning and deep learning scenarios.
Problem

Research questions and friction points this paper is trying to address.

Enhancing AI transparency via uncertainty decomposition in predictions
Distinguishing aleatoric and epistemic uncertainty for explanation selection
Improving explanation robustness using uncertainty-aware frameworks
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decompose uncertainty into aleatoric and epistemic types
Use epistemic uncertainty to reject unreliable explanations
Guide explanation choice via aleatoric uncertainty quantification
🔎 Similar Papers
No similar papers found.
Chenrui Zhu
Chenrui Zhu
PhD student, CNRS Heudiasyc, University of Technology of Compiègne
Explainable AIRobustnessClassification
L
Louenas Bounia
CNRS, Université Sorbonne Paris Nord, UMR CNRS 7030 LIPN, France
V
Vu Linh Nguyen
CNRS, Université de technologie de Compiègne, UMR CNRS 7253 Heudiasyc, France
Sébastien Destercke
Sébastien Destercke
Researcher in Computer Science, CNRS-Heudiasyc
Artificial intelligenceUncertaintyInformation FusionImprecise probabilityBelief functions
A
Arthur Hoarau
CNRS, Université de technologie de Compiègne, UMR CNRS 7253 Heudiasyc, France