DocVCE: Diffusion-based Visual Counterfactual Explanations for Document Image Classification

📅 2025-08-06
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Poor interpretability of black-box document image classifiers and the limited semantic clarity of existing feature importance maps motivate this work. We introduce, for the first time, generative counterfactual explanations to document image classification, proposing a classifier-guided visual counterfactual generation method based on Latent Diffusion Models (LDMs). Our approach performs hierarchical block-wise fine-grained search in the latent space to optimize local patches, yielding high-fidelity counterfactuals that closely adhere to the natural data distribution. The method is agnostic to backbone architectures—supporting ResNet, ConvNeXt, and DiT—and achieves significant improvements in counterfactual validity, proximity, and realism on RVL-CDIP, Tobacco3482, and DocLayNet. By producing semantically meaningful, human-interpretable, and empirically verifiable counterfactuals, our framework enhances transparency and trustworthiness in black-box document AI systems.

Technology Category

Application Category

📝 Abstract
As black-box AI-driven decision-making systems become increasingly widespread in modern document processing workflows, improving their transparency and reliability has become critical, especially in high-stakes applications where biases or spurious correlations in decision-making could lead to serious consequences. One vital component often found in such document processing workflows is document image classification, which, despite its widespread use, remains difficult to explain. While some recent works have attempted to explain the decisions of document image classification models through feature-importance maps, these maps are often difficult to interpret and fail to provide insights into the global features learned by the model. In this paper, we aim to bridge this research gap by introducing generative document counterfactuals that provide meaningful insights into the model's decision-making through actionable explanations. In particular, we propose DocVCE, a novel approach that leverages latent diffusion models in combination with classifier guidance to first generate plausible in-distribution visual counterfactual explanations, and then performs hierarchical patch-wise refinement to search for a refined counterfactual that is closest to the target factual image. We demonstrate the effectiveness of our approach through a rigorous qualitative and quantitative assessment on 3 different document classification datasets -- RVL-CDIP, Tobacco3482, and DocLayNet -- and 3 different models -- ResNet, ConvNeXt, and DiT -- using well-established evaluation criteria such as validity, closeness, and realism. To the best of the authors' knowledge, this is the first work to explore generative counterfactual explanations in document image analysis.
Problem

Research questions and friction points this paper is trying to address.

Explain decisions of document image classification models
Generate plausible visual counterfactual explanations for documents
Improve transparency of black-box AI in document processing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Leverages latent diffusion models for explanations
Uses classifier guidance for counterfactual generation
Performs hierarchical patch-wise refinement
🔎 Similar Papers
No similar papers found.
S
Saifullah Saifullah
Smarte Daten and Wissensdienste (SDS), Deutsches Forschungszentrum für Künstliche Intelligenz GmbH (DFKI), Trippstadter Straße 122, 67663 Kaiserslautern, Germany
S
Stefan Agne
DeepReader GmbH, 67663 Kaiserlautern, Germany
Andreas Dengel
Andreas Dengel
Professor of Computer Science, University of Kaiserslautern & Executive Director, DFKI
Artificial IntelligenceMachine LearningDocument AnalysisSemantic Technologies
Sheraz Ahmed
Sheraz Ahmed
German Research Center for Artificial Intelligence - DFKI GmbH