Safer Skin Lesion Classification with Global Class Activation Probability Map Evaluation and SafeML

📅 2025-08-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Although AI-based skin lesion diagnosis achieves higher accuracy than clinicians, its clinical adoption remains limited due to low trust—primarily stemming from inconsistencies and incomplete class coverage in existing explainability methods (e.g., LIME, CAM). To address this, we propose Global Class Activation Probability maps (GCAP), a pixel-level, multi-class, global explanation technique that enables fine-grained, cross-class interpretability of diagnostic decisions. GCAP is integrated with SafeML, a proactive uncertainty-aware mechanism for high-risk prediction alerting. Our framework combines MobileNetV2 and Vision Transformer (ViT) backbones, jointly optimizing class activation mapping and epistemic uncertainty quantification on the ISIC dataset. Experiments demonstrate that GCAP significantly improves explanation consistency and class discriminability, while SafeML reliably identifies high-uncertainty predictions. Together, they reduce misdiagnosis risk and enhance the clinical credibility and safety of AI-assisted dermatological diagnosis.

Technology Category

Application Category

📝 Abstract
Recent advancements in skin lesion classification models have significantly improved accuracy, with some models even surpassing dermatologists' diagnostic performance. However, in medical practice, distrust in AI models remains a challenge. Beyond high accuracy, trustworthy, explainable diagnoses are essential. Existing explainability methods have reliability issues, with LIME-based methods suffering from inconsistency, while CAM-based methods failing to consider all classes. To address these limitations, we propose Global Class Activation Probabilistic Map Evaluation, a method that analyses all classes' activation probability maps probabilistically and at a pixel level. By visualizing the diagnostic process in a unified manner, it helps reduce the risk of misdiagnosis. Furthermore, the application of SafeML enhances the detection of false diagnoses and issues warnings to doctors and patients as needed, improving diagnostic reliability and ultimately patient safety. We evaluated our method using the ISIC datasets with MobileNetV2 and Vision Transformers.
Problem

Research questions and friction points this paper is trying to address.

Addressing distrust in AI skin lesion classification models
Improving reliability of explainable diagnosis methods
Reducing misdiagnosis risks through probabilistic map evaluation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Global Class Activation Probabilistic Map Evaluation method
Pixel-level probabilistic analysis of activation maps
SafeML integration for false diagnosis detection
🔎 Similar Papers
No similar papers found.
Kuniko Paxton
Kuniko Paxton
University of Hull
AI FairnessExplainabilityPrune LearningFederated Learning
Koorosh Aslansefat
Koorosh Aslansefat
Assistant Professor of Computer Science, University of Hull
Reliability and SafetyAI SafetyTrustworthy AIExplainable AI
A
Amila Akagić
Faculty of Electrical Engineering, University of Sarajevo, Zmaja od Bosne bb, Sarajevo, 71000, Bosnia and Herzegovina
D
Dhavalkumar Thakker
School of Computer Science, University of Hull, Cottingham Road, Hull, HU6 7RX, United Kingdom
Yiannis Papadopoulos
Yiannis Papadopoulos
Professor of Computer Science, University of Hull
Safety of AISafe AutonomyMetaheuristicsOptimisationReliability Engineering