AnomalyExplainer Explainable AI for LLM-based anomaly detection using BERTViz and Captum

📅 2025-08-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the limited interpretability, high false-positive rates, and low analyst trust in LLM-based anomaly detection systems, this paper proposes a dual-visual explanation framework integrating BERTViz (for visualizing attention flows) and Captum (for feature attribution), coupled with automated natural language explanation generation. The method employs a RoBERTa-based detection model, achieving 99.6% accuracy on the HDFS log dataset—substantially outperforming Falcon-7B, DeBERTa, and Mistral-7B. Experimental results demonstrate significant improvements in explanation readability and credibility; user feedback confirms accelerated anomaly triage and reduced false-positive interference. The core contribution is the first integration of dual-visual attribution with natural language generation, delivering an end-to-end interpretable solution for log-based anomaly detection.

Technology Category

Application Category

📝 Abstract
Conversational AI and Large Language Models (LLMs) have become powerful tools across domains, including cybersecurity, where they help detect threats early and improve response times. However, challenges such as false positives and complex model management still limit trust. Although Explainable AI (XAI) aims to make AI decisions more transparent, many security analysts remain uncertain about its usefulness. This study presents a framework that detects anomalies and provides high-quality explanations through visual tools BERTViz and Captum, combined with natural language reports based on attention outputs. This reduces manual effort and speeds up remediation. Our comparative analysis showed that RoBERTa offers high accuracy (99.6 %) and strong anomaly detection, outperforming Falcon-7B and DeBERTa, as well as exhibiting better flexibility than large-scale Mistral-7B on the HDFS dataset from LogHub. User feedback confirms the chatbot's ease of use and improved understanding of anomalies, demonstrating the ability of the developed framework to strengthen cybersecurity workflows.
Problem

Research questions and friction points this paper is trying to address.

Detecting anomalies in cybersecurity using LLMs with explainability
Reducing false positives and manual effort in threat detection
Providing transparent AI decisions through visual and textual explanations
Innovation

Methods, ideas, or system contributions that make the work stand out.

BERTViz and Captum for explainable AI
RoBERTa model for high accuracy detection
Visual tools with natural language reports
🔎 Similar Papers
No similar papers found.
P
Prasasthy Balasubramanian
Faculty of Information Technology and Electrical Engineering, Center for Ubiquitous Computing, University of Oulu, Oulu, Finland 90570
D
Dumindu Kankanamge
Faculty of Information Technology and Electrical Engineering, Center for Machine Vision and Signal Analysis, University of Oulu, Oulu, Finland 90570
Ekaterina Gilman
Ekaterina Gilman
Researcher, Center for Ubiquitous Computing, University of Oulu
Ubiquitous computingcontext representation and reasoningdata analysisdata-intensive systems
Mourad Oussalah
Mourad Oussalah
University of Oulu
Social mediadata miningroboticsdata fusioncomputer vision