🤖 AI Summary
Existing text-attributed graph (TAG) anomaly detection methods predominantly rely on shallow text encoders, leading to insufficient semantic contextual modeling. Although large language models (LLMs) offer strong semantic understanding, their fixed input-length constraints hinder effective modeling of higher-order graph structures. To address this, we propose an LLM-GNN collaborative framework: for the first time, we employ multiple LLMs in concert to generate structure-aware semantic evidence; this is integrated with a graph neural network (GNN) to capture topological dependencies and aligned via a gated fusion mechanism that bridges structural and semantic representations. Evaluated on multiple benchmarks, our method achieves an average 13.37% improvement in Average Precision (AP) over state-of-the-art approaches. It delivers both high detection accuracy and model interpretability, empirically validating the effectiveness and necessity of jointly modeling textual semantics and graph structure.
📝 Abstract
The natural combination of intricate topological structures and rich textual information in text-attributed graphs (TAGs) opens up a novel perspective for graph anomaly detection (GAD). However, existing GAD methods primarily focus on designing complex optimization objectives within the graph domain, overlooking the complementary value of the textual modality, whose features are often encoded by shallow embedding techniques, such as bag-of-words or skip-gram, so that semantic context related to anomalies may be missed. To unleash the enormous potential of textual modality, large language models (LLMs) have emerged as promising alternatives due to their strong semantic understanding and reasoning capabilities. Nevertheless, their application to TAG anomaly detection remains nascent, and they struggle to encode high-order structural information inherent in graphs due to input length constraints. For high-quality anomaly detection in TAGs, we propose CoLL, a novel framework that combines LLMs and graph neural networks (GNNs) to leverage their complementary strengths. CoLL employs multi-LLM collaboration for evidence-augmented generation to capture anomaly-relevant contexts while delivering human-readable rationales for detected anomalies. Moreover, CoLL integrates a GNN equipped with a gating mechanism to adaptively fuse textual features with evidence while preserving high-order topological information. Extensive experiments demonstrate the superiority of CoLL, achieving an average improvement of 13.37% in AP. This study opens a new avenue for incorporating LLMs in advancing GAD.