🤖 AI Summary
Graph Neural Networks (GNNs) suffer from limited global interpretability in node classification, especially on large-scale, high-dimensional graphs with low subgraph redundancy—where existing subgraph-pattern-based explanation methods fail. To address this, we propose GnnXemplar: the first framework to leverage reverse k-nearest-neighbor coverage maximization for selecting representative nodes in the GNN embedding space, integrated with a large language model (LLM)-driven self-optimizing prompting strategy to generate human-readable natural-language rules that expose GNN decision logic. This end-to-end approach jointly models structure-attribute interactions while ensuring human comprehensibility. Extensive experiments on multiple benchmark datasets demonstrate superior explanation fidelity, scalability, and readability over state-of-the-art baselines. A user study with 60 participants further confirms the effectiveness and practical utility of GnnXemplar’s explanations.
📝 Abstract
Graph Neural Networks (GNNs) are widely used for node classification, yet their opaque decision-making limits trust and adoption. While local explanations offer insights into individual predictions, global explanation methods, those that characterize an entire class, remain underdeveloped. Existing global explainers rely on motif discovery in small graphs, an approach that breaks down in large, real-world settings where subgraph repetition is rare, node attributes are high-dimensional, and predictions arise from complex structure-attribute interactions. We propose GnnXemplar, a novel global explainer inspired from Exemplar Theory from cognitive science. GnnXemplar identifies representative nodes in the GNN embedding space, exemplars, and explains predictions using natural language rules derived from their neighborhoods. Exemplar selection is framed as a coverage maximization problem over reverse k-nearest neighbors, for which we provide an efficient greedy approximation. To derive interpretable rules, we employ a self-refining prompt strategy using large language models (LLMs). Experiments across diverse benchmarks show that GnnXemplar significantly outperforms existing methods in fidelity, scalability, and human interpretability, as validated by a user study with 60 participants.