🤖 AI Summary
Existing post-hoc image classification explanation methods struggle to simultaneously achieve high faithfulness and plausibility. This paper proposes a model-agnostic natural language explanation pipeline that requires no modification to the original model training. First, it localizes critical neurons via CNN feature attribution and generates activation maps to construct structured semantic representations. Subsequently, a large language model (LLM) translates these representations into human-readable textual explanations. To our knowledge, this is the first method to significantly improve faithfulness while preserving high plausibility. Neural intervention experiments demonstrate that its critical-neuron masking efficacy is three times that of baseline approaches. The method achieves state-of-the-art performance in both human evaluation and automated faithfulness metrics (e.g., deletion/insertion scores), establishing a novel paradigm for post-hoc interpretability in vision models.
📝 Abstract
Existing explanation methods for image classification struggle to provide faithful and plausible explanations. This paper addresses this issue by proposing a post-hoc natural language explanation method that can be applied to any CNN-based classifier without altering its training process or affecting predictive performance. By analysing influential neurons and the corresponding activation maps, the method generates a faithful description of the classifier's decision process in the form of a structured meaning representation, which is then converted into text by a language model. Through this pipeline approach, the generated explanations are grounded in the neural network architecture, providing accurate insight into the classification process while remaining accessible to non-experts. Experimental results show that the NLEs constructed by our method are significantly more plausible and faithful. In particular, user interventions in the neural network structure (masking of neurons) are three times more effective than the baselines.