🤖 AI Summary
To address the practical deployment challenges of model interpretation methods—namely, high query overhead and strong model dependency—this paper proposes a model-agnostic, end-to-end differentiable attribution framework. Methodologically, it reformulates explanation generation as a policy learning problem, representing attributions via a learnable probability distribution and optimizing inference efficiency through policy gradient methods. Crucially, the approach requires neither access to model internals nor gradients, nor does it rely on extensive forward queries. Evaluated on image and text classification tasks, it achieves over 97% faster inference and reduces memory consumption by 70%, while preserving high fidelity and cross-model generalizability. These results demonstrate a significant breakthrough in resolving the long-standing trade-off between computational efficiency and methodological generality in attribution-based explanation.
📝 Abstract
The challenge of delivering efficient explanations is a critical barrier that prevents the adoption of model explanations in real-world applications. Existing approaches often depend on extensive model queries for sample-level explanations or rely on expert's knowledge of specific model structures that trade general applicability for efficiency. To address these limitations, this paper introduces a novel framework Fast Explanation (FEX) that represents attribution-based explanations via probability distributions, which are optimized by leveraging the policy gradient method. The proposed framework offers a robust, scalable solution for real-time, large-scale model explanations, bridging the gap between efficiency and applicability. We validate our framework on image and text classification tasks and the experiments demonstrate that our method reduces inference time by over 97% and memory usage by 70% compared to traditional model-agnostic approaches while maintaining high-quality explanations and broad applicability.