🤖 AI Summary
Existing relation triple extraction (RTE) methods prioritize performance gains while neglecting intrinsic interpretability and relying on complex preprocessing, resulting in opaque systems and a disconnect between theoretical foundations and practical deployment. To address this, we propose the first slot-attention-driven intrinsically interpretable framework for RTE, formulated as a slot-based set prediction task. Our approach employs differentiable bipartite matching for end-to-end triplet generation and introduces a slot attention mechanism that explicitly links each predicted triplet to its corresponding slot representations and original input tokens, enabling token-level attribution explanations. Evaluated on NYT and WebNLG benchmarks, our method achieves state-of-the-art performance while simultaneously delivering strong interpretability—demonstrating, for the first time, that high accuracy and intrinsic explainability are not mutually exclusive but can be synergistically realized.
📝 Abstract
Relational Triple Extraction (RTE) is a fundamental task in Natural Language Processing (NLP). However, prior research has primarily focused on optimizing model performance, with limited efforts to understand the internal mechanisms driving these models. Many existing methods rely on complex preprocessing to induce specific interactions, often resulting in opaque systems that may not fully align with their theoretical foundations. To address these limitations, we propose SMARTe: a Slot-based Method for Accountable Relational Triple extraction. SMARTe introduces intrinsic interpretability through a slot attention mechanism and frames the task as a set prediction problem. Slot attention consolidates relevant information into distinct slots, ensuring all predictions can be explicitly traced to learned slot representations and the tokens contributing to each predicted relational triple. While emphasizing interpretability, SMARTe achieves performance comparable to state-of-the-art models. Evaluations on the NYT and WebNLG datasets demonstrate that adding interpretability does not compromise performance. Furthermore, we conducted qualitative assessments to showcase the explanations provided by SMARTe, using attention heatmaps that map to their respective tokens. We conclude with a discussion of our findings and propose directions for future research.