Relation Also Knows: Rethinking the Recall and Editing of Factual Associations in Auto-Regressive Transformer Language Models

📅 2024-08-27
🏛️ arXiv.org
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the overgeneralization problem in autoregressive Transformer language models during factual editing, which arises from excessive reliance on subject-centric explanations. We propose a relation-focused knowledge editing paradigm. Methodologically, we first identify the dominant role of relational information in factual recall, introduce the Relation-Specificity (R-Specificity) metric for quantitative evaluation, and design a relation-localization mechanism leveraging attention patterns and MLP activation profiles. Building upon this, we develop a single-fact editing framework integrating targeted weight tuning with constraint-based optimization. Evaluated on our newly constructed R-Specificity benchmark, our approach significantly mitigates overgeneralization while preserving high edit fidelity and generalization capability. Our contribution lies in shifting beyond the conventional subject-centric paradigm, establishing a relation-driven, interpretable, fine-grained, and low-side-effect knowledge editing methodology.

Technology Category

Application Category

📝 Abstract
The storage and recall of factual associations in auto-regressive transformer language models (LMs) have drawn a great deal of attention, inspiring knowledge editing by directly modifying the located model weights. Most editing works achieve knowledge editing under the guidance of existing interpretations of knowledge recall that mainly focus on subject knowledge. However, these interpretations are seriously flawed, neglecting relation information and leading to the over-generalizing problem for editing. In this work, we discover a novel relation-focused perspective to interpret the knowledge recall of transformer LMs during inference and apply it on single knowledge editing to avoid over-generalizing. Experimental results on the dataset supplemented with a new R-Specificity criterion demonstrate that our editing approach significantly alleviates over-generalizing while remaining competitive on other criteria, breaking the domination of subject-focused editing for future research.
Problem

Research questions and friction points this paper is trying to address.

Interpreting knowledge recall in transformer LMs
Avoiding over-generalizing in knowledge editing
Incorporating relation information in model editing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Relation-focused knowledge editing
Avoids over-generalizing problem
Uses R-Specificity criterion
🔎 Similar Papers
No similar papers found.
X
Xiyu Liu
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China; School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China
Z
Zhengxiao Liu
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China; School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China
N
Naibin Gu
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China; School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China
Z
Zheng Lin
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China; School of Cyber Security, University of Chinese Academy of Sciences, Beijing, China
W
Wanli Ma
University of Electronic Science and Technology of China, Chengdu, China
Ji Xiang
Ji Xiang
Institute of Information Engineering, Chinese Academy of Sciences, Beijing, China
Weiping Wang
Weiping Wang
School of Information Science and Engineering, Central South University
Computer NetworkNetwork Security