Beyond Local Edits: Embedding-Virtualized Knowledge for Broader Evaluation and Preservation of Model Editing

πŸ“… 2026-02-02
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Current evaluation methods for knowledge editing in large language models are constrained by limited datasets, making it difficult to comprehensively characterize the impact of edits on the model’s overall knowledge system. This work proposes the Embedding-Virtualized Knowledge (EVK) framework, which, for the first time, constructs virtualized knowledge regions in embedding space through controlled perturbations to quantify knowledge drift and enable broader assessment of editing effects. Building upon this, we introduce a plug-in EVK-Align module that significantly enhances knowledge retention without compromising editing accuracy. Leveraging the EVK framework, we develop EVK-Bench, a new evaluation benchmark that uncovers editing side effects overlooked by conventional metrics, thereby establishing a more comprehensive and fine-grained paradigm for evaluating knowledge editing.

Technology Category

Application Category

πŸ“ Abstract
Knowledge editing methods for large language models are commonly evaluated using predefined benchmarks that assess edited facts together with a limited set of related or neighboring knowledge. While effective, such evaluations remain confined to finite, dataset-bounded samples, leaving the broader impact of editing on the model's knowledge system insufficiently understood. To address this gap, we introduce Embedding-Virtualized Knowledge (EVK) that characterizes model knowledge through controlled perturbations in embedding space, enabling the exploration of a substantially broader and virtualized knowledge region beyond explicit data annotations. Based on EVK, we construct an embedding-level evaluation benchmark EVK-Bench that quantifies potential knowledge drift induced by editing, revealing effects that are not captured by conventional sample-based metrics. Furthermore, we propose a plug-and-play EVK-Align module that constrains embedding-level knowledge drift during editing and can be seamlessly integrated into existing editing methods. Experiments demonstrate that our approach enables more comprehensive evaluation while significantly improving knowledge preservation without sacrificing editing accuracy.
Problem

Research questions and friction points this paper is trying to address.

knowledge editing
large language models
knowledge preservation
evaluation benchmark
embedding space
Innovation

Methods, ideas, or system contributions that make the work stand out.

Embedding-Virtualized Knowledge
knowledge editing
knowledge drift
EVK-Bench
EVK-Align
πŸ”Ž Similar Papers
No similar papers found.