DOREMI: Optimizing Long Tail Predictions in Document-Level Relation Extraction

📅 2026-01-01
🏛️ Knowledge-Based Systems
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the challenge of long-tailed relation distributions in document-level relation extraction (DocRE), where scarce samples of rare relations hinder model generalization. To mitigate this issue, the authors propose a lightweight iterative active learning framework that selectively acquires the most informative samples for annotation with minimal labeling cost, without relying on large-scale noisy data or complex denoising mechanisms. The framework is designed to seamlessly integrate with any existing DocRE model, effectively alleviating long-tail bias. By iteratively refining the model through strategically chosen examples, the approach not only enhances performance on rare relations but also improves overall training efficiency.

Technology Category

Application Category

Problem

Research questions and friction points this paper is trying to address.

Document-Level Relation Extraction
long-tail distribution
relation types
data scarcity
underrepresented relations
Innovation

Methods, ideas, or system contributions that make the work stand out.

Document-Level Relation Extraction
Long-Tail Optimization
Active Learning
Iterative Framework
Data Efficiency