Effective Multi-Task Learning for Biomedical Named Entity Recognition

📅 2025-07-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Biomedical named entity recognition (BM-NER) faces challenges including terminological complexity, pervasive nested entity structures, and inconsistent annotations across datasets. To address these, we propose SRU-NER: a framework featuring a Slot-based Recurrent Unit (SRU) that explicitly models nested entities; multi-task learning to jointly train on heterogeneous biomedical and general-domain NER datasets; and a novel dynamic loss adjustment mechanism that automatically suppresses supervision signals for unannotated entity types—effectively mitigating negative transfer induced by annotation discrepancies. Crucially, SRU-NER enables unified modeling of both biomedical and general-domain NER. Experimental results demonstrate state-of-the-art or competitive performance across multiple benchmark datasets. Cross-domain evaluation and human assessment further validate its strong generalization capability and practical effectiveness.

Technology Category

Application Category

📝 Abstract
Biomedical Named Entity Recognition presents significant challenges due to the complexity of biomedical terminology and inconsistencies in annotation across datasets. This paper introduces SRU-NER (Slot-based Recurrent Unit NER), a novel approach designed to handle nested named entities while integrating multiple datasets through an effective multi-task learning strategy. SRU-NER mitigates annotation gaps by dynamically adjusting loss computation to avoid penalizing predictions of entity types absent in a given dataset. Through extensive experiments, including a cross-corpus evaluation and human assessment of the model's predictions, SRU-NER achieves competitive performance in biomedical and general-domain NER tasks, while improving cross-domain generalization.
Problem

Research questions and friction points this paper is trying to address.

Handling nested biomedical named entities
Integrating multiple datasets via multi-task learning
Improving cross-domain generalization in NER
Innovation

Methods, ideas, or system contributions that make the work stand out.

SRU-NER handles nested named entities
Dynamic loss adjustment for annotation gaps
Multi-task learning integrates multiple datasets
🔎 Similar Papers
No similar papers found.
J
João Ruano
Priberam Labs, Alameda D. Afonso Henriques, 41, 2º, 1000-123 Lisboa, Portugal
Gonçalo M. Correia
Gonçalo M. Correia
Priberam Labs
Machine LearningDeep LearningRepresentation Learning
L
Leonor Barreiros
Priberam Labs, Alameda D. Afonso Henriques, 41, 2º, 1000-123 Lisboa, Portugal
Afonso Mendes
Afonso Mendes
Priberam Informática
NLPMachine LearningDeep Learning