Improving the Robustness of Distantly-Supervised Named Entity Recognition via Uncertainty-Aware Teacher Learning and Student-Student Collaborative Learning

๐Ÿ“… 2023-11-14
๐Ÿ›๏ธ Annual Meeting of the Association for Computational Linguistics
๐Ÿ“ˆ Citations: 3
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
Distantly supervised named entity recognition (DS-NER) suffers from error propagation due to label noise, and existing teacher-student frameworks are limited by poor teacher model calibration. To address this, we propose an uncertainty-aware teacher learning and studentโ€“student collaborative learning framework. First, we design an uncertainty-driven pseudo-label filtering mechanism based on prediction entropy to enhance pseudo-label reliability. Second, we introduce a mutual distillation paradigm among students using only high-confidence labels, enabling bidirectional error correction. Third, we integrate adaptive pseudo-label weighting with noise-aware self-training. Evaluated on five DS-NER benchmarks, our method achieves new state-of-the-art performance, improving F1 scores by 2.1โ€“4.7 percentage points over prior work. It significantly mitigates error propagation and enhances robustness under weak supervision.
๐Ÿ“ Abstract
Distantly-Supervised Named Entity Recognition (DS-NER) is widely used in real-world scenarios. It can effectively alleviate the burden of annotation by matching entities in existing knowledge bases with snippets in the text but suffer from the label noise. Recent works attempt to adopt the teacher-student framework to gradually refine the training labels and improve the overall robustness. However, these teacher-student methods achieve limited performance because the poor calibration of the teacher network produces incorrectly pseudo-labeled samples, leading to error propagation. Therefore, we propose: (1) Uncertainty-Aware Teacher Learning that leverages the prediction uncertainty to reduce the number of incorrect pseudo labels in the self-training stage; (2) Student-Student Collaborative Learning that allows the transfer of reliable labels between two student networks instead of indiscriminately relying on all pseudo labels from its teacher, and further enables a full exploration of mislabeled samples rather than simply filtering unreliable pseudo-labeled samples. We evaluate our proposed method on five DS-NER datasets, demonstrating that our method is superior to the state-of-the-art DS-NER methods.
Problem

Research questions and friction points this paper is trying to address.

Reducing label noise in distantly-supervised NER via uncertainty-aware learning
Improving pseudo-label accuracy in teacher-student NER frameworks
Enhancing robustness by collaborative student learning and mislabel exploration
Innovation

Methods, ideas, or system contributions that make the work stand out.

Uncertainty-Aware Teacher Learning reduces incorrect pseudo labels
Student-Student Collaborative Learning transfers reliable labels
Method outperforms state-of-the-art DS-NER techniques
๐Ÿ”Ž Similar Papers
No similar papers found.
H
Helan Hu
National Key Laboratory for Multimedia Information Processing, Peking University; School of Software and Microelectronics, Peking University
Shuzheng Si
Shuzheng Si
Tsinghua University
Natural Language ProcessingLarge Language Models
H
Haozhe Zhao
National Key Laboratory for Multimedia Information Processing, Peking University; School of Software and Microelectronics, Peking University
Shuang Zeng
Shuang Zeng
Peking University, Georgia Institute of Technology
Self-supervised Contrastive LearningMedical Image SegmentationSuperpixelLarge Language Model
Kaikai An
Kaikai An
Peking University
Natural Language Processing
Zefan Cai
Zefan Cai
Student, Peking University
Inference AccelerationMulti-Modality
B
Baobao Chang
National Key Laboratory for Multimedia Information Processing, Peking University; Jiangsu Collaborative Innovation Center for Language Ability, Jiangsu Normal University, Xuzhou, China