Handling Label Noise via Instance-Level Difficulty Modeling and Dynamic Optimization

📅 2025-05-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the degradation of deep neural network generalization under label noise, this paper proposes a two-stage robust training framework. First, it introduces a novel “error event” metric to perform fine-grained modeling of sample cleanliness and difficulty. Second, it designs a probabilistic dynamic weighting loss function that enables hyperparameter-free, instance-level adaptive optimization. The method decouples noise identification from robust training, achieving a favorable balance among accuracy, efficiency, and scalability. Extensive experiments on five synthetic and real-world label-noise learning (LNL) benchmarks demonstrate consistent superiority over state-of-the-art methods: average test accuracy improves significantly, training time is reduced by approximately 75%, and model generalization—particularly under realistic deployment conditions—is markedly enhanced.

Technology Category

Application Category

📝 Abstract
Recent studies indicate that deep neural networks degrade in generalization performance under noisy supervision. Existing methods focus on isolating clean subsets or correcting noisy labels, facing limitations such as high computational costs, heavy hyperparameter tuning process, and coarse-grained optimization. To address these challenges, we propose a novel two-stage noisy learning framework that enables instance-level optimization through a dynamically weighted loss function, avoiding hyperparameter tuning. To obtain stable and accurate information about noise modeling, we introduce a simple yet effective metric, termed wrong event, which dynamically models the cleanliness and difficulty of individual samples while maintaining computational costs. Our framework first collects wrong event information and builds a strong base model. Then we perform noise-robust training on the base model, using a probabilistic model to handle the wrong event information of samples. Experiments on five synthetic and real-world LNL benchmarks demonstrate our method surpasses state-of-the-art methods in performance, achieves a nearly 75% reduction in computational time and improves model scalability.
Problem

Research questions and friction points this paper is trying to address.

Addressing noisy supervision in deep neural networks
Reducing computational costs and hyperparameter tuning
Improving model scalability and performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Dynamic weighted loss for instance-level optimization
Wrong event metric for noise modeling
Probabilistic model for robust training
🔎 Similar Papers
No similar papers found.
K
Kuan Zhang
Beijing Institute of Technology
Chengliang Chai
Chengliang Chai
Beijing Institute of Technology
Data cleaning and integration
J
Jingzhe Xu
Beijing Institute of Technology
C
Chi Zhang
Beijing Institute of Technology
Y
Ye Yuan
Beijing Institute of Technology
Guoren Wang
Guoren Wang
Beijing Institute of Technology
Lei Cao
Lei Cao
Assistant Professor, University of Arizona/Research Scientist, MIT CSAIL
DatabasesMachine learning