An Adaptive Cost-Sensitive Learning and Recursive Denoising Framework for Imbalanced SVM Classification

📅 2024-03-13
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the unreliability in minority-class recognition, model bias, and noise sensitivity caused by class imbalance in sentiment classification, this paper proposes a robust SVM framework. Methodologically, it introduces a dual-module architecture integrating adaptive cost-sensitive learning with recursive denoising. It pioneers a boundary-aware optimization strategy based on dynamic kernel distance and incorporates a minority-class neighborhood-driven noise identification mechanism, combining k-nearest-neighbor-based noise detection, stochastic optimization, and scalable ensemble techniques. Evaluated across multiple benchmark datasets and sentiment tasks with varying imbalance ratios, the framework achieves significant improvements in Accuracy, G-mean, Recall, and F1-score—consistently outperforming standard SVMs as well as mainstream resampling and cost-sensitive approaches. The work offers both theoretical novelty—particularly in boundary-aware kernel adaptation and noise-resilient minority-class modeling—and practical efficacy for imbalanced sentiment analysis.

Technology Category

Application Category

📝 Abstract
Category imbalance is one of the most popular and important issues in the domain of classification. Emotion classification model trained on imbalanced datasets easily leads to unreliable prediction. The traditional machine learning method tends to favor the majority class, which leads to the lack of minority class information in the model. Moreover, most existing models will produce abnormal sensitivity issues or performance degradation. We propose a robust learning algorithm based on adaptive cost-sensitivity and recursive denoising, which is a generalized framework and can be incorporated into most stochastic optimization algorithms. The proposed method uses the dynamic kernel distance optimization model between the sample and the decision boundary, which makes full use of the sample's prior information. In addition, we also put forward an effective method to filter noise, the main idea of which is to judge the noise by finding the nearest neighbors of the minority class. In order to evaluate the strength of the proposed method, we not only carry out experiments on standard datasets but also apply it to emotional classification problems with different imbalance rates (IR). Experimental results show that the proposed general framework is superior to traditional methods in Accuracy, G-mean, Recall and F1-score.
Problem

Research questions and friction points this paper is trying to address.

Imbalanced Data
Classification Accuracy
Minority Class
Innovation

Methods, ideas, or system contributions that make the work stand out.

Imbalanced Learning
Cost-sensitive Optimization
Noise Filtering
🔎 Similar Papers
No similar papers found.
Lu Jiang
Lu Jiang
Research Scientist @ Apple
Generative AIFoundation ModelRobust Deep LearningMultimediaVideo Generation
Q
Qi Wang
College of Sciences, Northeastern University, Shenyang, 110819, People’s Republic of China.; Key Laboratory of Data Analytics and Optimization for Smart Industry, Northeastern University, Shenyang, 110819, People’s Republic of China.
Y
Yuhang Chang
College of Sciences, Northeastern University, Shenyang, 110819, People’s Republic of China.
J
Jianing Song
School of Mechanical Engineering and Automation, Northeastern University, Shenyang, 110819, People’s Republic of China.
H
Haoyue Fu
College of Sciences, Northeastern University, Shenyang, 110819, People’s Republic of China.
Xiaochun Yang
Xiaochun Yang
Professor of Computer Science, Northeastern University, China
database