ULFine: Unbiased Lightweight Fine-tuning for Foundation-Model-Assisted Long-Tailed Semi-Supervised Learning

📅 2025-05-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the poor performance of tail classes in long-tailed semi-supervised learning (LTSSL). Existing CLIP-based linear probing or lightweight fine-tuning methods suffer from dual biases—pseudo-label bias and classifier bias—severely limiting tail-class accuracy. To tackle this, we propose ULFine, an unbiased lightweight fine-tuning framework. ULFine introduces a confidence-aware text prototype adaptive fitting mechanism that explicitly decouples and suppresses both biases. Furthermore, it employs a dual-logits complementary fusion strategy to enhance discriminability for tail classes. While incurring extremely low training overhead—over 10× less than state-of-the-art methods—ULFine significantly improves both tail-class accuracy and overall prediction performance. To our knowledge, it is the first lightweight fine-tuning paradigm for LTSSL that systematically mitigates inherent biases in foundation models.

Technology Category

Application Category

📝 Abstract
Based on the success of large-scale visual foundation models like CLIP in various downstream tasks, this paper initially attempts to explore their impact on Long-Tailed Semi-Supervised Learning (LTSSL) by employing the foundation model with three strategies: Linear Probing (LP), Lightweight Fine-Tuning (LFT), and Full Fine-Tuning (FFT). Our analysis presents the following insights: i) Compared to LTSSL algorithms trained from scratch, FFT results in a decline in model performance, whereas LP and LFT, although boosting overall model performance, exhibit negligible benefits to tail classes. ii) LP produces numerous false pseudo-labels due to extit{underlearned} training data, while LFT can reduce the number of these false labels but becomes overconfident about them owing to extit{biased fitting} training data. This exacerbates the pseudo-labeled and classifier biases inherent in LTSSL, limiting performance improvement in the tail classes. With these insights, we propose a Unbiased Lightweight Fine-tuning strategy, extbf{ULFine}, which mitigates the overconfidence via confidence-aware adaptive fitting of textual prototypes and counteracts the pseudo-labeled and classifier biases via complementary fusion of dual logits. Extensive experiments demonstrate that ULFine markedly decreases training costs by over ten times and substantially increases prediction accuracies compared to state-of-the-art methods.
Problem

Research questions and friction points this paper is trying to address.

Explores foundation models' impact on Long-Tailed Semi-Supervised Learning
Identifies biases in Lightweight Fine-Tuning and Linear Probing methods
Proposes ULFine to reduce biases and improve tail class performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Unbiased Lightweight Fine-tuning for LTSSL
Confidence-aware adaptive textual prototypes
Complementary fusion of dual logits
🔎 Similar Papers
No similar papers found.
Enhao Zhang
Enhao Zhang
University of Washington
databasesvideo analyticshuman-computer interaction
C
Chaohua Li
MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, China and College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics (NUAA), Nanjing 211106, China
Chuanxing Geng
Chuanxing Geng
Nanjing University of Aeronautics and Astronautics
Machine LearningPattern Recognition
Songcan Chen
Songcan Chen
Nanjing University of Aeronautics & Astronautics
Machine LearningPattern recognition