🤖 AI Summary
In semi-supervised learning (SSL), distribution mismatch between labeled and unlabeled data exacerbates classification bias induced by class imbalance, yet existing methods conflate class imbalance with intrinsic learning difficulty and apply only coarse-grained corrections. This paper proposes SC-SSL, the first unified SSL framework that explicitly decouples these two factors. SC-SSL introduces a two-stage sampling control mechanism—adaptive resampling in the feature space and explicit classifier expansion in the logits space—complemented by bias-vector-driven logits calibration at inference time. This fine-grained balancing strategy jointly mitigates representation-level and prediction-level biases for minority classes. Extensive experiments across multiple benchmark datasets and diverse distribution shift settings demonstrate that SC-SSL consistently improves minority-class accuracy and overall class-balanced performance, outperforming state-of-the-art methods.
📝 Abstract
Class imbalance remains a critical challenge in semi-supervised learning (SSL), especially when distributional mismatches between labeled and unlabeled data lead to biased classification. Although existing methods address this issue by adjusting logits based on the estimated class distribution of unlabeled data, they often handle model imbalance in a coarse-grained manner, conflating data imbalance with bias arising from varying class-specific learning difficulties. To address this issue, we propose a unified framework, SC-SSL, which suppresses model bias through decoupled sampling control. During training, we identify the key variables for sampling control under ideal conditions. By introducing a classifier with explicit expansion capability and adaptively adjusting sampling probabilities across different data distributions, SC-SSL mitigates feature-level imbalance for minority classes. In the inference phase, we further analyze the weight imbalance of the linear classifier and apply post-hoc sampling control with an optimization bias vector to directly calibrate the logits. Extensive experiments across various benchmark datasets and distribution settings validate the consistency and state-of-the-art performance of SC-SSL.