🤖 AI Summary
Medical image annotation is costly, and deep learning models suffer severe performance degradation under few-shot settings. To address this, we propose the Dual-Cooperative Self-Training (DCS-ST) framework, which innovatively integrates inter-teacher model consistency constraints with a region-level dynamic pseudo-label selection mechanism. Our method combines self-supervised pretraining, multi-teacher collaborative knowledge distillation, weak-strong data augmentation, and confidence-adaptive threshold adjustment. Evaluated on the Camelyon16/17 benchmarks, DCS-ST achieves 98.2% classification accuracy using only 5% labeled data—outperforming the current state-of-the-art by 3.7 percentage points. This demonstrates significantly improved few-shot generalization capability. The framework provides an efficient and reliable solution for low-resource computational pathology analysis, particularly where labeled data are scarce.
📝 Abstract
Deep learning methods have shown promise in classifying breast cancer histopathology images, but their performance often declines with limited annotated data, a critical challenge in medical imaging due to the high cost and expertise required for annotations.