🤖 AI Summary
To address poor model adaptability and overfitting in few-shot transfer learning, this paper proposes a structured output regularization framework. Specifically, we freeze the internal parameters of pretrained backbones (e.g., DenseNet121, EfficientNet-B4) and fine-tune only the top classification layer. Crucially, we jointly impose group Lasso and ℓ₁ regularization on the output-layer weights—thereby explicitly enforcing both inter-channel group sparsity and global weight sparsity—without significantly increasing the number of trainable parameters. This dual regularization enhances discriminative learning of domain-specific patterns while mitigating overfitting. Evaluated on three few-shot medical image classification benchmarks, our method achieves performance on par with state-of-the-art approaches. Results demonstrate its effectiveness, strong generalization capability, and architecture-agnostic scalability across diverse backbone networks.
📝 Abstract
Traditional transfer learning typically reuses large pre-trained networks by freezing some of their weights and adding task-specific layers. While this approach is computationally efficient, it limits the model's ability to adapt to domain-specific features and can still lead to overfitting with very limited data. To address these limitations, we propose Structured Output Regularization (SOR), a simple yet effective framework that freezes the internal network structures (e.g., convolutional filters) while using a combination of group lasso and $L_1$ penalties. This framework tailors the model to specific data with minimal additional parameters and is easily applicable to various network components, such as convolutional filters or various blocks in neural networks enabling broad applicability for transfer learning tasks. We evaluate SOR on three few shot medical imaging classification tasks and we achieve competitive results using DenseNet121, and EfficientNetB4 bases compared to established benchmarks.