🤖 AI Summary
To address the high computational cost and GPU memory consumption of self-supervised adaptation (SSA) in medical image analysis, this paper proposes Efficient Self-Supervised Adaptation (ESSA), the first systematic exploration of parameter-efficient fine-tuning (PEFT) for SSA. Its core innovation is Attention Projection Layer Adaptation (APLA), a lightweight module integrated into self-supervised pretraining frameworks such as MAE and SimMIM, synergistically combined with techniques like LoRA. Experiments across diverse medical imaging tasks demonstrate that ESSA outperforms both full-parameter SSA and supervised fine-tuning in accuracy. Moreover, it reduces GPU memory usage by 40.1% and increases training throughput by 25.2%, without compromising inference accuracy or efficiency. This work establishes a new paradigm for lightweight, deployable self-supervised transfer learning in medical imaging.
📝 Abstract
Self-supervised adaptation (SSA) improves foundation model transfer to medical domains but is computationally prohibitive. Although parameter efficient fine-tuning methods such as LoRA have been explored for supervised adaptation, their effectiveness for SSA remains unknown. In this work, we introduce efficient self-supervised adaptation (ESSA), a framework that applies parameter-efficient fine-tuning techniques to SSA with the aim of reducing computational cost and improving adaptation performance. Among the methods tested, Attention Projection Layer Adaptation (APLA) sets a new state-of-the-art, consistently surpassing full-parameter SSA and supervised fine-tuning across diverse medical tasks, while reducing GPU memory by up to 40.1% and increasing training throughput by 25.2%, all while maintaining inference efficiency.