🤖 AI Summary
Medical image segmentation often suffers from reduced accuracy due to highly variable lesion scales and uncertain spatial distributions. To address this, we propose a vision-language collaborative network integrated with multi-scale textual prompts and introduce the first retrieval-segmentation joint learning paradigm. During training, semantically relevant prompts are dynamically retrieved from a self-constructed medical text corpus to enhance visual feature representation; during inference, no textual input is required—enabling seamless cross-modal knowledge transfer and lightweight deployment. The method comprises multi-scale textual encoding, vision-language feature alignment, and end-to-end retrieval-augmented joint training. Extensive experiments on COVID-Xray, COVID-CT, and Kvasir-SEG demonstrate significant improvements over state-of-the-art methods, particularly in segmenting small-sized and irregularly shaped lesions.
📝 Abstract
Accurate segmentation of lesions plays a critical role in medical image analysis and diagnosis. Traditional segmentation approaches that rely solely on visual features often struggle with the inherent uncertainty in lesion distribution and size. To address these issues, we propose STPNet, a Scale-aware Text Prompt Network that leverages vision-language modeling to enhance medical image segmentation. Our approach utilizes multi-scale textual descriptions to guide lesion localization and employs retrieval-segmentation joint learning to bridge the semantic gap between visual and linguistic modalities. Crucially, STPNet retrieves relevant textual information from a specialized medical text repository during training, eliminating the need for text input during inference while retaining the benefits of cross-modal learning. We evaluate STPNet on three datasets: COVID-Xray, COVID-CT, and Kvasir-SEG. Experimental results show that our vision-language approach outperforms state-of-the-art segmentation methods, demonstrating the effectiveness of incorporating textual semantic knowledge into medical image analysis. The code has been made publicly on https://github.com/HUANGLIZI/STPNet.