Adapting Segment Anything Model 3 for Concept-Driven Lesion Segmentation in Medical Images: An Experimental Study

📅 2026-03-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limited generalization of existing medical image lesion segmentation methods across anatomical sites and imaging modalities. It presents the first systematic investigation of the Segment Anything Model 3 (SAM3) for lesion segmentation in multimodal medical imaging, leveraging flexible guidance through geometric box, textual, and image-concept prompts. To enhance robustness, the approach integrates inter-slice predictions and multi-parameter priors. Extensive experiments across 13 datasets encompassing 11 distinct lesion types demonstrate that SAM3 achieves strong cross-modal generalization, precise boundary delineation, and reliable semantics-driven segmentation performance without requiring extensive fine-tuning, thereby validating its practical potential in clinical settings.
📝 Abstract
Accurate lesion segmentation is essential in medical image analysis, yet most existing methods are designed for specific anatomical sites or imaging modalities, limiting their generalizability. Recent vision-language foundation models enable concept-driven segmentation in natural images, offering a promising direction for more flexible medical image analysis. However, concept-prompt-based lesion segmentation, particularly with the latest Segment Anything Model 3 (SAM3), remains underexplored. In this work, we present a systematic evaluation of SAM3 for lesion segmentation. We assess its performance using geometric bounding boxes and concept-based text and image prompts across multiple modalities, including multiparametric MRI, CT, ultrasound, dermoscopy, and endoscopy. To improve robustness, we incorporate additional prior knowledge, such as adjacent-slice predictions, multiparametric information, and prior annotations. We further compare different fine-tuning strategies, including partial module tuning, adapter-based methods, and full-model optimization. Experiments on 13 datasets covering 11 lesion types demonstrate that SAM3 achieves strong cross-modality generalization, reliable concept-driven segmentation, and accurate lesion delineation. These results highlight the potential of concept-based foundation models for scalable and practical medical image segmentation. Code and trained models will be released at: https://github.com/apple1986/lesion-sam3
Problem

Research questions and friction points this paper is trying to address.

lesion segmentation
medical image analysis
concept-driven segmentation
cross-modality generalization
foundation models
Innovation

Methods, ideas, or system contributions that make the work stand out.

concept-driven segmentation
Segment Anything Model 3 (SAM3)
medical image segmentation
cross-modality generalization
vision-language foundation model
🔎 Similar Papers
No similar papers found.
Guoping Xu
Guoping Xu
UTSW, WIT
Medical Image SegmentationDisease QuantificationComputer Vision
J
Jayaram K. Udupa
Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
Y
Yubing Tong
Medical Image Processing Group, Department of Radiology, University of Pennsylvania, Philadelphia, PA 19104, USA
X
Xin Long
Division of Digestive and Liver Diseases, Department of Internal Medicine, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA
Ying Zhang
Ying Zhang
Tsinghua University, Department of Industrial Engineering
Network optimizationVehicle RoutingHeuristic algorithm
Jie Deng
Jie Deng
Professor, University of Pennsylvania
lymphedemasymptom managementcancer survivorshiponcology nursing
W
Weiguo Lu
The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA
Y
You Zhang
The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA