StealthMark: Harmless and Stealthy Ownership Verification for Medical Segmentation via Uncertainty-Guided Backdoors

📅 2026-01-23
🏛️ IEEE Transactions on Image Processing
📈 Citations: 0
✨ Influential: 0
📄 PDF
🤖 AI Summary
This work proposes a novel, stealthy watermarking method for medical image segmentation models that simultaneously ensures imperceptibility, harmlessness, and black-box verifiability—capabilities lacking in existing ownership protection schemes. By leveraging uncertainty-guided backdoors, the approach modulates prediction uncertainty and integrates model-agnostic explanation techniques (e.g., LIME) to extract feature attributions, thereby embedding identifiable QR code watermarks without altering segmentation outputs. To the best of our knowledge, this is the first method to introduce harmless, covert watermarks into medical segmentation models, enabling efficient ownership verification in black-box settings. Extensive experiments across four medical datasets and five state-of-the-art models—including SAM—demonstrate over 95% watermark verification success rates, with negligible performance degradation (<1% drop in both Dice and AUC), significantly outperforming current backdoor-based watermarking approaches.

Technology Category

Application Category

📝 Abstract
Annotating medical data for training AI models is often costly and limited due to the shortage of specialists with relevant clinical expertise. This challenge is further compounded by privacy and ethical concerns associated with sensitive patient information. As a result, well-trained medical segmentation models on private datasets constitute valuable intellectual property requiring robust protection mechanisms. Existing model protection techniques primarily focus on classification and generative tasks, while segmentation models—crucial to medical image analysis—remain largely underexplored. In this paper, we propose a novel, stealthy, and harmless method, StealthMark, for verifying the ownership of medical segmentation models under closed-box conditions. Our approach subtly modulates model uncertainty without altering the final segmentation outputs, thereby preserving the model’s performance. To enable ownership verification, we incorporate model-agnostic explanation methods, e.g. LIME, to extract feature attributions from the model outputs. Under specific triggering conditions, these explanations reveal a distinct and verifiable watermark. We further design the watermark as a QR code to facilitate robust and recognizable ownership claims. We conducted extensive experiments across four medical imaging datasets (CMR dataset from UK Biobank, the SEG fundus dataset, the EchoNet echocardiography dataset, and the PraNet colonoscopy dataset) and five mainstream segmentation models. The results demonstrate the effectiveness, stealthiness, and harmlessness of our method on the original model’s segmentation performance. For example, when applied to the SAM model, StealthMark consistently achieved attack success rates (ASR) above 95% across various datasets while maintaining less than a 1% drop in Dice and AUC scores—significantly outperforming backdoor-based watermarking methods and highlighting its strong potential for practical deployment. Our implementation code is made available at https://github.com/Qinkaiyu/StealthMark
Problem

Research questions and friction points this paper is trying to address.

medical segmentation
model ownership verification
intellectual property protection
stealthy watermarking
black-box verification
Innovation

Methods, ideas, or system contributions that make the work stand out.

uncertainty-guided backdoor
model watermarking
medical image segmentation
black-box ownership verification
explainability-based watermark
🔎 Similar Papers
No similar papers found.
Qinkai Yu
Qinkai Yu
University of Exeter
Medical Image AnalysisComputer VisionLarge Language Models
Chong Zhang
Chong Zhang
University of Liverpool
LLM SafetyExplainability for LLMMulti-Agent Systems
Gaojie Jin
Gaojie Jin
Lecturer (Assistant Professor), University of Exeter
Machine LearningStatistical LearningTrustworthy AIHuman-GenAI-Alignment
Tianjin Huang
Tianjin Huang
Asst. Professor, CS@University of Exeter & Researcher Fellow, CS@TU/e
LLMsAdversarial examplesStable TrainingGraph Neural NetworkSparse Training
Wei Zhou
Wei Zhou
Huazhong University of Science and Technology
IoT SecuritySystem SecurityHardware Security
Wenhui Li
Wenhui Li
National Institute of Biological Sciences,Beijing
X
Xiaobo Jin
School of Advanced Technology, Xi’an Jiaotong-Liverpool University, Suzhou, China
Bo Huang
Bo Huang
Chongqing University
Computer VisionDeep Learning
Yitian Zhao
Yitian Zhao
Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences
Medical Imagingcomputer visionpattern recognition
G
Guang Yang
School of Bioengineering, Imperial College London, London, UK
G
Gregory Y. H. Lip
Liverpool Centre for Cardiovascular Science at University of Liverpool, Liverpool John Moores University and Liverpool Heart & Chest Hospital, Liverpool, United Kingdom
Yalin Zheng
Yalin Zheng
University of Liverpool
image processingcomputer visionmachine learning and medical image analysis
Aline Villavicencio
Aline Villavicencio
University of Exeter and University of Sheffield (UK)
Natural Language ProcessingMultiword ExpressionsLanguage Acquisition and Evolution
Yanda Meng
Yanda Meng
University of Exeter
Medical Image Analysis