Translation Consistent Semi-Supervised Segmentation for 3D Medical Images

📅 2022-03-28
🏛️ IEEE Transactions on Medical Imaging
📈 Citations: 15
Influential: 0
📄 PDF
🤖 AI Summary
To address the high cost of voxel-level annotations in 3D medical image segmentation, this paper proposes TraCoCo, a semi-supervised framework. Unlike mainstream consistency-based methods relying on spatially invariant perturbations (e.g., noise addition or cropping), TraCoCo introduces **spatial translation perturbations**, which deliberately disrupt background contextual consistency and thereby compel the model to focus on foreground structures. Furthermore, it designs a **translation-consistent co-training mechanism** and a **confidence-region cross-entropy (CRC) loss**, jointly improving pseudo-label quality and training stability. Evaluated on four major benchmarks—Left Atrium (LA), Pancreas, BraTS19, and ACDC—TraCoCo achieves state-of-the-art performance. The source code and pre-trained models are publicly available.
📝 Abstract
3D medical image segmentation methods have been successful, but their dependence on large amounts of voxel-level annotated data is a disadvantage that needs to be addressed given the high cost to obtain such annotation. Semi-supervised learning (SSL) solves this issue by training models with a large unlabelled and a small labelled dataset. The most successful SSL approaches are based on consistency learning that minimises the distance between model responses obtained from perturbed views of the unlabelled data. These perturbations usually keep the spatial input context between views fairly consistent, which may cause the model to learn segmentation patterns from the spatial input contexts instead of the foreground objects. In this paper, we introduce the Translation Consistent Co-training (TraCoCo) which is a consistency learning SSL method that perturbs the input data views by varying their spatial input context, allowing the model to learn segmentation patterns from foreground objects. Furthermore, we propose a new Confident Regional Cross entropy (CRC) loss, which improves training convergence and keeps the robustness to co-training pseudo-labelling mistakes. Our method yields state-of-the-art (SOTA) results for several 3D data benchmarks, such as the Left Atrium (LA), Pancreas-CT (Pancreas), and Brain Tumor Segmentation (BraTS19). Our method also attains best results on a 2D-slice benchmark, namely the Automated Cardiac Diagnosis Challenge (ACDC), further demonstrating its effectiveness. Our code, training logs and checkpoints are available at https://github.com/yyliu01/TraCoCo.
Problem

Research questions and friction points this paper is trying to address.

Reducing reliance on costly voxel-level annotated 3D medical images
Improving semi-supervised learning by perturbing spatial input context
Enhancing training convergence with Cross-model confident Binary Cross entropy loss
Innovation

Methods, ideas, or system contributions that make the work stand out.

Translation Consistent Co-training for spatial context variation
Cross-model confident Binary Cross entropy loss
3D SSL CutMix augmentation for improved generalization
🔎 Similar Papers
No similar papers found.
Y
Yuyuan Liu
Australian Institute for Machine Learning, University of Adelaide, SA 5000, AU
Y
Yu Tian
Harvard Ophtalmology AI Lab of Schepens Eye Research Institute, Harvard Medical School, Boston 02114, US
C
Chong Wang
College of Electronic and Information Engineering, Tongji University, Shanghai 201804, CN
Yuanhong Chen
Yuanhong Chen
University of Adelaide
Computer VisionMedical Image AnalysisMulti-modal
Fengbei Liu
Fengbei Liu
Cornell University
Computer VisionMedical Image Analysis
Vasileios Belagiannis
Vasileios Belagiannis
Professor, Friedrich-Alexander-Universität Erlangen-Nürnberg
Machine LearningComputer VisionRobotics
G
G. Carneiro
Centre for Vision, Speech and Signal Processing (CVSSP), University of Surrey, UK