🤖 AI Summary
To address the performance bottleneck of liver segmentation models caused by scarce annotated medical imaging data, this paper proposes a two-stage cross-modal image synthesis framework that unpairedly translates abdominal CT scans into MRI-like images to augment training data. We introduce EssNet—a deformation-invariant network—designed to overcome anatomical misalignment and distortion induced by bidirectional constraints in CycleGAN-based cross-modal synthesis. Furthermore, we propose, for the first time, a CT→MRI unidirectional enhancement paradigm specifically tailored for segmentation tasks. By fine-tuning a U-Net on a hybrid dataset comprising synthesized MRI and real CT images, our method achieves a 1.17% IoU improvement on public liver segmentation benchmarks. This demonstrates significant alleviation of the small-sample limitation and validates the efficacy and practicality of synthetically generated data for downstream segmentation tasks.
📝 Abstract
Deep learning-based computer-aided diagnosis (CAD) of medical images requires large datasets. However, the lack of large publicly available labeled datasets limits the development of deep learning-based CAD systems. Generative Adversarial Networks (GANs), in particular, CycleGAN, can be used to generate new cross-domain images without paired training data. However, most CycleGAN-based synthesis methods lack the potential to overcome alignment and asymmetry between the input and generated data. We propose a two-stage technique for the synthesis of abdominal MRI using cross-modality translation of abdominal CT. We show that the synthetic data can help improve the performance of the liver segmentation network. We increase the number of abdominal MRI images through cross-modality image transformation of unpaired CT images using a CycleGAN inspired deformation invariant network called EssNet. Subsequently, we combine the synthetic MRI images with the original MRI images and use them to improve the accuracy of the U-Net on a liver segmentation task. We train the U-Net on real MRI images and then on real and synthetic MRI images. Consequently, by comparing both scenarios, we achieve an improvement in the performance of U-Net. In summary, the improvement achieved in the Intersection over Union (IoU) is 1.17%. The results show potential to address the data scarcity challenge in medical imaging.