Scaling Tumor Segmentation: Best Lessons from Real and Synthetic Data

πŸ“… 2025-10-16
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Voxel-level annotation scarcity and heavy reliance on expert radiologists hinder abdominal multi-organ tumor segmentation. Method: We propose a real-synthetic data co-training paradigm and introduce AbdomenAtlas 2.0β€”the largest fine-grained annotated abdominal CT dataset to date (10,135 cases, 15,130 tumor instances), fully voxel-labeled by 23 radiologists. We rigorously validate the strong generalization gains of synthetic data in few-shot settings and establish a data-efficient scaling pathway. Results: Our approach achieves +7% Dice improvement on in-distribution test sets and +16% on out-of-distribution sets, substantially outperforming prior public benchmarks. AbdomenAtlas 2.0 establishes a high-quality, reproducible benchmark for cross-organ tumor segmentation and pioneers a novel data-algorithm co-development paradigm grounded in scalable, expert-validated annotation.

Technology Category

Application Category

πŸ“ Abstract
AI for tumor segmentation is limited by the lack of large, voxel-wise annotated datasets, which are hard to create and require medical experts. In our proprietary JHH dataset of 3,000 annotated pancreatic tumor scans, we found that AI performance stopped improving after 1,500 scans. With synthetic data, we reached the same performance using only 500 real scans. This finding suggests that synthetic data can steepen data scaling laws, enabling more efficient model training than real data alone. Motivated by these lessons, we created AbdomenAtlas 2.0--a dataset of 10,135 CT scans with a total of 15,130 tumor instances per-voxel manually annotated in six organs (pancreas, liver, kidney, colon, esophagus, and uterus) and 5,893 control scans. Annotated by 23 expert radiologists, it is several orders of magnitude larger than existing public tumor datasets. While we continue expanding the dataset, the current version of AbdomenAtlas 2.0 already provides a strong foundation--based on lessons from the JHH dataset--for training AI to segment tumors in six organs. It achieves notable improvements over public datasets, with a +7% DSC gain on in-distribution tests and +16% on out-of-distribution tests.
Problem

Research questions and friction points this paper is trying to address.

AI tumor segmentation limited by scarce annotated medical datasets
Synthetic data enhances training efficiency compared to real data
AbdomenAtlas 2.0 provides large-scale annotated dataset for multi-organ tumor segmentation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Combining real and synthetic data for tumor segmentation
Creating AbdomenAtlas 2.0 with 10,135 annotated CT scans
Achieving significant performance gains in distribution tests
πŸ”Ž Similar Papers
No similar papers found.
Q
Qi Chen
Johns Hopkins University
X
Xinze Zhou
Johns Hopkins University
C
Chen Liu
Johns Hopkins University, Hong Kong Polytechnic University
H
Hao Chen
University of Cambridge
Wenxuan Li
Wenxuan Li
Johns Hopkins University
Imaging InformaticsComputer-aided Diagnosis
Zekun Jiang
Zekun Jiang
College of Computer Science & West China Hospital, Sichuan University, China
Medical ImagingBiomedical SignalArtificial IntelligencePrecision Medicine
Z
Ziyan Huang
Shanghai Jiao Tong University, Shanghai AI Laboratory
Y
Yuxuan Zhao
Qilu Hospital of Shandong University
D
Dexin Yu
Qilu Hospital of Shandong University
Junjun He
Junjun He
Shanghai Jiao Tong University
Yefeng Zheng
Yefeng Zheng
Professor, Westlake University, Hangzhou, China, IEEE Fellow, AIMBE Fellow
AI in HealthMedical ImagingComputer VisionNatural Language ProcessingLarge Language Model
A
Alan L. Yuille
Johns Hopkins University
Z
Zongwei Zhou
Johns Hopkins University, UCAS-Terminus AI Lab, University of Chinese Academy of Sciences