A Fully Open and Generalizable Foundation Model for Ultrasound Clinical Applications

📅 2025-09-15
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Poor generalizability of clinical ultrasound AI models stems primarily from scarce high-quality annotated data and the task-specific nature of existing architectures. To address this, we propose the first open, general-purpose foundation model tailored for ultrasound imaging. It is pretrained via self-supervised learning on EchoCareData—a large-scale, publicly available dataset comprising multi-center, multi-device, and multi-ethnic ultrasound scans. Our model introduces a novel hierarchical classifier architecture that jointly captures pixel-level local features and representation-level global anatomical structures. The framework supports both downstream fine-tuning and localized adaptation. Evaluated across ten benchmark tasks—including diagnosis, segmentation, detection, quantitative analysis, and report generation—it consistently surpasses state-of-the-art methods, demonstrating plug-and-play, high-performance cross-scenario transferability.

Technology Category

Application Category

📝 Abstract
Artificial intelligence (AI) that can effectively learn ultrasound representations by integrating multi-source data holds significant promise for advancing clinical care. However, the scarcity of large labeled datasets in real-world clinical environments and the limited generalizability of task-specific models have hindered the development of generalizable clinical AI models for ultrasound applications. In this study, we present EchoCare, a novel ultrasound foundation model for generalist clinical use, developed via self-supervised learning on our curated, publicly available, large-scale dataset EchoCareData. EchoCareData comprises 4.5 million ultrasound images, sourced from over 23 countries across 5 continents and acquired via a diverse range of distinct imaging devices, thus encompassing global cohorts that are multi-center, multi-device, and multi-ethnic. Unlike prior studies that adopt off-the-shelf vision foundation model architectures, we introduce a hierarchical classifier into EchoCare to enable joint learning of pixel-level and representation-level features, capturing both global anatomical contexts and local ultrasound characteristics. With minimal training, EchoCare outperforms state-of-the-art comparison models across 10 representative ultrasound benchmarks of varying diagnostic difficulties, spanning disease diagnosis, lesion segmentation, organ detection, landmark prediction, quantitative regression, imaging enhancement and report generation. The code and pretrained model are publicly released, rendering EchoCare accessible for fine-tuning and local adaptation, supporting extensibility to additional applications. EchoCare provides a fully open and generalizable foundation model to boost the development of AI technologies for diverse clinical ultrasound applications.
Problem

Research questions and friction points this paper is trying to address.

Develops a generalizable AI foundation model for ultrasound clinical applications
Addresses scarcity of large labeled datasets and limited model generalizability
Enables diverse tasks like diagnosis, segmentation, and report generation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Self-supervised learning on large-scale ultrasound dataset
Hierarchical classifier for joint feature learning
Open foundation model for diverse clinical applications
🔎 Similar Papers
No similar papers found.
H
Hongyuan Zhang
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China
Yuheng Wu
Yuheng Wu
KAIST
Efficient AIEmbodied Intelligent SystemAutonomous Driving
Mingyang Zhao
Mingyang Zhao
Academy of Mathematics and System Sciences, CAS
Geometric computationArtificial intelligenceStatistics
Z
Zhiwei Chen
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China; Division of Electronic Engineering, Faculty of Engineering, The Chinese University of Hong Kong, Hong Kong, China
Rebecca Li
Rebecca Li
Accident and Emergency Medicine Academic Unit, The Chinese University of Hong Kong, Hong Kong, China
F
Fei Zhu
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China
H
Haohan Zhao
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China; City University of Hong Kong, Hong Kong, China
X
Xiaohua Yuan
Xiangya Hospital Central South University, Changsha, China
M
Meng Yang
Hunan Frontline Medical Technology Co., Ltd, Changsha, China
C
Chunli Qiu
Qilu Hospital of Shandong University, Jinan, China
X
Xiang Cong
Qilu Hospital of Shandong University, Jinan, China
H
Haiyan Chen
Zhongshan Hospital of Fudan University, Shanghai, China
L
Lina Luan
Shanghai Geriatric Medical Center, Shanghai, China
R
Randolph H.L. Wong
Division of Cardiothoracic Surgery, Department of Surgery, The Chinese University of Hong Kong, Hong Kong, China
H
Huai Liao
Department of Pulmonary and Critical Care Medicine, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou, China
C
Colin A Graham
Accident and Emergency Medicine Academic Unit, The Chinese University of Hong Kong, Hong Kong, China
Shi Chang
Shi Chang
Cornell University
active perceptionsensor motion planning
Guowei Tao
Guowei Tao
Qilu Hospital of Shandong University, Jinan, China
Dong Yi
Dong Yi
Hong Kong Institute of Science and Innovation, Chinese Academy of Sciences
Computer VisionPattern Recognition
Zhen Lei
Zhen Lei
Associate Professor, OSCO Research Chair in Off-site Construction
Offsite ConstructionConstruction Engineering and Management
Nassir Navab
Nassir Navab
Professor of Computer Science, Technische Universität München
Sebastien Ourselin
Sebastien Ourselin
Professor of Healthcare Engineering, King's College London
medical imagingmedical image computingmedical image analysisbiomedical image analysis
J
Jiebo Luo
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China; Department of Computer Science, University of Rochester, USA
H
Hongbin Liu
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China; School of Biomedical Engineering & Imaging Sciences, King’s College London, UK
G
Gaofeng Meng
Center for Artificial Intelligence and Robotics, Hong Kong Institute of Science & Innovation, Chinese Academy of Sciences, Hong Kong, China; University of Chinese Academy of Sciences, Beijing, China; State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing, China