A General-Purpose Multimodal Foundation Model for Dermatology

📅 2024-10-19
🏛️ arXiv.org
📈 Citations: 3
Influential: 1
📄 PDF
🤖 AI Summary
Existing deep learning models for dermatological multimodal clinical tasks suffer from poor generalizability, heavy reliance on labeled data, and limited robustness across diverse populations. To address these challenges, we propose PanDerm—the first foundation model for comprehensive dermatological multimodal analysis. PanDerm is pretrained on 2 million real-world clinical images spanning four imaging modalities, leveraging self-supervised learning, cross-modal feature alignment, and medical representation learning to construct a unified architecture. It supports 28 downstream tasks, achieving state-of-the-art performance under few-shot settings (using only 5–10% labeled data). For early- and mid-stage melanoma detection, PanDerm surpasses board-certified dermatologists by 10.2% in accuracy; in human-AI collaborative diagnosis, it improves accuracy by 11%. Crucially, PanDerm demonstrates strong generalization across institutions, imaging modalities, and demographic subgroups—including varied skin tones, ages, genders, and anatomical sites.

Technology Category

Application Category

📝 Abstract
Diagnosing and treating skin diseases require advanced visual skills across multiple domains and the ability to synthesize information from various imaging modalities. Current deep learning models, while effective at specific tasks such as diagnosing skin cancer from dermoscopic images, fall short in addressing the complex, multimodal demands of clinical practice. Here, we introduce PanDerm, a multimodal dermatology foundation model pretrained through self-supervised learning on a dataset of over 2 million real-world images of skin diseases, sourced from 11 clinical institutions across 4 imaging modalities. We evaluated PanDerm on 28 diverse datasets covering a range of clinical tasks, including skin cancer screening, phenotype assessment and risk stratification, diagnosis of neoplastic and inflammatory skin diseases, skin lesion segmentation, change monitoring, and metastasis prediction and prognosis. PanDerm achieved state-of-the-art performance across all evaluated tasks, often outperforming existing models even when using only 5-10% of labeled data. PanDerm's clinical utility was demonstrated through reader studies in real-world clinical settings across multiple imaging modalities. It outperformed clinicians by 10.2% in early-stage melanoma detection accuracy and enhanced clinicians' multiclass skin cancer diagnostic accuracy by 11% in a collaborative human-AI setting. Additionally, PanDerm demonstrated robust performance across diverse demographic factors, including different body locations, age groups, genders, and skin tones. The strong results in benchmark evaluations and real-world clinical scenarios suggest that PanDerm could enhance the management of skin diseases and serve as a model for developing multimodal foundation models in other medical specialties, potentially accelerating the integration of AI support in healthcare.
Problem

Research questions and friction points this paper is trying to address.

Develops a multimodal model for diverse dermatology tasks
Addresses limitations of single-task deep learning in clinical practice
Improves diagnostic accuracy across multiple skin conditions and modalities
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal foundation model for dermatology
Self-supervised learning on 2M+ images
Outperforms clinicians in multiple tasks
🔎 Similar Papers
No similar papers found.
Siyuan Yan
Siyuan Yan
Research Fellow@Monash University
AI for MedicineFoundation Model
Zhen Yu
Zhen Yu
School of Translational Medicine & Faculty of IT, Monash University
Digital HealthDermatology AIAging biomarker
Clare Primiero
Clare Primiero
The University of Queensland
C
Cristina Vico-Alonso
Victorian Melanoma Service, Alfred Hospital, Melbourne, Australia; School of Public Health and Preventive Medicine, Monash University, Melbourne, Australia
Z
Zhonghua Wang
AIM for Health Lab, Faculty of Information Technology, Monash University, Melbourne, Australia
L
Litao Yang
AIM for Health Lab, Faculty of Information Technology, Monash University, Melbourne, Australia; Faculty of Engineering, Monash University, Melbourne, Australia
Philipp Tschandl
Philipp Tschandl
Medical University of Vienna
M
Ming Hu
AIM for Health Lab, Faculty of Information Technology, Monash University, Melbourne, Australia; Faculty of Engineering, Monash University, Melbourne, Australia
G
Gin Tan
eResearch Centre, Monash University, Melbourne, Australia
Vincent Tang
Vincent Tang
Lawrence Livermore National Laboratory
Plasma Physics
A
Aik Beng Ng
NVIDIA AI Technology Center, Singapore
D
David Powell
eResearch Centre, Monash University, Melbourne, Australia
P
Paul Bonnington
The University of Queensland, Brisbane, Australia
Simon See
Simon See
nvidia
applied mathematicsAImachine learningHigh Performance ComputingSimulation
Monika Janda
Monika Janda
The University of Queensland and Queensland University of Technology
Cancer researchbehavioral research
Victoria Mar
Victoria Mar
Victorian Melanoma Service, Alfred Hospital, Melbourne, Australia; School of Public Health and Preventive Medicine, Monash University, Melbourne, Australia
Harald Kittler
Harald Kittler
Medical University of Vienna, Austria
Dermatology
H
H. Peter Soyer
Frazer Institute, The University of Queensland, Dermatology Research Centre, Brisbane, Australia; Dermatology Department, Princess Alexandra Hospital, Brisbane, Australia
Z
Zongyuan Ge
AIM for Health Lab, Faculty of Information Technology, Monash University, Melbourne, Australia; Faculty of Engineering, Monash University, Melbourne, Australia