Molecular-driven Foundation Model for Oncologic Pathology

📅 2025-01-28
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Current foundational models struggle to effectively process gigapixel whole-slide images (WSIs) and integrate multi-source molecular data, limiting computational pathology’s performance in diagnosis and prognostic prediction. To address this, we introduce Threads—the first molecular-driven foundation model for pathology—enabling end-to-end joint modeling of H&E-stained WSIs with genomic and transcriptomic data. Built upon the largest multimodal pretraining dataset to date (47,171 paired samples), Threads incorporates sliding-window adaptive encoding, cross-scale feature alignment, and gene–image joint embedding learning. Evaluated across 54 oncology tasks, Threads consistently outperforms state-of-the-art baselines, markedly improving label efficiency and generalization. Notably, it achieves breakthrough performance on clinically critical tasks—including somatic mutation prediction, immunohistochemistry interpretation, and survival analysis—demonstrating unprecedented integration of histomorphological and molecular information for precision pathology.

Technology Category

Application Category

📝 Abstract
Foundation models are reshaping computational pathology by enabling transfer learning, where models pre-trained on vast datasets can be adapted for downstream diagnostic, prognostic, and therapeutic response tasks. Despite these advances, foundation models are still limited in their ability to encode the entire gigapixel whole-slide images without additional training and often lack complementary multimodal data. Here, we introduce Threads, a slide-level foundation model capable of generating universal representations of whole-slide images of any size. Threads was pre-trained using a multimodal learning approach on a diverse cohort of 47,171 hematoxylin and eosin (H&E)-stained tissue sections, paired with corresponding genomic and transcriptomic profiles - the largest such paired dataset to be used for foundation model development to date. This unique training paradigm enables Threads to capture the tissue's underlying molecular composition, yielding powerful representations applicable to a wide array of downstream tasks. In extensive benchmarking across 54 oncology tasks, including clinical subtyping, grading, mutation prediction, immunohistochemistry status determination, treatment response prediction, and survival prediction, Threads outperformed all baselines while demonstrating remarkable generalizability and label efficiency. It is particularly well suited for predicting rare events, further emphasizing its clinical utility. We intend to make the model publicly available for the broader community.
Problem

Research questions and friction points this paper is trying to address.

Computational pathology
Large-scale whole-slide images
Multimodal data
Innovation

Methods, ideas, or system contributions that make the work stand out.

Threads model
large-scale dataset
predictive performance
🔎 Similar Papers
No similar papers found.
Anurag Vaidya
Anurag Vaidya
MIT
Multimodal frontier modelsagentsclinical deployment
Andrew Zhang
Andrew Zhang
PhD student, Harvard & MIT
computer visionartificial intelligencehealthcaremedical devicesneuroscience
Guillaume Jaume
Guillaume Jaume
Harvard Medical School, Brigham and Women's Hospital
Computational Pathology - Computer Vision
Andrew H. Song
Andrew H. Song
Postdoctoral fellow, Harvard Medical School
Computational pathologyStatistical signal processing
Tong Ding
Tong Ding
PhD student in Computer Science, Harvard University
Representation LearningComputer VisionMultimodal LearningMachine Learning for Health
Sophia J. Wagner
Sophia J. Wagner
Technical University Munich, Helmholtz AI
computational pathologydeep learningcomputer vision
Ming Y. Lu
Ming Y. Lu
MIT EECS, Harvard Medical School
Computational PathologyComputer VisionNatural Language Processing
Paul Doucet
Paul Doucet
Mahmood Lab, Harvard Medical School
H
Harry Robertson
Department of Pathology, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA; Sydney Precision Data Science Center, The University of Sydney, Camperdown, New South Wales, Australia
C
Cristina Almagro-Perez
Department of Pathology, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA; Department of Pathology, Massachusetts General Hospital, Harvard Medical School, Boston, MA; Cancer Program, Broad Institute of Harvard and MIT, Cambridge, MA; Health Sciences and Technology, Harvard-MIT, Cambridge, MA
Richard J. Chen
Richard J. Chen
Brigham and Women's Hospital, Harvard Medical School
Representation LearningComputer VisionComputational Pathology
D
Dina Elharouni
Cancer Program, Broad Institute of Harvard and MIT, Cambridge, MA; Department of Oncologic Pathology, Dana-Farber Cancer Institute, Boston, MA 02215, USA
G
Georges Ayoub
Cancer Program, Broad Institute of Harvard and MIT, Cambridge, MA; Department of Oncologic Pathology, Dana-Farber Cancer Institute, Boston, MA 02215, USA
C
Connor C Bossi
Cancer Program, Broad Institute of Harvard and MIT, Cambridge, MA; Department of Oncologic Pathology, Dana-Farber Cancer Institute, Boston, MA 02215, USA
Keith L. Ligon
Keith L. Ligon
Department of Pathology, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA; Cancer Program, Broad Institute of Harvard and MIT, Cambridge, MA; Department of Oncologic Pathology, Dana-Farber Cancer Institute, Boston, MA 02215, USA; Department of Pathology, Boston Children’s Hospital, Boston, MA 02115, USA
G
Georg K. Gerber
Department of Pathology, Brigham and Women’s Hospital, Harvard Medical School, Boston, MA
L
L. Le
Department of Pathology, Massachusetts General Hospital, Harvard Medical School, Boston, MA
Faisal Mahmood
Faisal Mahmood
Associate Professor, Harvard University