PAST: A multimodal single-cell foundation model for histopathology and spatial transcriptomics in cancer

πŸ“… 2025-07-08
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Current pathological foundation models lack deep integration with single-cell molecular data, limiting their utility in precision oncology. To address this, we propose the first cross-modal pan-cancer single-cell foundation model, enabling unified representation learning of tissue histopathology images and single-cell transcriptomes at cellular resolution. Our method employs a multimodal deep neural network to jointly encode 20 million image–gene expression pairs, integrating contrastive learning and cross-modal attention mechanisms for robust modality alignment. The model supports direct prediction of single-cell gene expression from routine H&E-stained whole-slide images, generation of virtual molecular staining maps, and multimodal survival analysis. It significantly outperforms state-of-the-art methods across multiple cancer types, demonstrating strong generalizability and label-free predictive capability. This work establishes a novel paradigm for high-resolution spatial omics and mechanistic investigation of tumorigenesis.

Technology Category

Application Category

πŸ“ Abstract
While pathology foundation models have transformed cancer image analysis, they often lack integration with molecular data at single-cell resolution, limiting their utility for precision oncology. Here, we present PAST, a pan-cancer single-cell foundation model trained on 20 million paired histopathology images and single-cell transcriptomes spanning multiple tumor types and tissue contexts. By jointly encoding cellular morphology and gene expression, PAST learns unified cross-modal representations that capture both spatial and molecular heterogeneity at the cellular level. This approach enables accurate prediction of single-cell gene expression, virtual molecular staining, and multimodal survival analysis directly from routine pathology slides. Across diverse cancers and downstream tasks, PAST consistently exceeds the performance of existing approaches, demonstrating robust generalizability and scalability. Our work establishes a new paradigm for pathology foundation models, providing a versatile tool for high-resolution spatial omics, mechanistic discovery, and precision cancer research.
Problem

Research questions and friction points this paper is trying to address.

Lack of integration between pathology images and molecular data
Need for unified cross-modal representation at cellular level
Improving precision oncology through multimodal single-cell analysis
Innovation

Methods, ideas, or system contributions that make the work stand out.

Multimodal model integrates histopathology and transcriptomics
Joint encoding of morphology and gene expression
Enables gene prediction and virtual staining
πŸ”Ž Similar Papers
No similar papers found.
Changchun Yang
Changchun Yang
KAUST
medical image analysisimaging science
H
Haoyang Li
Computer Science Program, Computer, Electrical and Mathematical Sciences and Engineering Division, King Abdullah University of Science and Technology (KAUST), Thuwal 23955-6900, Kingdom of Saudi Arabia
Y
Yushuai Wu
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University
Yilan Zhang
Yilan Zhang
King Abdullah University of Science and Technology
Computer VisionMedical Image Analysis
Y
Yifeng Jiao
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University
Y
Yu Zhang
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University
Rihan Huang
Rihan Huang
Computer Science Program, Computer, Electrical and Mathematical Sciences and Engineering Division, King Abdullah University of Science and Technology (KAUST), Thuwal 23955-6900, Kingdom of Saudi Arabia
Y
Yuan Cheng
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University
Y
Yuan Qi
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University; Zhongshan Hospital, Fudan University
X
Xin Guo
Shanghai Academy of Artificial Intelligence for Science; Artificial Intelligence Innovation and Incubation Institute, Fudan University
X
Xin Gao
Computer Science Program, Computer, Electrical and Mathematical Sciences and Engineering Division, King Abdullah University of Science and Technology (KAUST), Thuwal 23955-6900, Kingdom of Saudi Arabia; Shanghai Academy of Artificial Intelligence for Science