Interpretable Multimodal Cancer Prototyping with Whole Slide Images and Incompletely Paired Genomics

📅 2025-11-26
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
In clinical practice, incomplete or unavailable genomic data severely hinder multimodal precision oncology subtyping. To address this, we propose an interpretable fusion framework integrating whole-slide images (WSIs) with incompletely paired genomic data. Methodologically, our approach introduces biologically grounded prototypical representation, multi-view distribution- and sample-level alignment, bipartite fusion of shared and modality-specific features, and a semantic-prior-guided genomic imputation mechanism. This enables cross-modal collaborative learning and robust inference under missing modalities, significantly enhancing model interpretability and clinical applicability. Extensive experiments across multiple cancer typing and subtyping tasks—as well as genomic data imputation—demonstrate consistent superiority over state-of-the-art methods, validating the framework’s effectiveness, generalizability, and robustness to data incompleteness.

Technology Category

Application Category

📝 Abstract
Multimodal approaches that integrate histology and genomics hold strong potential for precision oncology. However, phenotypic and genotypic heterogeneity limits the quality of intra-modal representations and hinders effective inter-modal integration. Furthermore, most existing methods overlook real-world clinical scenarios where genomics may be partially missing or entirely unavailable. We propose a flexible multimodal prototyping framework to integrate whole slide images and incomplete genomics for precision oncology. Our approach has four key components: 1) Biological Prototyping using text prompting and prototype-wise weighting; 2) Multiview Alignment through sample- and distribution-wise alignments; 3) Bipartite Fusion to capture both shared and modality-specific information for multimodal fusion; and 4) Semantic Genomics Imputation to handle missing data. Extensive experiments demonstrate the consistent superiority of the proposed method compared to other state-of-the-art approaches on multiple downstream tasks. The code is available at https://github.com/helenypzhang/Interpretable-Multimodal-Prototyping.
Problem

Research questions and friction points this paper is trying to address.

Integrates histology and genomics for precision oncology despite data heterogeneity
Handles incomplete or missing genomic data in real-world clinical scenarios
Creates interpretable multimodal representations from whole slide images and genomics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Biological Prototyping with text prompting and prototype-wise weighting
Multiview Alignment through sample- and distribution-wise alignments
Semantic Genomics Imputation to handle missing data
Y
Yupei Zhang
Department of Clinical Neurosciences, University of Cambridge, UK
Y
Yating Huang
Department of Electrical & Electronic Engineering, The University of Manchester, UK
W
Wanming Hu
Department of Pathology, State Key Laboratory of Oncology in South China, Guangdong Provincial Clinical Research Center for Cancer, Sun Yat-sen University Cancer Center, China
Lequan Yu
Lequan Yu
Assistant Professor, The University of Hong Kong
Medical Image AnalysisMultimodal LearningComputational PathologyAI for Healthcare
Hujun Yin
Hujun Yin
School of Electrical and Electronic Engineering, The University of Manchester
Neural networksimage processingface recognitiondimension reductiontime series
C
Chao Li
Department of Clinical Neurosciences and Department of Applied Mathematics and Theoretical Physics, University of Cambridge; School of Science and Engineering and School of Medicine, University of Dundee, UK