DAFTED: Decoupled Asymmetric Fusion of Tabular and Echocardiographic Data for Cardiac Hypertension Diagnosis

📅 2025-09-19
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the challenge of effectively fusing echocardiographic time-series data with clinical tabular data for hypertension diagnosis in cardiac assessment, this paper proposes a decoupled asymmetric multimodal fusion framework. Echocardiography serves as the primary modality, while clinical tabular data act as the auxiliary modality. A feature disentanglement mechanism explicitly separates shared (cross-modal) representations from modality-specific features, and a cross-modal alignment strategy enables synergistic modeling of heterogeneous data. Evaluated on a cohort of 239 patients, the method achieves an AUC of 90.3%, significantly outperforming existing multimodal fusion approaches. Moreover, it offers strong interpretability—via disentangled feature attribution—and robust generalization across diverse clinical subgroups. This work provides a clinically deployable, principled solution for multimodal decision support in cardiovascular diagnostics.

Technology Category

Application Category

📝 Abstract
Multimodal data fusion is a key approach for enhancing diagnosis in medical applications. We propose an asymmetric fusion strategy starting from a primary modality and integrating secondary modalities by disentangling shared and modality-specific information. Validated on a dataset of 239 patients with echocardiographic time series and tabular records, our model outperforms existing methods, achieving an AUC over 90%. This improvement marks a crucial benchmark for clinical use.
Problem

Research questions and friction points this paper is trying to address.

Asymmetric fusion of multimodal medical data
Diagnosing cardiac hypertension from echocardiograms and tabular records
Disentangling shared and modality-specific information
Innovation

Methods, ideas, or system contributions that make the work stand out.

Asymmetric fusion strategy for multimodal data
Disentangling shared and modality-specific information
Primary modality integration with secondary modalities
J
Jérémie Stym-Popper
Sorbonne Université, CNRS, ISIR, MLIA, F-75005 Paris, France
N
Nathan Painchaud
INSA-Lyon, Université Claude Bernard Lyon 1, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69621, France
Clément Rambour
Clément Rambour
ISIR - The Institute of Intelligent Systems and Robotics
VisionVision-Language ModelsUncertaintyinverse problemSAR
P
Pierre-Yves Courand
INSA-Lyon, Université Claude Bernard Lyon 1, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69621, France
Nicolas Thome
Nicolas Thome
Professor of Computer Science, Sorbonne University, France
Machine LearningDeep LearningComputer Vision
Olivier Bernard
Olivier Bernard
CREATIS
Medical Image analysisPopulation characterizationUncertainty estimationUltrasound imaging