π€ AI Summary
This study addresses the challenges in retinal disease diagnosis posed by the heterogeneity, invasiveness, and registration difficulties of ophthalmic multimodal data. To circumvent the need for real multimodal acquisition, the authors propose a unified framework that constructs quasi-multimodal inputs by synthesizing fluorescein angiography (FFA), multispectral imaging (MSI), and saliency maps. A parallel deep network architecture learns modality-specific features, while a cross-modal adaptive calibration mechanism enables flexible fusion and information pruning, significantly enhancing lesion representation. Evaluated on two public datasets, the method achieves state-of-the-art performance: a multi-label classification F1-score of 0.683 (AUC 0.953) and a diabetic retinopathy grading accuracy of 0.842 (Cohenβs Kappa 0.861), demonstrating both high diagnostic accuracy and clinical practicality.