🤖 AI Summary
This work addresses the challenge of insufficient client personalization performance in multimodal federated learning, which arises from heterogeneous data modalities, task discrepancies, and inconsistent model architectures across clients. To tackle this issue, the authors propose FedAFD, a unified framework that integrates a dual-level adversarial alignment and granularity-aware fusion module at the client side, along with a similarity-guided ensemble knowledge distillation mechanism at the server. By jointly leveraging adversarial alignment, adaptive fusion, and distillation strategies, FedAFD effectively bridges modality and task gaps while accommodating model heterogeneity. Experimental results demonstrate that FedAFD consistently outperforms existing methods under both IID and non-IID settings, significantly enhancing both client-level personalization and global model efficiency.
📝 Abstract
Multimodal Federated Learning (MFL) enables clients with heterogeneous data modalities to collaboratively train models without sharing raw data, offering a privacy-preserving framework that leverages complementary cross-modal information. However, existing methods often overlook personalized client performance and struggle with modality/task discrepancies, as well as model heterogeneity. To address these challenges, we propose FedAFD, a unified MFL framework that enhances client and server learning. On the client side, we introduce a bi-level adversarial alignment strategy to align local and global representations within and across modalities, mitigating modality and task gaps. We further design a granularity-aware fusion module to integrate global knowledge into the personalized features adaptively. On the server side, to handle model heterogeneity, we propose a similarity-guided ensemble distillation mechanism that aggregates client representations on shared public data based on feature similarity and distills the fused knowledge into the global model. Extensive experiments conducted under both IID and non-IID settings demonstrate that FedAFD achieves superior performance and efficiency for both the client and the server.