๐ค AI Summary
In decentralized deep learning under distribution shift, collaborator selection remains challenging when no data exchange is permitted. Method: This paper conducts the first systematic empirical evaluation of three similarity metricsโgradient-, feature-, and output-distribution-based (e.g., cosine similarity, MMD, KL divergence)โfor model aggregation under non-IID data, within a Decentralized SGD framework and across heterogeneous datasets. Experiments span CIFAR-10/100, Tiny-ImageNet, and medical imaging benchmarks. Contribution/Results: Feature-space similarity demonstrates superior robustness, yielding average accuracy gains of 3.2โ7.8% and significantly reducing collaboration failure rates. Based on these findings, we propose an interpretable, reusable metric selection guideline that provides both theoretical grounding and practical guidance for partner identification in decentralized learning.
๐ Abstract
Decentralized Learning (DL) enables privacy-preserving collaboration among organizations or users to enhance the performance of local deep learning models. However, model aggregation becomes challenging when client data is heterogeneous, and identifying compatible collaborators without direct data exchange remains a pressing issue. In this paper, we investigate the effectiveness of various similarity metrics in DL for identifying peers for model merging, conducting an empirical analysis across multiple datasets with distribution shifts. Our research provides insights into the performance of these metrics, examining their role in facilitating effective collaboration. By exploring the strengths and limitations of these metrics, we contribute to the development of robust DL methods.