Zero-Shot Decentralized Federated Learning

📅 2025-09-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing federated prompt learning approaches (e.g., FedCoOp, FedTPG) suffer from poor generalization, high communication overhead, and strong reliance on a central server. To address these limitations, this paper proposes the first fully decentralized zero-shot federated prompt learning framework. Our method eliminates the central coordinator and instead enables asynchronous, low-overhead iterative prompt sharing and aggregation among clients, synergizing CLIP’s zero-shot transfer capability with CoOp-style prompt optimization to achieve adaptive prompt learning in distributed settings. The core innovation lies in the first decentralized prompt collaboration mechanism tailored for zero-shot learning—ensuring both privacy preservation and scalability. Extensive experiments across nine image classification benchmarks demonstrate that our approach matches or surpasses state-of-the-art methods in accuracy, while reducing communication cost by 118× compared to FedTPG.

Technology Category

Application Category

📝 Abstract
CLIP has revolutionized zero-shot learning by enabling task generalization without fine-tuning. While prompting techniques like CoOp and CoCoOp enhance CLIP's adaptability, their effectiveness in Federated Learning (FL) remains an open challenge. Existing federated prompt learning approaches, such as FedCoOp and FedTPG, improve performance but face generalization issues, high communication costs, and reliance on a central server, limiting scalability and privacy. We propose Zero-shot Decentralized Federated Learning (ZeroDFL), a fully decentralized framework that enables zero-shot adaptation across distributed clients without a central coordinator. ZeroDFL employs an iterative prompt-sharing mechanism, allowing clients to optimize and exchange textual prompts to enhance generalization while drastically reducing communication overhead. We validate ZeroDFL on nine diverse image classification datasets, demonstrating that it consistently outperforms--or remains on par with--state-of-the-art federated prompt learning methods. More importantly, ZeroDFL achieves this performance in a fully decentralized setting while reducing communication overhead by 118x compared to FedTPG. These results highlight that our approach not only enhances generalization in federated zero-shot learning but also improves scalability, efficiency, and privacy preservation--paving the way for decentralized adaptation of large vision-language models in real-world applications.
Problem

Research questions and friction points this paper is trying to address.

Enhancing zero-shot learning generalization in federated learning environments
Reducing communication costs and eliminating central server dependency
Enabling decentralized adaptation of vision-language models across distributed clients
Innovation

Methods, ideas, or system contributions that make the work stand out.

Decentralized framework eliminates central server dependency
Iterative prompt-sharing mechanism reduces communication overhead
Enables zero-shot adaptation across distributed clients
🔎 Similar Papers
No similar papers found.
A
Alessio Masano
PeRCeiVe Lab, University of Catania, Italy
M
Matteo Pennisi
PeRCeiVe Lab, University of Catania, Italy
F
Federica Proietto Salanitri
PeRCeiVe Lab, University of Catania, Italy
Concetto Spampinato
Concetto Spampinato
University of Catania
Deep LearningArtificial IntelligenceComputer VisionMedical Image Analysis
G
Giovanni Bellitto
PeRCeiVe Lab, University of Catania, Italy