🤖 AI Summary
In federated learning (FL), fixing the global training round count often leads to premature termination or redundant iterations, resulting in resource inefficiency. To address this, we propose the first generative-AI–based zero-shot synthetic validation method, enabling dynamic, client-level model performance assessment without access to real validation data and supporting adaptive early stopping. Our core innovation lies in leveraging pre-trained generative models—such as diffusion models or large language models—to synthesize high-fidelity, class-balanced validation samples locally in a zero-shot manner, coupled with a lightweight performance monitoring mechanism for real-time convergence detection. The method integrates seamlessly into mainstream FL frameworks. Evaluated on multi-label chest X-ray classification, it reduces total communication rounds by 74% while incurring less than 1% accuracy degradation, significantly enhancing training efficiency and deployment flexibility.
📝 Abstract
Federated Learning (FL) enables collaborative model training across decentralized devices while preserving data privacy. However, FL methods typically run for a predefined number of global rounds, often leading to unnecessary computation when optimal performance is reached earlier. In addition, training may continue even when the model fails to achieve meaningful performance. To address this inefficiency, we introduce a zero-shot synthetic validation framework that leverages generative AI to monitor model performance and determine early stopping points. Our approach adaptively stops training near the optimal round, thereby conserving computational resources and enabling rapid hyperparameter adjustments. Numerical results on multi-label chest X-ray classification demonstrate that our method reduces training rounds by up to 74% while maintaining accuracy within 1% of the optimal.