FedDuA: Doubly Adaptive Federated Learning

๐Ÿ“… 2025-05-16
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the slow convergence of FedAvg in federated learning under non-IID data and anisotropic parameter spaces, this paper proposes a dual-adaptive optimization framework. The method jointly models cross-client statistical heterogeneity and coordinate-wise gradient heterogeneity at the server side, and introducesโ€” for the first timeโ€”a dual adaptive step-size rule grounded in a mirror descent perspective for theoretical analysis. Without increasing communication rounds or client-side computational overhead, it achieves minimax-optimal step-size adaptation. Theoretically, it guarantees strict convergence for convex objectives. Empirically, it accelerates convergence by 30%โ€“50% over FedAvg across diverse non-IID settings, while exhibiting robustness to hyperparameter selection.

Technology Category

Application Category

๐Ÿ“ Abstract
Federated learning is a distributed learning framework where clients collaboratively train a global model without sharing their raw data. FedAvg is a popular algorithm for federated learning, but it often suffers from slow convergence due to the heterogeneity of local datasets and anisotropy in the parameter space. In this work, we formalize the central server optimization procedure through the lens of mirror descent and propose a novel framework, called FedDuA, which adaptively selects the global learning rate based on both inter-client and coordinate-wise heterogeneity in the local updates. We prove that our proposed doubly adaptive step-size rule is minimax optimal and provide a convergence analysis for convex objectives. Although the proposed method does not require additional communication or computational cost on clients, extensive numerical experiments show that our proposed framework outperforms baselines in various settings and is robust to the choice of hyperparameters.
Problem

Research questions and friction points this paper is trying to address.

Addresses slow convergence in federated learning due to data heterogeneity
Proposes adaptive global learning rate selection for client and coordinate variations
Ensures minimax optimality without extra communication or computational costs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adaptive global learning rate selection
Mirror descent optimization framework
Minimax optimal step-size rule
๐Ÿ”Ž Similar Papers
No similar papers found.