🤖 AI Summary
To address slow convergence, low accuracy, and excessive communication/computation overhead in federated learning for 6G-native AI—caused by client data heterogeneity and frequent participation—this paper proposes a long-horizon attention-driven client selection algorithm. The method innovatively employs an attention mechanism to quantify model similarity and individual client contribution, jointly modeling global model evolution dynamics; it adaptively adjusts the selection threshold based on a “late-stage prioritization” principle. This enables tight coordination between client selection and global optimization. Experiments on CIFAR-10 demonstrate that, compared to FedAvg, the proposed method reduces client participation by 32% while maintaining comparable accuracy; against state-of-the-art approaches, it improves accuracy by 3.5% and further reduces participating clients by 2%.
📝 Abstract
Native AI support is a key objective in the evolution of 6G networks, with Federated Learning (FL) emerging as a promising paradigm. FL allows decentralized clients to collaboratively train an AI model without directly sharing their data, preserving privacy. Clients train local models on private data and share model updates, which a central server aggregates to refine the global model and redistribute it for the next iteration. However, client data heterogeneity slows convergence and reduces model accuracy, and frequent client participation imposes communication and computational burdens. To address these challenges, we propose extit{FedABC}, an innovative client selection algorithm designed to take a long-term view in managing data heterogeneity and optimizing client participation. Inspired by attention mechanisms, extit{FedABC} prioritizes informative clients by evaluating both model similarity and each model's unique contributions to the global model. Moreover, considering the evolving demands of the global model, we formulate an optimization problem to guide extit{FedABC} throughout the training process. Following the ``later-is-better" principle, extit{FedABC} adaptively adjusts the client selection threshold, encouraging greater participation in later training stages. Extensive simulations on CIFAR-10 demonstrate that extit{FedABC} significantly outperforms existing approaches in model accuracy and client participation efficiency, achieving comparable performance with 32% fewer clients than the classical FL algorithm extit{FedAvg}, and 3.5% higher accuracy with 2% fewer clients than the state-of-the-art. This work marks a step toward deploying FL in heterogeneous, resource-constrained environments, thereby supporting native AI capabilities in 6G networks.