🤖 AI Summary
To address excessive communication overhead in bandwidth-constrained edge devices (e.g., mobile terminals and IoT nodes) under federated learning, this paper proposes a client-adaptive skipping mechanism leveraging a lightweight server-side digital twin. Innovatively, it deploys an LSTM-driven digital twin model on the server to jointly predict client update magnitudes and epistemic uncertainty by modeling historical gradient norm sequences. Based on these predictions, the server dynamically decides whether to skip the current local model upload. The mechanism incorporates a threshold-adaptation strategy, achieving 12–15.5% communication reduction under non-IID data while improving model accuracy by up to 0.5 percentage points—outperforming FedAvg significantly. This approach effectively balances communication efficiency and model performance.
📝 Abstract
Communication overhead remains a primary bottleneck in federated learning (FL), particularly for applications involving mobile and IoT devices with constrained bandwidth. This work introduces FedSkipTwin, a novel client-skipping algorithm driven by lightweight, server-side digital twins. Each twin, implemented as a simple LSTM, observes a client's historical sequence of gradient norms to forecast both the magnitude and the epistemic uncertainty of its next update. The server leverages these predictions, requesting communication only when either value exceeds a predefined threshold; otherwise, it instructs the client to skip the round, thereby saving bandwidth. Experiments are conducted on the UCI-HAR and MNIST datasets with 10 clients under a non-IID data distribution. The results demonstrate that FedSkipTwin reduces total communication by 12-15.5% across 20 rounds while simultaneously improving final model accuracy by up to 0.5 percentage points compared to the standard FedAvg algorithm. These findings establish that prediction-guided skipping is a practical and effective strategy for resource-aware FL in bandwidth-constrained edge environments.