🤖 AI Summary
To address the degradation of model generalization in Federated Edge Learning (FEEL) caused by data heterogeneity and resource constraints, this paper proposes a generalization-aware, parameter-efficient joint optimization framework. Methodologically, it innovatively leverages an information-theoretic generalization bound to guide convergence analysis and formulates, for the first time, a generalization-aware optimization problem targeting minimization of the upper bound on the average squared gradient norm—thereby enabling joint optimization of model pruning, client selection, and communication-computation resource allocation. A mixed-integer non-convex model is developed and solved via an alternating optimization algorithm under energy and latency constraints. Experimental results across multiple benchmarks demonstrate that the proposed method significantly reduces the generalization gap, accelerates model convergence, and improves resource utilization efficiency—outperforming state-of-the-art approaches.
📝 Abstract
Federated edge learning (FEEL) provides a promising foundation for edge artificial intelligence (AI) by enabling collaborative model training while preserving data privacy. However, limited and heterogeneous local datasets, as well as resource-constrained deployment, severely degrade both model generalization and resource utilization, leading to a compromised learning performance. Therefore, we propose a parameter-efficient FEEL framework that jointly leverages model pruning and client selection to tackle such challenges. First, we derive an information-theoretic generalization statement that characterizes the discrepancy between training and testing function losses and embed it into the convergence analysis. It reveals that a larger local generalization statement can undermine the global convergence. Then, we formulate a generalization-aware average squared gradient norm bound minimization problem, by jointly optimizing the pruning ratios, client selection, and communication-computation resources under energy and delay constraints. Despite its non-convexity, the resulting mixed-integer problem is efficiently solved via an alternating optimization algorithm. Extensive experiments demonstrate that the proposed design achieves superior learning performance than state-of-the-art baselines, validating the effectiveness of coupling generalization-aware analysis with system-level optimization for efficient FEEL.