🤖 AI Summary
This work addresses three key challenges in multi-objective optimization (MOO) for federated learning (FL): excessive communication overhead, limited convergence analysis under realistic assumptions, and inadequate preference modeling. To this end, we propose FedCMOO—a communication-efficient MOO algorithm for FL. FedCMOO is the first to achieve target-count-agnostic gradient compression, enabling collaborative optimization of multiple non-convex objectives within a single shared model. It introduces a preference-guided weighted update mechanism, allowing clients to explicitly specify desired trade-offs among objectives. Moreover, its convergence analysis is established under milder non-convex smoothness assumptions than prior works. Extensive experiments on image classification and recommendation tasks demonstrate that FedCMOO significantly improves Pareto front quality, reduces total communication volume by 62%, and decreases convergence error by 37%, consistently outperforming state-of-the-art baselines.
📝 Abstract
We study a federated version of multi-objective optimization (MOO), where a single model is trained to optimize multiple objective functions. MOO has been extensively studied in the centralized setting but is less explored in federated or distributed settings. We propose FedCMOO, a novel communication-efficient federated multi-objective optimization (FMOO) algorithm that improves the error convergence performance of the model compared to existing approaches. Unlike prior works, the communication cost of FedCMOO does not scale with the number of objectives, as each client sends a single aggregated gradient to the central server. We provide a convergence analysis of the proposed method for smooth and non-convex objective functions under milder assumptions than in prior work. In addition, we introduce a variant of FedCMOO that allows users to specify a preference over the objectives in terms of a desired ratio of the final objective values. Through extensive experiments, we demonstrate the superiority of our proposed method over baseline approaches.