Taming the Instability: A Robust Second-Order Optimizer for Federated Learning over Non-IID Data

๐Ÿ“… 2026-03-30
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work addresses the challenges of high computational overhead, numerical instability, and excessive communication costs that plague existing second-order optimization methods in non-IID federated learning. To overcome these limitations, the authors propose FedRCO, a novel framework that integrates efficient curvature approximation, real-time gradient anomaly monitoring, a fail-safe state reset mechanism, and a curvature-preserving adaptive aggregation strategy. This design effectively retains local curvature information while ensuring numerical stability. Experimental results demonstrate that FedRCO consistently outperforms state-of-the-art first- and second-order methods across diverse non-IID settings, achieving higher model accuracy, faster convergence, and reduced communication overhead.
๐Ÿ“ Abstract
In this paper, we present Federated Robust Curvature Optimization (FedRCO), a novel second-order optimization framework designed to improve convergence speed and reduce communication cost in Federated Learning systems under statistical heterogeneity. Existing second-order optimization methods are often computationally expensive and numerically unstable in distributed settings. In contrast, FedRCO addresses these challenges by integrating an efficient approximate curvature optimizer with a provable stability mechanism. Specifically, FedRCO incorporates three key components: (1) a Gradient Anomaly Monitor that detects and mitigates exploding gradients in real-time, (2) a Fail-Safe Resilience protocol that resets optimization states upon numerical instability, and (3) a Curvature-Preserving Adaptive Aggregation strategy that safely integrates global knowledge without erasing the local curvature geometry. Theoretical analysis shows that FedRCO can effectively mitigate instability and prevent unbounded updates while preserving optimization efficiency. Extensive experiments show that FedRCO achieves superior robustness against diverse non-IID scenarios while achieving higher accuracy and faster convergence than both state-of-the-art first-order and second-order methods.
Problem

Research questions and friction points this paper is trying to address.

Federated Learning
Non-IID Data
Second-Order Optimization
Numerical Instability
Statistical Heterogeneity
Innovation

Methods, ideas, or system contributions that make the work stand out.

second-order optimization
federated learning
non-IID data
curvature preservation
numerical stability
๐Ÿ”Ž Similar Papers
No similar papers found.
Y
Yuanqiao Zhang
Key Laboratory of Collaborative Intelligence Systems of Ministry of Education, Xidian University, Xiโ€™an, Shaanxi, China; State Key Laboratory of ISN, School of Cyber Engineering, Xidian University, Xiโ€™an, Shaanxi, China
Tiantian He
Tiantian He
PhD student, University College London
AI AgentProbabilistic modellingGraph learningSpatio-temporal modellingAI for Neuroscience
Yuan Gao
Yuan Gao
University of Science and Technology of China
Graph MiningAnomaly DetectionOut-of-distribution Generalization
Y
Yixin Wang
Key Laboratory of Collaborative Intelligence Systems of Ministry of Education, Xidian University, Xiโ€™an, Shaanxi, China
Yew-Soon Ong
Yew-Soon Ong
President Chair Professor of Computer Science, A*Star AI Chief Scientist, FIEEE
Artificial IntelligenceStatistical MLEvolutionary OptimizationBayesian Optimization
M
Maoguo Gong
Key Laboratory of Collaborative Intelligence Systems of Ministry of Education, Xidian University, Xiโ€™an, Shaanxi, China
A
A. K. Qin
Department of Computing Technologies, Swinburne University of Technology, Melbourne, Australia
Hui Li
Hui Li
Xidian University
wireless network Securitysecurity in cloud computinginformation theory