🤖 AI Summary
To address poor generalization and slow convergence caused by non-independent and identically distributed (Non-IID) data in autonomous driving federated learning, this paper proposes FedDSR—a novel framework that jointly leverages mutual information estimation and negentropy regularization over multi-layer intermediate features. FedDSR establishes a hierarchical intermediate supervision and collaborative regularization mechanism to optimize feature representations layer-wise. It is compatible with mainstream federated algorithms and network architectures, and applicable to tasks such as semantic segmentation. Experiments on multiple Non-IID autonomous driving datasets demonstrate that FedDSR achieves up to an 8.93% improvement in mean Intersection-over-Union (mIoU) over baseline methods while reducing communication rounds by 28.57%. These results confirm its effectiveness in significantly enhancing model generalization and training efficiency.
📝 Abstract
Federated Learning (FL) enables collaborative training of autonomous driving (AD) models across distributed vehicles while preserving data privacy. However, FL encounters critical challenges such as poor generalization and slow convergence due to non-independent and identically distributed (non-IID) data from diverse driving environments. To overcome these obstacles, we introduce Federated Deep Supervision and Regularization (FedDSR), a paradigm that incorporates multi-access intermediate layer supervision and regularization within federated AD system. Specifically, FedDSR comprises following integral strategies: (I) to select multiple intermediate layers based on predefined architecture-agnostic standards. (II) to compute mutual information (MI) and negative entropy (NE) on those selected layers to serve as intermediate loss and regularizer. These terms are integrated into the output-layer loss to form a unified optimization objective, enabling comprehensive optimization across the network hierarchy. (III) to aggregate models from vehicles trained based on aforementioned rules of (I) and (II) to generate the global model on central server. By guiding and penalizing the learning of feature representations at intermediate stages, FedDSR enhances the model generalization and accelerates model convergence for federated AD. We then take the semantic segmentation task as an example to assess FedDSR and apply FedDSR to multiple model architectures and FL algorithms. Extensive experiments demonstrate that FedDSR achieves up to 8.93% improvement in mIoU and 28.57% reduction in training rounds, compared to other FL baselines, making it highly suitable for practical deployment in federated AD ecosystems.