Semi-Supervised Federated Learning via Dual Contrastive Learning and Soft Labeling for Intelligent Fault Diagnosis

πŸ“… 2025-07-12
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the challenges of scarce labeled data, heterogeneous client data distributions, and privacy constraints in industrial intelligent fault diagnosis, this paper proposes SSFL-DCSL, a semi-supervised federated learning framework. Methodologically, SSFL-DCSL innovatively integrates dual contrastive learning (local and global) with a soft-labeling mechanism; introduces a Laplacian-weighted function to mitigate pseudo-label bias; and employs momentum-updated prototype aggregation to enable cross-client knowledge transfer. By jointly optimizing semi-supervised and federated learning objectives, the framework significantly enhances feature representation consistency and generalization under low-resource conditions. Extensive experiments on three public benchmark datasets demonstrate that, using only 10% labeled data, SSFL-DCSL achieves accuracy improvements of 1.15–7.85% over state-of-the-art methods. The framework thus effectively reconciles the tripartite challenges of few-shot learning, non-IID data, and privacy preservation in industrial fault diagnosis.

Technology Category

Application Category

πŸ“ Abstract
Intelligent fault diagnosis (IFD) plays a crucial role in ensuring the safe operation of industrial machinery and improving production efficiency. However, traditional supervised deep learning methods require a large amount of training data and labels, which are often located in different clients. Additionally, the cost of data labeling is high, making labels difficult to acquire. Meanwhile, differences in data distribution among clients may also hinder the model's performance. To tackle these challenges, this paper proposes a semi-supervised federated learning framework, SSFL-DCSL, which integrates dual contrastive loss and soft labeling to address data and label scarcity for distributed clients with few labeled samples while safeguarding user privacy. It enables representation learning using unlabeled data on the client side and facilitates joint learning among clients through prototypes, thereby achieving mutual knowledge sharing and preventing local model divergence. Specifically, first, a sample weighting function based on the Laplace distribution is designed to alleviate bias caused by low confidence in pseudo labels during the semi-supervised training process. Second, a dual contrastive loss is introduced to mitigate model divergence caused by different data distributions, comprising local contrastive loss and global contrastive loss. Third, local prototypes are aggregated on the server with weighted averaging and updated with momentum to share knowledge among clients. To evaluate the proposed SSFL-DCSL framework, experiments are conducted on two publicly available datasets and a dataset collected on motors from the factory. In the most challenging task, where only 10% of the data are labeled, the proposed SSFL-DCSL can improve accuracy by 1.15% to 7.85% over state-of-the-art methods.
Problem

Research questions and friction points this paper is trying to address.

Addresses data and label scarcity in federated fault diagnosis
Reduces model divergence from varied client data distributions
Enhances accuracy with semi-supervised contrastive learning techniques
Innovation

Methods, ideas, or system contributions that make the work stand out.

Semi-supervised federated learning with dual contrastive loss
Soft labeling using Laplace-based sample weighting
Knowledge sharing via weighted prototype aggregation
πŸ”Ž Similar Papers
No similar papers found.
Yajiao Dai
Yajiao Dai
Nanjing University of Science and Technology
Few-shot LearningRepresentation LearningMultimodal learning
J
Jun Li
School of Information Science and Engineering, Southeast University, Nanjing, 210096, China
Zhen Mei
Zhen Mei
Associate Professor, Nanjing University of Science and Technology
Information TheoryDeep LearningIn-Memory ComputingError Correction Code
Yiyang Ni
Yiyang Ni
JSSNU/NJUPT
RIS, Intelligent Communications
S
Shi Jin
School of Information Science and Engineering, Southeast University, Nanjing, 210096, China
Z
Zengxiang Li
Digital Research Institute, ENN Group, Langfang, 065001, China
Sheng Guo
Sheng Guo
Ant Group
Computer VisionDeep LearningLLM
W
Wei Xiang
School of Computing, Engineering and Mathematical Sciences, La Trobe University, Melbourne, VIC 3086, Australia, and also with the College of Science and Engineering, James Cook University, Cairns, QLD 4878, Australia