Infighting in the Dark: Multi-Labels Backdoor Attack in Federated Learning

πŸ“… 2024-09-29
πŸ›οΈ arXiv.org
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
Existing backdoor attacks in federated learning (FL) predominantly assume cooperative attackers with a unified objective (i.e., single-label attacks), overlooking non-cooperative, multi-target multi-label backdoor attacks (MBA), where conflicting backdoor mappings render conventional methods ineffective. This work introduces, for the first time, a non-collaborative MBA threat model and proposes Mirageβ€”a novel attack framework that mitigates multi-target conflicts via in-distribution (ID) backdoor mapping modeling, adversarial feature adaptation, and constrained optimization. Mirage further leverages robustness to FL aggregation dynamics to ensure stealthy and persistent backdoor injection. Experiments demonstrate an attack success rate exceeding 97%, sustaining over 90% after 900 communication rounds, while evading state-of-the-art defenses. The implementation is publicly available.

Technology Category

Application Category

πŸ“ Abstract
Federated Learning (FL), a privacy-preserving decentralized machine learning framework, has been shown to be vulnerable to backdoor attacks. Current research primarily focuses on the Single-Label Backdoor Attack (SBA), wherein adversaries share a consistent target. However, a critical fact is overlooked: adversaries may be non-cooperative, have distinct targets, and operate independently, which exhibits a more practical scenario called Multi-Label Backdoor Attack (MBA). Unfortunately, prior works are ineffective in the MBA scenario since non-cooperative attackers exclude each other. In this work, we conduct an in-depth investigation to uncover the inherent constraints of the exclusion: similar backdoor mappings are constructed for different targets, resulting in conflicts among backdoor functions. To address this limitation, we propose Mirage, the first non-cooperative MBA strategy in FL that allows attackers to inject effective and persistent backdoors into the global model without collusion by constructing in-distribution (ID) backdoor mapping. Specifically, we introduce an adversarial adaptation method to bridge the backdoor features and the target distribution in an ID manner. Additionally, we further leverage a constrained optimization method to ensure the ID mapping survives in the global training dynamics. Extensive evaluations demonstrate that Mirage outperforms various state-of-the-art attacks and bypasses existing defenses, achieving an average ASR greater than 97% and maintaining over 90% after 900 rounds. This work aims to alert researchers to this potential threat and inspire the design of effective defense mechanisms. Code has been made open-source.
Problem

Research questions and friction points this paper is trying to address.

Addresses vulnerability of Federated Learning to Multi-Label Backdoor Attacks
Proposes non-cooperative attack strategy without collusion among adversaries
Ensures persistent backdoors via in-distribution mapping and constrained optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Non-cooperative multi-label backdoor attack strategy
In-distribution backdoor mapping construction
Adversarial adaptation and constrained optimization
πŸ”Ž Similar Papers
No similar papers found.
Y
Ye Li
Nanjing University of Aeronautics and Astronautics
Yanchao Zhao
Yanchao Zhao
Nanjing University of Aeronautics and Astronautics
Computer Networks
C
Chengcheng Zhu
Yangzhou University
J
Jiale Zhang
Yangzhou University