Label Inference Attacks against Federated Unlearning

๐Ÿ“… 2025-08-08
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
This work identifies a novel privacy threat in federated unlearning (FU): adversaries can infer labels of forgotten user data by analyzing model parameter changes. To address this, we propose the first systematic label inference attack (ULIA), which establishes a gradientโ€“label mapping mechanism. ULIA models parameter differences and gradient dynamics to achieve high-accuracy label recovery across diverse unlearning levels (1%โ€“100%). It remains effective under both IID and non-IID data distributions, achieving 100% attack success rate in IID settings; even when only 1% of local data is unlearned, success rates range from 62.3% to 93%. Our findings fundamentally challenge the prevailing privacy-security assumptions underlying existing FU mechanisms. By exposing this vulnerability, we provide critical theoretical insights and an empirical benchmark for trustworthy evaluation and robust defense design in federated unlearning.

Technology Category

Application Category

๐Ÿ“ Abstract
Federated Unlearning (FU) has emerged as a promising solution to respond to the right to be forgotten of clients, by allowing clients to erase their data from global models without compromising model performance. Unfortunately, researchers find that the parameter variations of models induced by FU expose clients' data information, enabling attackers to infer the label of unlearning data, while label inference attacks against FU remain unexplored. In this paper, we introduce and analyze a new privacy threat against FU and propose a novel label inference attack, ULIA, which can infer unlearning data labels across three FU levels. To address the unique challenges of inferring labels via the models variations, we design a gradient-label mapping mechanism in ULIA that establishes a relationship between gradient variations and unlearning labels, enabling inferring labels on accumulated model variations. We evaluate ULIA on both IID and non-IID settings. Experimental results show that in the IID setting, ULIA achieves a 100% Attack Success Rate (ASR) under both class-level and client-level unlearning. Even when only 1% of a user's local data is forgotten, ULIA still attains an ASR ranging from 93% to 62.3%.
Problem

Research questions and friction points this paper is trying to address.

Label inference attacks against Federated Unlearning systems
Inferring unlearning data labels from model parameter variations
Addressing privacy threats across three Federated Unlearning levels
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gradient-label mapping mechanism design
Inferring labels via model variations
Attacking across three FU levels
๐Ÿ”Ž Similar Papers
No similar papers found.
W
Wei Wang
the Key Laboratory of Ethnic Language Intelligent Analysis and Security Management of MOE, Minzu University of China, Beijing, China
Xiangyun Tang
Xiangyun Tang
Minzu University of China
System Security
Yajie Wang
Yajie Wang
Beijing Institute of Technology
Y
Yijing Lin
School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing, China
T
Tao Zhang
School of Computer Science and Engineering, Beijing Jiaotong University, Beijing, China
M
Meng Shen
School of Cyberspace Science and Technology, Beijing Institute of Technology, Beijing, China
D
Dusit Niyato
School of Cyberspace Science and Technology, Nanyang Technological University, Singapore
L
Liehuang Zhu
School of Cyberspace Science and Technology, Beijing Institute of Technology, Beijing, China