Reconstructing Training Data from Adapter-based Federated Large Language Models

📅 2026-01-24
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the privacy risks inherent in federated large language models, where adapter-based parameter-efficient fine-tuning still exposes training data to leakage. The authors propose UTR, the first gradient inversion attack tailored to adapter architectures, which overcomes three key challenges—low-dimensional gradients, frozen backbone parameters, and combinatorial explosion—by analyzing attention patterns in the frozen backbone, performing sentence-level reconstruction within the low-rank adapter gradient subspace, and incorporating linguistic priors during decoding. Experiments demonstrate that UTR achieves remarkable reconstruction accuracy, exceeding 99% ROUGE-1/2 scores on models including GPT2-Large, BERT, and Qwen2.5-7B, substantially outperforming existing methods. These results expose a fundamental tension between parameter-efficient fine-tuning and data privacy in federated learning settings.

Technology Category

Application Category

📝 Abstract
Adapter-based Federated Large Language Models (FedLLMs) are widely adopted to reduce the computational, storage, and communication overhead of full-parameter fine-tuning for web-scale applications while preserving user privacy. By freezing the backbone and training only compact low-rank adapters, these methods appear to limit gradient leakage and thwart existing Gradient Inversion Attacks (GIAs). Contrary to this assumption, we show that low-rank adapters create new, exploitable leakage channels. We propose the Unordered-word-bag-based Text Reconstruction (UTR) attack, a novel GIA tailored to the unique structure of adapter-based FedLLMs. UTR overcomes three core challenges: low-dimensional gradients, frozen backbones, and combinatorially large reconstruction spaces by: (i) inferring token presence from attention patterns in frozen layers, (ii) performing sentence-level inversion within the low-rank subspace of adapter gradients, and (iii) enforcing semantic coherence through constrained greedy decoding guided by language priors. Extensive experiments across diverse models (GPT2-Large, BERT, Qwen2.5-7B) and datasets (CoLA, SST-2, Rotten Tomatoes) demonstrate that UTR achieves near-perfect reconstruction accuracy (ROUGE-1/2>99), even with large batch size settings where prior GIAs fail completely. Our results reveal a fundamental tension between parameter efficiency and privacy in FedLLMs, challenging the prevailing belief that lightweight adaptation inherently enhances security. Our code and data are available at https://github.com/shwksnshwowk-wq/GIA.
Problem

Research questions and friction points this paper is trying to address.

Gradient Inversion Attacks
Federated Learning
Large Language Models
Privacy Leakage
Adapter-based Fine-tuning
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gradient Inversion Attack
Adapter-based Federated Learning
Privacy Leakage
Low-rank Adaptation
Text Reconstruction
🔎 Similar Papers
No similar papers found.
S
Silong Chen
National University of Defense Technology
Y
Yuchuan Luo
National University of Defense Technology
G
Guilin Deng
National University of Defense Technology
Yi Liu
Yi Liu
Department of Computer Science, City University of Hong Kong
Security and PrivacyFederated LearningAI Security
Minjie Xu
Minjie Xu
Cohere
Machine LearningArtificial IntelligenceNatural Language Processing
S
Shaojing Fu
National University of Defense Technology
Xiaohua Jia
Xiaohua Jia
Chinese Academy of Science