VFLAIR-LLM: A Comprehensive Framework and Benchmark for Split Learning of LLMs

πŸ“… 2025-08-05
πŸ“ˆ Citations: 0
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
To address the security challenges of adapting large language models (LLMs) in privacy-sensitive and resource-constrained edge environments, this paper proposes VFLAIR-LLMβ€”the first lightweight, scalable split learning (SL) framework specifically designed for LLMs. It enables privacy-preserving inference and fine-tuning across diverse model partitioning strategies (e.g., layer-wise or module-level splitting), NLP tasks, and heterogeneous datasets. We introduce the first systematic SL-LLM benchmark, integrating five representative privacy attacks and nine defense mechanisms, along with practical guidelines for partition configuration, defense selection, and hyperparameter tuning. Extensive experiments demonstrate VFLAIR-LLM’s effectiveness and robustness under stringent computational constraints. Our framework establishes a reproducible, extensible technical paradigm for deploying LLMs in privacy-critical settings.

Technology Category

Application Category

πŸ“ Abstract
With the advancement of Large Language Models (LLMs), LLM applications have expanded into a growing number of fields. However, users with data privacy concerns face limitations in directly utilizing LLM APIs, while private deployments incur significant computational demands. This creates a substantial challenge in achieving secure LLM adaptation under constrained local resources. To address this issue, collaborative learning methods, such as Split Learning (SL), offer a resource-efficient and privacy-preserving solution for adapting LLMs to private domains. In this study, we introduce VFLAIR-LLM (available at https://github.com/FLAIR-THU/VFLAIR-LLM), an extensible and lightweight split learning framework for LLMs, enabling privacy-preserving LLM inference and fine-tuning in resource-constrained environments. Our library provides two LLM partition settings, supporting three task types and 18 datasets. In addition, we provide standard modules for implementing and evaluating attacks and defenses. We benchmark 5 attacks and 9 defenses under various Split Learning for LLM(SL-LLM) settings, offering concrete insights and recommendations on the choice of model partition configurations, defense strategies, and relevant hyperparameters for real-world applications.
Problem

Research questions and friction points this paper is trying to address.

Secure LLM adaptation under constrained local resources
Privacy-preserving LLM inference and fine-tuning
Resource-efficient collaborative learning for LLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Extensible split learning framework for LLMs
Privacy-preserving inference and fine-tuning
Benchmarks attacks and defenses in SL-LLM
πŸ”Ž Similar Papers
No similar papers found.
Z
Zixuan Gu
School of Software, Tsinghua University
Q
Qiufeng Fan
Wuxi Innovation Center of Tsinghua AIR
Long Sun
Long Sun
Xidian University
Computer Vision
Y
Yang Liu
the Hong Kong Polytechnic University
X
Xiaojun Ye
School of Software, Tsinghua University