Passive Inference Attacks on Split Learning via Adversarial Regularization

πŸ“… 2023-10-16
πŸ›οΈ Network and Distributed System Security Symposium
πŸ“ˆ Citations: 7
✨ Influential: 0
πŸ“„ PDF
πŸ€– AI Summary
This work addresses the challenge of passive inference attacks against Split Learning (SL) under an honest-but-curious server, specifically targeting U-shaped SL architectures. We propose SDARβ€”a novel framework that enables joint inference of both private client features and labels for the first time in such settings. SDAR leverages auxiliary data-driven modeling and adversarial regularization to construct a decodable private-model simulator, facilitating deep feature inversion and label prediction. Evaluated on CIFAR-10 with deep split at layer 7, SDAR achieves a private feature reconstruction MSE < 0.025 and label inference accuracy > 98%, matching the performance of active attacks and substantially outperforming existing passive methods. By breaking the capability ceiling of passive inference, SDAR establishes a new paradigm for rigorous privacy risk assessment in SL systems.
πŸ“ Abstract
Split Learning (SL) has emerged as a practical and efficient alternative to traditional federated learning. While previous attempts to attack SL have often relied on overly strong assumptions or targeted easily exploitable models, we seek to develop more capable attacks. We introduce SDAR, a novel attack framework against SL with an honest-but-curious server. SDAR leverages auxiliary data and adversarial regularization to learn a decodable simulator of the client's private model, which can effectively infer the client's private features under the vanilla SL, and both features and labels under the U-shaped SL. We perform extensive experiments in both configurations to validate the effectiveness of our proposed attacks. Notably, in challenging scenarios where existing passive attacks struggle to reconstruct the client's private data effectively, SDAR consistently achieves significantly superior attack performance, even comparable to active attacks. On CIFAR-10, at the deep split level of 7, SDAR achieves private feature reconstruction with less than 0.025 mean squared error in both the vanilla and the U-shaped SL, and attains a label inference accuracy of over 98% in the U-shaped setting, while existing attacks fail to produce non-trivial results.
Problem

Research questions and friction points this paper is trying to address.

Develop stronger passive attacks on Split Learning
Infer private features and labels via adversarial regularization
Achieve high accuracy in challenging reconstruction scenarios
Innovation

Methods, ideas, or system contributions that make the work stand out.

SDAR uses adversarial regularization for attacks
SDAR simulates client's private model effectively
SDAR achieves high inference accuracy
πŸ”Ž Similar Papers
No similar papers found.