DP^2-VL: Private Photo Dataset Protection by Data Poisoning for Vision-Language Models

📅 2026-03-25
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work identifies a novel identity-linkage privacy threat wherein fine-tuning vision-language models (VLMs) on a small set of private user images can inadvertently leak sensitive identity information and associated attributes. To address this, we formalize the threat model for the first time, construct the first identity-linkage benchmark dataset, and propose DP²-VL—the first adversarial data poisoning method specifically designed to protect privacy in VLMs. DP²-VL injects imperceptible perturbations that induce dataset-level shifts in the encoder’s embedding space, causing attackers to overfit to spurious features during fine-tuning. Extensive experiments demonstrate that DP²-VL achieves strong generalization, robustness, and consistent protection across diverse mainstream VLMs—including LLaVA, Qwen-VL, and MiniGPT-v2—while remaining effective against cross-model transfer and post-processing defenses.

Technology Category

Application Category

📝 Abstract
Recent advances in visual-language alignment have endowed vision-language models (VLMs) with fine-grained image understanding capabilities. However, this progress also introduces new privacy risks. This paper first proposes a novel privacy threat model named identity-affiliation learning: an attacker fine-tunes a VLM using only a few private photos of a target individual, thereby embedding associations between the target facial identity and their private property and social relationships into the model's internal representations. Once deployed via public APIs, this model enables unauthorized exposure of the target user's private information upon input of their photos. To benchmark VLMs' susceptibility to such identity-affiliation leakage, we introduce the first identity-affiliation dataset comprising seven typical scenarios appearing in private photos. Each scenario is instantiated with multiple identity-centered photo-description pairs. Experimental results demonstrate that mainstream VLMs like LLaVA, Qwen-VL, and MiniGPT-v2, can recognize facial identities and infer identity-affiliation relationships by fine-tuning on small-scale private photographic dataset, and even on synthetically generated datasets. To mitigate this privacy risk, we propose DP2-VL, the first Dataset Protection framework for private photos that leverages Data Poisoning. Though optimizing imperceptible perturbations by pushing the original representations toward an antithetical region, DP2-VL induces a dataset-level shift in the embedding space of VLMs'encoders. This shift separates protected images from clean inference images, causing fine-tuning on the protected set to overfit. Extensive experiments demonstrate that DP2-VL achieves strong generalization across models, robustness to diverse post-processing operations, and consistent effectiveness across varying protection ratios.
Problem

Research questions and friction points this paper is trying to address.

privacy risk
vision-language models
identity-affiliation leakage
data poisoning
private photo protection
Innovation

Methods, ideas, or system contributions that make the work stand out.

identity-affiliation learning
data poisoning
vision-language models
privacy protection
dataset-level embedding shift
🔎 Similar Papers
No similar papers found.
H
Hongyi Miao
Shandong University
J
Jun Jia
Shanghai Jiao Tong University
X
Xincheng Wang
Donghua University
Qianli Ma
Qianli Ma
Shanghai Jiao Tong University
Deep LearningGenerative AILLMsMLLMs
W
Wei Sun
East China Normal University
W
Wangqiu Zhou
Hefei University of Technology
D
Dandan Zhu
East China Normal University
Y
Yewen Cao
Shandong University
Z
Zhi Liu
Shandong University
Guangtao Zhai
Guangtao Zhai
Professor, IEEE Fellow, Shanghai Jiao Tong University
Multimedia Signal ProcessingVisual Quality AssessmentQoEAI EvaluationDisplays