Rethinking Visual Privacy: A Compositional Privacy Risk Framework for Severity Assessment with VLMs

📅 2026-03-23
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the limitation of existing visual privacy assessments, which predominantly rely on binary labels and fail to capture higher-order privacy risks arising from combinations of otherwise benign attributes. To overcome this, the authors propose a fine-grained, composition-based privacy evaluation framework. They introduce the first Composition-based Privacy Risk Taxonomy (CPRT), which establishes a hierarchical severity scale and an interpretable scoring function, and release the first large-scale image dataset aligned with this taxonomy. Leveraging vision-language models (VLMs) and supervised fine-tuning (SFT), they train an 8B-parameter deployable model that achieves performance on par with state-of-the-art VLMs in evaluating compositional privacy risks, significantly outperforming smaller models.

Technology Category

Application Category

📝 Abstract
Existing visual privacy benchmarks largely treat privacy as a binary property, labeling images as private or non-private based on visible sensitive content. We argue that privacy is fundamentally compositional. Attributes that are benign in isolation may combine to produce severe privacy violations. We introduce the Compositional Privacy Risk Taxonomy (CPRT), a regulation-aware framework that organizes visual attributes according to standalone identifiability and compositional harm potential. CPRT defines four graded severity levels and is paired with an interpretable scoring function that assigns continuous privacy severity scores. We further construct a taxonomy-aligned dataset of 6.7K images and derive ground-truth compositional risk scores. By evaluating frontier and open-weight VLMs we find that frontier models align well with compositional severity when provided structured guidance, but systematically underestimate composition-driven risks. Smaller models struggle to internalize graded privacy reasoning. To bridge this gap, we introduce a deployable 8B supervised fine-tuned (SFT) model that closely matches frontier-level performance on compositional privacy assessment.
Problem

Research questions and friction points this paper is trying to address.

visual privacy
compositional risk
privacy severity
attribute composition
VLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Compositional Privacy
Privacy Risk Assessment
Vision-Language Models (VLMs)
Supervised Fine-Tuning (SFT)
Interpretable Scoring
🔎 Similar Papers
No similar papers found.