Which private attributes do VLMs agree on and predict well?

📅 2026-02-08
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study evaluates the capacity of open-source vision-language models (VLMs) to identify privacy-related visual attributes in a zero-shot setting and examines their alignment with human annotations. By comparing model predictions against human-labeled data, the authors find that VLMs consistently exhibit greater sensitivity in detecting privacy-sensitive content than human annotators. Notably, when multiple VLMs converge on highly consistent predictions, these outputs not only effectively complement human judgments but also uncover privacy-relevant elements overlooked by humans. The findings demonstrate the potential of VLMs for large-scale privacy-aware data annotation and offer a novel pathway toward automated privacy perception in visual content.

Technology Category

Application Category

📝 Abstract
Visual Language Models (VLMs) are often used for zero-shot detection of visual attributes in the image. We present a zero-shot evaluation of open-source VLMs for privacy-related attribute recognition. We identify the attributes for which VLMs exhibit strong inter-annotator agreement, and discuss the disagreement cases of human and VLM annotations. Our results show that when evaluated against human annotations, VLMs tend to predict the presence of privacy attributes more often than human annotators. In addition to this, we find that in cases of high inter-annotator agreement between VLMs, they can complement human annotation by identifying attributes overlooked by human annotators. This highlights the potential of VLMs to support privacy annotations in large-scale image datasets.
Problem

Research questions and friction points this paper is trying to address.

Visual Language Models
privacy attributes
zero-shot detection
inter-annotator agreement
human annotation
Innovation

Methods, ideas, or system contributions that make the work stand out.

Visual Language Models
zero-shot evaluation
privacy attributes
inter-annotator agreement
human-VLM disagreement
🔎 Similar Papers
No similar papers found.
O
Olena Hrynenko
EPFL, Switzerland; Idiap Research Institute, Switzerland
D
Darya Baranouskaya
EPFL, Switzerland; Idiap Research Institute, Switzerland
Alina Elena Baia
Alina Elena Baia
Idiap Research Institute
Privacy in CVXAIAdversarial attacks
Andrea Cavallaro
Andrea Cavallaro
Director, Idiap Research Institute; Professor, EPFL
Machine LearningComputer VisionAudio ProcessingRobot PerceptionPrivacy