PRSM: A Measure to Evaluate CLIP's Robustness Against Paraphrases

📅 2025-11-14
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work investigates the robustness of CLIP models to paraphrasing—i.e., semantically preserving rewordings—in socially sensitive contexts, exposing critical fairness risks in real-world deployment. We propose the Paraphrase Ranking Stability Measure (PRSM), the first systematic metric quantifying CLIP’s consistency in ranking image–text alignments across diverse automated and human-generated paraphrases. Leveraging the Social Counterfactuals dataset, we conduct bias-aware correlation analysis. Experiments reveal pronounced sensitivity disparities across paraphrase types and uncover a stable, consistent robustness gap in gender-related queries: CLIP exhibits higher response stability for male-associated prompts but significantly greater vulnerability to paraphrase perturbations for female-associated ones. Our framework establishes a novel, reproducible paradigm for evaluating linguistic robustness in multimodal foundation models, providing actionable metrics for fairness auditing and trustworthy deployment.

Technology Category

Application Category

📝 Abstract
Contrastive Language-Image Pre-training (CLIP) is a widely used multimodal model that aligns text and image representations through large-scale training. While it performs strongly on zero-shot and few-shot tasks, its robustness to linguistic variation, particularly paraphrasing, remains underexplored. Paraphrase robustness is essential for reliable deployment, especially in socially sensitive contexts where inconsistent representations can amplify demographic biases. In this paper, we introduce the Paraphrase Ranking Stability Metric (PRSM), a novel measure for quantifying CLIP's sensitivity to paraphrased queries. Using the Social Counterfactuals dataset, a benchmark designed to reveal social and demographic biases, we empirically assess CLIP's stability under paraphrastic variation, examine the interaction between paraphrase robustness and gender, and discuss implications for fairness and equitable deployment of multimodal systems. Our analysis reveals that robustness varies across paraphrasing strategies, with subtle yet consistent differences observed between male- and female-associated queries.
Problem

Research questions and friction points this paper is trying to address.

Evaluating CLIP's robustness against linguistic paraphrasing variations
Measuring sensitivity to paraphrased queries using PRSM metric
Assessing interaction between paraphrase robustness and demographic biases
Innovation

Methods, ideas, or system contributions that make the work stand out.

Introducing PRSM metric for CLIP paraphrase robustness
Using Social Counterfactuals dataset for bias assessment
Evaluating robustness variation across paraphrasing strategies