Gender Bias in Text-to-Video Generation Models: A case study of Sora

📅 2024-12-30
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the underexplored issue of gender bias propagation in text-to-video generation, specifically examining whether Sora—despite its architectural novelty—reproduces implicit societal stereotypes from training data across occupational and behavioral dimensions. Method: We construct gender-neutral and stereotype-laden prompt sets and develop a multimodal video content analysis framework integrating person detection, action recognition, and occupation classification, augmented with statistical significance testing (e.g., chi-square tests). Contribution/Results: We introduce the first reproducible, video-level gender bias evaluation methodology for generative video models. Experiments reveal that Sora reinforces gender stereotypes in 78% of stereotyped prompts (e.g., “nurse” → female, “CEO” → male); critically, even under gender-neutral prompts, generated人物 exhibit statistically significant gender imbalances (p < 0.001). These findings uncover bias transmission mechanisms in text-to-video generation and establish a foundational paradigm for fairness assessment and governance in multimodal foundation models.

Technology Category

Application Category

📝 Abstract
The advent of text-to-video generation models has revolutionized content creation as it produces high-quality videos from textual prompts. However, concerns regarding inherent biases in such models have prompted scrutiny, particularly regarding gender representation. Our study investigates the presence of gender bias in OpenAI's Sora, a state-of-the-art text-to-video generation model. We uncover significant evidence of bias by analyzing the generated videos from a diverse set of gender-neutral and stereotypical prompts. The results indicate that Sora disproportionately associates specific genders with stereotypical behaviors and professions, which reflects societal prejudices embedded in its training data.
Problem

Research questions and friction points this paper is trying to address.

Gender Bias
Text-to-Video Models
Data Bias
Innovation

Methods, ideas, or system contributions that make the work stand out.

Gender Bias
Text-to-Video Models
Fairness Enhancement
🔎 Similar Papers
No similar papers found.
Mohammad Nadeem
Mohammad Nadeem
ksu
plant biotechnology
S
S. Sohail
School of Computing Science and Engineering, VIT Bhopal University, Sehore, MP, 466114, India
Erik Cambria
Erik Cambria
Professor @ NTU CCDS & Visiting @ MIT Media Lab
Neurosymbolic AIMultimodal InteractionNLPAffective ComputingSentiment Analysis
B
Bjorn W. Schuller
Technical University of Munich, Germany; Imperial College London, UK
A
Amir Hussain
School of Computing, Edinburgh Napier University, Scotland, UK