Which Demographics do LLMs Default to During Annotation?

📅 2024-10-11
🏛️ arXiv.org
📈 Citations: 2
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates implicit demographic biases—specifically gender, race, and age preferences—exhibited by large language models (LLMs) in politeness/impoliteness annotation tasks when no explicit demographic information is provided. Method: We introduce a “placebo prompt” controlled paradigm to isolate LLMs’ default annotation behavior, conducting three systematically designed experiments on the POPQUORN dataset: demographic-conditioned prompting, placebo prompting, and zero-shot prompting. Annotations are generated via multi-round LLM inference, followed by rigorous statistical significance testing. Contribution/Results: We demonstrate that LLMs are not demographically neutral in zero-shot settings; instead, they exhibit reproducible, statistically significant demographic biases—most prominently along gender and age dimensions. This work overcomes a key limitation of prior research—namely, the absence of appropriate controls for detecting latent biases—and establishes a novel methodological framework for fairness evaluation in LLM-based annotation systems.

Technology Category

Application Category

📝 Abstract
Demographics and cultural background of annotators influence the labels they assign in text annotation -- for instance, an elderly woman might find it offensive to read a message addressed to a"bro", but a male teenager might find it appropriate. It is therefore important to acknowledge label variations to not under-represent members of a society. Two research directions developed out of this observation in the context of using large language models (LLM) for data annotations, namely (1) studying biases and inherent knowledge of LLMs and (2) injecting diversity in the output by manipulating the prompt with demographic information. We combine these two strands of research and ask the question to which demographics an LLM resorts to when no demographics is given. To answer this question, we evaluate which attributes of human annotators LLMs inherently mimic. Furthermore, we compare non-demographic conditioned prompts and placebo-conditioned prompts (e.g.,"you are an annotator who lives in house number 5") to demographics-conditioned prompts ("You are a 45 year old man and an expert on politeness annotation. How do you rate {instance}"). We study these questions for politeness and offensiveness annotations on the POPQUORN data set, a corpus created in a controlled manner to investigate human label variations based on demographics which has not been used for LLM-based analyses so far. We observe notable influences related to gender, race, and age in demographic prompting, which contrasts with previous studies that found no such effects.
Problem

Research questions and friction points this paper is trying to address.

Investigates default demographics LLMs mimic in annotations
Compares demographic vs non-demographic prompt effects on labels
Analyzes bias in politeness/offensiveness annotations across demographics
Innovation

Methods, ideas, or system contributions that make the work stand out.

Studying LLM biases and inherent knowledge
Injecting diversity via demographic prompts
Comparing non-demographic and placebo-conditioned prompts
🔎 Similar Papers
No similar papers found.
C
Christopher Bagdon
Fundamentals of Natural Language Processing, University of Bamberg
A
Aidan Combs
Fundamentals of Natural Language Processing, University of Bamberg
L
Lynn Greschner
Fundamentals of Natural Language Processing, University of Bamberg
Roman Klinger
Roman Klinger
Professor for Fundamentals of Natural Language Processing, University of Bamberg
natural language processingemotion analysisbioNLPargument miningcomputational psychology
J
Jiahui Li
Fundamentals of Natural Language Processing, University of Bamberg
S
Sean Papay
Fundamentals of Natural Language Processing, University of Bamberg
N
Nadine Probol
Fundamentals of Natural Language Processing, University of Bamberg
Y
Yarik Menchaca Resendiz
Fundamentals of Natural Language Processing, University of Bamberg
J
Johannes Schafer
Fundamentals of Natural Language Processing, University of Bamberg
Aswathy Velutharambath
Aswathy Velutharambath
PhD student, University of Stuttgart
computational linguisticssociolinguisticstext style transferNLP + Psych
Sabine Weber
Sabine Weber
PhD Student, ILCC, School of Informatics, University of Edinburgh
Natural Language ProcessingComputational LinguisticsArtificial IntelligenceMachine Learning
A
Amelie Wuhrl
Fundamentals of Natural Language Processing, University of Bamberg