Adapting Psycholinguistic Research for LLMs: Gender-inclusive Language in a Coreference Context

📅 2025-02-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates whether large language models (LLMs) exhibit gender neutrality in coreference resolution involving gender-inclusive language (e.g., non-binary or gender-neutral expressions) and uncovers latent cross-linguistic gender biases. Methodologically, it innovatively adapts a French psycholinguistic paradigm—first extended to English and German—combining prompt engineering with controlled cloze tasks across Llama, GPT, and Claude models, validated via statistical significance testing. Results reveal that while English LLMs generally preserve antecedent gender, they exhibit an underlying male bias; in German, this bias is markedly stronger, systematically overriding diverse gender-neutral strategies. Crucially, the study demonstrates how grammatical gender systems amplify implicit biases in LLMs—a previously undocumented phenomenon. It thus establishes a novel methodology and empirical benchmark for cross-linguistic fairness evaluation in NLP.

Technology Category

Application Category

📝 Abstract
Gender-inclusive language is often used with the aim of ensuring that all individuals, regardless of gender, can be associated with certain concepts. While psycholinguistic studies have examined its effects in relation to human cognition, it remains unclear how Large Language Models (LLMs) process gender-inclusive language. Given that commercial LLMs are gaining an increasingly strong foothold in everyday applications, it is crucial to examine whether LLMs in fact interpret gender-inclusive language neutrally, because the language they generate has the potential to influence the language of their users. This study examines whether LLM-generated coreferent terms align with a given gender expression or reflect model biases. Adapting psycholinguistic methods from French to English and German, we find that in English, LLMs generally maintain the antecedent's gender but exhibit underlying masculine bias. In German, this bias is much stronger, overriding all tested gender-neutralization strategies.
Problem

Research questions and friction points this paper is trying to address.

Examines LLMs' processing of gender-inclusive language
Investigates LLMs' biases in gender coreference resolution
Adapts psycholinguistic methods to assess LLM performance
Innovation

Methods, ideas, or system contributions that make the work stand out.

Adapts psycholinguistic methods for LLMs
Examines gender-inclusive language in LLMs
Assesses coreferent terms for gender bias
🔎 Similar Papers
No similar papers found.
M
Marion Bartl
Insight SFI Research Centre for Data Analytics, School of Information and Communication Studies
T
Thomas Brendan Murphy
Insight SFI Research Centre for Data Analytics, School of Mathematics and Statistics
Susan Leavy
Susan Leavy
University College Dublin, Insight Centre for Data Analytics
AI EthicsArtificial IntelligenceNatural Lanugage ProcessingAlgorithmic BiasDigital Humanities