DAIQ: Auditing Demographic Attribute Inference from Question in LLMs

📅 2025-08-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work exposes a covert fairness and privacy risk: large language models (LLMs) can implicitly infer users’ sensitive demographic attributes—such as gender and race—from question phrasing alone, even without explicit demographic input. To systematically study this phenomenon, the authors introduce the Demographic Attribute Inference from Questions (DAIQ) task and a comprehensive evaluation framework, conducting the first cross-model audit across major open- and closed-source LLMs. Methodologically, they construct a neutral benchmark query set and employ both quantitative metrics and qualitative analysis, empirically demonstrating pervasive demographic inference biases across diverse LLMs. Furthermore, they propose a prompt-engineering–based mitigation strategy that effectively suppresses such inference without modifying model parameters. This work advances understanding of implicit social biases in LLMs and delivers a practical, deployable approach to enhance model fairness.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) are known to reflect social biases when demographic attributes, such as gender or race, are explicitly present in the input. But even in their absence, these models still infer user identities based solely on question phrasing. This subtle behavior has received far less attention, yet poses serious risks: it violates expectations of neutrality, infers unintended demographic information, and encodes stereotypes that undermine fairness in various domains including healthcare, finance and education. We introduce Demographic Attribute Inference from Questions (DAIQ), a task and framework for auditing an overlooked failure mode in language models: inferring user demographic attributes from questions that lack explicit demographic cues. Our approach leverages curated neutral queries, systematic prompting, and both quantitative and qualitative analysis to uncover how models infer demographic information. We show that both open and closed source LLMs do assign demographic labels based solely on question phrasing. Prevalence and consistency of demographic inferences across diverse models reveal a systemic and underacknowledged risk: LLMs can fabricate demographic identities, reinforce societal stereotypes, and propagate harms that erode privacy, fairness, and trust posing a broader threat to social equity and responsible AI deployment. To mitigate this, we develop a prompt-based guardrail that substantially reduces identity inference and helps align model behavior with fairness and privacy objectives.
Problem

Research questions and friction points this paper is trying to address.

LLMs infer user demographics from questions lacking explicit cues
This poses risks to privacy, fairness, and trust in AI systems
Models reinforce stereotypes across healthcare, finance, and education domains
Innovation

Methods, ideas, or system contributions that make the work stand out.

Auditing demographic inference from neutral queries
Leveraging systematic prompting and quantitative analysis
Developing prompt-based guardrail to reduce identity inference
🔎 Similar Papers
No similar papers found.