Beyond Sociodemographic Prompting: Using Supervision to Align LLMs with Human Response Distributions

📅 2025-07-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This paper addresses the challenge of accurately predicting response distributions across diverse demographic groups for subjective questions using large language models (LLMs). We propose a lightweight supervised alignment mechanism that replaces complex sociodemographic prompting with simple, universal group labels as supervision signals, guiding LLMs to learn consistent response distributions across groups. Our method is adaptable across multiple LLMs and prompting strategies on multi-topic datasets and supports quantitative evaluation of distributional alignment. Experiments demonstrate significant improvements in cross-group response distribution prediction accuracy across multiple benchmarks, with strong generalizability and model-agnostic performance. We open-source all code, data, and evaluation tools, establishing the first reproducible benchmark for cross-group response distribution alignment. This work introduces a novel paradigm for fairness-aware modeling and socially aware AI.

Technology Category

Application Category

📝 Abstract
The ability to accurately predict how different population groups would answer subjective questions would have great value. In this work, we show that use of relatively simple supervision can greatly improve language model alignment with diverse population groups, as measured over three datasets spanning various topics. Beyond evaluating average performance, we also report how alignment varies across specific groups. The simplicity and generality of our approach promotes easy adoption, while our broad findings provide useful guidance for when to use or not use our approach in practice. By conducting evaluation over many LLMs and prompting strategies, along with open-sourcing our work, we provide a useful benchmark to stimulate future research.
Problem

Research questions and friction points this paper is trying to address.

Improving LLM alignment with diverse human response distributions
Evaluating alignment performance across specific population groups
Providing practical guidance for adopting the supervision approach
Innovation

Methods, ideas, or system contributions that make the work stand out.

Simple supervision improves LLM alignment
Evaluates alignment across diverse population groups
Open-sourced benchmark for future research
🔎 Similar Papers
No similar papers found.
G
Gauri Kambhatla
The University of Texas at Austin
Sanjana Gautam
Sanjana Gautam
University of Texas at Austin
Human-Centered AISocio-technical NLP
A
Angela Zhang
The University of Texas at Austin
Alex Liu
Alex Liu
University of Washington
AI in educationStrategic Teacher EngagementK-12 education policy
R
Ravi Srinivasan
The University of Texas at Austin
Junyi Jessy Li
Junyi Jessy Li
Associate Professor, The University of Texas at Austin
Computational LinguisticsNatural Language Processing
M
Matthew Lease
The University of Texas at Austin