Reasoning Boosts Opinion Alignment in LLMs

📅 2026-03-01
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This work addresses the susceptibility of large language models to inherent biases when generating political opinions under direct prompting, which often impedes their ability to accurately reflect individuals’ or groups’ genuine political preferences. To mitigate this limitation, the paper introduces— for the first time—a reinforcement learning–based structured reasoning mechanism that guides the model to perform stepwise inference grounded in user profiles, thereby producing more consistent and better-aligned political viewpoints. The proposed approach is rigorously evaluated on three major political datasets from the United States, Europe, and Switzerland, demonstrating significant improvements in opinion consistency across diverse political contexts. It achieves performance comparable to strong baselines while highlighting the potential of structured reasoning as a promising avenue for mitigating bias in language models.

Technology Category

Application Category

📝 Abstract
Opinion modeling aims to capture individual or group political preferences, enabling applications such as digital democracies, where models could help shape fairer and more popular policies. Given their versatility, strong generalization capabilities, and demonstrated success across diverse text-to-text applications, large language models (LLMs) are natural candidates for this task. However, due to their statistical nature and limited causal understanding, they tend to produce biased opinions when prompted naively. In this work, we study whether reasoning can improve opinion alignment. Motivated by the recent advancement in mathematical reasoning enabled by reinforcement learning (RL), we train models to produce profile-consistent answers through structured reasoning. We evaluate our approach on three datasets covering U.S., European, and Swiss politics. Results indicate that reasoning enhances opinion modeling and is competitive with strong baselines, but does not fully remove bias, highlighting the need for additional mechanisms to build faithful political digital twins using LLMs. By releasing both our method and datasets, we establish a solid baseline to support future research on LLM opinion alignment.
Problem

Research questions and friction points this paper is trying to address.

opinion alignment
large language models
political bias
opinion modeling
digital twins
Innovation

Methods, ideas, or system contributions that make the work stand out.

reasoning
opinion alignment
large language models
structured reasoning
political digital twins
🔎 Similar Papers
No similar papers found.