Political Ideology Shifts in Large Language Models

📅 2025-08-21
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates the ideological plasticity of large language models (LLMs) when prompted with synthetic personas. Using the Political Compass test, we systematically evaluate the political orientation responses of seven LLMs—spanning diverse parameter scales—to multiple role-based prompts, including right-authoritarian and left-libertarian personas. Our findings reveal: (1) scaling up model size significantly expands the implicit ideological spectrum and intensifies polarization, particularly under right-authoritarian prompts; (2) persona descriptions induce systematic, directional ideological shifts; and (3) ideological expression is highly context-dependent and malleable via prompt engineering. Crucially, this work provides the first empirical evidence of a synergistic interaction between model scale and persona prompting. It establishes a foundational framework for modeling LLM political behavior, advancing safety alignment strategies, and informing transparency-aware governance in high-stakes deployment scenarios.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) are increasingly deployed in politically sensitive settings, raising concerns about their potential to encode, amplify, or be steered toward specific ideologies. We investigate how adopting synthetic personas influences ideological expression in LLMs across seven models (7B-70B+ parameters) from multiple families, using the Political Compass Test as a standardized probe. Our analysis reveals four consistent patterns: (i) larger models display broader and more polarized implicit ideological coverage; (ii) susceptibility to explicit ideological cues grows with scale; (iii) models respond more strongly to right-authoritarian than to left-libertarian priming; and (iv) thematic content in persona descriptions induces systematic and predictable ideological shifts, which amplify with size. These findings indicate that both scale and persona content shape LLM political behavior. As such systems enter decision-making, educational, and policy contexts, their latent ideological malleability demands attention to safeguard fairness, transparency, and safety.
Problem

Research questions and friction points this paper is trying to address.

Investigating ideological expression shifts in LLMs under synthetic personas
Assessing scale-related susceptibility to explicit political cues
Examining systematic ideological shifts induced by persona thematic content
Innovation

Methods, ideas, or system contributions that make the work stand out.

Using synthetic personas to probe ideological expression
Leveraging Political Compass Test as standardized measurement
Analyzing scale-dependent susceptibility to ideological priming
🔎 Similar Papers
No similar papers found.
P
Pietro Bernardelle
The University of Queensland, Australia
S
Stefano Civelli
The University of Queensland, Australia
L
Leon Fröhling
GESIS, Germany
R
Riccardo Lunardi
University of Udine, Italy
Kevin Roitero
Kevin Roitero
University of Udine
Artificial IntelligenceNatural Language ProcessingCrowdsourcingInformation Retrieval
Gianluca Demartini
Gianluca Demartini
Professor at the University of Queensland
Information RetrievalSemantic WebHuman ComputationCrowdsourcing