🤖 AI Summary
This work addresses the vulnerability of large language models (LLMs) to covert ideological manipulation along the political spectrum—a risk inadequately captured by traditional left–right binary bias frameworks. We introduce the first systematic assessment of LLMs’ ideological plasticity across a continuous axis from progressive left to conservative right. To this end, we propose a multi-task, multi-dimensional ideological alignment evaluation paradigm and construct a fine-grained, human-annotated benchmark comprising ideological question answering, statement ranking, manifesto cloze, and U.S. congressional bill comprehension. Using Phi-2, Mistral, and Llama-3, we conduct supervised fine-tuning and compare its efficacy against explicit prompt-based control. Results demonstrate that fine-tuning substantially enhances ideological alignment, whereas prompt engineering yields only marginal shifts—confirming a tangible risk of gradual, non-explicit ideological manipulation in current LLMs.
📝 Abstract
Large Language Models (LLMs) have transformed natural language processing, but concerns have emerged about their susceptibility to ideological manipulation, particularly in politically sensitive areas. Prior work has focused on binary Left-Right LLM biases, using explicit prompts and fine-tuning on political QA datasets. In this work, we move beyond this binary approach to explore the extent to which LLMs can be influenced across a spectrum of political ideologies, from Progressive-Left to Conservative-Right. We introduce a novel multi-task dataset designed to reflect diverse ideological positions through tasks such as ideological QA, statement ranking, manifesto cloze completion, and Congress bill comprehension. By fine-tuning three LLMs-Phi-2, Mistral, and Llama-3-on this dataset, we evaluate their capacity to adopt and express these nuanced ideologies. Our findings indicate that fine-tuning significantly enhances nuanced ideological alignment, while explicit prompts provide only minor refinements. This highlights the models' susceptibility to subtle ideological manipulation, suggesting a need for more robust safeguards to mitigate these risks.