Position of Uncertainty: A Cross-Linguistic Study of Positional Bias in Large Language Models

📅 2025-05-22
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study investigates the interaction between positional bias and linguistic diversity in large language models (LLMs), challenging the prevailing assumption that earlier tokens are more reliable. Method: Using entropy analysis, controlled positional prompting experiments, cross-lingual syntactic alignment evaluation, and systematic prompt perturbations, we examine Qwen2.5-7B across five typologically diverse languages—English, Russian, German, Hindi, and Vietnamese. Contribution/Results: We identify a late-position preference in Qwen2.5-7B, contrary to dominant early-position assumptions; demonstrate that explicit positional cues degrade accuracy; and reveal that LLMs impose rigid dominant word order even in free-word-order languages—e.g., enforcing SOV in Hindi despite its flexibility. We quantify language-specific positional bias patterns, empirically falsifying the “earlier is better” consensus. Moreover, we establish a non-monotonic relationship between uncertainty and positional bias, offering novel empirical foundations for robust prompt engineering and model calibration.

Technology Category

Application Category

📝 Abstract
Large language models exhibit positional bias -- systematic neglect of information at specific context positions -- yet its interplay with linguistic diversity remains poorly understood. We present a cross-linguistic study across five typologically distinct languages (English, Russian, German, Hindi, Vietnamese), examining how positional bias interacts with model uncertainty, syntax, and prompting. Key findings: (1) Positional bias is model-driven, with language-specific variations -- Qwen2.5-7B favors late positions, challenging assumptions of early-token bias; (2) Explicit positional guidance (e.g., correct context is at position X) reduces accuracy across languages, undermining prompt-engineering practices; (3) Aligning context with positional bias increases entropy, yet minimal entropy does not predict accuracy. (4) We further uncover that LLMs differently impose dominant word order in free-word-order languages like Hindi.
Problem

Research questions and friction points this paper is trying to address.

Examines positional bias in LLMs across diverse languages
Investigates how positional bias interacts with model uncertainty and syntax
Challenges assumptions about early-token bias and prompt-engineering practices
Innovation

Methods, ideas, or system contributions that make the work stand out.

Cross-linguistic study on positional bias in LLMs
Explicit positional guidance reduces model accuracy
Aligning context with bias increases entropy unpredictably
🔎 Similar Papers
No similar papers found.
M
Menschikov Mikhail
ITMO, Saint-Petersburg, Russia; Skoltech, Moscow, Russia
Alexander Kharitonov
Alexander Kharitonov
Natural Language Processing Researcher
Natural Language ProcessingArtificial Intelligence
M
Maiia Kotyga
MIPT, Moscow, Russia
V
Vadim Porvatov
Skoltech, Moscow, Russia; Sber, Moscow, Russia
A
Anna Zhukovskaya
Lomonosov MSU, Moscow, Russia
D
David Kagramanyan
HSE University, Moscow, Russia
E
Egor Shvetsov
Skoltech, Moscow, Russia
Evgeny Burnaev
Evgeny Burnaev
Skoltech, Full Professor, Head of AI center, Head of research group, AIRI
Generative ModelingManifold LearningSurrogate Modeling3D Deep Learning