Position of Uncertainty: A Cross-Linguistic Study of Positional Bias in Large Language Models
Position of Uncertainty: A Cross-Linguistic Study of Positional Bias in Large Language Models
Large language models exhibit positional bias -- systematic neglect of information at specific context positions -- yet its interplay with linguistic diversity remains poorly understood. We present a cross-linguistic study across five typologically distinct languages (English, Russian, German, Hindi, Vietnamese), examining how positional bias interacts with model uncertainty, syntax, and prompting. Key findings: (1) Positional bias is model-driven, with language-specific variations -- Qwen2.5-7B favors late positions, challenging assumptions of early-token bias; (2) Explicit positional guidance (e.g., correct context is at position X) reduces accuracy across languages, undermining prompt-engineering practices; (3) Aligning context with positional bias increases entropy, yet minimal entropy does not predict accuracy. (4) We further uncover that LLMs differently impose dominant word order in free-word-order languages like Hindi.
Menschikov Mikhail、Alexander Kharitonov、Maiia Kotyga、Vadim Porvatov、Anna Zhukovskaya、David Kagramanyan、Egor Shvetsov、Evgeny Burnaev
语言学印欧语系南亚语系(澳斯特罗-亚细亚语系)
Menschikov Mikhail,Alexander Kharitonov,Maiia Kotyga,Vadim Porvatov,Anna Zhukovskaya,David Kagramanyan,Egor Shvetsov,Evgeny Burnaev.Position of Uncertainty: A Cross-Linguistic Study of Positional Bias in Large Language Models[EB/OL].(2025-05-21)[2025-06-29].https://arxiv.org/abs/2505.16134.点此复制
评论