Mitigate Position Bias in Large Language Models via Scaling a Single Dimension

Abstract

Large Language Models (LLMs) are increasingly applied in various real-world scenarios due to their excellent generalization capabilities and robust generative abilities. However, they exhibit position bias, also known as "lost in the middle", a phenomenon that is especially pronounced in long-context scenarios, which indicates the placement of the key information in different positions of a prompt can significantly affect accuracy. This paper first explores the micro-level manifestations of position bias, concluding that attention weights are a micro-level expression of position bias. It further identifies that, in addition to position embeddings, causal attention mask also contributes to position bias by creating position-specific hidden states. Based on these insights, we propose a method to mitigate position bias by scaling this positional hidden states. Experiments on the NaturalQuestions Multi-document QA, KV retrieval, LongBench and timeline reorder tasks, using various models including RoPE models, context window-extended models, and Alibi models, demonstrate the effectiveness and generalizability of our approach. Our method can improve performance by up to 15.2% by modifying just one dimension of hidden states.

Cite

Text

Yu et al. "Mitigate Position Bias in Large Language Models via Scaling a Single Dimension." ICML 2024 Workshops: LCFM, 2024.

Markdown

[Yu et al. "Mitigate Position Bias in Large Language Models via Scaling a Single Dimension." ICML 2024 Workshops: LCFM, 2024.](https://mlanthology.org/icmlw/2024/yu2024icmlw-mitigate/)

BibTeX

@inproceedings{yu2024icmlw-mitigate,
  title     = {{Mitigate Position Bias in Large Language Models via Scaling a Single Dimension}},
  author    = {Yu, Yijiong and Jiang, Huiqiang and Luo, Xufang and Wu, Qianhui and Lin, Chin-Yew and Li, Dongsheng and Yang, Yuqing and Huang, Yongfeng and Qiu, Lili},
  booktitle = {ICML 2024 Workshops: LCFM},
  year      = {2024},
  url       = {https://mlanthology.org/icmlw/2024/yu2024icmlw-mitigate/}
}