Mitigate Position Bias in Large Language Models via Scaling a Single Dimension
Abstract
Large Language Models (LLMs) are increasingly applied in various real-world scenarios due to their excellent generalization capabilities and robust generative abilities. However, they exhibit position bias, also known as "lost in the middle", a phenomenon that is especially pronounced in long-context scenarios, which indicates the placement of the key information in different positions of a prompt can significantly affect accuracy. This paper first explores the micro-level manifestations of position bias, concluding that attention weights are a micro-level expression of position bias. It further identifies that, in addition to position embeddings, causal attention mask also contributes to position bias by creating position-specific hidden states. Based on these insights, we propose a method to mitigate position bias by scaling this positional hidden states. Experiments on the NaturalQuestions Multi-document QA, KV retrieval, LongBench and timeline reorder tasks, using various models including RoPE models, context window-extended models, and Alibi models, demonstrate the effectiveness and generalizability of our approach. Our method can improve performance by up to 15.2% by modifying just one dimension of hidden states.
Cite
Text
Yu et al. "Mitigate Position Bias in Large Language Models via Scaling a Single Dimension." ICML 2024 Workshops: LCFM, 2024.Markdown
[Yu et al. "Mitigate Position Bias in Large Language Models via Scaling a Single Dimension." ICML 2024 Workshops: LCFM, 2024.](https://mlanthology.org/icmlw/2024/yu2024icmlw-mitigate/)BibTeX
@inproceedings{yu2024icmlw-mitigate,
title = {{Mitigate Position Bias in Large Language Models via Scaling a Single Dimension}},
author = {Yu, Yijiong and Jiang, Huiqiang and Luo, Xufang and Wu, Qianhui and Lin, Chin-Yew and Li, Dongsheng and Yang, Yuqing and Huang, Yongfeng and Qiu, Lili},
booktitle = {ICML 2024 Workshops: LCFM},
year = {2024},
url = {https://mlanthology.org/icmlw/2024/yu2024icmlw-mitigate/}
}