Lost but Not Only in the Middle Positional Bias in Retrieval Augmented Generation

Open Access
Authors
Publication date 2025
Host editors
  • C. Hauff
  • C. Macdonald
  • D. Jannach
  • G. Kazai
  • F.M. Nardini
  • F. Pinelli
  • F. Silvestri
  • N. Tonellotto
Book title Advances in Information Retrieval
Book subtitle 47th European Conference on Information Retrieval, ECIR 2025, Lucca, Italy, April 6–10, 202 : proceedings
ISBN
  • 9783031887079
ISBN (electronic)
  • 9783031887086
Series Lecture Notes in Computer Science
Event 47th European Conference on Information Retrieval
Volume | Issue number I
Pages (from-to) 247-261
Number of pages 15
Publisher Cham: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
Large language models (LLMs) are known to exhibit positional bias, the tendency of models to perform differently based on where relevant information appears within the input context. Understanding this bias is important in a retrieval augmented generation (RAG) setting, as it impacts how retrieved passages are taken into account by the model. We systematically investigate positional bias in a RAG setting, by evaluating four LLMs using three different types of distractor documents, assessing their ability to extract relevant information from the input context. Our findings reveal significant positional bias depending on the type of context documents used and the total amount of documents in the context. Furthermore, the results show that positional bias in state-of-the-art LLMs is not limited to information located in the middle of the input context. By analyzing the models’ attention, we identified patterns between the model’s accuracy in responding to questions and the correct attribution of attention to relevant information in the context. Our code is available at https://github.com/Janhutter/LBNOITM.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-031-88708-6_16
Other links https://github.com/Janhutter/LBNOITM
Downloads
978-3-031-88708-6_16 (Final published version)
Permalink to this page
Back