Lost but Not Only in the Middle Positional Bias in Retrieval Augmented Generation
| Authors | |
|---|---|
| Publication date | 2025 |
| Host editors |
|
| Book title | Advances in Information Retrieval |
| Book subtitle | 47th European Conference on Information Retrieval, ECIR 2025, Lucca, Italy, April 6–10, 202 : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Lecture Notes in Computer Science |
| Event | 47th European Conference on Information Retrieval |
| Volume | Issue number | I |
| Pages (from-to) | 247-261 |
| Number of pages | 15 |
| Publisher | Cham: Springer |
| Organisations |
|
| Abstract |
Large language models (LLMs) are known to exhibit positional bias, the tendency of models to perform differently based on where relevant information appears within the input context. Understanding this bias is important in a retrieval augmented generation (RAG) setting, as it impacts how retrieved passages are taken into account by the model. We systematically investigate positional bias in a RAG setting, by evaluating four LLMs using three different types of distractor documents, assessing their ability to extract relevant information from the input context. Our findings reveal significant positional bias depending on the type of context documents used and the total amount of documents in the context. Furthermore, the results show that positional bias in state-of-the-art LLMs is not limited to information located in the middle of the input context. By analyzing the models’ attention, we identified patterns between the model’s accuracy in responding to questions and the correct attribution of attention to relevant information in the context. Our code is available at https://github.com/Janhutter/LBNOITM.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-031-88708-6_16 |
| Other links | https://github.com/Janhutter/LBNOITM |
| Downloads |
978-3-031-88708-6_16
(Final published version)
|
| Permalink to this page | |
