Query Generation Using Large Language Models A Reproducibility Study of Unsupervised Passage Reranking
| Authors | |
|---|---|
| Publication date | 2024 |
| Host editors |
|
| Book title | Advances in Information Retrieval |
| Book subtitle | 46th European Conference on Information Retrieval, ECIR 2024, Glasgow, UK, March 24–28, 2024 : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Lecture Notes in Computer Science |
| Event | 46th European Conference on Information Retrieval |
| Volume | Issue number | IV |
| Pages (from-to) | 226-239 |
| Number of pages | 14 |
| Publisher | Cham: Springer |
| Organisations |
|
| Abstract |
Existing passage retrieval techniques predominantly emphasize classification or dense matching strategies. This is in contrast with classic language modeling approaches focusing on query or question generation. Recently, Sachan et al. introduced an Unsupervised Passage Retrieval (UPR) approach that resembles this by exploiting the inherent generative capabilities of large language models. In this replicability study, we revisit the concept of zero-shot question generation for re-ranking and focus our investigation on the ranking experiments, validating the UPR findings, particularly on the widely recognized BEIR benchmark. Furthermore, we extend the original work by evaluating the proposed method additionally on the TREC Deep Learning track benchmarks of 2019 and 2020. To enhance our understanding of the technique’s performance, we introduce novel experiments exploring the influence of different prompts on retrieval outcomes. Our comprehensive analysis provides valuable insights into the robustness and applicability of zero-shot question generation as a re-ranking strategy in passage retrieval.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-031-56066-8_19 |
| Downloads |
978-3-031-56066-8_19
(Final published version)
|
| Permalink to this page | |
