Squeezing Water from a Stone: A Bag of Tricks for Further Improving Cross-Encoder Effectiveness for Reranking

Open Access
Authors
  • R. Pradeep
  • Y. Liu
  • X. Zhang
  • Y. Li
Publication date 2022
Host editors
  • M. Hagen
  • S. Verberne
  • C. Macdonald
  • C. Seifert
  • K. Balog
  • K. Nørvåg
  • V. Setty
Book title Advances in Information Retrieval
Book subtitle 44th European Conference on IR Research, ECIR 2022, Stavanger, Norway, April 10–14, 2022 : proceedings
ISBN
  • 9783030997359
ISBN (electronic)
  • 9783030997366
Series Lecture Notes in Computer Science
Event 44th European Conference on IR Research
Volume | Issue number I
Pages (from-to) 655–670
Publisher Cham: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
While much recent work has demonstrated that hard negative mining can be used to train better bi-encoder models, few have considered it in the context of cross-encoders, which are key ingredients in modern retrieval pipelines due to their high effectiveness. One noteworthy exception comes from Gao et al. [13], who propose to train cross-encoders by adapting the well-known NCE loss and augmenting it with a “localized” selection of hard negative examples from the first-stage retriever, which they call the Localized Contrastive Estimation (LCE) loss. In this work, we present a replication study of LCE on a different task and combine it with several other “tricks” (e.g., replacing BERTBase with ELECTRABase and replacing BM25 with TCT-ColBERTv2) to substantially improve ranking effectiveness. We attempt to more systematically explore certain parts of the hyperparameter space, including the choice of losses and the group size in the LCE loss. While our findings, for the most part, align with those from the original paper, we observe that for MS MARCO passage, orienting the retriever used for hard negative mining with the first-stage retriever used for inference is not as critical for improving effectiveness across all settings. Our code and documentation can be found in: https://github.com/castorini/replicate-lce.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-030-99736-6_44
Other links https://github.com/castorini/replicate-lce.
Downloads
978-3-030-99736-6_44 (Final published version)
Permalink to this page
Back