Validating query simulators: an experiment using commercial searches and purchases

Authors
Publication date 2010
Host editors
  • M. Agosti
  • N. Ferro
  • C. Peters
  • M. de Rijke
  • A. Smeaton
Book title Multilingual and Multimodal Information Access Evaluation
Book subtitle international conference of the Cross-Language Evaluation Forum, CLEF 2010, Padua, Italy, September 20-23, 2010 : proceedings
ISBN
  • 9783642159978
ISBN (electronic)
  • 9783642159985
Series Lecture Notes in Computer Science
Event Multilingual and Multimodal Information Access Evaluation: International conference of the Cross-Language Evaluation Forum (CLEF 2010), Padua, Italy
Pages (from-to) 40-51
Publisher Berlin: Springer
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
We design and validate simulators for generating queries and relevance judgments for retrieval system evaluation. We develop a simulation framework that incorporates existing and new simulation strategies. To validate a simulator, we assess whether evaluation using its output data ranks retrieval systems in the same way as evaluation using real-world data. The real-world data is obtained using logged commercial searches and associated purchase decisions. While no simulator reproduces an ideal ranking, there is a large variation in simulator performance that allows us to distinguish those that are better suited to creating artificial testbeds for retrieval experiments. Incorporating knowledge about document structure in the query generation process helps create more realistic simulators.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-642-15998-5_6
Permalink to this page
Back