Validating query simulators: an experiment using commercial searches and purchases
| Authors |
|
|---|---|
| Publication date | 2010 |
| Host editors |
|
| Book title | Multilingual and Multimodal Information Access Evaluation |
| Book subtitle | international conference of the Cross-Language Evaluation Forum, CLEF 2010, Padua, Italy, September 20-23, 2010 : proceedings |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Lecture Notes in Computer Science |
| Event | Multilingual and Multimodal Information Access Evaluation: International conference of the Cross-Language Evaluation Forum (CLEF 2010), Padua, Italy |
| Pages (from-to) | 40-51 |
| Publisher | Berlin: Springer |
| Organisations |
|
| Abstract |
We design and validate simulators for generating queries and relevance judgments for retrieval system evaluation. We develop a simulation framework that incorporates existing and new simulation strategies. To validate a simulator, we assess whether evaluation using its output data ranks retrieval systems in the same way as evaluation using real-world data. The real-world data is obtained using logged commercial searches and associated purchase decisions. While no simulator reproduces an ideal ranking, there is a large variation in simulator performance that allows us to distinguish those that are better suited to creating artificial testbeds for retrieval experiments. Incorporating knowledge about document structure in the query generation process helps create more realistic simulators.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-642-15998-5_6 |
| Permalink to this page | |
