Overview of the INEX 2010 book track: Scaling up the evaluation using crowdsourcing
| Authors |
|
|---|---|
| Publication date | 2011 |
| Host editors |
|
| Book title | Comparative Evaluation of Focused Retrieval |
| Book subtitle | 9th International Workshop of the Inititative for the Evaluation of XML Retrieval, INEX 2010, Vugh, The Netherlands, December 13-15, 2010 : revised selected papers |
| ISBN |
|
| ISBN (electronic) |
|
| Series | Lecture Notes in Computer Science |
| Event | 9th International Workshop of the INitiative for the Evaluation of XML retrieval (INEX 2010), Vught, the Netherlands |
| Pages (from-to) | 98-117 |
| Publisher | Heidelberg: Springer |
| Organisations |
|
| Abstract |
The goal of the INEX Book Track is to evaluate approaches for supporting users in searching, navigating and reading the full texts of digitized books. The investigation is focused around four tasks: 1) Best Books to Reference, 2) Prove It, 3) Structure Extraction, and 4) Active Reading. In this paper, we report on the setup and the results of these tasks in 2010. The main outcome of the track lies in the changes to the methodology for constructing the test collection for the evaluation of the Best Books and Prove It search tasks. In an effort to scale up the evaluation, we explored the use of crowdsourcing both to create the test topics and then to gather the relevance labels for the topics over a corpus of 50k digitized books. The resulting test collection construction methodology combines editorial judgments contributed by INEX participants with crowdsourced relevance labels. We provide an analysis of the crowdsourced data and conclude that – with appropriate task design – crowdsourcing does provide a suitable framework for the evaluation of book search approaches.
|
| Document type | Conference contribution |
| Language | English |
| Published at | https://doi.org/10.1007/978-3-642-23577-1_9 |
| Permalink to this page | |
