Overview of the INEX 2010 book track: Scaling up the evaluation using crowdsourcing

Authors
  • M. Landoni
Publication date 2011
Host editors
  • S. Geva
  • J. Kamps
  • R. Schenkel
  • A. Trotman
Book title Comparative Evaluation of Focused Retrieval
Book subtitle 9th International Workshop of the Inititative for the Evaluation of XML Retrieval, INEX 2010, Vugh, The Netherlands, December 13-15, 2010 : revised selected papers
ISBN
  • 9783642235764
ISBN (electronic)
  • 9783642235771
Series Lecture Notes in Computer Science
Event 9th International Workshop of the INitiative for the Evaluation of XML retrieval (INEX 2010), Vught, the Netherlands
Pages (from-to) 98-117
Publisher Heidelberg: Springer
Organisations
  • Interfacultary Research - Institute for Logic, Language and Computation (ILLC)
Abstract
The goal of the INEX Book Track is to evaluate approaches for supporting users in searching, navigating and reading the full texts of digitized books. The investigation is focused around four tasks: 1) Best Books to Reference, 2) Prove It, 3) Structure Extraction, and 4) Active Reading. In this paper, we report on the setup and the results of these tasks in 2010. The main outcome of the track lies in the changes to the methodology for constructing the test collection for the evaluation of the Best Books and Prove It search tasks. In an effort to scale up the evaluation, we explored the use of crowdsourcing both to create the test topics and then to gather the relevance labels for the topics over a corpus of 50k digitized books. The resulting test collection construction methodology combines editorial judgments contributed by INEX participants with crowdsourced relevance labels. We provide an analysis of the crowdsourced data and conclude that – with appropriate task design – crowdsourcing does provide a suitable framework for the evaluation of book search approaches.
Document type Conference contribution
Language English
Published at https://doi.org/10.1007/978-3-642-23577-1_9
Permalink to this page
Back