Relative confidence sampling for efficient on-line ranker evaluation

Authors
Publication date 2014
Book title WSDM '14: proceedings of the 7th ACM International Conference on Web Search and Data Mining: February 24-28, 2014, New York, New York, USA
ISBN
  • 9781450323512
Event 7th ACM International Conference on Web Search and Data Mining (WSDM2014)
Pages (from-to) 73-82
Publisher New York: ACM
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
A key challenge in information retrieval is that of on-line ranker evaluation: determining which one of a finite set of rankers performs the best in expectation on the basis of user clicks on presented document lists. When the presented lists are constructed using interleaved comparison methods, which interleave lists proposed by two different candidate rankers, then the problem of minimizing the total regret accumulated while evaluating the rankers can be formalized as a K-armed dueling bandits problem. In this paper, we propose a new method called relative confidence sampling (RCS) that aims to reduce cumulative regret by being less conservative than existing methods in eliminating rankers from contention. In addition, we present an empirical comparison between RCS and two state-of-the-art methods, relative upper confidence bound and SAVAGE. The results demonstrate that RCS can substantially outperform these alternatives on several large learning to rank datasets.
Document type Conference contribution
Language English
Published at https://doi.org/10.1145/2556195.2556256
Permalink to this page
Back