Contextual Bandits for Information Retrieval

Authors
Publication date 2011
Book title NIPS 2011: Proceedings of the Conference on Neural Information Processing Systems, Workshop on Bayesian Optimization, Experimental Design and Bandits: Theory and Applications
Pages (from-to) 1-5
Publisher NIPS
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
In this paper we give an overview of and outlook on research at the intersection of information retrieval (IR) and contextual bandit problems. A critical problem in information retrieval is online learning to rank, where a search engine strives to improve the quality of the ranked result lists it presents to users on the basis of those users’ interactions with those result lists. Recently, researchers have started to model interactions between users and search engines as contextual bandit problems, and initial methods for learning in this setting have been devised. Our research focuses on two aspects: balancing exploration and exploitation and inferring preferences from implicit user interactions. This paper summarizes our recent work on online learning to rank for information retrieval and points out challenges that are characteristic of this application area.
Document type Conference contribution
Language English
Published at http://staff.science.uva.nl/~whiteson/pubs/hofmannnips11.pdf
Permalink to this page
Back