Abstract
Online ranker evaluation is a key challenge in information retrieval. An important task in the online evaluation of rankers is using implicit user feedback for inferring preferences between rankers. Interleaving methods have been found to be efficient and sensitive, i.e. they can quickly detect even small differences in quality. It has recently been shown that multileaving methods exhibit similar sensitivity but can be more efficient than interleaving methods. This paper presents empirical results demonstrating that existing multileaving methods either do not scale well with the number of rankers, or, more problematically, can produce results which substantially differ from evaluation measures like NDCG. The latter problem is caused by the fact that they do not correctly account for the similarities that can occur between rankers being multileaved. We propose a new multileaving method for handling this problem and demonstrate that it substantially outperforms existing methods, in some cases reducing errors by as much as 50%.
Originalsprog | Engelsk |
---|---|
Titel | Proceedings of the 39th International ACM SIGIR conference on Research and Development in Information Retrieval : SIGIR '16 |
Antal sider | 4 |
Forlag | Association for Computing Machinery |
Publikationsdato | 7 jul. 2016 |
Sider | 745-748 |
ISBN (Trykt) | 978-1-4503-4069-4 |
DOI | |
Status | Udgivet - 7 jul. 2016 |
Begivenhed | International ACM SIGIR conference on Research and Development in Information Retrieval 2016: SIGIR '16 - Pisa, Italien Varighed: 17 jul. 2016 → 21 jul. 2016 Konferencens nummer: 39 http://sigir.org/sigir2016/ |
Konference
Konference | International ACM SIGIR conference on Research and Development in Information Retrieval 2016 |
---|---|
Nummer | 39 |
Land/Område | Italien |
By | Pisa |
Periode | 17/07/2016 → 21/07/2016 |
Internetadresse |