Probabilistic multileave gradient descent

18Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Online learning to rank methods aim to optimize ranking models based on user interactions. The dueling bandit gradient descent (DBGD) algorithm is able to effectively optimize linear ranking models solely from user interactions. We propose an extension of DBGD, called probabilistic multileave gradient descent (P-MGD) that builds on probabilistic multileave, a recently proposed highly sensitive and unbiased online evaluation method. We demonstrate that P-MGD significantly outperforms state-of-the-art online learning to rank methods in terms of online performance, without sacrificing offline performance and at greater learning speed.

Cite

CITATION STYLE

APA

Oosterhuis, H., Schuth, A., & de Rijke, M. (2016). Probabilistic multileave gradient descent. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9626, pp. 661–668). Springer Verlag. https://doi.org/10.1007/978-3-319-30671-1_50

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free