Revisiting offline evaluation for implicit-feedback recommender systems

33Citations
Citations of this article
54Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Recommender systems are typically evaluated in an offline setting. A subset of the available user-item interactions is sampled to serve as test set, and some model trained on the remaining data points is then evaluated on its performance to predict which interactions were left out. Alternatively, in an online evaluation setting, multiple versions of the system are deployed and various metrics for those systems are recorded. Systems that score better on these metrics, are then typically preferred. Online evaluation is effective, but inefficient for a number of reasons. Offline evaluation is much more efficient, but current methodologies often fail to accurately predict online performance. In this work, we identify three ways to improve and extend current work on offline evaluation methodologies. More specifically, we believe there is much room for improvement in temporal evaluation, off-policy evaluation, and moving beyond using just clicks to evaluate performance.

Cite

CITATION STYLE

APA

Jeunen, O. (2019). Revisiting offline evaluation for implicit-feedback recommender systems. In RecSys 2019 - 13th ACM Conference on Recommender Systems (pp. 596–600). Association for Computing Machinery, Inc. https://doi.org/10.1145/3298689.3347069

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free