Recommender systems are typically evaluated in an offline setting. A subset of the available user-item interactions is sampled to serve as test set, and some model trained on the remaining data points is then evaluated on its performance to predict which interactions were left out. Alternatively, in an online evaluation setting, multiple versions of the system are deployed and various metrics for those systems are recorded. Systems that score better on these metrics, are then typically preferred. Online evaluation is effective, but inefficient for a number of reasons. Offline evaluation is much more efficient, but current methodologies often fail to accurately predict online performance. In this work, we identify three ways to improve and extend current work on offline evaluation methodologies. More specifically, we believe there is much room for improvement in temporal evaluation, off-policy evaluation, and moving beyond using just clicks to evaluate performance.
CITATION STYLE
Jeunen, O. (2019). Revisiting offline evaluation for implicit-feedback recommender systems. In RecSys 2019 - 13th ACM Conference on Recommender Systems (pp. 596–600). Association for Computing Machinery, Inc. https://doi.org/10.1145/3298689.3347069
Mendeley helps you to discover research relevant for your work.