Monte Carlo Estimates of Evaluation Metric Error and Bias

  • Tian M
  • Ekstrand M
N/ACitations
Citations of this article
10Readers
Mendeley users who have this article in their library.

Abstract

Traditional offline evaluations of recommender systems apply met-rics from machine learning and information retrieval in settings where their underlying assumptions no longer hold. This results in significant error and bias in measures of top-N recommendation performance, such as precision, recall, and nDCG. Several of the specific causes of these errors, including popularity bias and mis-classified decoy items, are well-explored in the existing literature. In this paper we survey a range of work on identifying and addressing these problems, and report on our work in progress to simulate the recommender data generation and evaluation processes to quantify the extent of evaluation metric errors and assess their sensitivity to various assumptions.

Cite

CITATION STYLE

APA

Tian, M., & Ekstrand, M. (2018). Monte Carlo Estimates of Evaluation Metric Error and Bias. Boise State University. https://doi.org/10.18122/cs_facpubs/148/boisestate

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free