Bias and Unfairness of Collaborative Filtering Based Recommender Systems in MovieLens Dataset

13Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

Recommender Systems have become one of the most important tools for streaming and marketplace systems in recent years. Their increased use has revealed clear bias and unfairness against minorities and underrepresented groups. This paper seeks the origin of these biases and unfairness. To this end, it analyzes the demographic characteristics of a gold standard dataset and its prediction performance when used in a multitude of Recommender Systems. In addition, this paper proposes Soft Matrix Factorization (SoftMF), which tries to balance the predictions of different types of users to reduce the present inequality. The experimental results show that those biases and unfairness are not introduced by the different recommendation models and that they come from the socio-psychological and demographic characteristics of the used dataset.

Cite

CITATION STYLE

APA

Gonzalez, A., Ortega, F., Perez-Lopez, D., & Alonso, S. (2022). Bias and Unfairness of Collaborative Filtering Based Recommender Systems in MovieLens Dataset. IEEE Access, 10, 68429–68439. https://doi.org/10.1109/ACCESS.2022.3186719

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free