Learning Robust Recommenders through Cross-Model Agreement

58Citations
Citations of this article
16Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Learning from implicit feedback is one of the most common cases in the application of recommender systems. Generally speaking, interacted examples are considered as positive while negative examples are sampled from uninteracted ones. However, noisy examples are prevalent in real-world implicit feedback. A noisy positive example could be interacted but it actually leads to negative user preference. A noisy negative example which is uninteracted because of user unawareness could also denote potential positive user preference. Conventional training methods overlook these noisy examples, leading to sub-optimal recommendations. In this work, we propose a general framework to learn robust recommenders from implicit feedback. Through an empirical study, we find that different models make relatively similar predictions on clean examples which denote the real user preference, while the predictions on noisy examples vary much more across different models. Motivated by this observation, we propose denoising with cross-model agreement (DeCA) which minimizes the KL-divergence between the real user preference distributions parameterized by two recommendation models while maximizing the likelihood of data observation. We instantiate DeCA on four representative recommendation models, empirically demonstrating its superiority over normal training and existing denoising methods. Codes are available at https://github.com/wangyu-ustc/DeCA.

Cite

CITATION STYLE

APA

Wang, Y., Xin, X., Meng, Z., Jose, J. M., Feng, F., & He, X. (2022). Learning Robust Recommenders through Cross-Model Agreement. In WWW 2022 - Proceedings of the ACM Web Conference 2022 (pp. 2015–2025). Association for Computing Machinery, Inc. https://doi.org/10.1145/3485447.3512202

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free