Multimodal review generation for recommender systems

64Citations
Citations of this article
80Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Key to recommender systems is learning user preferences, which are expressed through various modalities. In online reviews, for instance, this manifests in numerical rating, textual content, as well as visual images. In this work, we hypothesize that modelling these modalities jointly would result in a more holistic representation of a review towards more accurate recommendations. Therefore, we propose Multimodal Review Generation (MRG), a neural approach that simultaneously models a rating prediction component and a review text generation component. We hypothesize that the shared user and item representations would augment the rating prediction with richer information from review text, while sensitizing the generated review text to sentiment features based on user and item of interest. Moreover, when review photos are available, visual features could inform the review text generation further. Comprehensive experiments on real-life datasets from several major US cities show that the proposed model outperforms comparable multimodal baselines, while an ablation analysis establishes the relative contributions of the respective components of the joint model.

Cite

CITATION STYLE

APA

Truong, Q. T., & Lauw, H. W. (2019). Multimodal review generation for recommender systems. In The Web Conference 2019 - Proceedings of the World Wide Web Conference, WWW 2019 (pp. 1864–1874). Association for Computing Machinery, Inc. https://doi.org/10.1145/3308558.3313463

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free