The rating game: Sentiment rating reproducibility from text

6Citations
Citations of this article
94Readers
Mendeley users who have this article in their library.

Abstract

Sentiment analysis models often use ratings as labels, assuming that these ratings reflect the sentiment of the accompanying text. We investigate (i) whether human readers can infer ratings from review text, (ii) how human performance compares to a regression model, and (iii) whether model performance is affected by the rating "source" (i.e. original author vs. annotator). We collect IMDb movie reviews with author-provided ratings, and have them re-annotated by crowdsourced and trained annotators. Annotators reproduce the original ratings better than a model, but are still far off in more than 5% of the cases. Models trained on annotator-labels outperform those trained on author-labels, questioning the usefulness of author-rated reviews as training data for sentiment analysis.

Cite

CITATION STYLE

APA

Borgholt, L., Simonsen, P., & Hovy, D. (2015). The rating game: Sentiment rating reproducibility from text. In Conference Proceedings - EMNLP 2015: Conference on Empirical Methods in Natural Language Processing (pp. 2527–2532). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d15-1301

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free