Context-dependent sentiment analysis in user-generated videos

692Citations
Citations of this article
465Readers
Mendeley users who have this article in their library.

Abstract

Multimodal sentiment analysis is a developing area of research, which involves the identification of sentiments in videos. Current research considers utterances as independent entities, i.e., ignores the inter-dependencies and relations among the utterances of a video. In this paper, we propose a LSTM-based model that enables utterances to capture contextual information from their surroundings in the same video, thus aiding the classification process. Our method shows 5-10% performance improvement over the state of the art and high robustness to generalizability.

Cite

CITATION STYLE

APA

Poria, S., Mazumder, N., Cambria, E., Hazarika, D., Morency, L. P., & Zadeh, A. (2017). Context-dependent sentiment analysis in user-generated videos. In ACL 2017 - 55th Annual Meeting of the Association for Computational Linguistics, Proceedings of the Conference (Long Papers) (Vol. 1, pp. 873–883). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/P17-1081

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free