Real-time speech emotion and sentiment recognition for interactive dialogue systems

77Citations
Citations of this article
189Readers
Mendeley users who have this article in their library.

Abstract

In this paper, we describe our approach of enabling an interactive dialogue system to recognize user emotion and sentiment in real-time. These modules allow otherwise conventional dialogue systems to have “empathy” and answer to the user while being aware of their emotion and intent. Emotion recognition from speech previously consists of feature engineering and machine learning where the first stage causes delay in decoding time. We describe a CNN model to extract emotion from raw speech input without feature engineering. This approach even achieves an impressive average of 65.7% accuracy on six emotion categories, a 4.5% improvement when compared to the conventional feature based SVM classification. A separate, CNN-based sentiment analysis module recognizes sentiments from speech recognition results, with 82.5 F-measure on human-machine dialogues when trained with out-of-domain data.

Cite

CITATION STYLE

APA

Bertero, D., Siddique, F. B., Wu, C. S., Wan, Y., Chan, R. H. Y., & Fung, P. (2016). Real-time speech emotion and sentiment recognition for interactive dialogue systems. In EMNLP 2016 - Conference on Empirical Methods in Natural Language Processing, Proceedings (pp. 1042–1047). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/d16-1110

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free