Bimodal emotion recognition based on speech signals and facial expression

11Citations
Citations of this article
8Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Voice signals and facial expression changes are synchronized under the different emotions, the recognition algorithm based audio-visual feature fusion is proposed to identify emotional states more accurately. Prosodic features were extracted for speech emotional features, and local Gabor binary patterns were adopted for facial expression features. Two types of features were modeled with SVM respectively to obtain the probabilities of anger, disgust fear, happiness, sadness and surprise, and then fused the probabilities to gain the final decision. Simulation results demonstrate that the average recognition rates of the single modal classifier based on speech signals and based on facial expression reach 60% and 57% respectively, while the multimodal classifier with the feature fusion of speech signals and facial expression achieves 72%. © 2011 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Tu, B., & Yu, F. (2011). Bimodal emotion recognition based on speech signals and facial expression. In Advances in Intelligent and Soft Computing (Vol. 122, pp. 691–696). https://doi.org/10.1007/978-3-642-25664-6_81

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free