Semi-coupled hidden Markov model with state-based alignment strategy for audio-visual emotion recognition

2Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

This paper presents an approach to bi-modal emotion recognition based on a semi-coupled hidden Markov model (SC-HMM). A simplified state-based bi-modal alignment strategy in SC-HMM is proposed to align the temporal relation of states between audio and visual streams. Based on this strategy, the proposed SC-HMM can alleviate the problem of data sparseness and achieve better statistical dependency between states of audio and visual HMMs in most real world scenarios. For performance evaluation, audio-visual signals with four emotional states (happy, neutral, angry and sad) were collected. Each of the invited seven subjects was asked to utter 30 types of sentences twice to generate emotional speech and facial expression for each emotion. Experimental results show the proposed bi-modal approach outperforms other fusion-based bi-modal emotion recognition methods. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Lin, J. C., Wu, C. H., & Wei, W. L. (2011). Semi-coupled hidden Markov model with state-based alignment strategy for audio-visual emotion recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6974 LNCS, pp. 185–194). https://doi.org/10.1007/978-3-642-24600-5_22

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free