Unsupervised temporal segmentation of talking faces using visual cues to improve emotion recognition

0Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

The mouth region of human face possesses highly discriminative information regarding the expressions on the face. Facial expression analysis to infer the emotional state of a user becomes very challenging when the user talks, as most of the mouth actions while uttering certain words match with mouth shapes expressing various emotions. We introduce a novel unsupervised method to temporally segment talking faces from the faces displaying only emotions, and use the knowledge of talking face segments to improve emotion recognition. The proposed method uses integrated gradient histogram of local binary patterns to represent mouth features suitably and identifies temporal segments of talking faces online by estimating the uncertainties of mouth movements over a period of time. The algorithm accurately identifies talking face segments on a real-world database where talking and emotion happens naturally. Also, the emotion recognition system, using talking face cues, showed considerable improvement in recognition accuracy. © 2011 Springer-Verlag.

Cite

CITATION STYLE

APA

Velusamy, S., Gopalakrishnan, V., Navathe, B., Kannan, H., Anand, B., & Sharma, A. (2011). Unsupervised temporal segmentation of talking faces using visual cues to improve emotion recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 6974 LNCS, pp. 417–426). https://doi.org/10.1007/978-3-642-24600-5_45

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free