SeLF: A deep neural network based multimodal sequential late fusion approach for human emotion recognition

1Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Computer vision domain consists of algorithms and techniques to enhance computers with the ability to see and perceive. Human emotion recognition using computer vision is a challenging research area. Facial expression may not always give accurate judgment of emotion hence needs to be combined with other modalities such as voice, text and physiological signals. Several fusion approaches such as direct, early and late were introduced but the problem still persists. This paper focuses on deep neural network (NN) based sequential late fusion approach to identify emotions from various available modalities. Modalities are integrated into the system sequentially at the decision level. A deep CNN was trained to identify face emotions. Short videos were analyzed to recognize emotions. Further, frames were extracted and the emotions were analyzed. The voice channel was processed and transcripts were generated. Each channel outcome was compared for accuracy. The opinion was recorded manually for conformance of results. The opinion matched with the emotion classified by the system.

Cite

CITATION STYLE

APA

Modi, A., & Sharma, P. (2019). SeLF: A deep neural network based multimodal sequential late fusion approach for human emotion recognition. In Communications in Computer and Information Science (Vol. 1045, pp. 275–283). Springer Verlag. https://doi.org/10.1007/978-981-13-9939-8_25

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free