Emotion recognition with spatial attention and temporal softmax pooling

17Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Video-based emotion recognition is a challenging task because it requires to distinguish the small deformations of the human face that represent emotions, while being invariant to stronger visual differences due to different identities. State-of-the-art methods normally use complex deep learning models such as recurrent neural networks (RNNs, LSTMs, GRUs), convolutional neural networks (CNNs, C3D, residual networks) and their combination. In this paper, we propose a simpler approach that combines a CNN pre-trained on a public dataset of facial images with (1) a spatial attention mechanism, to localize the most important regions of the face for a given emotion, and (2) temporal softmax pooling, to select the most important frames of the given video. Results on the challenging EmotiW dataset show that this approach can achieve higher accuracy than more complex approaches.

Cite

CITATION STYLE

APA

Aminbeidokhti, M., Pedersoli, M., Cardinal, P., & Granger, E. (2019). Emotion recognition with spatial attention and temporal softmax pooling. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11662 LNCS, pp. 323–331). Springer Verlag. https://doi.org/10.1007/978-3-030-27202-9_29

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free