Facial expression recognition with multi-scale convolution neural network

20Citations
Citations of this article
13Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We present a deep convolutional neural network (CNN) architecture for facial expression recognition. Inspired by the fact that regions located around certain facial parts (e.g. mouth, nose, eyes, and brows) contain the most representative information of expressions, an architecture extracts features at different scale from intermediate layers is designed to combine both local and global information. In addition, noticing that in specific to facial expression recognition, traditional face alignment would distort the images and lose expression information. To avoid this side effect, we apply batch normalization to the architecture instead of face alignment and feed the network with original images. Moreover, considering the tiny differences between classes caused by the same facial movements, a triplet-loss learning method is used to train the architecture, which improves the discrimination of deep features. Experiments show that the proposed architecture achieves superior performance to other state-of-the-art methods on the FER2013 dataset.

Cite

CITATION STYLE

APA

Wang, J., & Yuan, C. H. (2016). Facial expression recognition with multi-scale convolution neural network. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 9916 LNCS, pp. 376–385). Springer Verlag. https://doi.org/10.1007/978-3-319-48890-5_37

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free