Enhancing feature correlation for bi-modal group emotion recognition

4Citations
Citations of this article
5Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Group emotion recognition in the wild has received much attention in computer vision community. It is a very challenge issue, due to interactions taking place between various numbers of people, different occlusions. According to human cognitive and behavioral researches, background and facial expression play a dominating role in the perception of group’s mood. Hence, in this paper, we propose a novel approach that combined these two features for image-based group emotion recognition with feature correlation enhancement. The feature enhancement is mainly reflected in two parts. For facial expression feature extraction, we plug non-local blocks into Xception network to enhance the feature correlation of different positions in low-level, which can avoid the fast loss of position information of the traditional CNNs and effectively enhance the network’s feature representation capability. For global scene information, we build a bilinear convolutional neural network (B-CNN) consisting of VGG16 networks to model local pairwise feature interactions in a translationally invariant manner. The experimental results show that the fused feature could effectively improve the performance.

Cite

CITATION STYLE

APA

Liu, N., Fang, Y., & Guo, Y. (2018). Enhancing feature correlation for bi-modal group emotion recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11165 LNCS, pp. 24–34). Springer Verlag. https://doi.org/10.1007/978-3-030-00767-6_3

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free