Multimodal Emotion Recognition Based on Ensemble Convolutional Neural Network

49Citations
Citations of this article
58Readers
Mendeley users who have this article in their library.

This article is free to access.

Abstract

In recent years, emotional recognition based on Electrophysiological (EEG) signals has become more and more popular. But the researchers ignored the fact that peripheral physiological signals can also reflect changes in mood. We propose an Ensemble Convolutional Neural Network (ECNN) model, which is used to automatically mine the correlation between multi-channel EEG signals and peripheral physiological signals in order to improve the emotion recognition accuracy. First, we design five convolution networks and use global average pooling (GAP) layers instead of fully connected layers; and then the plurality voting strategy is adopted to establish the ensemble model; eventually this model divides emotions into four categories. Based on the simulations on DEAP dataset, the experimental results demonstrate the superiority of the ECNN compared with other methods.

Cite

CITATION STYLE

APA

Huang, H., Hu, Z., Wang, W., & Wu, M. (2020). Multimodal Emotion Recognition Based on Ensemble Convolutional Neural Network. IEEE Access, 8, 3265–3271. https://doi.org/10.1109/ACCESS.2019.2962085

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free