The impact of attention mechanisms on speech emotion recognition

34Citations
Citations of this article
31Readers
Mendeley users who have this article in their library.

Abstract

Speech emotion recognition (SER) plays an important role in real-time applications of hu-man-machine interaction. The Attention Mechanism is widely used to improve the performance of SER. However, the applicable rules of attention mechanism are not deeply discussed. This paper discussed the difference between Global-Attention and Self-Attention and explored their applicable rules to SER classification construction. The experimental results show that the Global-Attention can improve the accuracy of the sequential model, while the Self-Attention can improve the accuracy of the parallel model when conducting the model with the CNN and the LSTM. With this knowledge, a classifier (CNN-LSTM×2+Global-Attention model) for SER is proposed. The experi-ments result show that it could achieve an accuracy of 85.427% on the EMO-DB dataset.

Cite

CITATION STYLE

APA

Chen, S., Zhang, M., Yang, X., Zhao, Z., Zou, T., & Sun, X. (2021). The impact of attention mechanisms on speech emotion recognition. Sensors, 21(22). https://doi.org/10.3390/s21227530

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free