Salient cross-lingual acoustic and prosodic features for English and German emotion recognition

0Citations
Citations of this article
2Readers
Mendeley users who have this article in their library.
Get full text

Abstract

While approaches on automatic recognition of human emotion from speech have already achieved reasonable results, a lot of room for improvement still remains there. In our research, we select the most essential features by applying a self-adaptive multi-objective genetic algorithm. The proposed approach is evaluated using data from different languages (English and German) with two different feature sets consisting of 37 and 384 dimensions, respectively. The obtained results of the developed technique have increased the emotion recognition performance by up to 49.8% relative improvement in accuracy. Furthermore, in order to identify salient features across speech data from different languages, we analysed the selection count of the features to generate a feature ranking. Based on this, a feature set for speechbased emotion recognition based on the most salient features has been created. By applying this feature set, we achieve a relative improvement of up to 37.3% without the need of time-consuming feature selection using a genetic algorithm.

Cite

CITATION STYLE

APA

Sidorov, M., Brester, C., Ultes, S., & Schmitt, A. (2017). Salient cross-lingual acoustic and prosodic features for English and German emotion recognition. In Lecture Notes in Electrical Engineering (Vol. 427 427 LNEE, pp. 159–169). Springer Verlag. https://doi.org/10.1007/978-981-10-2585-3_12

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free