On the influence of phonetic content variation for acoustic emotion recognition

16Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Acoustic Modeling in today's emotion recognition engines employs general models independent of the spoken phonetic content. This seems to work well enough given sufficient instances to cover for a broad variety of phonetic structures and emotions at the same time. However, data is usually sparse in the field and the question arises whether unit specific models as word emotion models could outperform the typical general models. In this respect this paper tries to answer the question how strongly acoustic emotion models depend on the textual and phonetic content. We investigate the influence on the turn and word level by use of state-of-the-art techniques for frame and word modeling on the well-known public Berlin Emotional Speech and Speech Under Simulated and Actual Stress databases. In the result it is clearly shown that the phonetic structure does strongly influence the accuracy of emotion recognition. © 2008 Springer-Verlag Berlin Heidelberg.

Cite

CITATION STYLE

APA

Vlasenko, B., Schuller, B., Wendemuth, A., & Rigoll, G. (2008). On the influence of phonetic content variation for acoustic emotion recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5078 LNCS, pp. 217–220). https://doi.org/10.1007/978-3-540-69369-7_24

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free