Acoustic Modeling in today's emotion recognition engines employs general models independent of the spoken phonetic content. This seems to work well enough given sufficient instances to cover for a broad variety of phonetic structures and emotions at the same time. However, data is usually sparse in the field and the question arises whether unit specific models as word emotion models could outperform the typical general models. In this respect this paper tries to answer the question how strongly acoustic emotion models depend on the textual and phonetic content. We investigate the influence on the turn and word level by use of state-of-the-art techniques for frame and word modeling on the well-known public Berlin Emotional Speech and Speech Under Simulated and Actual Stress databases. In the result it is clearly shown that the phonetic structure does strongly influence the accuracy of emotion recognition. © 2008 Springer-Verlag Berlin Heidelberg.
CITATION STYLE
Vlasenko, B., Schuller, B., Wendemuth, A., & Rigoll, G. (2008). On the influence of phonetic content variation for acoustic emotion recognition. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 5078 LNCS, pp. 217–220). https://doi.org/10.1007/978-3-540-69369-7_24
Mendeley helps you to discover research relevant for your work.