In speaker recognition applications, the changes of emotional states are main causes of errors. The ongoing work described in this contribution attempts to enhance the performance of automatic speaker recognition (ASR) systems on emotional speech. Two procedures that only need a small quantity of affective training data are applied to ASR task, which is very practical in real-world situations. The method includes classifying the emotional states by acoustical features and generating emotion-added model based on the emotion grouping. Experimental works are performed on Emotional Prosody Speech (EPS) corpus and show significant improvement in EERs and IRs compared with baseline and comparative experiments. © Springer-Verlag Berlin Heidelberg 2005.
CITATION STYLE
Wu, T., Yang, Y., & Wu, Z. (2005). Improving speaker recognition by training on emotion-added models. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 3784 LNCS, pp. 382–389). Springer Verlag. https://doi.org/10.1007/11573548_49
Mendeley helps you to discover research relevant for your work.