Modeling multimodal behaviors from speech prosody

22Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Head and eyebrow movements are an important communication mean. They are highly synchronized with speech prosody. Endowing virtual agent with synchronized verbal and nonverbal behavior enhances their communicative performance. In this paper, we propose an animation model for the virtual agent based on a statistical model linking speech prosody and facial movement. A fully parameterized Hidden Markov Model is proposed first to capture the tight relationship between speech and facial movement of a human face extracted from a video corpus and then to drive automatically virtual agent's behaviors from speech signals. The correlation between head and eyebrow movements is also taken into account during the building of the model. Subjective and objective evaluations were conducted to validate this model. © 2013 Springer-Verlag.

Cite

CITATION STYLE

APA

Ding, Y., Pelachaud, C., & Artières, T. (2013). Modeling multimodal behaviors from speech prosody. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8108 LNAI, pp. 217–228). https://doi.org/10.1007/978-3-642-40415-3_19

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free