Multimodal Speech Synthesis

0Citations
Citations of this article
15Readers
Mendeley users who have this article in their library.
Get full text

Abstract

Summary: Speech output generation in the SmartKom system is realized by a corpus-based unit selection strategy that preserves many properties of the human voice. When the system's avatar "Smartakus" is present on the screen, the synthetic speech signal is temporally synchronized with Smartakus visible speech gestures and prosodically adjusted to his pointing gestures to enhance multimodal communication. The unit selection voice was formally evaluated and found to be very well accepted and reasonably intelligible in SmartKom- specific scenarios.

Cite

CITATION STYLE

APA

Schweitzer, A., Braunschweiler, N., Dogil, G., Klankert, T., Möbius, B., Möhler, G., … Thomae, M. (2006). Multimodal Speech Synthesis. Cognitive Technologies, 7, 411–435. https://doi.org/10.1007/3-540-36678-4_27

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free