Comic-guided speech synthesis

17Citations
Citations of this article
39Readers
Mendeley users who have this article in their library.

Abstract

We introduce a novel approach for synthesizing realistic speeches for comics. Using a comic page as input, our approach synthesizes speeches for each comic character following the reading flow. It adopts a cascading strategy to synthesize speeches in two stages: Comic Visual Analysis and Comic Speech Synthesis. In the first stage, the input comic page is analyzed to identify the gender and age of the characters, as well as texts each character speaks and corresponding emotion. Guided by this analysis, in the second stage, our approach synthesizes realistic speeches for each character, which are consistent with the visual observations. Our experiments show that the proposed approach can synthesize realistic and lively speeches for different types of comics. Perceptual studies performed on the synthesis results of multiple sample comics validate the efficacy of our approach.

Author supplied keywords

Cite

CITATION STYLE

APA

Wang, Y., Wang, W., Liang, W., & Yu, L. F. (2019). Comic-guided speech synthesis. ACM Transactions on Graphics, 38(6). https://doi.org/10.1145/3355089.3356487

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free