Multimodal Persona Based Generation of Comic Dialogs

10Citations
Citations of this article
22Readers
Mendeley users who have this article in their library.
Get full text

Abstract

We focus on the novel problem of persona based dialogue generation for comic strips. Dialogs in comic strips is a unique and unexplored area where every strip contains utterances from various characters with each one building upon the previous utterances and the associated visual scene. Previous works like DialoGPT, PersonaGPT and other dialog generation models encode two-party dialogues and do not account for the visual information. To the best of our knowledge we are the first to propose the paradigm of multimodal persona based dialogue generation. We contribute a novel dataset, COMSET, consisting of 54K strips, harvested from 13 popular comics available online. Further, we propose a multimodal persona-based architecture, MPDIALOG, to generate dialogues for the next panel in the strip which decreases the perplexity score by ∼10 points over strong dialogue generation baseline models. We demonstrate that there is still ample opportunity for improvement, highlighting the importance of building stronger dialogue systems that are able to generate persona-consistent dialogues and understand the context through various modalities.

Cite

CITATION STYLE

APA

Agrawal, H., Mishra, A. M., Gupta, M., & Mausam. (2023). Multimodal Persona Based Generation of Comic Dialogs. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (Vol. 1, pp. 14150–14164). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.acl-long.791

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free