Visualyre: Multimodal visualization of lyrics

0Citations
Citations of this article
9Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we present Visualyre, a web application that synthesizes images based on the semantics of the lyrics, and the mood of the music. We use a multimodal approach, generating initial images with a text-to-image generative models from the lyrics (text) of a song, followed by a style transfer model conditioned to the mood of the music (audio). Our target user base is the independent music artist community, by providing a means for composers and songwriters to generate suitable images for their music (such as album covers). We discuss the possible usage of such application, as well as the possible improvements in future iterations.

Cite

CITATION STYLE

APA

Azuaje, G., Liew, K., Epure, E., Yada, S., Wakamiya, S., & Aramaki, E. (2021). Visualyre: Multimodal visualization of lyrics. In ACM International Conference Proceeding Series (pp. 130–134). Association for Computing Machinery. https://doi.org/10.1145/3478384.3478420

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free