Abstract
In this paper, we present Visualyre, a web application that synthesizes images based on the semantics of the lyrics, and the mood of the music. We use a multimodal approach, generating initial images with a text-to-image generative models from the lyrics (text) of a song, followed by a style transfer model conditioned to the mood of the music (audio). Our target user base is the independent music artist community, by providing a means for composers and songwriters to generate suitable images for their music (such as album covers). We discuss the possible usage of such application, as well as the possible improvements in future iterations.
Author supplied keywords
Cite
CITATION STYLE
Azuaje, G., Liew, K., Epure, E., Yada, S., Wakamiya, S., & Aramaki, E. (2021). Visualyre: Multimodal visualization of lyrics. In ACM International Conference Proceeding Series (pp. 130–134). Association for Computing Machinery. https://doi.org/10.1145/3478384.3478420
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.