Using a Language Model to Generate Music in Its Symbolic Domain While Controlling Its Perceived Emotion

4Citations
Citations of this article
19Readers
Mendeley users who have this article in their library.

Abstract

This work proposes a transformer-based model capable of generating music in its symbolic domain, in a controllable fashion. The ultimate goal of this is to build a system with which people can compose music collaboratively with a computer. Using an NLP model as a base (GPT-2), we take advantage of the similarities across symbolic music representation and written language to build a model capable of conditionally predicting musical sequences. Controllability is achieved without explicit programming for it, and does not require extensive retraining of the model. A study with 939 participants was performed to evaluate this controllability. The results of this suggest the proposed method is indeed effective and can be used to control the generation of music in its symbolic domain. The method itself is flexible to any desired 'control', but this work focuses specifically on the emotion conveyed when one listens to a piece of music.

Cite

CITATION STYLE

APA

Imasato, N., Miyazawa, K., Duncan, C., & Nagai, T. (2023). Using a Language Model to Generate Music in Its Symbolic Domain While Controlling Its Perceived Emotion. IEEE Access, 11, 52412–52428. https://doi.org/10.1109/ACCESS.2023.3280603

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free