Abstract
Tempo and genre are two inter-leaved aspects of music, genres are often associated to rhythm patterns which are played in specific tempo ranges. In this article, we focus on the Deep Rhythm system based on a harmonic representation of rhythm used as an input to a convolutional neural network. To consider the relationships between frequency bands, we process complex-valued inputs through complex-convolutions. We also study the joint estimation of tempo/genre using a multitask learning approach. Finally, we study the addition of a second input convolutional branch to the system applied to a mel-spectrogram input dedicated to the timbre. This multi-input approach allows to improve the performances for tempo and genre estimation.
Author supplied keywords
Cite
CITATION STYLE
Aarabi, H. F., & Peeters, G. (2022). Extending Deep Rhythm for Tempo and Genre Estimation Using Complex Convolutions, Multitask Learning and Multi-input Network. Journal of Creative Music Systems, 1(1). https://doi.org/10.5920/jcms.887
Register to see more suggestions
Mendeley helps you to discover research relevant for your work.