A single predominant instrument recognition of polyphonic music using CNN-based timbre analysis

10Citations
Citations of this article
6Readers
Mendeley users who have this article in their library.

Abstract

Classifying musical instrument from polyphonic music is a challenging but important task in music information retrieval. This work enables to automatically tag music information, such as genre classification. In previous, almost every work of spectrogram analysis has been used Short Time Fourier Transform (STFT) and Mel Frequency Cepstral Coefficient (MFCC). Recently, sparkgram is researched and used in audio source analysis. Moreover, for deep learning approach, modified convolutional neural networks (CNN) widely have been researched, but many results have not been improved drastically. Instead of improving backbone networks, we have researched on preprocessing process. In this paper, we use CNN and Hilbert Spectral Analysis (HSA) to solve the polyphonic music problem. The HSA is performed at the fixed length of polyphonic music, and a predominant instrument is labeled at its result. As result, we have achieved the state-of-the-art result in IRMAS dataset and 3% performance improvement in individual instruments.

Cite

CITATION STYLE

APA

Kim, D., Sung, T. T., Cho, S. Y., Lee, G., & Sohn, C. B. (2018). A single predominant instrument recognition of polyphonic music using CNN-based timbre analysis. International Journal of Engineering and Technology(UAE), 7(3.34 Special Issue  34), 590–593. https://doi.org/10.14419/ijet.v7i3.34.19388

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free