Audio-Based Music Classification with DenseNet and Data Augmentation

16Citations
Citations of this article
41Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In recent years, deep learning technique has received intense attention owing to its great success in image recognition. A tendency of adaption of deep learning in various information processing fields has formed, including music information retrieval (MIR). In this paper, we conduct a comprehensive study on music audio classification with improved convolutional neural networks (CNNs). To the best of our knowledge, this the first work to apply Densely Connected Convolutional Networks (DenseNet) to music audio tagging, which has been demonstrated to perform better than Residual neural network (ResNet). Additionally, two specific data augmentation approaches of time overlapping and pitch shifting have been proposed to address the deficiency of labelled data in the MIR. Moreover, an ensemble learning of stacking is employed based on SVM. We believe that the proposed combination of strong representation of DenseNet and data augmentation can be adapted to other audio processing tasks.

Cite

CITATION STYLE

APA

Bian, W., Wang, J., Zhuang, B., Yang, J., Wang, S., & Xiao, J. (2019). Audio-Based Music Classification with DenseNet and Data Augmentation. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 11672 LNAI, pp. 56–65). Springer Verlag. https://doi.org/10.1007/978-3-030-29894-4_5

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free