Music Genre Classification is one of the most active tasks in Music Information Retrieval (MIR). Many successful approaches can be found in literature. Most of them are based on Machine Learning algorithms applied to different audio features automatically computed for a specific database. But there is no computational model that explains how musical features are combined in order to yield genre decision in humans. In this work we present a listening experiment where audio has been altered in order to preserve some properties of music (rhythm, harmony, etc) but at the same time degrading other ones. Results are compared with a series of state-of-the-art genre classifiers based on these musical properties and we draw some lessons from that comparison.
CITATION STYLE
Guaus, E., & Herrera, P. (2006). Music genre categorization in humans and machines. In Audio Engineering Society - 121st Convention Papers 2006 (Vol. 3, pp. 1388–1398).
Mendeley helps you to discover research relevant for your work.