Improving piano music transcription by Elman dynamic neural networks

0Citations
Citations of this article
3Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper, we present two methods based on neural networks for the automatic transcription of polyphonic piano music. The input to these methods consists in live piano music acquired by a microphone, while the pitch of all the notes in the corresponding score forms the output. The aim of this work is to compare the accuracy achieved using a feed-forward neural network, such as the MLP (MultiLayer Perceptron), with that supplied by a recurrent neural network, such as the ENN (Elman Neural Network). Signal processing techniques based on the CQT (Constant-Q Transform) are used in order to create a time-frequency representation of the input signals. The processing phases involve non-negative matrix factorization (NMF) for onset detection. Since large scale tests were required, the whole process (synthesis of audio data generated starting from MIDI files, comparison of the results with the original score) has been automated. Test, validation and training sets have been generated with reference to three different musical styles respectively represented by J. S. Bach's inventions, F. Chopin's nocturnes and C. Debussy's preludes. © 2010 Springer Science+Business Media B.V.

Cite

CITATION STYLE

APA

Costantini, G., Todisco, M., & Carota, M. (2010). Improving piano music transcription by Elman dynamic neural networks. In Lecture Notes in Electrical Engineering (Vol. 54 LNEE, pp. 387–390). https://doi.org/10.1007/978-90-481-3606-3_78

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free