Dynamic cortex memory: Enhancing recurrent neural networks for gradient-based sequence learning

19Citations
Citations of this article
23Readers
Mendeley users who have this article in their library.
Get full text

Abstract

In this paper a novel recurrent neural network (RNN) model for gradient-based sequence learning is introduced. The presented dynamic cortex memory (DCM) is an extension of the well-known long short term memory (LSTM) model. The main innovation of the DCM is the enhancement of the inner interplay of the gates and the error carousel due to several new and trainable connections. These connections enable a direct signal transfer from the gates to one another. With this novel enhancement the networks are able to converge faster during training with back-propagation through time (BPTT) than LSTM under the same training conditions. Furthermore, DCMs yield better generalization results than LSTMs. This behaviour is shown for different supervised problem scenarios, including storing precise values, adding and learning a context-sensitive grammar. © 2014 Springer International Publishing Switzerland.

Cite

CITATION STYLE

APA

Otte, S., Liwicki, M., & Zell, A. (2014). Dynamic cortex memory: Enhancing recurrent neural networks for gradient-based sequence learning. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8681 LNCS, pp. 1–8). Springer Verlag. https://doi.org/10.1007/978-3-319-11179-7_1

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free