Shuffling Recurrent Neural Networks

35Citations
Citations of this article
17Readers
Mendeley users who have this article in their library.

Abstract

We propose a novel recurrent neural network model, where the hidden state ht is obtained by permuting the vector elements of the previous hidden state ht−1 and adding the output of a learned function β (xt) of the input xt at time t. In our model, the prediction is given by a second learned function, which is applied to the hidden state s (ht). The method is easy to implement, extremely efficient, and does not suffer from vanishing nor exploding gradients. In an extensive set of experiments, the method shows competitive results, in comparison to the leading literature baselines. We share our implementation at https://github.com/rotmanmi/SRNN.

Cite

CITATION STYLE

APA

Rotman, M., & Wolf, L. (2021). Shuffling Recurrent Neural Networks. In 35th AAAI Conference on Artificial Intelligence, AAAI 2021 (Vol. 11A, pp. 9428–9435). Association for the Advancement of Artificial Intelligence. https://doi.org/10.1609/aaai.v35i11.17136

Register to see more suggestions

Mendeley helps you to discover research relevant for your work.

Already have an account?

Save time finding and organizing research with Mendeley

Sign up for free