12
35

Shuffling Recurrent Neural Networks

Abstract

We propose a novel recurrent neural network model, where the hidden state hth_t is obtained by permuting the vector elements of the previous hidden state ht1h_{t-1} and adding the output of a learned function b(xt)b(x_t) of the input xtx_t at time tt. In our model, the prediction is given by a second learned function, which is applied to the hidden state s(ht)s(h_t). The method is easy to implement, extremely efficient, and does not suffer from vanishing nor exploding gradients. In an extensive set of experiments, the method shows competitive results, in comparison to the leading literature baselines.

View on arXiv
Comments on this paper