Recurrent Normalization Propagation

César Laurent, Nicolas Ballas, Pascal Vincent

Feb 15, 2017 (modified: Feb 15, 2017) ICLR 2017 workshop submission readers: everyone
  • Abstract: We propose an LSTM parametrization that preserves the means and variances of the hidden states and memory cells across time. While having training benefits similar to Recurrent Batch Normalization and Layer Normalization, it does not need to estimate statistics at each time step, therefore, requiring fewer computations overall. We also investigate the parametrization impact on the gradient flows and present a way of initializing the weights accordingly. We evaluate our proposal on language modelling and image generative modelling tasks. We empirically show that it performs similarly or better than other recurrent normalization approaches, while being faster to execute.
  • TL;DR: Extension of Normalization Propagation to the LSTM.
  • Conflicts: umontreal.ca
  • Keywords: Deep learning, Optimization

Loading