Clare S. Y. Huang Data Scientist | Atmospheric Dynamicist

Common issues in RNN training

Exploding gradients and vanishing gradients are two common issues with the training of RNN.

To avoid exploding gradients, one may use:

  • Truncated Back-propagation through time (BPTT)
  • Clip gradients at threshold
  • RMSprop to adjust learning rate

Vanishing gradients are harder to detect. To avoid it, one may use:

  • Weight initialization
  • ReLu activation functions
  • RMSprop
  • LSTM, GRUs
<< Previous Page