vanishing gradients

Concept

A problem in training deep neural networks, particularly RNNs, where gradients become too small to effectively update weights in earlier layers.

Mentioned in 1 video