$\color{black}\rule{365px}{3px}$
Table of Contents
$\color{black}\rule{365px}{3px}$
Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU) are advanced RNN variants designed to address these limitations and improve sequence modeling.
Motivations
Vanishing Gradient Problem: Standard RNNs struggle to learn long-term dependencies because gradients diminish as they are backpropagated through time.
Long-Term Dependency: As the context length grows, useful information get lost as towards the end of the sequence.

Efficient Memory Control: A need for mechanisms to selectively retain or forget information during sequence processing.
Contributions
LSTM: