LSTM
Long short-term memory (LSTM) is a type of recurrent neural network designed to learn long-range dependencies in sequence data. LSTMs mitigate vanishing and exploding gradient problems that impede standard recurrent networks during learning, by maintaining a memory cell that can preserve information over many time steps.
An LSTM unit uses a memory cell c_t and a hidden state h_t at each time step
LSTM networks are typically trained by backpropagation through time (BPTT). They can model long-range dependencies more
Origins: LSTM was introduced in 1997 by Sepp Hochreiter and Jürgen Schmidhuber. Applications span speech recognition,