BPTT
Backpropagation Through Time (BPTT) is a gradient-based learning algorithm used to train recurrent neural networks (RNNs) by unrolling the network in time and applying backpropagation to compute weight updates across multiple time steps.
During training, the RNN is unrolled for a sequence of length T, creating a feedforward network with
Because BPTT backpropagates through many steps, it can require substantial memory and computation. It is also
To manage cost and memory, truncated BPTT limits the unrolling to a fixed window of recent time
In modern practice, BPTT is commonly used with architectures such as long short-term memory (LSTM) and gated