Convergence Speed of LSTM vs. Simple RNN
Investigate and determine, via a rigorous mathematical analysis of backpropagation dynamics, whether Long Short-Term Memory (LSTM) networks converge faster than simple recurrent neural networks (RNNs) under standard training protocols.
References
It seems that LSTM can preserve gradients across long time intervals, which mitigates the vanishing gradient problem in standard RNNs. We have not analyzed the expression of backpropagation from a mathematical perspective to show that they converge faster than simple RNN.
                — The algebra and the geometry aspect of Deep learning
                
                (2510.18862 - Aristide, 21 Oct 2025) in Section 7. Recurrent Neural Network (Long Short-Term Memory)