LSTM: Why It Was Born, How It Fixes RNN, and Why It Changed Sequence Learning
Sequence data is messy. Words depend on earlier words, notes depend on earlier notes, signals depend on earlier signals, and the past has this annoying habit of refusing to stay irrelevant. That is exactly why recurrent models exist in the first place….