RNN Part1
RNN Part1
Dr.Tamam Alsarhan
Rumelhart, D. E., Hinton, G. E., & Williams, R. J. (1986). Learning representations by back-propagating errors. nature,
323(6088), 533-536.
RECURRENT NEURAL NETWORK
AN UNROLLED RECURRENT NEURAL
NETWORK
CONT…
RNNs can use their internal state (memory) to process sequences of
inputs.
First, it takes the X(0) from the sequence of input and then it outputs h(0)
which together with X(1) is the input for the next step. So, the h(0) and
X(1) is the input for the next step. Similarly, h(1) from the next is the
input with X(2) for the next step and so on. This way, it keeps
remembering the context while training.
WHY LONG SHORT TERM MEMORY
(LSTM)?
If we are trying to predict the last word in “the clouds are in the sky,” we
don’t need any further context.
On the contrary, predicting the last word in the text “I grew up in France
… I speak fluent French.” needs context from further back!