I am trying to understand sequence-to-sequence learning with a RNN. What I understand so far, is that the output of the encoder is used to condition the decoder.
Yet, I have two sources which, in my opinion, do the conditioning differently and I would like to know which way is valid (may be both) or if I miss something.
Source: Neural Network Methods for Natural Language Processing by Yoav Goldberg
As far as I understand the author, the decoder operates in every step with a state vector AND the decoder output AND the next part of the sequence. Thus, the state vector of the decoder is seperated from the result of the encoder.
Source: A ten-minute introduction to sequence-to-sequence learning in Keras by Francois Chollet
As far as I understand the author and the source, the decoder is provided with the encoder state as intial state. Thus, the state vector of the decoder is the output of the decoder. The decoder steps only depend on the encoder output through the state vector.