You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Aug 18, 2021. It is now read-only.
Thank you very much for the awesome work, I need a clarification in the decoder part of seq2seq-translation. #Combine embedded input word and last context, run through RNN rnn_input = torch.cat((word_embedded, last_context.unsqueeze(0)), 2)
Yes it is, though looking back at it I'm missing one layer between the context vector c_t and the softmax layer, to create the "attentional hidden state" ~h_t, which is what they use for input feeding.
Thank you very much for the awesome work, I need a clarification in the decoder part of seq2seq-translation.
#Combine embedded input word and last context, run through RNN
rnn_input = torch.cat((word_embedded, last_context.unsqueeze(0)), 2)
Is the above code an implementation of Input-feeding Approach in the Effective Approaches to Attention-based Neural Machine Translation paper?
The text was updated successfully, but these errors were encountered: