awd-lstm-lm
awd-lstm-lm copied to clipboard
Attention Model
Any ideas on how to incorporate attention model from http://pytorch.org/tutorials/intermediate/seq2seq_translation_tutorial.html ?
Every PyTorch implementation of the attention model I have come across uses encoder-decoder architecture adds attention model to the decoder. Is it possible to have attention in your implementation?