Transformer
Transformer copied to clipboard
Transformer seq2seq model, program that can build a language translator from parallel corpus
Hi. Please i would like to know how to add a new layer in your Transformer model between the Encoder and Decoder Layers so that the outputs coming from the...
While trying out the repo, I came across this argument **_create_valset_** which was being parsed from the terminal, but I not not sure if it is being used anywhere else....
I read your blog post in TowardsDataScience on this model, and I think there may be a computational error in line 27 of Transformer/Embed.py. In the paper and in other...
Hello, Many thanks for sharing project, unfortunatelly getting `IndexError: index 0 is out of bounds for dimension 0 with size 0` when running on floydhub `python translate.py -load_weights weights -src_lang...
# the whole error is as follows:, x = x + pe got different size? creating dataset and iterator... model weights will be saved every 20 minutes and at end...
I got the following error while compiling python train.py -src_data data/europarl-v7_de.txt -trg_data data/europarl-v7_en.txt -src_lang de -trg_lang en -SGDR -epochs 10 -checkpoint 10 -batchsize 128 -load_weights weights loading spacy tokenizers... loading...
Hi, I think you misinterpreted the multi head attention in Vaswani's Attention is all you need paper. What you do, is (assume only one query) projecting the query and keys...
Hello~ may I ask a question? In this line of code -- ys = trg[:, 1:].contiguous().view(-1),why do we have to discard the first seq?
Here, in Training and Translating chapter. What is 8bg? Maybe it is 8GB? 