Transformer icon indicating copy to clipboard operation
Transformer copied to clipboard

Transformer seq2seq model, program that can build a language translator from parallel corpus

Results 29 Transformer issues
Sort by recently updated
recently updated
newest added

Hi. Please i would like to know how to add a new layer in your Transformer model between the Encoder and Decoder Layers so that the outputs coming from the...

While trying out the repo, I came across this argument **_create_valset_** which was being parsed from the terminal, but I not not sure if it is being used anywhere else....

I read your blog post in TowardsDataScience on this model, and I think there may be a computational error in line 27 of Transformer/Embed.py. In the paper and in other...

Hello, Many thanks for sharing project, unfortunatelly getting `IndexError: index 0 is out of bounds for dimension 0 with size 0` when running on floydhub `python translate.py -load_weights weights -src_lang...

# the whole error is as follows:, x = x + pe got different size? creating dataset and iterator... model weights will be saved every 20 minutes and at end...

I got the following error while compiling python train.py -src_data data/europarl-v7_de.txt -trg_data data/europarl-v7_en.txt -src_lang de -trg_lang en -SGDR -epochs 10 -checkpoint 10 -batchsize 128 -load_weights weights loading spacy tokenizers... loading...

Hi, I think you misinterpreted the multi head attention in Vaswani's Attention is all you need paper. What you do, is (assume only one query) projecting the query and keys...

Hello~ may I ask a question? In this line of code -- ys = trg[:, 1:].contiguous().view(-1),why do we have to discard the first seq?

Here, in Training and Translating chapter. What is 8bg? Maybe it is 8GB? ![image](https://github.com/user-attachments/assets/6ff488ca-ebd1-40ee-b87d-02bac84a4382)