annotated-transformer
annotated-transformer copied to clipboard
An annotated implementation of the Transformer paper.
On the Figure, there `Multi-Head Attention`s and `Masked Multi-Head Attention`s. Are all Multi-Head Attention "Masked" really? 
I downloaded the notebook from the Colab link in the readme. I uploaded it to a fresh Colab environment. I ran the first cell and got: > ERROR: torch-0.3.0.post4-cp36-cp36m-linux_x86_64.whl is...
Hi all, First, I want to express my gratitude for the amazing work on The Annotated Transformer. It has been an invaluable resource for the AI community and a fantastic...
https://github.com/harvardnlp/annotated-transformer/blob/debc9fd747bb2123160a98046ad1c2d4da44a567/the_annotated_transformer.py#L868 Is there a problem with the parameters passed in here? It should be **out = test_model.decode(ys,memory, src_mask, subsequent_mask(ys.size(1)).type_as(src.data))** , not **out = test_model.decode(memory, src_mask, ys, subsequent_mask(ys.size(1)).type_as(src.data))**
when run !pip install -q torchdata==0.3.0 torchtext==0.12 spacy==3.2 altair GPUtil !python -m spacy download de_core_news_sm !python -m spacy download en_core_web_sm it returns: ERROR: Could not find a version that satisfies...
Thank you for providing such a well-organized and comprehensive Transformer tutorial. As a beginner, I’ve learned a lot from this repository☺️! When I was building the positional encoding block, I...