transformer
transformer copied to clipboard
A Pytorch Implementation of "Attention is All You Need" and "Weighted Transformer Network for Machine Translation"
self.w_q = Linear([d_model, d_k * n_heads]) --------------------------------------------------------------------------- TypeError Traceback (most recent call last) in 2 d_k = 16 3 n_heads = 6 ----> 4 w_q = Linear([d_model, d_k * n_heads])...
This error occurs when I run **train.py** Traceback (most recent call last): File "train.py", line 208, in main(opt) File "train.py", line 72, in main train_loss, train_sents = train(model, criterion, optimizer,...
In the original paper(weighted transformer), the author mentioned that "all bounds are respected during each training step by projection." I have no idea what "by project" means and don't know...
When I run the colder, there is an error in model.py (line 53): `self.layers = nn.ModuleList( [self.layer_type(d_k, d_v, d_model, d_ff, n_heads, dropout) for _ in range(n_layers)])` only integer tensors of...
Please i would like to know how to use this repository because i am getting errors that i don't understand will running it.