Transformer
Transformer copied to clipboard
Transformer seq2seq model, program that can build a language translator from parallel corpus
Thanks for sharing so great repo, and one question: what's the shape of the padding mask? and how i construct my padding mask with custom dataset? such as now my...
``` pe[pos, i] = math.sin(pos / (10000 ** ((2 * i) / d_model))) pe[pos, i + 1] = math.cos(pos / (10000 ** ((2 * (i + 1)) / d_model))) ```...
Im getting this error when following the readme instructions: RuntimeError: Expected object of device type cuda but got device type cpu for argument #3 'index' in call to _th_index_select any...
Traceback (most recent call last): ] 0% loss = ... File "train.py", line 183, in main() File "train.py", line 111, in main train_model(model, opt) File "train.py", line 34, in train_model...
if mask is not None: mask = mask.unsqueeze(1) scores = scores.masked_fill(mask == 0, -1e9) scores = F.softmax(scores, dim=-1) runtime error in scores.masked_fill(mask == 0, -1e9) Same for this: x =...
Hello everyone, I am a new researcher on the field of deep learning. Please help/example/guide me on feeding embeddings from XLM- RoBERTa to transformer seq2seq model? Thanks!!!
l wann't to know if lt support to load pretrained model. like bert as the encoder?
Hi, I have been using this transformer implementation for so long and everything worked well. I have some trained model (saved in checkpoint) but now, after some time when I...
how do i run this on my local pc? it always shows me that error: file not found. even the files are in that folder. i tried all slash versions...