Results 13 comments of Hieu Hoang

out of interest, do you know if you're likely to get overflows when using fp16, and if you're doing anything about it * Looking for MT/NLP opportunities * Hieu Hoang...

feedback from my own work with fp16 in amun. When running on a P100 (wilkes) it gives about a 20% speedup over using fp32. Most of the speedup is in...

Probably cos it's set up to read text On Sun, Apr 12, 2020, 6:41 PM Marcin Junczys-Dowmunt < [email protected]> wrote: > @frankseide Do you remember what this was > about?...

you're right, it's always been opened as binary. Maybe some other bug in the previous code. I see no reason why it shouldn't work now Hieu Hoang http://statmt.org/hieu On Sun,...

You might want to take a look at Amun's nth_element.cu which has been changed to use the actual (target) vocab size. There's also Amun regression tests to make sure that...

There is some truth in this religion ![image](https://user-images.githubusercontent.com/691732/36370406-b0483aea-1556-11e8-9890-71c87712ec4c.png) Also, make sure --maxi-batch is a multiple of mini-batch, and add --maxi-batch-sort src

ah, there's a difference between marian and amun's maxi-batch definition then. Marian's is better imo

@emjotde are you using --maxi-batch to buffer input too during training? Otherwise I see no point in not sorting maxi-batches

I've seen unit test break on Moses when the boost library is non-standard. However, kenlm is building fine on my Ubuntu 16.04, boost Version: 1.58.0.1ubuntu1 Try uninstalling and deleting every...

Just had the same problem and debugged it. You need to put torch.cuda.set_device(rank) before dist.init_process_group()