Chaitanya
Chaitanya
I'm on pytorch 1.4.0. It is the training script that fails
> I have the same problem. This is happening in AudioDictionary.dummy_sentence() where (I think) the code is creating a dummy audio segment, and trying to initialise it with `uniform_(self.nspecial +...
> Actually, I'm not sure that the training of @Chaitanya-git was good. I've never seen a perplexity of 1 in a translation task. @Chaitanya-git can you please tell us how...
@mattiadg , I ran the training script with pytorch 1.1.0 for a while and I was able to train for 7 epochs total. Here's what those results look like: ```...
So pytorch 1.4 also works with the original workaround? Edit: The translations seem just as bad with the pytorch 1.4 model as before. I guess I'll have to wait for...
Yes, the same issue persists as before. The translations all very bad, consisting of the same output for all inputs
The issue persists with pytorch 1.1 as well. Is this expected without ASR pretraining? I'll try ASR pretraining as well and see how it goes
@mattiadg I have tried training with two and three GPUs. Could you tell me what parameters need to be changed as the number of GPUs change and how they vary...
I just trained from scratch again with 4 GPUs for around 7 epochs and now the translations seem to be much better. Earlier the translations were indeed one word translations...
Alright, thanks!