WaveRNN
WaveRNN copied to clipboard
how long did you train the WaveRNN model
hello,i have been training the waveRNN model.However, the training speed of this model remains at 0.1step/s.Is there anythong wrong?thank you
@doctor-xiang Something is wrong if you are only getting 0.1 steps/s - normally I get around 3 steps/s on a GTX1080. What GPU and pytorch version are you running?
@fatchord ,i use the TITAN 1080.But the batch size of my model is 32,so i get that rate.If I reduce the batch size, the voice quality will decline. What is the best batch size for training this model ? Thank you
@doctor-xiang I'm getting 3 steps/s for a batch size of 32. I've found that a batch size of 16 or 32 works well (I haven't experimented outside of that). What version of pytorch are you using?
@fatchord the version of 1.0.0
@fatchord i had solved that.thank you
@doctor-xiang Great to hear! Just curious what happened? In case anyone else runs into the same problem.
@fatchord There was an impact between tasks.
@fatchord Hello, may i ask what is the speed of speech synthesis when you use GTX1080.
@fatchord There was an impact between tasks.
hello, I met the same problem as you, and the training speed of wavernn only 0.3step/s , how did you fix it? Thank you!
@kanatazhan you might want @doctor-xiang to get his attention as I've no idea how he solved the problem :)
@kanatazhan There may be multiple tasks running that take up resources on the GPU, resulting in slow speeds.
Using a K80 GPU and batch size of 32 I have about 0.8step/s, and the GPU seems to be around 90% utilisation
I'm using sample_rate=22050
so perhaps this could explain some difference in the speed
@maelp your sample_rate is fine. I wonder - are you accidentally using cpu? Can you try watch nvidia-smi
and run training - does the mem and gpu-util increase when starting off?
I'm using the GPU, it shows the use as close to 90% so I think it is working
And python3.7 and pytorch 1.0
K80s are pretty ancient gpus, ~and almost entirely catered to double precision floating point operations~. To be honest, I am not surprised by this training rate. A simple consumer level (GTX 1070) graphics card will blow it away
I'm getting 2.3 steps/s on an RTX 3080. CUDA GPU usage hovers around 50%. Is that right?
I'm facing the same issue. Can anyone explain this issue clearly?