tensorflow-wavenet icon indicating copy to clipboard operation
tensorflow-wavenet copied to clipboard

Training tensorflow-wavenet

Open arieltci opened this issue 6 years ago • 4 comments

What are the time scales for training wavenet? testing? Did someone try it on AWS? What were the specs of the cloud?

arieltci avatar Jun 14 '18 14:06 arieltci

I am running it on paperspace (GPU enabled, let me know if you want to be more specific, but believe me, this virtual machine is a beast). Took 2,5 days to train with 1e5 (100000 steps). To generate 3 minutes of audio takes like 6 hours or something. Result: a sort of average turd. I am a beginner in coding and one must tweak this script somehow.

zegenerative avatar Dec 07 '18 21:12 zegenerative

I am running it on paperspace (GPU enabled, let me know if you want to be more specific, but believe me, this virtual machine is a beast). Took 2,5 days to train with 1e5 (100000 steps). To generate 3 minutes of audio takes like 6 hours or something. Result: a sort of average turd. I am a beginner in coding and one must tweak this script somehow.

Could you please tell me how you generate and get the voice? My training steps are only 10,000 and it takes 2 days, because 100,000 steps are expected to take more than ten days. Is it because my training time and step size are insufficient that the speech generated by running "generator.py" is 0 seconds?

Ahapy avatar Feb 14 '20 03:02 Ahapy

You have to use gpu for this, cpu is too slow.

If that is not the problem, try changing the training parameters. Hope it helps.

Op vr 14 feb. 2020 04:10 schreef Ahapy [email protected]:

I am running it on paperspace (GPU enabled, let me know if you want to be more specific, but believe me, this virtual machine is a beast). Took 2,5 days to train with 1e5 (100000 steps). To generate 3 minutes of audio takes like 6 hours or something. Result: a sort of average turd. I am a beginner in coding and one must tweak this script somehow.

Could you please tell me how you generate and get the voice? My training steps are only 10,000 and it takes 2 days, because 100,000 steps are expected to take more than ten days. Is it because my training time and step size are insufficient that the speech generated by running "generator.py" is 0 seconds?

— You are receiving this because you commented. Reply to this email directly, view it on GitHub https://github.com/ibab/tensorflow-wavenet/issues/345?email_source=notifications&email_token=AIWQVUTBXXZRFMWBGSRDJ7LRCYDT3A5CNFSM4FFAPUM2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOELXNHJA#issuecomment-586077092, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIWQVUSQX5IID6DE5IQLQMDRCYDT3ANCNFSM4FFAPUMQ .

zegenerative avatar Feb 14 '20 08:02 zegenerative

You have to use gpu for this, cpu is too slow. If that is not the problem, try changing the training parameters. Hope it helps. Op vr 14 feb. 2020 04:10 schreef Ahapy [email protected]: I am running it on paperspace (GPU enabled, let me know if you want to be more specific, but believe me, this virtual machine is a beast). Took 2,5 days to train with 1e5 (100000 steps). To generate 3 minutes of audio takes like 6 hours or something. Result: a sort of average turd. I am a beginner in coding and one must tweak this script somehow. Could you please tell me how you generate and get the voice? My training steps are only 10,000 and it takes 2 days, because 100,000 steps are expected to take more than ten days. Is it because my training time and step size are insufficient that the speech generated by running "generator.py" is 0 seconds? — You are receiving this because you commented. Reply to this email directly, view it on GitHub <#345?email_source=notifications&email_token=AIWQVUTBXXZRFMWBGSRDJ7LRCYDT3A5CNFSM4FFAPUM2YY3PNVWWK3TUL52HS4DFVREXG43VMVBW63LNMVXHJKTDN5WW2ZLOORPWSZGOELXNHJA#issuecomment-586077092>, or unsubscribe https://github.com/notifications/unsubscribe-auth/AIWQVUSQX5IID6DE5IQLQMDRCYDT3ANCNFSM4FFAPUMQ .

Thanks for your suggestions, but I'm sure I have used a GPU. When running "train.py", there was such a paragraph: 2020-02-14 17:09:16.715677: I tensorflow/core/common_runtime/bfc_allocator.cc:929] Stats: Limit: 1422723891 InUse: 1344603648 MaxInUse: 1416595456 NumAllocs: 2844 MaxAllocSize: 193528576

And I tested these two lines of codefrom tensorflow.python.client import device_lib print(device_lib.list_local_devices())to detect whether the GPU or CPU is used at runtime. The results are shown below: [name: "/device:CPU:0" device_type: "CPU" memory_limit: 268435456 locality { } incarnation: 8034237779902744267 , name: "/device:GPU:0" device_type: "GPU" memory_limit: 1422723891 locality { bus_id: 1 links { } } incarnation: 4308487621035411117 physical_device_desc: "device: 0, name: GeForce MX150, pci bus id: 0000:01:00.0, compute capability: 6.1" ]

Limit: 1422723891 in "train.py" refers to device_type: "GPU" memory_limit: 1422723891 in the test results, so I think "train.py" uses a GPU instead of a CPU.

Ahapy avatar Feb 14 '20 09:02 Ahapy