nonparaSeq2seqVC_code
nonparaSeq2seqVC_code copied to clipboard
Multi-GPU training
Hello,
Could you please specify the steps to enable multi-GPU training, please?
I set distributed_run=True
in hparams.py
and then set --n_gpus=2
and CUDA_VISIBLE_DEVICES=0,3
in file run.sh
to select GPUs 0 and 3, respectively. I did this and the code seems to enter some kind of deadlock because it does not start training.
Thank you.
The use of multi-GPU training is basically the same as in https://github.com/NVIDIA/tacotron2. First create a directory named "logs", then run python -m multiproc train.py --output_directory=outdir --log_directory=logdir --n_gpus=2 --hparams=distributed_run=True
Thanks your impressive work.
when I use multi-GPU training, such as python -m multiproc train.py --output_directory=outdir --log_directory=logdir --n_gpus=2 --hparams=distributed_run=True
I run into the error, as shown below:
Traceback (most recent call last):
File "train.py", line 369, in
Hi, as the error message says, when using multi-GPU training, you need to set up shuffle=False in dataloader.