MDT icon indicating copy to clipboard operation
MDT copied to clipboard

Training Time

Open mapengsen opened this issue 10 months ago • 1 comments

Dear Authors,

when i use one gpu to train the model : python -m torch.distributed.launch --nproc_per_node=1 scripts/image_train.py

and use 4 gpu to train the model: python -m torch.distributed.launch --nproc_per_node=4 scripts/image_train.py

The number of training steps is the same under the same training time (the final modelXXX.pt obtained by 4 GPUs will not be more/faster than that obtained by 1 GPU). Is this normal?

mapengsen avatar Apr 22 '24 07:04 mapengsen