DAFormer icon indicating copy to clipboard operation
DAFormer copied to clipboard

why not support Multi-GPU? is it not necessary for UDA?

Open fuweifu-vtoo opened this issue 2 years ago • 1 comments

Have you read code for SoftTeacher? This code seems to have nothing to do with ddp_train_gans, but it still use multi-gpu training UDA model. https://github.com/lhoyer/DAFormer/issues/9

fuweifu-vtoo avatar Sep 09 '22 07:09 fuweifu-vtoo

It feels like your code supports multi-gpu training natively because you use the DDP wrapper. You just need to start the command with OMP_NUM_THREADS=1 python3 -m torch.distributed.launch ?

fuweifu-vtoo avatar Sep 09 '22 10:09 fuweifu-vtoo

Have you read code for SoftTeacher? This code seems to have nothing to do with ddp_train_gans, but it still use multi-gpu training UDA model. https://github.com/lhoyer/DAFormer/issues/9

We have supported the Multi-GPU training procedure. You can follow our work and repo (https://github.com/xiaoachen98/DDB).

xiaoachen98 avatar Sep 23 '22 14:09 xiaoachen98

@fuweifu-vtoo DAFormer can be trained on a single GPU within 16 hours. As this is a very reasonable training time, we opted for simplicity and did not use Multi-GPU training. The instructions on how to generally use Multi-GPU training can be found in the mmsegmentation repository. Please, also check that the model update of the momentum teacher is compatible.

@xiaoachen98 Thanks for sharing your code with Multi-GPU training.

lhoyer avatar Sep 27 '22 06:09 lhoyer