Results 1 comments of Yiwei

@xavierdawn `torchrun --nproc_per_node 4` should be enough to kick off distribued training. I only changed this setting without explictly indicating `CUDA_VISIBLE_DEVICES` to enable distribued training. You can use `nvidia-smi` to...