memcnn icon indicating copy to clipboard operation
memcnn copied to clipboard

Does your memcnn support muti-gpu training ?

Open MrLinNing opened this issue 4 years ago • 3 comments

MrLinNing avatar Oct 07 '19 06:10 MrLinNing

Yes, it should support multi-gpu training. Please, let me know if you run into any issues.

silvandeleemput avatar Oct 07 '19 07:10 silvandeleemput

Can your method balance the GPU memory usage ? I find this problem. image

MrLinNing avatar Oct 07 '19 07:10 MrLinNing

I assume that what you are showing is the output of the nvidia-smi command. The ReversibleBlock of MemCNN by default frees memory after assigning the memory. So the memory might actually be available, but what you see is that it is still reserved by PyTorch. Hence, the actual available memory is not reflected by the nvidia-smi command. To see the actual available memory you can use the torch.cuda.memory_allocated() command per device.

See also: https://pytorch.org/docs/stable/notes/cuda.html#cuda-memory-management

silvandeleemput avatar Oct 07 '19 12:10 silvandeleemput