ResNeXt.pytorch
ResNeXt.pytorch copied to clipboard
GPU memory usage during training
It seems to me that each image uses ~5GB of GPU memory (ResNeXt-152), making it only possible to train with 2 images per GPU (TITAN X). Is that normal? I would appreciate if someone could be able to point out where I can start debugging for this?
I have never tried with 152 layers, but it seems normal that memory blows up when using so many layers. Did you find any possible answer?