RepDistiller
RepDistiller copied to clipboard
teacher model is too big to run with batch_size 64
when I try to train a teacher model on cub200(200 classes), I use resnet50 and batch size 64, It will out of memory, I use 16G GPU. I could run when i set the batch size 8. Why resnet50 is so big ?