TransFG icon indicating copy to clipboard operation
TransFG copied to clipboard

Batch_size is 16 or 64?

Open qilong-zhang opened this issue 3 years ago • 1 comments

Hi @TACJu, I notice you apply DDP with 4 GPUs in train.py. Therefore, if the batch_size in args is set to 16, then the overall batch_size will be 16x4=64. However, in your paper, you say that the batch_size is 16. I also try batch_size 16x4 on Tesla V100, but OOM will be raised, so I wonder batch_size is 16 means 16 or 64? thanks!

qilong-zhang avatar Jul 12 '21 02:07 qilong-zhang

Maybe the batch_size is 16*4 = 64. I run the code with batch_size=4*4 , and the accuracy on the CUB_200_2011 dataset is only 90.9%. After changing the batch_size to 4*8 (limited by memory, 4*16 cause OOM on my server with 4 RTX3090 GPUs), the accuracy raises up to 91.4%.

slothfulxtx avatar Sep 21 '21 09:09 slothfulxtx