HAT icon indicating copy to clipboard operation
HAT copied to clipboard

Should I change the lr if I use 4 GPUs while retaining the same total batchsize, i.e., 8 for each GPU? Thanks

Open c-yn opened this issue 2 years ago • 3 comments

c-yn avatar Jun 25 '22 20:06 c-yn

By the way, could you pls tell me if you try to train the network with 4 A100 gpus? If so, can each gpu accommodate 8 batchsize? Thanks.

c-yn avatar Jun 25 '22 20:06 c-yn

@morgen-star You don't need to change the learning rate when keeping the total batch size the same. For basic version of HAT, 4 A100 GPUs with batch size per gpu of 8 is OK.

chxy95 avatar Jun 26 '22 04:06 chxy95

If I use 2 GPUs with 8 Batch per GPU (total Batch = 16), does the learning rate need to be adjusted? Thank you for your answer!

Zency-Sun avatar Apr 15 '23 07:04 Zency-Sun