pytorch-retinanet icon indicating copy to clipboard operation
pytorch-retinanet copied to clipboard

Why do you freeze batch norm parameters when training?

Open leao1995 opened this issue 7 years ago • 4 comments

would it be better to let batch norm parameters adapt to your current data?

leao1995 avatar Nov 09 '17 21:11 leao1995

it's a common practice.First, Because the pretrain network's bn layers have been trained. Second,Object Detection 's batchsize is small, hard to make bn parameter stable.

xichangzun avatar Dec 19 '17 07:12 xichangzun

use Group norm instead of batch norm . it is more stable.

prakashjayy avatar Apr 26 '18 06:04 prakashjayy

Use synchronized batch normalization

lxtGH avatar May 02 '18 14:05 lxtGH

Use synchronized batch normalization

Using sync batch norm does not help with single GPU training and low batch sizes though.

PhilipMay avatar Apr 21 '20 19:04 PhilipMay