TinyImageNet-Benchmarks
TinyImageNet-Benchmarks copied to clipboard
Batch norm in Mobilenet v2.
As per the architecture definition provided here, it is shown that the batch normalization is used in inverted-residual blocks as below.
- Bottleneck layer which is normal convolutional layer is equipped with batch norm
- Depthwise layer is also equipped with batch norm
- Pointwise layer is also equipped with batch norm. Even in their paper they mentioned that they use batch norm after every layer.
But when we download a pretrained model from tensorflow and visualize it in Netron, it is as per below.
- Bottleneck layer doesnt use batch norm instead use bias.
- Depthwise is equipped with batch norm.
- Pointwise layer doesnt use batch norm instead use bias.
So this make a huge difference in number of parameters and final accuracy.