Batch norm in Mobilenet v2.
meet-minimalist opened this issue · 0 comments
meet-minimalist commented
As per the architecture definition provided here, it is shown that the batch normalization is used in inverted-residual blocks as below.
- Bottleneck layer which is normal convolutional layer is equipped with batch norm
- Depthwise layer is also equipped with batch norm
- Pointwise layer is also equipped with batch norm.
Even in their paper they mentioned that they use batch norm after every layer.
But when we download a pretrained model from tensorflow and visualize it in Netron, it is as per below.
- Bottleneck layer doesnt use batch norm instead use bias.
- Depthwise is equipped with batch norm.
- Pointwise layer doesnt use batch norm instead use bias.
So this make a huge difference in number of parameters and final accuracy.