yuanyuanli85/Stacked_Hourglass_Network_Keras

about bottleneck_block

Opened this issue · 1 comments

I notice that in bottleneck_block, the structure is Conv2D->ReLU->BN, I wonder weather the network works well because all residual_block I used to see is Conv2D->BN->ReLUs?

There are some discussion on that.
https://www.reddit.com/r/MachineLearning/comments/67gonq/d_batch_normalization_before_or_after_relu/

I think both work, but don't know which one is better without experimenting