- 知乎上面有关各种Normalization算法理解
- 简书上关于BatchNorm2d参数解释
- note 11:BatchNorm2D官方手册,具体解析,如下:
How to set learning rate as 0 in BN layer中所讲有关的参数affine理解如下:
Setting affine=False will remove the gamma and beta terms from the calculation,
thus only using the running mean and var. So that’s basically, what you want.
I don’t know, how Caffe works, but setting the learning rate