Definition

Batch normalization (batch norm) make training of neural network faster through normalization of each layers’ input by re-centering and re-scaling where:
- is the mean vector of the mini-batch
- is the variance vector of the mini-batch
- is the batch size.
- , are the learnable scaling parameters.
- is a small constant to prevent division by zero.
In test stage, the training stage’s moving average of the and are used.