VL_NNBNORM - CNN batch normalisation.
Y = VL_NNBNORM(X,G,B) applies batch normalization to the input X. Batch normalization is defined as:
Y(i,j,k,t) = G(k) * (X(i,j,k,t) - mu(k)) / sigma(k) + B(k)
mu(k) = mean_ijt X(i,j,k,t), sigma2(k) = mean_ijt (X(i,j,k,t) - mu(k))^2, sigma(k) = sqrt(sigma2(k) + EPSILON)
are respectively the per-channel mean, variance, and standard deviation of each feature channel in the data X. The parameters G(k) and B(k) are multiplicative and additive constants use to scale each data channel.
Means and variances are accumulated across all the data items (images) stored in the 4D tensor X (from which the name batch normalization is derived). The constant EPSILON is used to regularize the computation of sigma(k) and to avoid division by zero.
[DZDX,DZDG,DZDB] = VL_NNBNORM(X,G,B,DZDY) computes the derviatives of the block projected onto DZDY. DZDX, DZDG, DZDB and DZDY have the same dimensions as X, G, B, and Y respectivey.
Optionally, [Y,MOMENTS] = VL_NNBNORM(...) and [DZDX,DZDG,DZDB,MOMENTS] = VL_NNBNORM(...,DZDY) return the values of the vectors mu and sigma in the formulas above. Here, MOMENTS is a DEPTH x 2 array [MU, SIGMA].
VL_NNBNROM(..., 'Option', value) takes the following options:
Specifies the constant EPSILON in the formuals above.
Specifies an array MOMENTS with the values of mu and sigma to use instead of computing them according to the equations above. This is useful to disable batch normalization during testing.
If specified, turns on CuDNN. CuDNN is on by default. This option can be useful to undo the effect of a previous
NoCuDNNoption in the argument list.
If specified, turns off CuDNN.
See also: VL_NNNORMALIZE().