VL_NNBNORM - CNN batch normalisation.

Y = VL_NNBNORM(X,G,B) applies batch normalization to the input X. Batch normalization is defined as:

   Y(i,j,k,t) = G(k) * (X(i,j,k,t) - mu(k)) / sigma(k) + B(k)

where:

   mu(k) = mean_ijt X(i,j,k,t),
   sigma2(k) = mean_ijt (X(i,j,k,t) - mu(k))^2,
   sigma(k) = sqrt(sigma2(k) + EPSILON)

are respectively the per-channel mean, variance, and standard deviation of each feature channel in the data X. The parameters G(k) and B(k) are multiplicative and additive constants use to scale each data channel.

Means and variances are accumulated across all the data items (images) stored in the 4D tensor X (from which the name batch normalization is derived). The constant EPSILON is used to regularize the computation of sigma(k) and to avoid division by zero.

[DZDX,DZDG,DZDB] = VL_NNBNORM(X,G,B,DZDY) computes the derviatives of the block projected onto DZDY. DZDX, DZDG, DZDB and DZDY have the same dimensions as X, G, B, and Y respectivey.

Optionally, [Y,MOMENTS] = VL_NNBNORM(...) and [DZDX,DZDG,DZDB,MOMENTS] = VL_NNBNORM(...,DZDY) return the values of the vectors mu and sigma in the formulas above. Here, MOMENTS is a DEPTH x 2 array [MU, SIGMA].

VL_NNBNROM(..., 'Option', value) takes the following options:

See also: VL_NNNORMALIZE().