### Table of Contents

This page describes the *Fisher Vector* (FV) of [23] [24] . See Fisher Vector encoding (FV) for an overview of the C API and Fisher kernel for its relation to the more general notion of Fisher kernel.

The FV is an image representation obtained by pooling local image features. It is frequently used as a global image descriptor in visual classification.

While the FV can be derived as a special, approximate, and improved case of the general Fisher Kernel framework, it is easy to describe directly. Let \(I = (\bx_1,\dots,\bx_N)\) be a set of \(D\) dimensional feature vectors (e.g. SIFT descriptors) extracted from an image. Let \(\Theta=(\mu_k,\Sigma_k,\pi_k:k=1,\dots,K)\) be the parameters of a Gaussian Mixture Model fitting the distribution of descriptors. The GMM associates each vector \(\bx_i\) to a mode \(k\) in the mixture with a strength given by the posterior probability:

\[ q_{ik} = \frac {\exp\left[-\frac{1}{2}(\bx_i - \mu_k)^T \Sigma_k^{-1} (\bx_i - \mu_k)\right]} {\sum_{t=1}^K \exp\left[-\frac{1}{2}(\bx_i - \mu_t)^T \Sigma_k^{-1} (\bx_i - \mu_t)\right]}. \]

For each mode \(k\), consider the mean and covariance deviation vectors

\begin{align*} u_{jk} &= {1 \over {N \sqrt{\pi_k}}} \sum_{i=1}^{N} q_{ik} \frac{x_{ji} - \mu_{jk}}{\sigma_{jk}}, \\ v_{jk} &= {1 \over {N \sqrt{2 \pi_k}}} \sum_{i=1}^{N} q_{ik} \left[ \left(\frac{x_{ji} - \mu_{jk}}{\sigma_{jk}}\right)^2 - 1 \right]. \end{align*}

where \(j=1,2,\dots,D\) spans the vector dimensions. The FV of image \(I\) is the stacking of the vectors \(\bu_k\) and then of the vectors \(\bv_k\) for each of the \(K\) modes in the Gaussian mixtures:

\[ \Phi(I) = \begin{bmatrix} \vdots \\ \bu_k \\ \vdots \\ \bv_k \\ \vdots \end{bmatrix}. \]

# Normalization and improved Fisher vectors

The *improved* Fisher Vector [24] (IFV) improves the classification performance of the representation by using to ideas:

*Non-linear additive kernel.*The Hellinger's kernel (or Bhattacharya coefficient) can be used instead of the linear one at no cost by signed squared rooting. This is obtained by applying the function \(|z| \sign z\) to each dimension of the vector \(\Phi(I)\). Other additive kernels can also be used at an increased space or time cost.*Normalization.*Before using the representation in a linear model (e.g. a support vector machine), the vector \(\Phi(I)\) is further normalized by the \(l^2\) norm (note that the standard Fisher vector is normalized by the number of encoded feature vectors).

After square-rooting and normalization, the IFV is often used in a linear classifier such as an SVM.

# Faster computations

In practice, several data to cluster assignments \(q_{ik}\) are likely to be very small or even negligible. The *fast* version of the FV sets to zero all but the largest assignment for each input feature \(\bx_i\).