Authors Top

If you have a few years of experience in Computer Science or research, and you’re interested in sharing that experience with the community, have a look at our Contribution Guidelines.

1. Introduction

In this tutorial, we’ll review the Principal Component Analysis (PCA), and we’ll describe how to use it by choosing the optimal number of principal components.

Finally, we illustrate the application of such a technique to a popular machine learning dataset.

2. Dimensionality Reduction

Dimensionality reduction is the task of transforming data from a high dimensional space into a new space whose dimensionality is much smaller. Dimensionality reduction is closely related to the process of data compression in the signal processing field.

In machine learning, dimensionality reduction is generally used to speed up the training of a model. Moreover, in some cases reducing the dimensionality of data may filter out noisy features, and models with higher accuracy are obtained. Dimensionality reduction is also useful for data visualization and for finding a meaningful interpretation of the data. In fact, reducing the dimensions down to two or three allows to represent high dimensional data on a plot and obtain relevant information.

3. PCA Overview

PCA is the most popular dimensionality reduction technique. In PCA, the compression and the recovery are performed by means of linear transformations; the algorithm finds the linear transformations for which the mean squared error computed over a set of instances is minimal.

Given a set of M feature vectors \mathbf{x}_1, \dots, \mathbf{x}_M in \mathbf{R}^D, a real matrix T of size N \times D, where N<D, is used to reduce the dimensionality of \mathbf{x} by performing the linear transformation \mathbf{y} = T \mathbf{x}. A real matrix U of size D \times N is used to approximately reconstruct each original vector \mathbf{x} from the compressed version \mathbf{y} by performing the transformation  \tilde{\mathbf{x}} = U \mathbf{y}; hence, \tilde{\mathbf{x}} is the recovered version of the original feature vector \mathbf{x} and it has the same dimensionality DThe PCA consists in finding the compression matrix T and the recover matrix U such that the mean squared error computed over the set of instances is minimal, i.e., the following problem is solved:

(1)   \begin{equation*} \operatorname{argmin}_{(T , U)} \sum_{i=1}^{M}\left\|\mathbf{x}_{i}- \tilde{\mathbf{x}}_{i}\right\|_{2}^{2} = \operatorname{argmin}_{(T , U)} \sum_{i=1}^{M}\left\|\mathbf{x}_{i} -  U T \mathbf{x}_{i} \right\|_{2}^{2}. \end{equation*}

It can be demonstrated that if T and W are solutions of Eq. (1), then it must be W=T^{\top}. Hence, the recover matrix W is obtained by transposing the compression matrix T. Moreover it can be demonstrated that the columns of the matrix T, solution of Eq. (1), correspond to the eigenvectors of the sample covariance matrix of the data \{ \mathbf{x}_i \}^M_{i=1}.
We remind that the covariance matrix is defined as:

(2)   \begin{equation*} C =\frac{1}{M-1} \sum_{i=1}^{M}\left(\mathbf{x}_i- \boldsymbol{\mu}_{x} \right)\left(\mathbf{x}_i-\boldsymbol{\mu}_{x} \right)^{\top} \end{equation*}

where \boldsymbol{\mu}_{x} is a vector whose components are the mean values of each feature:

(3)   \begin{equation*} \boldsymbol{\mu}_{x} =\left[\begin{array}{c} \frac{1}{M} \sum_{i=1}^{M} {(\mathbf{x}_{i})}_1 \\ \vdots \\ \frac{1}{M} \sum_{i=1}^{M} {(\mathbf{x}_{i})}_d \end{array}\right] = \frac{1}{M} \sum_{i=1}^{M} \left[\begin{array}{c} {(\mathbf{x}_{i})}_1 \\ \vdots \\ {(\mathbf{x}_{i})}_d \end{array}\right] = \frac{1}{M} \sum_{i=1}^{M} \mathbf{x}_i \end{equation*}

Hence we need to compute the eigenvectors \mathbf{u}_i of the covariance matrix C and sort them so that the corresponding eigenvalues \lambda_i are sorted in descending order, i.e.:

(4)   \begin{equation*} \mathbf{u}_1, \mathbf{u}_2, \cdots, \mathbf{u}_N \ \ \ \hbox{s.t.}  \ \ \ \lambda_1 > \lambda_2  > \cdots > \lambda_N \end{equation*}

Finally, the compression matrix T of PCA can be expressed as:

(5)   \begin{equation*} T = \left[\begin{array}{c} \mathbf{u}_1 \cdots \mathbf{u}_N \end{array}\right] . \end{equation*}

The eigenvectors of the covariance matrix are called principal components.

4. Choosing the Right Number of Principal Components

The eigenvalue \lambda_i represents the variance of the data along the direction of the corresponding principal component. Hence the components with the lowest eigenvalues contain the least information, so they can be dropped. The importance of each component is represented by the so-called explained variance ratio, which indicates the portion of the variance that lies along each principal component:

(6)   \begin{equation*} \hbox{explained variance ratio of the i-th component} = \frac{\lambda_i}{\sum^D_{j=1} \lambda_j} \ . \end{equation*}

Note that sum at the denominator is performed over the maximum number of principal components.
By summing the explained variance ratio of the first N_{pc}} components, we obtain the so-called cumulative explained variance:

(7)   \begin{equation*} \hbox{cumulative explained variance of the first $N_{pc}$ components} = \frac{\sum^{N_{pc}}_{j=i}\lambda_j}{\sum^D_{j=1} \lambda_j} \end{equation*}

A good strategy is to choose the number of dimensions for which the cumulative explained variance exceeds a threshold, e.g., 0.95 (95%).

5. Examples

Let’s now apply the PCA to the MNIST dataset by exploiting the implementation provided in the Python package scikit-learn.

We load the MNIST dataset using a keras function, and we normalize the training set in the range [0, 1]:

import numpy as np
from keras.datasets import mnist

(x_train, y_train), (x_test, y_test) = mnist.load_data()
x_train = x_train.reshape(-1, 28*28) / 255.0

We perform the PCA on the training set by considering all principal components (28 \times 28 = 784), and then we compute the cumulative explained variance for each number of dimension:

from sklearn.decomposition import PCA
pca = PCA()
cumsum = np.cumsum(pca.explained_variance_ratio_)

Looking at the plot of the explained variance as a function of the number of principal components, we observe an elbow in the curve. The optimal number of principal components is reached when the cumulative variance stops growing fast:

Figure 1-2

The minimum number of principal components required to preserve the 95% of the data’s variance can be computed with the following command:

d = np.argmax(cumsum >= 0.95) + 1

We found that the number of dimensions can be reduced from 784 to 150 while preserving 95% of its variance. Hence, the compressed dataset is now 19% of its original size! If we want to preserve 99% of the variance, the number of dimensions can be reduced to 330, i.e., 42% of the original size:

Figure 2-2

6. Conclusion

In this article, we reviewed the Principal Component Analysis, a popular dimensionality reduction technique. Then, we described how to use it by choosing the optimal number of principal components. Finally, we showed the application of the PCA to the MNIST dataset using a Python implementation.

Authors Bottom

If you have a few years of experience in Computer Science or research, and you’re interested in sharing that experience with the community, have a look at our Contribution Guidelines.

Comments are closed on this article!