Review of Linear Algebra

Matrix

Def: Matrix

Example application, we can create a matrix where we index the rows by companies, and columns by dates. Each entry in the matrix represents the stock price of a given company on a given date.

Ok, so we can represent data in a matrix, but what makes it so powerful?

From a theoretical point of view, a matrix is an operation! That is, an matrix represents a Linear Transformation.

Now let’s try to reconcile the two points of views. That is, what does it mean for a dataset to be a linear transformation?

Eigenvalues and Eigenvectors

The most important concepts for us are the Eigenvector and Eigenvalues of a matrix. Recall that, due to the Characteristic Equation we find that all square matrices have eigenvectors and eigenvalues which may be in the complex plane.

We also learn about the geometrical meaning of eigenvalues and eigenvectors.

As matrices are linear transformations which simply transform some vector into a new vector . We note that an eigenvector is some special vector that, when transformed by the matrix simply results in a scaled version of itself. The scaling factor is the eigenvalue.

Diagonalizable Matrices

A matrix is orthonormally diagonalizable if there exists an orthonormal matrix such that:

Where is a diagonal matrix

Geometrical Meaning

This means that we can decompose the transformation into three simpler transformations:

  1. : Rotate the vector into the eigenvector basis
  2. : Scale each component by the corresponding eigenvalue
  3. : Rotate back to the original basis

This decomposition is powerful because it allows us to understand and compute the effect of by breaking it down into simpler operations.

Symmetric Matrices

It is known that all Symmetric Matrices are orthonormally diagonalizable and have real eigenvalues.

This is by the Spectral Theorem

What about General Matrices?

First of all, not all matrices are diagonalizable.

But we still want to understand the matrix through “simple” operations such as scaling.

But there are still tools we can apply to all matrices, that is the Singular Value Decomposition (SVD)!

Comparing EVD with SVD

EVD only takes in square diagonalizable matrices. Whereas SVD can take in any matrix. The reason why EVD is still more powerful is because

EVD gives you a set of vectors in which applying will simply scale them.

Whereas SVD gives you a set of vectors, tells you how they map and scale onto a set of vectors. What happens to the vectors left? They get mapped to the zero vector!

Looking at our data matrix example, we can assume that we will have a lot more columns than rows if columns are dates and rows are companies. Thus, we will often use the Simplified SVD form as it saves an enormous amount of computation.

Bringing it back

Let’s reconsider a data matrix with companies and dates, where each row represents the returns of that company.

Then is a covariance matrix, where entry represents the covariance between companies and .

An eigenvector of this matrix represents a weighted combination of companies (a portfolio or factor), and its corresponding eigenvalue measures the amount of variance explained by that factor.