Matrix diagonalization

A diagonalizable matrix A can be written as

A=PΔP1

where Δ is a diagonal matrix (the eigenvalues at the diagonal) and P is a matrix whose columns are the eigenvectors.

Idea

The idea is that we have a basis of the vector space in which the transformation given by A is simply made of scale changes (even negative or null) in the main axes. When the basis change is given by means of an orthogonal matrix then the matrix A is symmetric. That is, we have

It is related to singular value decomposition. Indeed they are equal when the matrix is symmetric and positive-semidefinite. See MSE.

Diagonalization method

  1. Solve det(AλI)=0. The solutions λi are called eigenvalues.
  2. For every λi we look for a basis of the subspace Vλi
(AλiI)(x1x2)=(00)

They are the eigenvectors associated to the eigenvalue λi.
3. If we have enough eigenvectors to complete a basis of RN then the matrix A is diagonalizable, and Λ is made with the eigenvalues (repeated if necessary, according to the dimension of Vλi). The P matrix is made with the eigenvectors. If we don't have enough eigenvectors, the matrix is not diagonalizable and we look for the Jordan canonical form of a matrix. The reasons could be:
a. There are complex solutions in step 1.
b. Even with enough solutions, or even in the case of a complex matrix, any Vλi has not dimension enough "to fill" the multiplicity of λi in step 1.

If a matrix is diagonalized, its diagonal form is unique, up to a permutation of the diagonal entries. This is because the entries on the diagonal must be all the eigenvalues. For instance,

[100020001] and [100010002]

are examples of two different ways to diagonalize the same matrix.

Important fact: common eigenvectors.