A diagonal square matrix is a matrix whose only nonzero entries are on the diagonal: A square matrix is said to be diagonalizable if it is similar to a diagonal matrix. That is, is diagonalizable if there is an invertible matrix and a diagonal matrix such that
The matrix is diagonalizable: On the other hand, the matrix is not diagonalizable, as we will see below.
By the change of basis theorem, an matrix with entries in a field is diagonalizable if and only if there is a basis of consisting of eigenvectors of In what follows, we will mostly assume or but the definition is valid over an arbitrary field. This extends immediately to a definition of diagonalizability for linear transformations: if is a finite-dimensional vector space, we say that a linear transformation is diagonalizable if there is a basis of consisting of eigenvectors for
So the process of diagonalizing a matrix involves computing its eigenvectors and following the recipe of the change-of-basis theorem to compute the matrices and Matrix diagonalization is useful in many computations involving matrices, because multiplying diagonal matrices is quite simple compared to multiplying arbitrary square matrices. In particular, the powers of a diagonalizable matrix can be easily computed once the matrices and are known, as can the matrix exponential.
Review Eigenvalues and Eigenvectors.
The first theorem about diagonalizable matrices shows that a large class of matrices is automatically diagonalizable.
If is an matrix with distinct eigenvalues, then is diagonalizable.
Explicitly, let be these eigenvalues. Then where the column of is an eigenvector of with eigenvalue
Let be an eigenvector with eigenvalue Then the key fact is that the are linearly independent. To see this, let be the largest positive integer such that are linearly independent. If then there is a dependence relation for some coefficients Now multiply both sides on the left by to get Multiplying both sides of the original equation by instead gives and subtracting these two equations gives but this is impossible because are linearly independent. Note that it is very important that the are distinct, because at least one of the are nonzero, so the coefficient is nonzero as well--if the were not distinct, the coefficients of the left side might all be zero even if some of the were nonzero.
So and are linearly independent. So this gives a basis of eigenvectors of and hence is diagonalizable. Indeed, if is the matrix whose column vectors are the then let be the column of the identity matrix; then for all So But multiplying a matrix by just gives its column. So and have the same column, for all So they're the same matrix: or
The intuition from the theorem in the previous section is that there are two ways that a matrix can fail to be diagonalizable. One is that its eigenvalues can "live" in some other, larger field. This is in some sense a cosmetic issue, which can be corrected by passing to the larger field. For instance, if the matrix has real entries, its eigenvalues may be complex, so that the matrix may be diagonalizable over without being diagonalizable over
The rotation matrix is not diagonalizable over Indeed, it has no real eigenvalues: if is a vector in then equals rotated counterclockwise by But it is not hard to check that it has two distinct eigenvalues over since the characteristic polynomial is So is diagonalizable over
The second way in which a matrix can fail to be diagonalizable is more fundamental. The fundamental theorem of algebra applied to the characteristic polynomial shows that there are always complex eigenvalues, counted with multiplicity. But this does not mean that every square matrix is diagonalizable over the complex numbers. The multiplicity of each eigenvalue is important in deciding whether the matrix is diagonalizable: as we have seen, if each multiplicity is the matrix is automatically diagonalizable. Here is an example where an eigenvalue has multiplicity and the matrix is not diagonalizable:
Let Then the characteristic polynomial of is so there is only one eigenvalue, Since similar matrices have the same eigenvalues (indeed, the same characteristic polynomial), if were diagonalizable, it would be similar to a diagonal matrix with as its only eigenvalue, namely the identity matrix. But the only matrix similar to the identity matrix is the identity matrix: for all So cannot be diagonalizable.
There are other ways to see that is not diagonalizable, e.g. by computing the size of the eigenspace corresponding to and showing that there is no basis of eigenvalues of
Note that having repeated roots in the characteristic polynomial does not imply that the matrix is not diagonalizable: to give the most basic example, the identity matrix is diagonalizable (diagonal, in fact), but it has only one eigenvalue with multiplicity
More generally, there are two concepts of multiplicity for eigenvalues of a matrix. The multiplicity we have referred to above, the exponent of the factor in the characteristic polynomial, is known as the algebraic multiplicity. The dimension of the eigenspace corresponding to is called the geometric multiplicity. It is not hard to prove that the algebraic multiplicity is always the geometric multiplicity, so is diagonalizable if and only if these multiplicities are equal for every eigenvalue See the wiki on Jordan canonical form for more details.
The eigenvalues are the roots of the characteristic polynomial: We find eigenvectors for both eigenvalues:
- One solution, the one we will pick, is
- One solution, the one we will pick, is .
The prescription in the change-of-basis theorem leads us immediately to the matrices and :
Then we have
The process starts in the same way:
There are only two eigenvalues, and the eigenvalue has algebraic multiplicity since the characteristic polynomial factors as So it is not clear whether is diagonalizable until we know whether there are enough eigenvectors in the -eigenspace i.e. whether the geometric multiplicity of is or
We find eigenvectors for these eigenvalues:
which can be computed by Gauss-Jordan elimination: which has nullspace spanned by the vector
which can be computed by Gauss-Jordan elimination: which has a two-dimensional nullspace, spanned by, for instance, the vectors and
So this shows that is indeed diagonalizable, because there are "enough" eigenvectors to span
And we can write down the matrices and :
It is straightforward to check that as desired.
Note that the matrices and are not unique. is unique up to a rearrangement of the diagonal terms, but has much more freedom: while the column vectors from the -dimensional eigenspaces are determined up to a constant multiple, the column vectors from the larger eigenspaces can be chosen completely arbitrarily as long as they form a basis for their eigenspace.
Diagonal matrices are relatively easy to compute with, and similar matrices share many properties, so diagonalizable matrices are well-suited for computation. In particular, many applications involve computing large powers of a matrix, which is easy if the matrix is diagonal. But if then because the terms in the middle all collapse.
Find a closed-form formula for the Fibonacci number by looking at powers of the matrix
Looking at the first few powers of we have so the natural conjecture is that which is easy to prove by induction. The base case is clear, and the inductive step is as desired.
So the only thing left to do is to compute The characteristic polynomial is whose roots are and where is the golden ratio and is its conjugate.
The -eigenspace is the nullspace of which is one-dimensional and spanned by The same holds with all the 's replaced by 's. So the conclusion is that where
Putting this all together gives (we don't really care about the second column, although it's not much harder to compute).
In particular, the bottom left entry, which is by induction, equals which is Binet's formula for
More applications to exponentiation and solving differential equations are in the wiki on matrix exponentiation.