Definition. Let . The characteristic polynomial of A is
(I is the identity matrix.)
A root of the characteristic polynomial is called an eigenvalue (or a characteristic value) of A.
While the entries of A come from the field F, it makes sense to ask for the roots of in an extension field E of F. For example, if A is a matrix with real entries, you can ask for the eigenvalues of A in or in .
Example. Consider the matrix
The characteristic polynomial is . Hence, A has no eigenvalues in . Its eigenvalues in are .
Example. Let
You can use row and column operations to simplify the computation of :
(Adding a multiple of a row or a column to a row or column, respectively, does not change the determinant.) Now expand by cofactors of the second row:
The eigenvalues are , (double).
Example. A matrix is upper triangular if for . Thus, the entries below the main diagonal are zero. ( Lower triangular matrices are defined in an analogous way.)
The eigenvalues of a triangular matrix
are just the diagonal entries . (You can prove this by induction on n.)
Remark. To find the eigenvalues of a matrix, you need to find the roots of the characteristic polynomial.
There are formulas for finding the roots of polynomials of degree . (For example, the quadratic formula gives the roots of a quadratic equation .) However, Abel showed in the early part of the 19-th century that the general quintic is not solvable by radicals. (For example, is not solvable by radicals over .) In the real world, the computation of eigenvalues often requires numerical approximation.
If is an eigenvalue of A, then . Hence, the matrix is not invertible. It follows that must row reduce to a row reduced echelon matrix R with fewer than n leading coefficients. Thus, the system has at least one free variable, and hence has more than one solution. In particular, --- and therefore, --- has at least one nonzero solution.
Definition. Let , and let be an eigenvalue of A. An eigenvector (or a characteristic vector) of A for is a nonzero vector such that
Equivalently,
Example. Let
The eigenvalues are , (double).
First, I'll find an eigenvector for .
I want such that
You can solve the system by row reduction. Since the column of zeros on the right will never change, it's enough to row reduce the matrix on the right.
This says
Therefore, , , and the eigenvector is
Notice that this is the usual algorithm for finding a basis for the solution space of a homogeneous system (or the null space of a matrix).
I can set c to any nonzero number. For example, gives the eigenvector . Notice that there are infinitely many eigenvectors for this eigenvalue, but all of these eigenvectors are multiples of .
Likewise,
Hence, the eigenvectors are
Taking , gives ; taking , gives . This eigenvalue gives rise to two independent eigenvectors.
Note, however, that a double root of the characteristic polynomial need not give rise to two independent eigenvectors.
Definition. Matrices are similar if there is an invertible matrix such that .
Lemma. Similar matrices have the same characteristic polynomial (and hence the same eigenvalues).
Proof.
Therefore, the matrices and are similar. Hence, they have the same determinant. The determinant of is the characteristic polynomial of A and the determinant of is the characteristic polynomial of .
Definition. Let be a linear transformation, where V is a finite-dimensional vector space. The characteristic polynomial of T is the characteristic polynomial of a matrix of T relative to a basis of V.
The preceding lemma shows that this is independent of the choice of basis. For if and are bases for V, then
Therefore, and are similar, so they have the same characteristic polynomial.
This shows that it makes sense to speak of the eigenvalues and eigenvectors of a linear transformation T.
Definition. A matrix is diagonalizable if A has n independent eigenvectors --- that is, if there is a basis for consisting of eigenvectors of A.
Proposition. is diagonalizable if and only if it is similar to a diagonal matrix.
Proof. Let be n independent eigenvectors for A corresponding to eigenvalues . Let T be the linear transformation corresponding to A:
Since for all i, the matrix of T relative to the basis is
Now A is the matrix of T relative to the standard basis, so
The matrix is obtained by building a matrix using the as the columns. Then .
Hence,
Conversely, if D is diagonal, P is invertible, and , the columns of P are independent eigenvectors for A. In fact, if
then says
Hence, .
Example. Consider the matrix matrix
In an earlier example, I showed that A has 3 independent eigenvectors , , . Therefore, A is diagonalizable.
To find a diagonalizing matrix, build a matrix using the eigenvectors as the columns:
You can check by finding and doing the multiplication that you get a diagonal matrix:
Of course, I knew this was the answer! I should get a diagonal matrix with the eigenvalues on the main diagonal, in the same order that I put the corresponding eigenvectors into P.
You can put the eigenvectors in as the columns of P in any order: A different order will give a diagonal matrix with the eigenvalues on the main diagonal in a different order.
Example. Let
Find the eigenvalues and, for each eigenvalue, a complete set of eigenvectors. If A is diagonalizable, find a matrix P such that is a diagonal matrix.
The eigenvalue is .
Now
Thinking of this as the coefficient matrix of a homogeneous linear system with variables a, b, and c, I obtain the equations
Then , so
is an eigenvector. Since there's only one independent eigenvector --- as opposed to 3 --- the matrix A is not diagonalizable.
Example. The following matrixhas eigenvalue (a triple root):
Now
Thinking of this as the coefficient matrix of a homogeneous linear system with variables a, b, and c, I obtain the equations
Set . This gives and . Thus, the only eigenvectors are the nonzero multiples of . Since there is only one independent eigenvectors, B is not diagonalizable.
Proposition. Let be a linear transformation on an n dimensional vector space. If are eigenvectors corresponding to the distinct eigenvalues , then is independent.
Proof. Suppose to the contrary that is dependent. Let p be the smallest number such that the subset is dependent. Then there is a nontrivial linear relation
Note that , else
This would contradict minimality of p.
Hence, I can rewrite the equation above in the form
Apply T to both sides, and use :
On the other hand,
Subtract the last equation from the one before it to obtain
Since the eigenvalues are distinct, the terms are nonzero. Hence, this is a linear relation in which contradicts minimality of p --- unless .
In this case, , which contradicts the fact that is an eigenvector. Therefore, the original set must in fact be independent.
Example. Let A be an real matrix. The complex eigenvalues of A always come in conjugate pairs and .
Moreover, if v is an eigenvector for , then the conjugate is an eigenvector for .
For suppose . Taking complex conjugates, I get
( because A is a real matrix.)
In practical terms, this means that once you've found an eigenvector for one complex eigenvalue, you can get an eigenvector for the conjugate eigenvalue by taking the conjugate of the eigenvector. You don't need to do a separate eigenvector computation.
For example, suppose
The characteristic polynomial is . The eigenvalues are .
Find an eigenvector for :
I knew that the second row must be a multiple of the first row, because I know the system has nontrivial solutions. So I don't have to work out what multiple it is; I can just zero out the second row on general principles.
This only works for matrices, and only for those which are 's in eigenvector computations.
Next, there's no point in going all the way to row reduced echelon form. I just need some nonzero vector such that
That is, I want
I can find an a and b that work by swapping and -1, and negating one of them. For example, take (-1 negated) and . This checks:
So is an eigenvector for .
By the discussion at the start of the example, I don't need to do a computation for . Just conjugate the previous eigenvector: must be an eigenvector for .
Since there are 2 independent eigenvectors, you can use them construct a diagonalizing matrix for A:
Notice that you get a diagonal matrix with the eigenvalues on the main diagonal, in the same order in which you listed the eigenvectors.
Example. For the following matrix, find the eigenvalues over , and for each eigenvalue, a complete set of independent eigenvectors.
Find a diagonalizing matrix and the corresponding diagonal matrix.
The characteristic polynomial is
Now
The eigenvalues are and .
For , I have
With variables a, b, and c, the corresponding homogeneous system is and . This gives the solution vector
Taking , I obtain the eigenvector .
For , I have
I multiplied the first row by , then divided it by 5. This made it the same as the third row.
I divided the second row by .
(I knew the the first and third rows had to be multiples, since they're clearly independent of the second row. Thus, if they weren't multiples, the three rows would be independent, the eigenvector matrix would be invertible, and there would be no eigenvectors [which must be nonzero].)
Now I can wipe out the third row by subtracting the first:
With variables a, b, and c, the corresponding homogeneous system is
There will only be one parameter (c), so there will only be one independent eigenvector. To get one, switch the "-5" and " " and negate the "-5" to get "5". This gives , , and . You can see that these values for a and c work:
Thus, my eigenvector is .
Hence, an eigenvector for is the conjugate .
A diagonalizing matrix is given by
With this diagonalizing matrix, I have
Copyright 2011 by Bruce Ikenaga