* Definition.* Let . The * characteristic polynomial* of A is

(I is the identity matrix.)

A root of the characteristic polynomial is called an * eigenvalue* (or a * characteristic
value*) of A.

While the entries of A come from the field F, it makes sense to ask for the roots of in an extension field E of F. For example, if A is a matrix with real entries, you can ask for the eigenvalues of A in or in .

* Example.* Consider the matrix

The characteristic polynomial is . Hence, A has no eigenvalues in . Its eigenvalues in are .

* Example.* Let

You can use row and column operations to simplify the computation of :

(Adding a multiple of a row or a column to a row or column, respectively, does not change the determinant.) Now expand by cofactors of the second row:

The eigenvalues are , (double).

* Example.* A matrix is * upper triangular* if for . Thus, the entries below the main
diagonal are zero. (* Lower triangular* matrices
are defined in an analogous way.)

The eigenvalues of a triangular matrix

are just the diagonal entries . (You can prove this by induction on n.)

* Remark.* To find the eigenvalues of a matrix,
you need to find the roots of the characteristic polynomial.

There are formulas for finding the roots of polynomials of degree . (For example, the quadratic formula gives the roots of a quadratic equation .) However, Abel showed in the early part of the 19-th century that the general quintic is not solvable by radicals. (For example, is not solvable by radicals over .) In the real world, the computation of eigenvalues often requires numerical approximation.

If is an eigenvalue of A, then . Hence, the matrix is not invertible. It follows that must row reduce to a row reduced echelon matrix R
with fewer than n leading coefficients. Thus, the system has at least one free variable, and hence has more
than one solution. In particular, --- and therefore,
--- has at least *one nonzero
solution*.

* Definition.* Let , and let
be an eigenvalue of A. An *
eigenvector* (or a * characteristic vector*)
of A for is a *nonzero* vector such that

Equivalently,

* Example.* Let

The eigenvalues are , (double).

First, I'll find an eigenvector for .

I want such that

You can solve the system by row reduction. Since the column of zeros on the right will never change, it's enough to row reduce the matrix on the right.

This says

Therefore, , , and the eigenvector is

Notice that this is the usual algorithm for finding a basis for the solution space of a homogeneous system (or the null space of a matrix).

I can set c to any nonzero number. For example, gives the eigenvector . Notice that there are infinitely many eigenvectors for this eigenvalue, but all of these eigenvectors are multiples of .

Likewise,

Hence, the eigenvectors are

Taking , gives ; taking , gives . This eigenvalue gives rise to two independent eigenvectors.

Note, however, that a double root of the characteristic polynomial
*need not* give rise to two independent eigenvectors.

* Definition.* Matrices are * similar* if there is an
invertible matrix such that .

* Lemma.* Similar matrices have the same
characteristic polynomial (and hence the same eigenvalues).

* Proof.*

Therefore, the matrices and are similar. Hence, they have the same determinant. The determinant of is the characteristic polynomial of A and the determinant of is the characteristic polynomial of .

* Definition.* Let be a
linear transformation, where V is a finite-dimensional vector space.
The * characteristic polynomial* of T is the
characteristic polynomial of a matrix of T relative to a basis of V.

The preceding lemma shows that this is independent of the choice of basis. For if and are bases for V, then

Therefore, and are similar, so they have the same characteristic polynomial.

This shows that it makes sense to speak of the *
eigenvalues* and * eigenvectors* of a
*linear transformation* T.

* Definition.* A matrix is * diagonalizable* if A has n
independent eigenvectors --- that is, if there is a basis for
consisting of eigenvectors of A.

* Proposition.* is
diagonalizable if and only if it is similar to a diagonal matrix.

* Proof.* Let be n
independent eigenvectors for A corresponding to eigenvalues . Let T be the linear
transformation corresponding to A:

Since for all i, the matrix of T relative to the basis is

Now A is the matrix of T relative to the standard basis, so

The matrix is obtained by building a matrix using the as the columns. Then .

Hence,

Conversely, if D is diagonal, P is invertible, and , the columns of P are independent eigenvectors for A. In fact, if

then says

Hence, .

* Example.* Consider the matrix matrix

In an earlier example, I showed that A has 3 independent eigenvectors , , . Therefore, A is diagonalizable.

To find a diagonalizing matrix, build a matrix using the eigenvectors as the columns:

You can check by finding and doing the multiplication that you get a diagonal matrix:

Of course, I knew this was the answer! I should get a diagonal matrix
with the eigenvalues on the main diagonal, *in the same order that
I put the corresponding eigenvectors into* P.

You can put the eigenvectors in as the columns of P in any order: A different order will give a diagonal matrix with the eigenvalues on the main diagonal in a different order.

* Example.* Let

Find the eigenvalues and, for each eigenvalue, a complete set of eigenvectors. If A is diagonalizable, find a matrix P such that is a diagonal matrix.

The eigenvalue is .

Now

Thinking of this as the coefficient matrix of a homogeneous linear system with variables a, b, and c, I obtain the equations

Then , so

is an eigenvector. Since there's only one independent eigenvector --- as opposed to 3 --- the matrix A is not diagonalizable.

* Example.* The following matrixhas eigenvalue
(a triple root):

Now

Thinking of this as the coefficient matrix of a homogeneous linear system with variables a, b, and c, I obtain the equations

Set . This gives and . Thus, the only eigenvectors are the nonzero multiples of . Since there is only one independent eigenvectors, B is not diagonalizable.

* Proposition.* Let be a linear transformation on an n dimensional vector
space. If are eigenvectors corresponding to
the *distinct* eigenvalues , then is independent.

* Proof.* Suppose to the contrary that is dependent. Let p be the smallest number
such that the subset is dependent. Then
there is a nontrivial linear relation

Note that , else

This would contradict minimality of p.

Hence, I can rewrite the equation above in the form

Apply T to both sides, and use :

On the other hand,

Subtract the last equation from the one before it to obtain

Since the eigenvalues are distinct, the terms are nonzero. Hence, this is a linear relation in which contradicts minimality of p --- unless .

In this case, , which contradicts the fact that is an eigenvector. Therefore, the original set must in fact be independent.

* Example.* Let A be an *real* matrix. The complex eigenvalues of A
always come in conjugate pairs and .

Moreover, if v is an eigenvector for , then the conjugate is an eigenvector for .

For suppose . Taking complex conjugates, I get

( because A is a real matrix.)

In practical terms, this means that once you've found an eigenvector for one complex eigenvalue, you can get an eigenvector for the conjugate eigenvalue by taking the conjugate of the eigenvector. You don't need to do a separate eigenvector computation.

For example, suppose

The characteristic polynomial is . The eigenvalues are .

Find an eigenvector for :

I knew that the second row must be a multiple of the
first row, because I know the system has nontrivial solutions. So I
don't have to work out *what* multiple it is; I can just zero
out the second row on general principles.

This *only* works for matrices, and only for
those which are 's in eigenvector computations.

Next, there's no point in going all the way to row reduced echelon form. I just need some nonzero vector such that

That is, I want

I can find an a and b that work by swapping and -1, and negating one of them. For example, take (-1 negated) and . This checks:

So is an eigenvector for .

By the discussion at the start of the example, I don't need to do a computation for . Just conjugate the previous eigenvector: must be an eigenvector for .

Since there are 2 independent eigenvectors, you can use them construct a diagonalizing matrix for A:

Notice that you get a diagonal matrix with the eigenvalues on the main diagonal, in the same order in which you listed the eigenvectors.

* Example.* For the following matrix, find the
eigenvalues over , and for each eigenvalue, a
complete set of independent eigenvectors.

Find a diagonalizing matrix and the corresponding diagonal matrix.

The characteristic polynomial is

Now

The eigenvalues are and .

For , I have

With variables a, b, and c, the corresponding homogeneous system is and . This gives the solution vector

Taking , I obtain the eigenvector .

For , I have

I multiplied the first row by , then divided it by 5. This made it the same as the third row.

I divided the second row by .

(I knew the the first and third rows had to be multiples, since they're clearly independent of the second row. Thus, if they weren't multiples, the three rows would be independent, the eigenvector matrix would be invertible, and there would be no eigenvectors [which must be nonzero].)

Now I can wipe out the third row by subtracting the first:

With variables a, b, and c, the corresponding homogeneous system is

There will only be one parameter (c), so there will only be one independent eigenvector. To get one, switch the "-5" and " " and negate the "-5" to get "5". This gives , , and . You can see that these values for a and c work:

Thus, my eigenvector is .

Hence, an eigenvector for is the conjugate .

A diagonalizing matrix is given by

With this diagonalizing matrix, I have

Copyright 2011 by Bruce Ikenaga