Section 4.1 Eigenvalues and Eigenvectors
We jump right into the definition, which you have probably seen previously in your first course in linear algebra.
Remark 4.1.2.
You might reasonably wonder: where does this definition come from? And why should I care? We are assuming that you saw at least a basic introduction to eigenvalues in your first course on linear algebra, but that course probably focused on mechanics. Possibly you learned that diagonalizing a matrix lets you compute powers of that matrix.
But why should we be interested in computing powers (in particular, large powers) of a matrix? An important context comes from the study of discrete linear dynamical systems 1 , as well as Markov chains 2 , where the evolution of a state is modelled by repeated multiplication of a vector by a matrix.
When we’re able to diagonalize our matrix using eigenvalues and eigenvectors, not only does it become easy to compute powers of a matrix, it also enables us to see that the entire process is just a linear combination of geometric sequences! If you have completed Worksheet 2.5, you probably will not be surprised to learn that the polynomial roots you found are, in fact, eigenvalues of a suitable matrix.
Remark 4.1.3.
Eigenvalues and eigenvectors can just as easily be defined for a general linear operator In this context, an eigenvector is sometimes referred to as a characteristic vector (or characteristic direction) for since the property simply states that the transformed vector is parallel to the original vector Some linear algebra textbooks that focus more on general linear transformations frame this topic in the context of invariant subspaces for a linear operator.
Exercise 4.1.4.
Hint.
Use Definition 4.1.1.
where denotes the identity matrix. Thus, if is an eigenvalue of any corresponding eigenvector is an element of
Definition 4.1.5.
Since we know that the null space of any matrix is a subspace, it follows that eigenspaces are subspaces of
Note that can be defined for any real number whether or not is an eigenvalue. However, the eigenvalues of are distinguished by the property that there is a nonzero solution to (4.1.1). Furthermore, we know that (4.1.1) can only have nontrivial solutions if the matrix is not invertible. We also know that is non-invertible if and only if This gives us the following theorem.
Theorem 4.1.6.
Strategy.
To prove a theorem involving a “the following are equivalent” statement, a good strategy is to show that the first implies the second, the second implies the third, and the third implies the first. The ideas needed for the proof are given in the paragraph preceding the theorem. See if you can turn them into a formal proof.
The polynomial is called the characteristic polynomial of (Note that We choose this order so that the coefficient of is always 1.) The equation
is called the characteristic equation of The solutions to this equation are precisely the eigenvalues of
Remark 4.1.7.
A careful study of eigenvalues and eigenvectors relies heavily on polynomials. An interesting fact is that we can plug any square matrix into a polynomial! Given the polynomial and an matrix we define
Note the use of the identity matrix in the first term, since it doesn’t make sense to add a scalar to a matrix.
One interesting aspect of this is the relationship between the eigenvalues of and the eigenvalues of For example, if has the eigenvalue see if you can prove that has the eigenvalue
Exercise 4.1.8.
Recall that a matrix is said to be similar to a matrix if there exists an invertible matrix such that Much of what follows concerns the question of whether or not a given matrix is diagonalizable.
Definition 4.1.9.
The following results will frequently be useful.
Theorem 4.1.10.
Proof.
The first two follow directly from properties of the determinant and trace. For the last, note that if then
so and therefore
Example 4.1.11.
Determine the eigenvalues and eigenvectors of
Solution.
We begin with the characteristic polynomial. We have
The roots of the characteristic polynomial are our eigenvalues, so we have and Note that the first eigenvalue comes from a repeated root. This is typically where things get interesting. If an eigenvalue does not come from a repeated root, then there will only be one (independent) eigenvector that corresponds to it. (That is, ) If an eigenvalue is repeated, it could have more than one eigenvector, but this is not guaranteed.
We find that which has reduced row-echelon form Solving for the nullspace, we find that there are two independent eigenvectors:
so
For the second eigenvector, we have which has reduced row-echelon form An eigenvector in this case is given by
In general, if the characteristic polynomial can be factored as
where is not divisible by then we say that is an eigenvalue of multiplicity In the example above, has multiplicty 2, and has multiplicty 1.
The
eigenvects
command in SymPy takes a square matrix as input, and outputs a list of lists (one list for each eigenvalue). For a given eigenvalue, the corresponding list has the form (eigenvalue, multiplicity, eigenvectors)
. Using SymPy to solve Example 4.1.11 looks as follows:xxxxxxxxxx
from sympy import Matrix, init_printing
init_printing()
A = Matrix([[0,1,1],[1,0,1],1,1,0])
A.eigenvects()
An important result about multiplicity is the following.
Theorem 4.1.12.
To prove Theorem 4.1.12 we need the following lemma, which we’ve borrowed from Section 5.5 of Nicholson’s textbook.
Lemma 4.1.13.
Proof.
We have
For we have
But is the th column of which proves the result.
We can use Lemma 4.1.13 to prove that as follows. Suppose is a basis for Then this is a linearly independent set of eigenvectors, so our lemma guarantees the existence of a matrix such that
This shows that is divisible by Since is the largest integer such that is divisible by we must have
Another important result is the following. The proof is a bit tricky: it requires mathematical induction, and a couple of clever observations.
Theorem 4.1.14.
Let be eigenvectors corresponding to distinct eigenvalues of a matrix Then is linearly independent.
Proof.
The proof is by induction on the number of distinct eigenvalues. Since eigenvectors are nonzero, any set consisting of a single eigenvector is independent. Suppose, then, that a set of eigenvectors corresponding to distinct eigenvalues is independent, and let be eigenvectors corresponding to distinct eigenvalues
Consider the equation
for scalars Multiplying both sides by the matrix we have
On the other hand, we can also multiply both sides by the eigenvalue giving
By hypothesis, the set of eigenvectors is linearly independent. We know that for since the eigenvalues are all distinct. Therefore, the only way this linear combination can equal zero is if This leaves us with but so as well.
Theorem 4.1.14 tells us that vectors from different eigenspaces are independent. In particular, a union of bases from each eigenspace will be an independent set. Therefore, Theorem 4.1.12 provides an initial criterion for diagonalization: if the dimension of each eigenspace is equal to the multiplicity of then is diagonalizable.
Our focus in the next section will be on diagonalization of symmetric matrices, and soon we will see that for such matrices, eigenvectors corresponding to different eigenvalues are not just independent, but orthogonal.
Exercises Exercises
1.
Find the characteristic polynomial of the matrix
2.
Find the three distinct real eigenvalues of the matrix
3.
The matrix has two real eigenvalues, one of multiplicity and one of multiplicity Find the eigenvalues and a basis for each eigenspace.
4.
5.
6.
The matrix
has two real eigenvalues Find these eigenvalues, their multiplicities, and the dimensions of their corresponding eigenspaces.
The smaller eigenvalue has multiplicity and the dimension of its corresponding eigenspace is .
The larger eigenvalue has multiplicity and the dimension of its corresponding eigenspace is .
7.
Supppose is an invertible matrix and is an eigenvector of with associated eigenvalue Convince yourself that is an eigenvector of the following matrices, and find the associated eigenvalues.
- The eigenvalue of the matrix
- The eigenvalue of the matrix
- The eigenvalue of the matrix
- The eigenvalue of the matrix
8.
9.
Recall that similarity of matrices is an equivalence relation; that is, the relation is reflexive, symmetric and transitive.
You have attempted 1 of 12 activities on this page.