Skip to main content

Section 4.1 Eigenvalues and Eigenvectors

We jump right into the definition, which you have probably seen previously in your first course in linear algebra.

Definition 4.1.1.

Let A be an n×n matrix. A number λ is called an eigenvalue of A if there exists a nonzero vector x such that
Ax=λx.
Any such vector x is called an eigenvector associated to the eigenvalue λ.

Remark 4.1.2.

You might reasonably wonder: where does this definition come from? And why should I care? We are assuming that you saw at least a basic introduction to eigenvalues in your first course on linear algebra, but that course probably focused on mechanics. Possibly you learned that diagonalizing a matrix lets you compute powers of that matrix.
But why should we be interested in computing powers (in particular, large powers) of a matrix? An important context comes from the study of discrete linear dynamical systems 1 , as well as Markov chains 2 , where the evolution of a state is modelled by repeated multiplication of a vector by a matrix.
When we’re able to diagonalize our matrix using eigenvalues and eigenvectors, not only does it become easy to compute powers of a matrix, it also enables us to see that the entire process is just a linear combination of geometric sequences! If you have completed Worksheet 2.5, you probably will not be surprised to learn that the polynomial roots you found are, in fact, eigenvalues of a suitable matrix.

Remark 4.1.3.

Eigenvalues and eigenvectors can just as easily be defined for a general linear operator T:VV. In this context, an eigenvector x is sometimes referred to as a characteristic vector (or characteristic direction) for T, since the property T(x)=λx simply states that the transformed vector T(x) is parallel to the original vector x. Some linear algebra textbooks that focus more on general linear transformations frame this topic in the context of invariant subspaces for a linear operator.
A subspace UV is invariant with respect to T if T(u)U for all uU. Note that if x is an eigenvector of T, then span{x} is an invariant subspace. To see this, note that if T(x)=λx and y=kx, then
T(y)=T(kx)=kT(x)=k(λx)=λ(kx)=λy.

Exercise 4.1.4.

Note that if x is an eigenvector of the matrix A, then we have
(4.1.1)(AλIn)x=0,
where In denotes the n×n identity matrix. Thus, if λ is an eigenvalue of A, any corresponding eigenvector is an element of null(AλIn).

Definition 4.1.5.

For any real number λ and n×n matrix A, we define the eigenspace Eλ(A) by
Eλ(A)=null(AλIn).
Since we know that the null space of any matrix is a subspace, it follows that eigenspaces are subspaces of Rn.
Note that Eλ(A) can be defined for any real number λ, whether or not λ is an eigenvalue. However, the eigenvalues of A are distinguished by the property that there is a nonzero solution to (4.1.1). Furthermore, we know that (4.1.1) can only have nontrivial solutions if the matrix AλIn is not invertible. We also know that AλIn is non-invertible if and only if det(AλIn)=0. This gives us the following theorem.

Strategy.

To prove a theorem involving a “the following are equivalent” statement, a good strategy is to show that the first implies the second, the second implies the third, and the third implies the first. The ideas needed for the proof are given in the paragraph preceding the theorem. See if you can turn them into a formal proof.
The polynomial cA(x)=det(xInA) is called the characteristic polynomial of A. (Note that det(xInA)=(1)ndet(AxIn). We choose this order so that the coefficient of xn is always 1.) The equation
(4.1.2)det(xInA)=0
is called the characteristic equation of A. The solutions to this equation are precisely the eigenvalues of A.

Remark 4.1.7.

A careful study of eigenvalues and eigenvectors relies heavily on polynomials. An interesting fact is that we can plug any square matrix into a polynomial! Given the polynomial p(x)=a0+a1x+a2x2++anxn and an n×n matrix A, we define
p(A)=a0In+a1A+a2A2++anAn.
Note the use of the identity matrix in the first term, since it doesn’t make sense to add a scalar to a matrix.
One interesting aspect of this is the relationship between the eigenvalues of A and the eigenvalues of p(A). For example, if A has the eigenvalue λ, see if you can prove that Ak has the eigenvalue λk.

Exercise 4.1.8.

Recall that a matrix B is said to be similar to a matrix A if there exists an invertible matrix P such that B=P1AP. Much of what follows concerns the question of whether or not a given n×n matrix A is diagonalizable.

Definition 4.1.9.

An n×n matrix A is said to be diagonalizable if A is similar to a diagonal matrix.
The following results will frequently be useful.

Proof.

The first two follow directly from properties of the determinant and trace. For the last, note that if B=P1AP, then
P1(xInA)P=P1(xIn)PP1AP=xInB,
so xInBxInA, and therefore det(xInB)=det(xInA).

Example 4.1.11.

Determine the eigenvalues and eigenvectors of A=[011101110].
Solution.
We begin with the characteristic polynomial. We have
det(xInA)=det[x111x111x]=x|x11x|+1|111x|1|1x11|=x(x21)+(x1)(1+x)x(x1)(x+1)2(x+1)(x+1)[x2x2](x+1)2(x2).
The roots of the characteristic polynomial are our eigenvalues, so we have λ1=1 and λ2=2. Note that the first eigenvalue comes from a repeated root. This is typically where things get interesting. If an eigenvalue does not come from a repeated root, then there will only be one (independent) eigenvector that corresponds to it. (That is, dimEλ(A)=1.) If an eigenvalue is repeated, it could have more than one eigenvector, but this is not guaranteed.
We find that A(1)In=[111111111], which has reduced row-echelon form [111000000]. Solving for the nullspace, we find that there are two independent eigenvectors:
x1,1=[110], and x1,2=[101],
so
E1(A)=span{[110],[101]}.
For the second eigenvector, we have A2I=[211121112], which has reduced row-echelon form [101011000]. An eigenvector in this case is given by
x2=[111].
In general, if the characteristic polynomial can be factored as
cA(x)=(xλ)mq(x),
where q(x) is not divisible by xλ, then we say that λ is an eigenvalue of multiplicity m. In the example above, λ1=1 has multiplicty 2, and λ2=2 has multiplicty 1.
The eigenvects command in SymPy takes a square matrix as input, and outputs a list of lists (one list for each eigenvalue). For a given eigenvalue, the corresponding list has the form (eigenvalue, multiplicity, eigenvectors). Using SymPy to solve Example 4.1.11 looks as follows:
An important result about multiplicity is the following.
To prove Theorem 4.1.12 we need the following lemma, which we’ve borrowed from Section 5.5 of Nicholson’s textbook.

Proof.

We have
P1AP=P1A[x1xn]=[(P1A)x1(P1A)xn].
For 1ik, we have
(P1A)(xi)=P1(Axi)=P1(λixi)=λi(P1xi).
But P1xi is the ith column of P1P=In, which proves the result.
We can use Lemma 4.1.13 to prove that dimEλ(A)m as follows. Suppose {x1,,xk} is a basis for Eλ(A). Then this is a linearly independent set of eigenvectors, so our lemma guarantees the existence of a matrix P such that
P1AP=[λIkB0A1].
Let A~=P1AP. On the one hand, since A~A, we have cA(x)=pA~(x). On the other hand,
det(xInA~)=det[(xλ)IkB0xInkA1]=(xλ)kdet(xInkA1).
This shows that cA(x) is divisible by (xλ)k. Since m is the largest integer such that cA(x) is divisible by (xλ)m, we must have dimEλ(A)=km.
Another important result is the following. The proof is a bit tricky: it requires mathematical induction, and a couple of clever observations.

Proof.

The proof is by induction on the number k of distinct eigenvalues. Since eigenvectors are nonzero, any set consisting of a single eigenvector v1 is independent. Suppose, then, that a set of eigenvectors corresponding to k1 distinct eigenvalues is independent, and let v1,,vk be eigenvectors corresponding to distinct eigenvalues λ1,,λk.
Consider the equation
c1v1+c2v2++ckvk=0,
for scalars c1,,ck. Multiplying both sides by the matrix A, we have
(4.1.3)A(c1v1+c2v2++ckvk)=A0(4.1.4)c1Av1+c2Av2++ckAvk=0(4.1.5)c1λ1v1+c2λ2v2++ckλkvk=0.
On the other hand, we can also multiply both sides by the eigenvalue λ1, giving
(4.1.6)0=c1λ1v1+c2λ1v2++ckλ1vk.
Subtracting (4.1.6) from (4.1.5), the first temrs cancel, and we get
c2(λ2λ1)v2++ck(λkλ1)vk=0.
By hypothesis, the set {v2,,vk} of k1 eigenvectors is linearly independent. We know that λjλ10 for j=2,,k, since the eigenvalues are all distinct. Therefore, the only way this linear combination can equal zero is if c2=0,,ck=0. This leaves us with c1v1=0, but z10, so c1=0 as well.
Theorem 4.1.14 tells us that vectors from different eigenspaces are independent. In particular, a union of bases from each eigenspace will be an independent set. Therefore, Theorem 4.1.12 provides an initial criterion for diagonalization: if the dimension of each eigenspace Eλ(A) is equal to the multiplicity of λ, then A is diagonalizable.
Our focus in the next section will be on diagonalization of symmetric matrices, and soon we will see that for such matrices, eigenvectors corresponding to different eigenvalues are not just independent, but orthogonal.

Exercises Exercises

1.

Find the characteristic polynomial of the matrix A=[120044310].

2.

Find the three distinct real eigenvalues of the matrix B=[147048007].

3.

The matrix A=[84124812448] has two real eigenvalues, one of multiplicity 1 and one of multiplicity 2. Find the eigenvalues and a basis for each eigenspace.

4.

The matrix A=[52142215211411174] has two distinct real eigenvalues λ1<λ2. Find the eigenvalues and a basis for each eigenspace.

5.

The matrix
A=[210941k00]
has three distinct real eigenvalues if and only if
<k< .

6.

The matrix
A=[4484224222420000]
has two real eigenvalues λ1<λ2. Find these eigenvalues, their multiplicities, and the dimensions of their corresponding eigenspaces.
The smaller eigenvalue λ1= has multiplicity and the dimension of its corresponding eigenspace is .
The larger eigenvalue λ2= has multiplicity and the dimension of its corresponding eigenspace is .

7.

Supppose A is an invertible n×n matrix and v is an eigenvector of A with associated eigenvalue 3. Convince yourself that v is an eigenvector of the following matrices, and find the associated eigenvalues.
  1. The eigenvalue of the matrix A8.
  2. The eigenvalue of the matrix A1.
  3. The eigenvalue of the matrix A3In.
  4. The eigenvalue of the matrix 3A.

8.

Let
v1=[031],v2=[330],v3=[101]
be eigenvectors of the matrix A which correspond to the eigenvalues λ1=1, λ2=0, and λ3=4, respectively, and let
x=[233].
Express x as a linear combination x=av1+bv2+cv3, and find Ax.

9.

Recall that similarity of matrices is an equivalence relation; that is, the relation is reflexive, symmetric and transitive.
Verify that A=[0111] is similar to itself by finding a T such that A=T1AT.
We know that A and B=[1112] are similar since A=P1BP where P=[1123].
Verify that BA by finding an S such that B=S1AS.
We also know that B and C=[3512] are similar since B=Q1CQ where Q=[1110].
Verify that AC by finding an R such that A=R1CR.
You have attempted 1 of 12 activities on this page.