Section 5.4 Direct Sums and Invariant Subspaces
This section continues the discussion of direct sums (from Section 1.8) and invariant subspaces (from Section 4.1), to better understand the structure of linear operators.
Subsection 5.4.1 Invariant subspaces
Exercise 5.4.2.
Hint.
In each case, choose an element of the subspace. What does the definition of the space tell you about that element? (For example, if what is the value of ) Then show that also fits the defintion of that space.
A subspace is -invariant if does not map any vectors in outside of Notice that if we shrink the domain of to then we get an operator from to since the image is contained in
Definition 5.4.3.
Let be a linear operator, and let be a -invariant subspace. The restriction of to denoted is the operator defined by for all
Exercise 5.4.4.
True.
- The definition of a function includes its domain and codomain. Since the domain of
is different from that of they are not the same function. False.
- The definition of a function includes its domain and codomain. Since the domain of
is different from that of they are not the same function.
A lot can be learned by studying the restrictions of an operator to invariant subspaces. Indeed, the textbook by Axler does almost everything from this point of view. One reason to study invariant subspaces is that they allow us to put the matrix of into simpler forms.
Theorem 5.4.5.
Subsection 5.4.2 Eigenspaces
An important source of invariant subspaces is eigenspaces. Recall that for any real number and any operator we define
For most values of we’ll have The values of for which is non-trivial are precisely the eigenvalues of Note that since similar matrices have the same characteristic polynomial, any matrix representation will have the same eigenvalues. They do not generally have the same eigenspaces, but we do have the following.
Theorem 5.4.6.
In other words, the two eigenspaces are isomorphic, although the isomorphism depends on a choice of basis.
Subsection 5.4.3 Direct Sums
are subspaces of Saying that means that can be written as a sum of a vector in and a vector in However, this sum may not be unique. If and then we can write giving two different representations of a vector as an element of
We proved in Theorem 1.8.9 in Section 1.8 that for any there exist unique vectors and such that if and only if
Typically we are interested in the case that the two subspaces sum to Recall from Definition 1.8.11 that if we say that is a complement of We also say that is a direct sum decomposition of Of course, the orthogonal complement of a subspace is a complement in this sense, if is equipped with an inner product. (Without an inner product we have no concept of “orthogonal”.) But even if we don’t have an inner product, finding a complement is not too difficult, as the next example shows.
Example 5.4.7. Finding a complement by extending a basis.
The easiest way to determine a direct sum decomposition (or equivalently, a complement) is through the use of a basis. Suppose is a subspace of with basis and extend this to a basis
which gives
is a basis for Indeed, spans since every element of can be written as with Independence follows by reversing the argument above: if
then and equality is only possible if both sides belong to Since is independent, the have to be zero, and since is independent, the have to be zero.
The argument given in the second part of Example 5.4.7 has an immediate, but important consequence.
Theorem 5.4.8.
Example 5.4.9.
Suppose where and are -invariant subspaces for some operator Let and let be bases for and respectively. Determine the matrix of with respect to the basis of
Solution.
Since we don’t know the map or anything about the bases we’re looking for a fairly general statement here. Since is -invariant, we must have for each Similarly, for each This means that we have
for some scalars If we set and then we have
Moreover, we can also see that and
You have attempted 1 of 3 activities on this page.