Skip to main content

Section 1.8 New subspaces from old

Let V be a finite-dimensional vector space, and let U,W be subspaces of V. In what ways can we combine U and W to obtain new subspaces?
At first, we might try set operations: union, intersection, and difference. The set difference we can rule out right away: since U and W must both contain the zero vector, UW cannot.
What about the union, UW? Before trying to understand this in general, let’s try a concrete example: take V=R2, and let U={(x,0)|,xR} (the x axis, essentially), and W={(0,y)|yR} (the y axis). Is their union a subspace?

Exercise 1.8.1.

    The union of the “x axis” and “y axis” in R2 is a subspace of R2.
  • True.

  • Any subspace has to be closed under addition. If we add the vector (1,0) (which lies along the x axis) to the vector (0,1) (which lies along the y axis), we get the vector (1,1), which does not lie along either axis.
  • False.

  • Any subspace has to be closed under addition. If we add the vector (1,0) (which lies along the x axis) to the vector (0,1) (which lies along the y axis), we get the vector (1,1), which does not lie along either axis.
With a motivating example under our belts, we can try to tackle the general result. (Note that this result remains true even if V is infinite-dimensional!)

Strategy.

We have an “if and only” if statement, which means we have to prove two directions:
  1. If UW or WU, then UW is a subspace.
  2. If UW is a subspace, then UW or WU.
The first direction is the easy one: if UW, what can you say about UW?
For the other direction, it’s not clear how to get started with our hypothesis. When a direct proof seems difficult, remember that we can also try proving the contrapositive: If UW and WU, then UW is not a subspace.
Now we have more to work with: negation turns the “or” into an “and”, and proving that something is not a subspace is easier: we just have to show that one part of the subspace test fails. As our motivating example suggests, we should expect closure under addition to be the condition that fails.
To get started, we need to answer one more question: if U is not a subset of W, what does that tell us?
An important point to keep in mind with this proof: closure under addition means that if a subspace contains u and w, then it must contain u+w. But if a subspace contains u+w, that does not mean it has to contain u and w. As an example, consider the subspace {(x,x)|xR} of R2. It contains the vector (1,1)=(1,0)+(0,1), but it does not contain (1,0) or (0,1).

Proof.

Suppose UW or WU. In the first case, UW=W, and in the second case, UW=U. Since both U and W are subspaces, UW is a subspace.
Now, suppose that UW, and WU. Since UW, there must be some element uU such that uW. Since WU, there must be some element wW such that wU. We know that u,wUW, so we consider the sum, u+w.
If u+wUW, then u+wU, or u+wW. Suppose u+wU. Since uU and U is a subspace, uU. Since u,u+wU and U is a subspace,
u+(u+w)=(u+u)+w=0+w=wU.
But we assumed that wU, so it must be that u+wU.
By a similar argument, if u+wW, we can conclude that uW, contradicting the assumption that uW. So u+w does not belong to U or W, so it cannot belong to UW. Since UW is not closed under addition, it is not a subspace.
This leaves us with intersection. Will it fail as well? Fortunately, the answer is no: this operation actually gives us a subspace.

Strategy.

The key here is that the intersection contains only those vectors that belong to both subspaces. So any operation (addition, scalar multiplication) that we do in UW can be viewed as taking place in either U or W, and we know that these are subspaces. After this observation, the rest is the Subspace Test.

Proof.

Let U and W be subspaces of V. Since 0U and 0W, we have 0UW. Now, suppose x,yUW. Then x,yU, and x,yW. Since x,yU and U is a subspace, x+yU. Similarly, x+yW, so x+yUW. If c is any scalar, then cx is in both U and W, since both sets are subspaces, and therefore, cxUW. By the Subspace Test, UW is a subspace.
The intersection of two subspaces gives us a subspace, but it is a smaller subspace, contained in the two subspaces we’re intersecting. Given subspaces U and W, is there a way to construct a larger subspace that contains them? We know that UW doesn’t work, because it isn’t closed under addition. But what if we started with UW, and threw in all the missing sums? This leads to a definition:

Definition 1.8.4.

Let U and W be subspaces of a vector space V. We define the sum U+W of these subspaces by
U+W={u+w|uU and wW}.
It turns out that this works! Not only is U+W a subspace of V, it is the smallest subspace containing both U and W.

Strategy.

The key to working with U+W is to understand how to work with the definition. If we say that xU+W, then we are saying there exist vectors uU and wW such that u+w=x.
We prove that U+W is a subspace using this observation and the subspace test.
To prove the second part, we assume that UX and WX. We then choose an element xU+W, and using the idea above, show that xX.

Proof.

Let U,W be subspaces. Since 0=0+0, with 0U and 0W, we see that 0U+W.
Suppose that x,yU+W. Then there exist u1,u2U, and w1,w2W, with u1+w1=x, and u2+w2=y. Then
x+y=(u1+w1)+(u2+w2)=(u1+u2)+(w1+w2),
and we know that u1+u2U, and w1+w2W, since U and W are subspaces. Since x+y can be written as the sum of an element of U and an element of W, we have x+yU+W.
If c is any scalar, then
cx=c(u1+w1)=cu1+cw1U+W,
since cu1U and cw1W.
Since U+W contains 0, and is closed under both addition and scalar multiplication, it is a subspace.
Now, suppose X is a subspace of V such that UX and WX. Let xU+W. Then x=u+w for some uU and wW. Since uU and UX, uX. Similarly, wX. Since X is a subspace, it is closed under addition, so u+w=xX. Therefore, U+WX.
By choosing bases for two subspaces U and W of a finite-dimensional vector space, we can obtain the following cool dimension-counting result:

Strategy.

This is a proof that would be difficult (if not impossible) without using a basis. Your first thought might be to choose bases for the subspaces U and W, but this runs into trouble: some of the basis vectors for U might be in W, and vice-versa.
Of course, those vectors will be in UW, but it gets hard to keep track: without more information (and we have none, since we want to be completely general), how do we tell which basis vectors are in the intersection, and how many?
Instead, we start with a basis for UW. This is useful, because UW is a subspace of both U and W. So any basis for UW can be extended to a basis of U, and it can also be extended to a basis of W.
The rest of the proof relies on making sure that neither of these extensions have any vectors in common, and that putting everything together gives a basis for U+W. (This amounts to going back to the definition of a basis: we need to show that it’s linearly independent, and that it spans U+W.)

Proof.

Let B1={x1,,xk} be a basis for UW. Extend B1 to a basis B2={x1,,xk,u1,,um} of U, and to a basis B3={x1,,xk,w1,,wn} of W. Note that we have dim(UW)=k, dimU=k+m, and dimW=k+n.
Now, consider the set B={x1,,xk,u1,,um,w1,,wn}. We claim that B is a basis for U+W. We know that B2 is linearly independent, since it’s a basis for U, and that B=B2{w1,,wn}. It remains to show that none of the wi are in the span of B2; if so, then B is independent by Lemma 1.7.11.
Since spanB2=U, it suffices to show that none of the wi belong to U. But we know that wiW, so if wiU, then wiUW. But if wiUW, then wispanB1, which would imply that B3 is linearly dependent, and since B3 is a basis, this is impossible.
Next, we need to show that spanB=U+W. Let vU+W; then v=u+w for some uU and wW. Since uU, there exist scalars a1,,ak,b1,,bm such that
u=a1x1++akxk+b1u1++bmum,
and since wW, there exist scalars c1,,ck,d1,,dn such that
w=c1x1++ckxk+d1w1++dnwn.
Thus,
v=u+w=(a1+c1)x1++(ak+ck)xk+b1u1++bmum+d1w1++dnwn,
which shows that vspanB.
Finally, we check that this gives the dimension as claimed. We have
dimU+dimWdim(UW)=(k+m)+(k+n)k=k+m+n=dim(U+W),
since there are k vectors in B1, k+m vectors in B2, k+n vectors in B3, and k+m+n vectors in B.
Notice how a vector vU+W can be written as a sum of a vector in U and a vector W, but not uniquely, in general: in the above proof, we can change the values of the coefficients ai and ci, as long as the sum ai+ci remains unchanged. Note that these are the coefficients of the basis vectors for UW, so we can avoid this ambiguity if U and W have no nonzero vectors in common.

Exercise 1.8.7.

Let V=R3, and let U={(x,y,0)|x,y,R},W={(0,y,z)|y,zR} be two subspaces.

(a)

Determine the intersection UW.

(b)

Write the vector v=(1,1,1) in the form v=u+w, where uU and wW, in at least two different ways.

Definition 1.8.8.

Let U and W be subspaces of a vector space V. If UW={0}, we say that the sum U+W is a direct sum, which we denote by UW.
If the sum is direct, then we have simply dim(UW)=dimU+dimW. The other reason why direct sums are preferable, is that any vUW can be written uniquely as v=u+w where uU and wW, since we no longer have the ambiguity resulting from the basis vectors in UW.

Proof.

Suppose that UW={0}, and suppose that we have v=u1+w1=u2+w2, for u1,u2U,w1,w2W. Then 0=(u1u2)+(w1w2), which implies that
w1w2=(u1u2).
Now, u=u1u2U, since U is a subspace, and similarly, w=w1w2W. But we also have w=u, which implies that wU. (Since u is in U, and this is the same vector as w.) Therefore, wUW, which implies that w=0, so w1=w2. But we must also then have u=0, so u1=u2.
Conversely, suppose that every vU+W can be written uniquely as v=u+w, with uU and wW. Suppose that aUW. Then aU and aW, so we also have aW, since W is a subspace. But then 0=a+(a), where aU and aW. On the other hand, 0=0+0, and 0 belongs to both U and W. It follows that a=0. Since a was arbitrary, UW={0}.
We end with one last application of the theory we’ve developed on the existence of a basis for a finite-dimensional vector space. As we continue on to later topics, we’ll find that it is often useful to be able to decompose a vector space into a direct sum of subspaces. Using bases, we can show that this is always possible.

Proof.

Let {u1,,um} be a basis of U. Since UV, the set {u1,,um} is a linearly independent subset of V. Since any linearly independent set can be extended to a basis of V, there exist vectors w1,,wn such that
{u1,,um,w1,,wn}
is a basis of V.
Now, let W=span{w1,,wn}. Then W is a subspace, and {w1,,wn} is a basis for W. (It spans, and must be independent since it’s a subset of an independent set.)
Clearly, U+W=V, since U+W contains the basis for V we’ve constructed. To show the sum is direct, it suffices to show that UW={0}. To that end, suppose that vUW. Since vU, we have
v=a1u1++amum
for scalars a1,,am. Since vW, we can write
v=b1w1++bnwn
for scalars b1,,bn. But then
0=vv=a1u1+amumb1w1bnwn.
Since {u1,,um,w1,,wn} is a basis for V, it’s independent, and therefore, all of the ai,bj must be zero, and therefore, v=0.
The subspace W constructed in the theorem above is called a complement of U. It is not unique; indeed, it depends on the choice of basis vectors. For example, if U is a one-dimensional subspace of R2; that is, a line, then any other non-parallel line through the origin provides a complement of U. Later we will see that an especially useful choice of complement is the orthogonal complement.

Definition 1.8.11.

Let U be a subspace of a vector space V. We say that a subspace W of V is a complement of U if UW=V.

Exercises Exercises

1.

Let U be the subspace of P3(R) consisting of all polynomials p(x) with p(1)=0.
(a)
Determine a basis for U.
Hint.
Use the factor theorem.
(b)
Find a complement of U.
Hint.
What is the dimension of U? (So what must be the dimension of its complement?) What condition ensures that a polynomial does not belong to U?

2.

Let U be the subspace of R5 define by
U={(x1,x2,x3,x4,x5)|x1=3x3, and 3x25x4=x5}.
(a)
Determine a basis for U.
Hint.
Try plugging in the given conditions, and then decomposing the vector into pieces with one variable each.
(b)
Find a complement of U.
Hint.
One way to solve this is to ask yourself, what vectors are not in the span of the basis you found above? You can do this by solving an appropriate system of equations.

3.

    Suppose U and W are 4-dimensional subspaces of R6. What are all possible dimensions of UW?
  • 1
  • What would Theorem 1.8.6 say about dimU+W in this case? Why is that not possible?
  • 2
  • Good job! If dimU+W=6 (the largest it possibly can), then dimUW=dimU+dimWdim(U+W)=4+46=2.
  • 3
  • Yes! This will be the case if dimU+W=5.
  • 4
  • Correct! If UW, then U=W=U+W=UW, all with dimension 4.
  • 5
  • Since U+W contains both U and W, its dimension cannot be less than 4.
Hint.

4.

Let U=span{2x21,2x4x2} and W=span{4x32x2,24x28x4} be subspaces of the vector space V=P3(R).
(a)
Is {2x21,2x4x2,4x32x2,24x28x4} a basis for V?
(b)
What is the dimension of U+W?
(c)
What is the dimension of UW?
You have attempted 1 of 10 activities on this page.