Skip to main content

Appendix C Solutions to Selected Exercises

1 Vector spaces
1.2 Properties

Exercise 1.2.2.

1.2.2.a
Solution.
Suppose u+v=u+w. By adding u on the left of each side, we obtain:
u+(u+v)=u+(u+w)(u+u)+v=(u+u)+w by A30+v=0+w by A5v=w by A4,
which is what we needed to show.
1.2.2.b
Solution.
We have c0=c(0+0)=c0+c0, by A4 and S2, respectively. Adding c0 to both sides gives us
c0+c0=c0+(c0+c0).
Using associativity (A3), this becomes
c0+c0=(c0+c0)+c0,
and since c0+c0=0 by A5, we get 0=0+c0. Finally, we apply A4 on the right hand side to get 0=c0, as required.
1.2.2.c
Solution.
Suppose there are two vectors 01,02 that act as additive identities. Then
01=01+02 since v+02=v for any v=02+01 by axiom A202 since v+01=v for any v
So any two vectors satisfying the property in A4 must, in fact, be the same.
1.2.2.d
Solution.
Let vV, and suppose there are vectors w1,w2V such that v+w1=0 and v+w2=0. Then
w1=w1+0 by A4=w1+(v+w2) by assumption=(w1+v)+w2 by A3=(v+w1)+w2 by A2=0+w2 by assumptionw2 by A4.

1.6 Linear Independence

Exercise 1.6.7.

Solution.
We set up a matrix and reduce:
Notice that this time we don’t get a unique solution, so we can conclude that these vectors are not independent. Furthermore, you can probably deduce from the above that we have 2v1+3v2v3=0. Now suppose that wspan{v1,v2,v3}. In how many ways can we write w as a linear combination of these vectors?

Exercise 1.6.8.

Solution.
In each case, we set up the defining equation for independence, collect terms, and then analyze the resulting system of equations. (If you work with polynomials often enough, you can probably jump straight to the matrix. For now, let’s work out the details.)
Suppose
r(x2+1)+s(x+1)+tx=0.
Then rx2+(s+t)x+(r+s)=0=0x2+0x+0, so
r=0s+t=0r+s=0.
And in this case, we don’t even need to ask the computer. The first equation gives r=0 right away, and putting that into the third equation gives s=0, and the second equation then gives t=0.
Since r=s=t=0 is the only solution, the set is independent.
Repeating for S2 leads to the equation
(r+2s+t)x2+(r+s+5t)x+(3r+5s+t)1=0.
This gives us:

Exercise 1.6.9.

Solution.
We set a linear combination equal to the zero vector, and combine:
a[1001]+b[1111]+c[1111]+d[0110]=[0000][a+b+cb+cdb+cda+b+c]=[0000].
We could proceed, but we might instead notice right away that equations 1 and 4 are identical, and so are equations 2 and 3. With only two distinct equations and 4 unknowns, we’re certain to find nontrivial solutions.

1.7 Basis and dimension

Exercise 1.7.5.

Solution.
Let X=[abcd]. Then AX=[a+cb+d00], and XA=[aacc], so the condition AX=XA requires:
a+c=ab+d=a0=c0=c.
So c=0, in which case the first equation a=a is trivial, and we are left with the single equation a=b+d. Thus, our matrix X must be of the form
X=[b+db0d]=b[1100]+d[1001].
Since the matrices [1100] and [1001] are not scalar multiples of each other, they must be independent, and therefore, they form a basis for U. (Why do we know these matrices span U?)

Exercise 1.7.7.

1.7.7.a
Solution.
We need to show that the set is independent, and that it spans.
The set is independent if the equation
x(1,1,0)+y(1,0,1)+z(0,1,1)=(0,0,0)
has x=y=z=0 as its only solution. This equation is equivalent to the system
x+y=0x+z=0y+z=0.
We know that the solution to this system is unique if the coefficient matrix A=[110101011] is invertible. Note that the columns of this matrix are vectors in our set.
We can determine invertibility either by showing that the RREF of A is the identity, or by showing that the determinant of A is nonzero. Either way, this is most easily done by the computer:
Our set of vectors is therefore linearly independent. Now, to show that it spans, we need to show that for any vector (a,b,c), the equation
x(1,1,0)+y(1,0,1)+z(0,1,1)=(a,b,c)
has a solution. But we know that this system has the same coefficient matrix as the one above, and that existence of a solution again follows from invertibility of A, which we have already established.
Note that for three vectors in R3, once independence has been confirmed, span is automatic. We will soon see that this is not a coincidence.
1.7.7.b
Solution.
Based on what we learned from the first set, determining whether or not this set is a basis is equivalent to determining whether or not the matrix A whose columns consist of the vectors in the set is invertible. We form the matrix
A=[111111111]
and then check invertibility using the computer.
Since the determinant is nonzero, our set is a basis.

Exercise 1.7.10.

1.7.10.a
Solution.
By definition, U1=span{1+x,x+x2}, and these vectors are independent, since neither is a scalar multiple of the other. Since there are two vectors in this basis, dimU1=2.
1.7.10.b
Solution.
If p(1)=0, then p(x)=(x1)q(x) for some polynomial q. Since U2 is a subspace of P2, the degree of q is at most 2. Therefore, q(x)=ax+b for some a,bR, and
p(x)=(x1)(ax+b)=a(x2x)+b(x1).
Since p was arbitrary, this shows that U2=span{x2x,x1}.
The set {x2x,x1} is also independent, since neither vector is a scalar multiple of the other. Therefore, this set is a basis, and dimU2=2.
1.7.10.c
Solution.
If p(x)=p(x), then p(x) is an even polynomial, and therefore p(x)=a+bx2 for a,bR. (If you didn’t know this it’s easily verified: if
a+bx+cx2=a+b(x)+c(x)2,
we can immediately cancel a from each side, and since (x)2=x2, we can cancel cx2 as well. This leaves bx=bx, or 2bx=0, which implies that b=0.)
It follows that the set {1,x2} spans U3, and since this is a subset of the standard basis {1,x,x2} of P2, it must be independent, and is therefore a basis of U3, letting us conclude that dimU3=2.

Exercise 1.7.14.

Solution.
By the previous theorem, we can form a basis by adding vectors from the standard basis
{[1000],[0100],[0010],[0001]}.
It’s easy to check that [1000] is not in the span of {v,w}. To get a basis, we need one more vector. Observe that all three of our vectors so far have a zero in the (2,1)-entry. Thus, [0010] cannot be in the span of the first three vectors, and adding it gives us our basis.

Exercise 1.7.15.

Solution.
Again, we only need to add one vector from the standard basis {1,x,x2,x3}, and it’s not too hard to check that any of them will do.

1.8 New subspaces from old

Exercise 1.8.7.

1.8.7.a
Solution.
If (x,y,z)U, then z=0, and if (x,y,z)W, then x=0. Therefore, (x,y,z)UW if and only if x=z=0, so UW={(0,y,0)|yR}.
1.8.7.b
Solution.
There are in fact infinitely many ways to do this. Three possible ways include:
v=(1,1,0)+(0,0,1)=(1,0,0)+(0,1,1)=(1,12,0)+(0,12,1).

2 Linear Transformations
2.1 Definition and examples

Exercise 2.1.11.

Solution.
We need to find scalars a,b,c such that
2x+3x2=a(x+2)+b(1)+c(x2+x).
We could set up a system and solve, but this time it’s easy enough to just work our way through. We must have c=3, to get the correct coefficient for x2. This gives
2x+3x2=a(x+2)+b(1)+3x2+3x.
Now, we have to have 3x+ax=x, so a=4. Putting this in, we get
2x+3x2=4x8+b+3x2+3x.
Simiplifying this leaves us with b=10. Finally, we find:
T(2x+3x2)=T(4(x+2)+10(1)+3(x2+x))=4T(x+2)+10T(1)+3T(x2+x)=4(1)+10(5)+3(0)=46.

2.2 Kernel and Image

Exercise 2.2.10.

2.2.10.a
Solution.
We have T(0)=0 since 0T=0. Using proerties of the transpose and matrix algebra, we have
T(A+B)=(A+B)(A+B)T=(AAT)+(BBT)=T(A)+T(B)
and
T(kA)=(kA)(kA)T=kAkAT=k(AAT)=kT(A).
2.2.10.b
Solution.
It’s clear that if AT=A, then T(A)=0. On the other hand, if T(A)=0, then AAT=0, so A=AT. Thus, the kernel consists of all symmetric matrices.
2.2.10.c
Solution.
If B=T(A)=AAT, then
BT=(AAT)T=ATA=B,
so certainly every matrix in imA is skew-symmetric. On the other hand, if B is skew-symmetric, then B=T(12B), since
T(12B)=12T(B)=12(BBT)=12(B(B))=B.

Exercise 2.2.17.

2.2.17.a
Solution.
Suppose T:VW is injective. Then kerT={0}, so
dimV=0+dimimTdimW,
since imT is a subspace of W.
Conversely, suppose dimVdimW. Choose a basis {v1,,vm} of V, and a basis {w1,,wn} of W, where mn. By Theorem 2.1.8, there exists a linear transformation T:VW with T(vi)=wi for i=1,,m. (The main point here is that we run out of basis vectors for V before we run out of basis vectors for W.) This map is injective: if T(v)=0, write v=c1v1++cmvm. Then
0=T(v)=T(c1v1++cmvm)=c1T(v1)++cmT(vm)=c1w1++cmwm.
Since {w1,,wm} is a subset of a basis, it’s independent. Therefore, the scalars ci must all be zero, and therefore v=0.
2.2.17.b
Solution.
Suppose T:VW is surjective. Then dimimT=dimW, so
dimV=dimkerT+dimWdimW.
Conversely, suppose dimVdimW. Again, choose a basis {v1,,vm} of V, and a basis {w1,,wn} of W, where this time, mn. We can define a linear transformation as follows:
T(v1)=w1,,T(vn)=wn, and T(vj)=0 for j>n.
It’s easy to check that this map is a surjection: given wW, we can write it in terms of our basis as w=c1w1++cnwn. Using these same scalars, we can define v=c1v1++cnvnV such that T(v)=w.
Note that it’s not important how we define T(vj) when j>n. The point is that this time, we run out of basis vectors for W before we run out of basis vectors for V. Once each vector in the basis of W is in the image of T, we’re guaranteed that T is surjective, and we can define the value of T on any remaining basis vectors however we want.

2.3 Isomorphisms, composition, and inverses
2.3.2 Composition and inverses

Exercise 2.3.7.

Solution.
Suppose we have linear maps UTVSW, and let u1,u2U. Then
ST(u1+u2)=S(T(u1+u2))=S(T(u1)+T(u2))=S(T(u1))+S(T(u2))=ST(u1)+ST(u2),
and for any scalar c,
ST(cu1)=S(T(cu1))=S(cT(u1))=cS(T(u1))=c(ST(u1)).

Exercise 2.3.9.

Solution.
Let w1,w2W. Then there exist v1,v2V with w1=T(v1),w2=T(v2). We then have
T1(w1+w2)=T1(T(v1)+T(v2))=T1(T(v1+v2))=v1+v2=T1(w1)+T1(w2).
For any scalar c, we similarly have
T1(cw1)=T1(cT(v1))=T1(T(cv1))=cv1=cT1(w1).

3 Orthogonality and Applications
3.1 Orthogonal sets of vectors
3.1.1 Basic definitions and properties

Exercise 3.1.5.

Solution.
This is simply an exercise in properties of the dot product. We have
x+y2=(x+y)(x+y)=xx+xy+yx+yy=x2+2xy+y2.

Exercise 3.1.6.

Solution.
If x=0, then the result follows immediately from the dot product formula in Definition 3.1.1. Conversely, suppose xvi=0 for each i. Since the vi span Rn, there must exist scalars c1,c2,,ck such that x=c1v1+c2v2++ckvk. But then
xx=x(c1v1+c2v2++ckvk)=c1(xv1)+c2(xv2)++ck(xvk)=c1(0)+c2(0)++ck(0)=0.

3.1.2 Orthogonal sets of vectors

Exercise 3.1.13.

Solution.
Clearly, all three vectors are nonzero. To confirm the set is orthogonal, we simply compute dot products:
(1,0,1,0)(1,0,1,1)=1+0+1+0=0(1,0,1,1)(1,1,1,2)=1+01+2=0(1,0,1,0)(1,1,1,2)=1+01+0=0.
To find a fourth vector, we proceed as follows. Let x=(a,b,c,d). We want x to be orthogonal to the three vectors in our set. Computing dot products, we must have:
(a,b,c,d)(1,0,1,0)=a+c=0(a,b,c,d)(1,0,1,1)=a+c+d=0(a,b,c,d)(1,1,1,2)=a+bc+2d=0.
This is simply a homogeneous system of three equations in four variables. Using the Sage cell below, we find that our vector must satisfy a=12d,b=3d,c=12d.
One possible nonzero solution is to take d=2, giving x=(1,6,1,2). We’ll leave the verification that this vector works as an exercise.

Exercise 3.1.18.

Solution.
We compute
(vx1x12)x1+(vx2x22)x2+(vx3x32)x3=42x1+93x2+287x3=2(1,0,1,0)3(1,0,1,1)4(1,1,1,2)=(1,4,3,11)=v,
so vspan{x1,x2,x3}.
On the other hand, repeating the same calculation with w, we find
(vx1x12)x1+(vx2x22)x2+(vx3x32)x3=12(1,0,1,0)53(1,0,1,1)+47(1,1,1,2)=(7342,47,11542,1121)w,
so wspan{x1,x2,x3}.
Soon, we’ll see that the quantity we computed when showing that wspan{x1,x2,x3} is, in fact, the orthogonal projection of w onto the subspace span{x1,x2,x3}.

4 Diagonalization
4.2 Diagonalization of symmetric matrices

Exercise 4.2.1.

Solution.
Take x=ei and y=ej, where {e1,,en} is the standard basis for Rn. Then with A=[aij] we have
aij=ei(Aej)=(Aei)ej=aji,
which shows that AT=A.

4.4 Diagonalization of complex matrices
4.4.2 Complex matrices

Exercise 4.4.8.

Solution.
We have A¯=[41+i23i1i57i2+3i7i4], so
AH=(A¯)T=[41i2+3i1+i57i23i7i4]=A,
and
BBH=14[1+i21i2i][1i1+i22i]=14[(1+i)(1i)+2(1+i)(1+i)2i(1i)(1i)+2i(1i)(1+i)+2]=14[4004]=[1001],
so that BH=B1.

Exercise 4.4.12.

Solution.
Confirming that AH=A is almost immediate. We will use the computer below to compute the eigenvalues and eigenvectors of A, but it’s useful to attempt this at least once by hand. We have
det(zIA)=det[z43+i3iz1](z4)(z1)(3i)(3+i)z25z+410(z+1)(z6),
so the eigenvalues are λ1=1 and λ2=6, which are both real, as expected.
Finding eigenvectors can seem trickier than with real numbers, mostly because it is no longer immediately apparent when one row or a matrix is a multiple of another. But we know that the rows of AλI must be parallel for a 2×2 matrix, which lets proceed nonetheless.
For λ1=1, we have
A+I=[53i3+i2].
There are two ways one can proceed from here. We could use row operations to get to the reduced row-echelon form of A. If we take this approach, we multiply row 1 by 15, and then take 3i times the new row 1 and add it to row 2, to create a zero, and so on.
Easier is to realize that if we haven’t made a mistake calculating our eigenvalues, then the above matrix can’t be invertible, so there must be some nonzero vector in the kernel. If (A+I)[ab]=[00], then we must have
5a+(3i)b=0,
when we multiply by the first row of A. This suggests that we take a=3i and b=5, to get z=[3i5] as our first eigenvector. To make sure we’ve done things correctly, we multiply by the second row of A+I:
(3+i)(3i)+2(5)=1010=0.
Success! Now we move onto the second eigenvalue.
For λ2=6, we get
A6I=[23i3+i5].
If we attempt to read off the answer like last time, the first row of A6I suggests the vector w=[3i2]. Checking the second row to confirm, we find:
(3+i)(3i)5(2)=1010=0,
as before.
Finally, we note that
z,w=(3i)(3i)+(5)(2)=(3i)(3+i)10=0,
so the two eigenvectors are orthogonal, as expected. We have
z=10+25=35 and w=10+4=14,
so our orthogonal matrix is
U=[3i353i14535214].
With a bit of effort, we can finally confirm that
UHAU=[1006],
as expected.

5 Change of Basis
5.1 The matrix of a linear transformation

Exercise 5.1.2.

Solution.
It’s clear that CB(0)=0, since the only way to write the zero vector in V in terms of B (or, indeed, any independent set) is to set all the scalars equal to zero.
If we have two vectors v,w given by
v=a1e1++anenw=b1e1++bnen,
then
v+w=(a1+b1)e1++(an+bn)en,
so
CB(v+w)=[a1+b1an+bn]=[a1an]+[b1bn]=CB(v)+CB(w).
Finally, for any scalar c, we have
CB(cv)=CB((ca1)e1++(can)en)=[ca1can]=c[a1an]=cCB(v).
This shows that CB is linear. To see that CB is an isomorphism, we can simply note that CB takes the basis B to the standard basis of Rn. Alternatively, we can give the inverse: CB1:RnV is given by
CB1[c1cn]=c1e1++cnen.

Exercise 5.1.4.

Solution.
We have
T(1)=(1,0)=1(1,0)+0(1,1)T(1x)=(1,2)=1(1,0)+2(1,1)T((1x)2)=T(12+x2)=(2,4)=2(1,0)+4(1,1).
Thus,
MDB(T)=[CD(T(1))CD(T(1x))CD(T((1x)2))]=[112024].
To confirm, note that
MDB(T)CB(a+bx+cx2)=MDB(T)CB((a+b+c)(b+2c)(1x)+c(1x)2)=[112024][a+b+cb2cc][(a+b+c)+(b+2c)2c02(b+2c)+4c]=[a+2b+c2b],
while on the other hand,
CD(T(a+bx+cx2))=CD(a+c,2b)=CD((a+2b+c)(1,0)2b(1,1))=[a+2b+c2b].

Exercise 5.1.8.

Solution.
We must first write our general input in terms of the given basis. With respect to the standard basis
B0={[1000],[0100],[0010],[0001]},
we have the matrix P=[1001011000100101], representing the change from the basis B the basis B0. The basis D of P2(R) is already the standard basis, so we need the matrix MDB(T)P1:
For a matrix X=[abcd] we find
MDB(T)P1CB0(X)=[222103811121][abcd]=[2a2b+2c+d3b8c+da+b+2cd].
But this is equal to CD(T(X)), so
T([abcd])=CD1[2a2b+2c+d3b8c+da+b+2cd]=(2a2b+2c+d)+(3b8c+d)x+(a+b+2cd)x2.

5.2 The matrix of a linear operator

Exercise 5.2.4.

Solution.
With respect to the standard basis, we have
M0=MB0(T)=[324150027],
and the matrix P is given by P=[131212025]. Thus, we find
MB(T)=P1M0P=[9563671515104633].

5.7 Jordan Canonical Form

Exercise 5.7.7.

Solution.
With respect to the standard basis of R4, the matrix of T is
M=[1100010001201111].
We find (perhaps using the Sage cell provided below, and the code from the example above) that
cT(x)=(x1)3(x2),
so T has eigenvalues 1 (of multiplicity 3), and 2 (of multiplicity 1).
We tackle the repeated eigenvalue first. The reduced row-echelon form of MI is given by
R1=[1000010000100000],
so
E1(M)=span{x1}, where x1=[0001].
We now attempt to solve (MI)x=x1. We find
(0100000001101110|0001)RREF(1000010000100000|1000),
so x=tx1+x2, where x2=[1000]. We take x2 as our first generalized eigenvector. Note that (MI)2x2=(MI)x1=0, so x2null(MI)2, as expected.
Finally, we look for an element of null(MI)3 of the form x3, where (MI)x3=x2. We set up and solve the system (MI)x=x2 as follows:
(0100000001101110|1000)RREF(1000010000100000|0110),
so x=tx1+x3, where x3=[0110].
Finally, we deal with the eigenvalue 2. The reduced row-echelon form of M2I is
R2=[1000010000110000],
so
E2(M)=span{y}, where y=[0011].
Our basis of column vectors is therefore B={x1,x2,x3,y}. Note that by design,
Mx1=x1Mx2=x1+x2Mx3=x2+x3My=2y.
The corresponding Jordan basis for R4 is
{(0,0,0,1),(1,0,0,0),(0,1,1,0),(0,0,1,1)},
and with respect to this basis, we have
MB(T)=[1100011000100002].