We begin with a proof by induction (Proof Technique
I) of the first statement in the conclusion of the theorem. We use induction on the dimension of
to show that if
is a linear transformation, then there is a basis
for
such that the matrix representation of
relative to
is an upper triangular matrix.
To start suppose that
Choose any nonzero vector
and realize that
Then
for some
which determines
uniquely (
Theorem LTDB). This description of
also gives us a matrix representation relative to the basis
as the
matrix with lone entry equal to
And this matrix representation is upper triangular (
Definition UTM).
For the induction step let
and assume the theorem is true for every linear transformation defined on a vector space of dimension less than
By
Theorem EMHE (suitably converted to the setting of a linear transformation),
has at least one eigenvalue, and we denote this eigenvalue as
(We will remark later about how critical this step is.) We now consider properties of the linear transformation
Let be an eigenvector of for By definition Then
Let be the subspace of that is the range of and define We define a new linear transformation on
This does not look we have accomplished much, since the action of is identical to the action of For our purposes this will be a good thing. What is different is the domain and codomain. is defined on a vector space with dimension less than and so is susceptible to our induction hypothesis. Verifying that is really a linear transformation is almost entirely routine, with one exception. Employing in our definition of raises the possibility that the outputs of will not be contained within (but instead will lie inside but outside ). To examine this possibility, suppose that We have
Since
is the range of
And by
Property SC,
Finally, applying
Property AC we see by closure that the sum is in
and so we conclude that
This argument convinces us that it is legitimate to define
as we did with
as the codomain.
is a linear transformation defined on a vector space with dimension less than so we can apply the induction hypothesis and conclude that has a basis, such that the matrix representation of relative to is an upper triangular matrix.
Beginning with the linearly independent set
repeatedly apply
Theorem ELIS to add vectors to
maintaining a linearly independent set and spanning ever larger subspaces of
This process will end with the addition of
vectors, which together with
will span all of
Denote these vectors as
Then
is a basis for
and is the basis we desire for the conclusion of the theorem. So we now consider the matrix representation of
relative to
Since the definition of and agree on the first columns of will have the upper triangular matrix representation of in the first rows. The remaining rows of these first columns will be all zeros since the outputs of for basis vectors from are all contained in and hence are linear combinations of the basis vectors in The situation for on the basis vectors in is not quite as pretty, but it is close.
In the penultimate equality, we have rewritten an element of the range of as a linear combination of the basis vectors, for the range of using the scalars If we incorporate these column vectors into the matrix representation we find occurrences of on the diagonal, and any nonzero entries lying only in the first rows. Together with the upper triangular representation in the upper left-hand corner, the entire matrix representation for is clearly upper triangular. This completes the induction step. So for any linear transformation there is a basis that creates an upper triangular matrix representation.
We have one more statement in the conclusion of the theorem to verify. The eigenvalues of
and their multiplicities, can be computed with the techniques of
Chapter E relative to any matrix representation (
Theorem EER). We take this approach with our upper triangular matrix representation
Let
be the diagonal entry of
in row
and column
Then the characteristic polynomial, computed as a determinant (
Definition CP) with repeated expansions about the first column, is
The roots of the polynomial equation
are the eigenvalues of the linear transformation (
Theorem EMRCP). So each diagonal entry is an eigenvalue, and is repeated on the diagonal exactly
times (
Definition AME).