Recall that in Example 1.3.4, we had to take care to insist that the vectors spanning our plane were not parallel. Otherwise, what we thought was a plane would, in fact, be only a line. Similarly, we said that a line is given by the set of all vectors of the form , where is a scalar, and is not the zero vector. Otherwise, if , we would have for all , and our “line” would be the trivial subspace.
When we define a subspace as the span of a set of vectors, we want to have an idea of the size (or perhaps complexity) of the subspace. Certainly the number of vectors we use to generate the span gives a measure of this, but it is not the whole story: we also need to know how many of these vectors “depend” on other vectors in the generating set. As Theorem 1.4.10 tells us, when one of the vectors in our generating set can be written as a linear combination of the others, we can remove it as a generator without changing the span.
Given a set of vectors , an important question is therefore: can any of these vectors be written as a linear combination of other vectors in the set? If the answer is no, we say that is linearly independent. This is a difficult condition to check, however: first, we would have to show that cannot be written as a linear combination of . Then, that cannot be written in terms of , and so on.
This could amount to solving different systems of equations in variables! But the systems are not all unrelated. The equation can be rewritten as , where we happen to have set .
In fact, we can do the same thing for each of these systems, and in each case we end up with the same thing: a single homogeneous system with one extra variable. (We get back each of the systems we started with by setting one of the variables equal to .) This not only is far more efficient, it changes the question: it is no longer a question of existence of solutions to a collection of non-homogeneous systems, but a question of uniqueness for the solution of a single homogeneous system.
True or false: if , where , then is linearly independent.
True.
The definition of independence is a conditional statement: if,then. It is important to get the order of the logic correct here, as the converse is always true.
False.
The definition of independence is a conditional statement: if,then. It is important to get the order of the logic correct here, as the converse is always true.
Note that the definition of independence asserts that there can be no “non-trivial” linear combinations that add up to the zero vector. Indeed, if even one scalar can be nonzero, then we can solve for the corresponding vector. Say, for example, that we have a solution to with . Then we can move all other vectors to the right-hand side, and multiply both sides by to give
When we want to conclude that a set of vectors is linearly independent, we should assume that for some , and then try to show that the scalars must be zero. It’s important that we do not assume anything about the scalars to begin with.
If the hypothesis of a statement includes the assumption that a set of vectors is independent, we know that if we can get a linear combination of those vectors equal to the zero vector, then the scalars in that linear combination are automatically zero.
When looking for vectors that span a subspace, it is useful to find a spanning set that is also linearly independent. Otherwise, as Theorem 1.4.10 tells us, we will have some “redundant” vectors, in the sense that removing them as generators does not change the span.
This time, we will outline the strategy, and leave the execution to you. Both parts are about linear combinations. What does independence look like for a single vector? We would need to show that if for some scalar , then . Now recall that in Exercise 1.2.4, we showed that if , either or . We’re assuming , so what does that tell you about ?
In the second part, if we have a linear combination involving the zero vector, does the value of the scalar in front of matter? (Can it change the value of the linear combination?) If not, is there any reason that scalar would have to be zero?
The definition of linear independence tells us that if is an independent set of vectors, then there is only one way to write as a linear combination of these vectors; namely,
Computationally, questions about linear independence are just questions about homogeneous systems of linear equations. For example, suppose we want to know if the vectors
We now apply some basic theory from linear algebra. A unique (and therefore, trivial) solution to this system is guaranteed if the matrix is invertible, since in that case we have .
The approach in Remark 1.6.6 is problematic, however, since it won’t work if we have 2 vectors, or 4. In general, we should look at the reduced row-echelon form. A unique solution corresponds to having a leading 1 in each column of . Let’s check this condition.
One observation is useful here, and will lead to a better understanding of independence. First, it would be impossible to have 4 or more linearly independent vectors in . Why? (How many leading ones can you have in a matrix?) Second, having two or fewer vectors makes it more likely that the set is independent.
The largest set of linearly independent vectors possible in contains three vectors. You might have also observed that the smallest number of vectors needed to span is 3. Hmm. Seems like there’s something interesting going on here. But first, some more computation. (For the first two exercises, once you’ve tried it yourself, you can find a solution using a Sage cell for computation at the end of the book.)