If you take a set of vectors in a vector space, there's no reason why that set of vectors should be a subspace. For instance, consider the following vectors in the -vector space :
Now , which is in the set. But is not in the set, so the set is not closed under vector addition. And is not in the set, so the set is not closed under scalar multiplication. The set is not a subspace of .
Suppose we try to fix this by finding a subspace of which contains the three vectors. Well, that's easy: We could take ! This would work for any set of vectors in , but for that reason, it isn't a very interesting answer.
It might be better to ask for the "smallest" subspace of which contains the three vectors. This set will be called the span of the given set of vectors. We might approach this by asking: What vectors do we need to add to the set of three vectors to make it a subspace? Since subspaces are closed under vector addition, we'll need to throw in all the vectors which you get by adding the three vectors. You can check that we'd need to throw in two more vectors, so we now have 5. But subspaces are also closed under scalar multiplication, so we need to throw in multiples of these 5 vectors by every element of . Now we have an infinite number of vectors. Oops! We need to ensure that the set is closed under addition, so we need to go back and throw in all sums of the infinitely many vectors we have so far. Then, since we threw in some additional vectors, we need to take scalar multiples of those, and ... .
Our approach is a reasonable one, but we clearly need a more systematic way of carrying it out. We will see how to do it in this section.
Building subspaces involves forming sums and scalar multiples. Rather than thinking of doing these two operations separately, we'll define a concept which does both at once.
Definition. If , , ..., are vectors in a vector space V over a field F, a linear combination of the v's is a vector
Notice that if you take two vectors and and take the scalars to be and , you get , which is vector addition. And if you just take one vector , then represents a scalar multiple of that vector. Thus, the idea of a linear combination contains the operations of vector addition and scalar multiplication, but allows us to do many such operations all at once.
Let's see a numerical example. Take and in . Here is a linear combination of u and v:
is also a linear combination of u and v. And u and v are themselves linear combinations of u and v, as is the zero vector:
On the other hand, there are vectors in which are not linear combinations of and . Do you see how this pair is different from the first?
Definition. If S is a set of vectors in a vector space V, the span of S is the set of all linear combinations of vectors in S.
Remark. is also referred to as the subspace generated by S.
If S is a subset of a subspace W, then S spans W (or S is a spanning set for W, or S generates W) if .
If S is a finite set of vectors, we could just describe as the set of all linear combinations of the elements of S. If S is infinite, a particular linear combination of vectors from S involves finitely many vectors from S --- we "grab" finitely many vectors from S at a time --- but we're doing this in all possible ways.
Note that . For if s is a vector in S, then is a linear combination of a vector from S, so it is in .
Here's a pictorial example in . Take two vectors u and v:
It turns out that the span of u and v is all of ; that is, if , then w can be written as a linear combination of u and v. To see why this is reasonable, take a vector w and "project" w onto the lines containing the vectors u and v:
I can scale u up by multiplying by an appropriate number to get a vector which is the projection of w on the line of u. Likewise, I can scale v up by multiplying by an appropriate number to get a vector which is the projection of w on the line of v.
As you can see in the picture, w is the diagonal of the parallelogram whose sides are and , so .
Try this with some other vectors in place of w. Of course, a picture isn't a proof, but this example should help you get an idea of what the span of a set of vectors means geometrically.
Theorem. If S is a subset of a vector space V, the span of S is a subspace of V which contains S.
Proof. We already observed that .
Here are typical elements of the span of S:
Here the j's and k's are scalars and the u's and v's are elements of S.
Take two elements of the span and add them:
This sum is an element of the span, because it's a sum of vectors in S, each multiplied by a scalar --- that is, a linear combination of elements of S. Thus, the span is closed under taking sums.
Take an element of the span and multiply it by a scalar:
This is an element of the span, because it's a linear combination of elements of S. Thus, the span is closed under scalar multiplication.
Therefore, the span is a subspace.
Example. Prove that the span of and in is
To show that two sets are equal, you need to show that each is contained in the other. To do this, take a typical element of the first set and show that it's in the second set. Then take a typical element of the second set and show that it's in the first set.
Let W be the span of and in . A typical element of W is a linear combination of the two vectors:
Since the sum is a vector of the form for , it is in V. This proves that .
Now let . I have to show that this vector is a linear combination of and . This means that I have to find real numbers x and y such that
If I write the vectors as column vectors, I get
I can write this as the matrix equation
If I solve this system by row reduction, I get
The solution is
In other words,
Since is a linear combination of and , it follows that . This proves that .
Since and , I have .
Example. Let
(a) Prove or disprove: is in the span of S.
(b) Prove or disprove: is in the span of S.
(a) The vector is in the span of and if it can be written as a linear combination of and . So I try to find numbers a and b such that
I'll convert this to a matrix equation by writing the vectors as column vectors:
By using column vectors (rather than row vectors), we get a familiar kind of matrix equation:
We can solve for a and b by row reducing the augmented matrix:
The solution is , . That is,
The vector is in the span of S.
(b) Following the approach we took in part (a), we'll convert the vectors to column vectors to get the equation
This is equivalent to the matrix equation
Row reduce to solve the system:
The last matrix says " ", a contradiction. The system is inconsistent, so there are no such numbers a and b. Therefore, is not in the span of S.
Thus, to determine whether the vector is in the span of , , ..., in , form the augmented matrix
If the system has a solution, b is in the span, and coefficients of a linear combination of the v's which add up to b are given by a solution to the system. If the system has no solutions, then b is not in the span of the v's.
(In a general vector space where vectors may not be "numbers in slots", you have to got back to the definition of spanning set.)
Example. Consider the following set of vectors in :
Prove that the span of the set is all of .
I have to show that any vector can be written as a linear combination of the vectors in the set. That is, I must find real numbers a, b, c such that
The matrix equation is
Row reduce to solve:
I get the ugly solution
This shows that, given any vector , I can find a linear combination of the original three vectors which equals .
Thus, the span of the original set of three vectors is all of .
Example. Let
(a) Determine whether .
(b) Determine whether .
(a) I want to determine whether there are numbers such that
This gives the system
Form the augmented matrix and row reduce:
The last matrix says and . Therefore, is in the span of S:
(b) I want to determine whether there are numbers such that
This gives the system
Form the augmented matrix and row reduce:
The last row of the row reduced echelon matrix says " ". This contradiction implies that the system is has no solutions. Therefore, is not in the span of S.
In the next example, we go back to the definition of the span of a set of vectors, rather than writing down a matrix equation.
Example. is the vector space over of continuous functions . Let
Show that h is not in , the span of f and g.
We'll give a proof by contradiction.
Suppose there are numbers such that
Set . We get
The equation becomes
Set . We get
The equation is now
Finally, set . We get
This contradiction shows there are no numbers which make the initial equation true. Hence, h is not in the span of f and g.
Another definition of the span of a set
There is another way to define the span of a set of vectors. The construction is a little more subtle than the one we gave earlier. It uses the intersection of a (possibly) infinite collection of subspaces, which we know by an earlier result is a subspace.
Let V be a vector space over a field F, and let S be a set of vectors in V. Let
That is, to form we intersect all subspaces of V which contain S. Note that there is at least one subspace that contains S, namely V itself. We will show that is just the span of S, as constructed earlier: The set of all linear combinations of elements of S.
(By the way, " " is just temporary notation I'm using for this discussion. After this, I'll just use to denote the span of S.)
Theorem. .
Proof. First, the span is a subspace of V which contains S. So is one of the subspaces being intersected to construct , and hence
To complete the proof, I'll show that . Take an element of , namely a linear combination of elements of S:
Let W be a subspace of V which contains S. Then , and since W is a subspace,
Thus, is contained in every subspace which contains S --- that is, it is contained in every subspace being intersected to construct . Hence, it is contained in . This proves that .
Therefore, .
If you want to construct an object in math using some "building blocks", there are often two general ways to do so. The first way is to start with the building blocks, and combine them to build the object "from the inside out". This direct approach is what we did in the definition of as all linear combinations of elements of S. It requires that you know how to put the blocks together to do the "building".
Another approach is to take all objects of the "right kind" which contain the building blocks and find the "smallest". You find the "smallest" such object by intersecting all the objects of the "right kind" containing the building blocks. In this approach, you will usually have to intersect an infinite number of sets. And you will need to show that the intersection of sets of the "right kind" is still a set of the "right kind". This is the approach we used in intersecting all subspaces containing S.
Of course, if you've done things correctly, the two approaches give the same object at the end, as they did here. Both approaches are useful in mathematics.
Copyright 2022 by Bruce Ikenaga