Let V be a vector space and let be a basis for V. Every vector can be uniquely expressed as a linear combination of elements of :
(Let me remind you of why this is true. Since a basis spans, every can be written in this way. On the other hand, if are two ways of writing a given vector, then , and by independence , ..., --- that is, , ..., . So the representation of a vector in this way is unique.)
Consider the situation where is a finite ordered basis --- that is, fix a numbering of the elements of . If , the ordered list of coefficients is uniquely associated with v. The are the components of v with respect to the (ordered) basis ; I will use the notation
It is easy to confuse a vector with the representation of the vector in terms of its components relative to a basis. This confusion arises because representation of vectors which is most familiar is that of a vector as an ordinary n-tuple in :
This amounts to identifying the elements of with their representation relative to the standard basis
Example. (a) Show that
is a basis for .
These are three vectors in , which has dimension 3. Hence, it suffices to check that they're independent. Form the matrix with the elements of as its rows and row reduce:
The vectors are independent. Three independent vectors in must form a basis.
(b) Find the components of relative to .
I must find numbers a, b, and c such that
This is equivalent to the matrix equation
Set up the matrix for the system and row reduce to solve:
This says , , and . Therefore, .
(c) Write in terms of the standard basis.
I'll write for v relative to and for v relative to the standard basis. The matrix equation in (b)
says
In (b), I knew and I wanted ; this time it's the other way around. So I simply put into the spot and multiply:
Let me generalize the observation I made in (c).
I'll write for M, and call it a translation matrix. Again, translates vectors written in terms of to vectors written in terms of the standard basis.
The inverse of a square matrix M is a matrix such that , where I is the identity matrix. If I multiply the last equation on the left by , I get
In words, this means:
This means that . Dispensing with M, I can say that
In the example above, left multiplication by the following matrix translates vectors from to the standard basis:
The inverse of is
Left multiplication by this matrix translates vectors from the standard basis to .
Example. (Translating vectors from one basis to another) The translation analogy is a useful one, since it makes it easy to see how to set up arbitrary changes of basis.
For example, suppose
is another basis for .
Here's how to translate vectors from to :
Remember that the product is read from right to left! Thus, the composite operation translates a vector to a standard vector, and then translates the resulting standard vector to a vector. Moreover, I have matrices which perform each of the right-hand operations.
This matrix translates vectors from to the standard basis:
This matrix translates vectors from the standard basis to :
Therefore, multiplication by the following matrix will translate vectors from to :
Copyright 2008 by Bruce Ikenaga