Let V be a vector space and let
be a basis for V.
Every vector
can be uniquely expressed as a linear
combination of elements of
:
(Let me remind you of why this is true. Since a basis spans, every
can be written in this way. On the other hand, if
are two
ways of writing a given vector, then
, and by independence
, ...,
--- that is,
, ...,
. So the representation of
a vector in this way is unique.)
Consider the situation where
is a finite ordered basis --- that is, fix a numbering
of the elements of
. If
, the
ordered list of coefficients
is uniquely
associated with v. The
are the
components of v with respect to the (ordered) basis
; I will use the notation
It is easy to confuse a vector with the representation of
the vector in terms of its components relative to a basis. This
confusion arises because representation of vectors which is most
familiar is that of a vector as an ordinary n-tuple in
:
This amounts to identifying the elements of
with their representation relative to the standard basis
Example. (a) Show that
is a basis for
.
These are three vectors in
, which has
dimension 3. Hence, it suffices to check that they're independent.
Form the matrix with the elements of
as its rows and
row reduce:
The vectors are independent. Three independent vectors in
must form a basis.
(b) Find the components of
relative to
.
I must find numbers a, b, and c such that
This is equivalent to the matrix equation
Set up the matrix for the system and row reduce to solve:
This says
,
, and
. Therefore,
.
(c) Write
in terms of the standard basis.
I'll write
for v relative to
and
for v relative to the
standard basis. The matrix equation in (b)
says
In (b), I knew
and I wanted
; this time it's the other way around. So I simply
put
into the
spot and multiply:
Let me generalize the observation I made in (c).
I'll write
for M, and call it a translation matrix. Again,
translates vectors written in terms of
to vectors written in terms of the standard basis.
The inverse of a square matrix M is a matrix
such that
,
where I is the identity matrix. If I multiply the last equation on
the left by
, I get
In words, this means:
This means that
. Dispensing
with M, I can say that
In the example above, left multiplication by the following matrix
translates vectors from
to the standard basis:
The inverse of is
Left multiplication by this matrix translates vectors from the
standard basis to
.
Example. (Translating vectors from one basis to another) The translation analogy is a useful one, since it makes it easy to see how to set up arbitrary changes of basis.
For example, suppose
is another basis for
.
Here's how to translate vectors from
to
:
Remember that the product
is read from right
to left! Thus, the composite operation
translates a
vector to a standard vector, and then translates the
resulting standard vector to a
vector. Moreover,
I have matrices which perform each of the right-hand operations.
This matrix translates vectors from
to the standard
basis:
This matrix translates vectors from the standard basis to
:
Therefore, multiplication by the following matrix will translate
vectors from
to
:
Copyright 2008 by Bruce Ikenaga