Scientific Computing Using Python - PHYS:4905
Lecture #09 - Prof. Kaaret
Linear operators and basis vectors
These notes borrow from Linear Algebra
by Cherney, Denton, Thomas, and Waldron.
Linear Transformations
Now that we have the concept of vector spaces, we can introduce a
new way to think about functions - that they are transformations of
one vector space to another vector space. For example, we may
have the vector spaces V and W and the function L
which would then transform
.
Since this is a course on linear algebra, we are interested in
linear transformations. The function L is linear if
for all
and
.
Sometimes linear functions are called linear maps, linear operators,
or linear transformations. The name "map" comes from the idea
of mapping one vector space into another vector space.
Note that we could instead have required
and
These separate statements of additivity and homogeneity are
equivalent.
Why Are Linear Operators so Special?
In general, specifying a real function of one variable requires an
infinite number of real variables. This is clearly true if we
allow any real function without any requirements like
continuity. Such a function is a map from the set of real
numbers to another set of real numbers, so we need to specify
the output real number separately for each input real number.
It also holds true if we require continuity. In that case, we
have a choice of what the derivative is at each point, so again, we
need an infinite number of real numbers.
In contrast, linear functions are completely specified by just a few
numbers. Let's say that L is a linear operator and
and
Can you figure out what
is?
We can go back to the requirement of additivity given above,
.
Then,
We wrote down definitions for the two operations on the right hand
side of that equation, so we can substitute those in and do a little
math,
If you think about it for a little while, you may conclude
(correctly) that you can do this for any vector. Any arbitrary
vector can be written in terms of our 'basis' vectors
We can then use that sum of vectors and do the same calculations
that we did before
Therefore, we can figure out what L does to any vector in
the vector space that L
is defined over.
Exactly how much information, or how many numbers, did we need to
specify a linear function on ℝ2?
Looking back at the definition above, we see that we need two column
vectors each with two elements. This is four numbers. We
could also write the four numbers as a matrix
In general, a linear transformation operating on the vector
space is
completely specified by how it acts on n vectors each with
exactly one non-zero component called the basis vectors. The
result on each basis vector will be a vector with n
components. Hence, we need
numbers to completely specify the function. Those
numbers can be written as an matrix.
Basis vectors
We just saw that a linear operator acting on
is
completely specified by how it acts on the two vectors
and .
Do we need to use those particular two vectors?
The crucial steps in our little 'proof' above were using additivity
and homogeneity to break down the operation of L on an arbitrary
vector into operations on our basis vectors.
We could do this because any arbitrary vector can be written in
terms of the basis vectors
Could we write any arbitrary vector in terms of some other set can
be written in terms of the basis vectors?
How about
and
?
The question is whether we can find a and b such
that
Yes, we can! You can set this up as a system of linear
equations. The solution is a = (x+y)/2,
b = (x-y)/2. Substituting in, we can
check that this solution is valid.
How about
and
?
How about
and
?
How about
and
and
?
In general, the number of basis vectors needed is equal to the
number of dimensions of the vector space on which the linear
operator acts. Any more creates redundancy in how we
'decompose' vectors.
Also the basis vectors must span the entire space, meaning
that it must be possible to write any arbitrary vector as a linear
combination of the basis vectors.
Do
and
span ?
No, because those two vectors are not linearly independent.
You can get the second vector by multiplying the first vector by a
constant.
Linear independence is easy for the two dimensional case, but how do
we generalize to three dimension? Is this set of vectors
linearly independent and do they span ?
You cannot get any of these vectors by multiplying another vector by
a constant.
However, you can get the third vector by adding the first two.
A set of vectors are linearly dependent if you can get one vector
(or more) in the set from a linear combination of other vectors in
the set. Or, in math speak, the vectors v1,
v2, ..., vn are linearly
dependent if there exist scalars c1, c2,
..., cn not all zero such that
c1v1 + c2v2
+ ... + cnvn = 0
Once we do determinants, we see how to efficiently check if a set of
vectors are linearly dependent.
A basis is a set of vectors that can be used to uniquely
express any other vector in the vector space. The vectors must
be linearly independent. If the set of vectors is finite, then
the number of vectors in the basis is the dimension of the
vector space.
Basis Notation
In physics, we often write 3-vectors in a notation like
The set of unit vectors
are the basis. In this notation, we need to write down the
basis vectors whenever we write a vector. The vector written
above makes sense, by trying to write a vector without the basis
vectors, e.g.
would not.
In matrix notation, we define a set of vectors as a basis. The
'standard' basis for is
To use matrix notation for vectors, we need to assign an order to
the vectors in our basis. We denote that by using parentheses
instead of curly brackets. This basis given is then and we can write
the vector in column vector form or in algebraic form,
The notation with the subscript E gives "the vector v in the
basis E". This is equivalent to writing out the vector
in algebraic form with the basis vectors given explicitly and also
to the product of the set of basis vectors with the column
vector. To evaluate the latter, do the multiplication in the
same fashion as multiplying a matrix times a column vector.
We can use a different basis to describe the vectors. For
example, to form the basis B, we could pick a new set of
vectors,
We can write the vector in this basis,
Note that this changes the values of the components.
Note that only in the standard basis do the components equal the
components of the column vector. Converting from one basis to
another is equivalent to solving a system of linear equations,
Many physics problem are much simpler if you change to an
appropriate basis. Indeed, some problems, like finding the
energy levels of a quantum mechanical system, are essentially
exercises in figuring out an appropriate basis that simplifies the
matrix multiplications.
From Linear Operators to Matrices
Linear functions are completely defined by how they act on their
basis vectors. One needs to compute what the linear
transformation does to every input basis vector and then write the
answers in terms of the output basis vectors.
More formally, if L is a linear operator from the vector space V
to the vector space W and we define a basis
for V
and a basis
for W,
then L is completely specified by a set of numbers
that give
Note that this equation is defined for all i from 1 to the
dimension of the basis, n, and that the sum goes over the
same range. If both bases are standard bases, then the numbers
give the matrix representation of the operator. If not, one
can compute the matrix representation of the operator.
Assignment
The assignment for today's class involves linear operators and
basis vectors.
HW #9 is due on 10/1.