Applied linear operators and spectral methods/Lecture 2

Norms in inner product spaces edit

Inner product spaces have   norms which are defined as

 

When  , we get the   norm

 

When  , we get the   norm

 

In the limit as   we get the   norm or the sup norm

 

The adjacent figure shows a geometric interpretation of the three norms.

 
Geomtric interpretation of various norms

If a vector space has an inner product then the norm

 

is called the induced norm. Clearly, the induced norm is nonnegative and zero only if  . It is also linear under multiplication by a positive vector. You can think of the induced norm as a measure of length for the vector space.

So useful results that follow from the definition of the norm are discussed below.

Schwarz inequality edit

In an inner product space

 

Proof

This statement is true if  .

If   we have

 

Now

 

Therefore,

 

Let us choose   such that it minimizes the left hand side above. This value is clearly

 

which gives us

 

Therefore,

 

Triangle inequality edit

The triangle inequality states that

 

Proof

 

From the Schwarz inequality

 

Hence

 

Angle between two vectors edit

In   or   we have

 

So it makes sense to define   in this way for any real vector space.

We then have

 

Orthogonality edit

In particular, if   we have an analog of the Pythagoras theorem.

 

In that case the vectors are said to be orthogonal.

If   then the vectors are said to be orthogonal even in a complex vector space.

Orthogonal vectors have a lot of nice properties.

Linear independence of orthogonal vectors edit

  • A set of nonzero orthogonal vectors is linearly independent.

If the vectors   are linearly dependent

 

and the   are orthogonal, then taking an inner product with   gives

 

since

 

Therefore the only nontrivial case is that the vectors are linearly independent.

Expressing a vector in terms of an orthogonal basis edit

If we have a basis   and wish to express a vector   in terms of it we have

 

The problem is to find the  s.

If we take the inner product with respect to  , we get

 

In matrix form,

 

where   and  .

Generally, getting the  s involves inverting the   matrix  , which is an identity matrix  , because  , where   is the Kronecker delta.

Provided that the  s are orthogonal then we have

 

and the quantity

 

is called the projection of   onto  .

Therefore the sum

 

says that   is just a sum of its projections onto the orthogonal basis.

 
Projection operation.

Let us check whether   is actually a projection. Let

 

Then,

 

Therefore   and   are indeed orthogonal.

Note that we can normalize   by defining

 

Then the basis   is called an orthonormal basis.

It follows from the equation for   that

 

and

 

You can think of the vectors   as orthogonal unit vectors in an  -dimensional space.

Biorthogonal basis edit

However, using an orthogonal basis is not the only way to do things. An alternative that is useful (for instance when using wavelets) is the biorthonormal basis.

The problem in this case is converted into one where, given any basis  , we want to find another set of vectors   such that

 

In that case, if

 

it follows that

 

So the coefficients   can easily be recovered. You can see a schematic of the two sets of vectors in the adjacent figure.

 
Biorthonomal basis

Gram-Schmidt orthogonalization edit

One technique for getting an orthogonal baisis is to use the process of Gram-Schmidt orthogonalization.

The goal is to produce an orthogonal set of vectors   given a linearly independent set  .

We start of by assuming that  . Then   is given by subtracting the projection of   onto   from  , i.e.,

 

Thus   is clearly orthogonal to  . For   we use

 

More generally,

 

If you want an orthonormal set then you can do that by normalizing the orthogonal set of vectors.

We can check that the vectors   are indeed orthogonal by induction. Assume that all   are orthogonal for some  . Pick  . Then

 

Now   unless  . However, at  ,   because the two remaining terms cancel out. Hence the vectors are orthogonal.

Note that you have to be careful while numerically computing an orthogonal basis using the Gram-Schmidt technique because the errors add up in the terms under the sum.

Linear operators edit

The object   is a linear operator from   onto   if

 

A linear operator satisfies the properties

  1.  .
  2.  .

Note that   is independent of basis. However, the action of   on a basis   determines   completely since

 

Since   we can write

 

where   is the   matrix representing the operator   in the basis  .

Note the location of the indices here which is not the same as what we get in matrix multiplication. For example, in  , we have

 

We will get into more details in the next lecture.

  Resource type: this resource contains a lecture or lecture notes.