Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 21/refcontrol



Eigenvalues and eigenvectors

For a reflection at an axis in the plane, certain vectors behave particularly simply. The vectors on the axis are sent to themselves, and the vectors which are orthogonal to the axis are sent to their negatives. For all these vectors, the image under this linear mapping lies on the line spanned by these vectors. In the theory of eigenvalues and eigenvectors, we want to know whether, for a given linear mapping, there exist lines (one-dimensional linear subspaces), which are mapped to themselves. The goal is to find, for the linear mapping, a basis such that the describing matrix is quite simple. Here, an important application is to find solutions for a system of linear differential equations.

A reflection at an axis has two eigenlines, the axis with eigenvalue and the orthogonal line with eigenvalue .

Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a linear mapping.MDLD/linear mapping Then an element , , is called an eigenvector of (for the eigenvalueMDLD/eigenvalue ), if

for some

holds.

To paraphrase, an eigenvector is a vector which is linearly dependentMDLD/linearly dependent to .

A shear mapping has one eigenline for the eigenvalue , and no further eigenvalues.

Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a linear mapping.MDLD/linear mapping Then an element is called an eigenvalue for , if there exists a vector , such that

The set of all eigenvalues is called the spectrum of .


Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a linear mapping.MDLD/linear mapping For , we denote by

the eigenspace of for the value .


Thus we allow arbitrary values (not only eigenvalues) in the definition of an eigenspace. We will see in fact that they are linear subspaces.MDLD/linear subspaces In particular, belongs to every eigenspace, though it is never an eigenvector. The linear subspace generated by an eigenvector is called an eigenline. For most (in fact all up to finitely many, in case the vector space has finite dimension) , the eigenspace is just the zero space.


In case of a homothetyMDLD/homothety with factor , every vector is an eigenvectorMDLD/eigenvector for the eigenvalueMDLD/eigenvalue . The eigenspaceMDLD/eigenspace of the eigenvalue is the total space. Also, we can restrict an endomorphism to any eigenspace (as source and target), yielding the mapping

This mapping is just the homothety with factor .

For matrices, we use the same concepts. If is a linear mapping, and is a describing matrix with respect to a basis, then for an eigenvalue and an eigenvector with corresponding coordinate tuple with respect to the basis, we have the relation


We consider the linear mappingMDLD/linear mapping

given by the diagonal matrixMDLD/diagonal matrix

The diagonal entries are the eigenvaluesMDLD/eigenvalues of , and the -th standard vector is a corresponding eigenvector.MDLD/eigenvector The eigenspaces are

These spaces are not if and only if equals one of the diagonal entries. The dimension of the eigenspace is given by the number how often the value occurs in the diagonal. The sum of all these dimension gives .


We consider the linear mappingMDLD/linear mapping

given by the matrix

The question whether this mapping has eigenvalues,MDLD/eigenvalues leads to the question whether there exists some , such that the equation

has a nontrivial solution . For a given , this is a linear problem and can be solved with the elimination algorithm. However, the question whether there exist eigenvalues at all, leads, due to the variable "eigenvalue parameter“ , to a nonlinear problem. The system of equations above is

For , we get , but the null vector is not an eigenvector. Hence, suppose that . Both equations combined yield the condition

hence . But in , the number does not have a square root,MDLD/square root therefore there is no solution, and that means that has no eigenvalues and no eigenvectors.MDLD/eigenvectors

Now we consider the matrix as a real matrix, and look at the corresponding mapping

The same computations as above lead to the condition , and within the real numbers, we have the two solutions

For both values, we have now to find the eigenvectors. First, we consider the case , which yields the linear system

We write this as

and as

This system can be solved easily, the solution space has dimension one, and

is a basic solution.

For , we do the same steps, and the vector

is a basic solution. Thus over , the numbers and are eigenvalues, and the corresponding eigenspacesMDLD/eigenspaces are


Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a

linear mapping.MDLD/linear mapping Then the following statements hold.
  1. Every eigenspaceMDLD/eigenspace

    is a linear subspaceMDLD/linear subspace of .

  2. is an eigenvalueMDLD/eigenvalue for , if and only if the eigenspace is not the null space.MDLD/null space
  3. A vector , is an eigenvectorMDLD/eigenvector for , if and only if .

(1). Let and let . Then

(2) and (3) follow directly from the definitions.



Kernel and fixed space

Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a linear mapping.MDLD/linear mapping Then

In particular, is an

eigenvalueMDLD/eigenvalue of if and only if is not

injective.MDLD/injective

Proof



Beside the eigenspaceMDLD/eigenspace for , which is the kernelMDLD/kernel (linear mapping) of the linear mapping, the eigenvalues and are in particular interesting. The eigenspace for consists of all vectors which are sent to themselves. Restricted to this linear subspace, the mapping is just the identity, it is called the fixed space. The eigenspace for consists in all vector which are sent to their negative. On this linear subspace, the mapping acts like the reflection at the origin.


Let be a field,MDLD/field a -vector spaceMDLD/vector space and

a linear mapping.MDLD/linear mapping The fixed space of is the eigenspaceMDLD/eigenspace for the eigenvalueMDLD/eigenvalue

, that is, the set .



Eigenvalues under base change

Let

denote an endomorphismMDLD/endomorphism on a -vector spaceMDLD/vector space , and let

denote an isomorphismMDLD/isomorphism (vs) of -vector spaces. Set

Then the following hold.
  1. A vector is an eigenvectorMDLD/eigenvector of for the eigenvalueMDLD/eigenvalue if and only if is an eigenvector of for the eigenvalue .
  2. and have the same eigenvalues.
  3. The mapping induces for every an isomorphism

(1). Let be an eigenvector of for the eigenvalue . Set

Then

The reverse statement holds in the same way. (2) and (3) follow directly from (1).


If we have an endomorphism on a finite-dimensional vector space, which is described by the matrix with respect to a given basis, then the eigenvalues and the eigenvectors correspond to each other. Then eigenvector of the matrix is the Coordinate tuple of the corresponding eigenvector with respect to the basis. The eigenvalues do not depend on the chosen basis, but the eigenvectors do.


Let be an endomorphismMDLD/endomorphism on the finite-dimensionalMDLD/finite-dimensional (vs) -vector spaceMDLD/vector space , and let denote a basisMDLD/basis (vs) of . Let be the describing matrixMDLD/describing matrix of with respect to the basis. Then is an eigenvectorMDLD/eigenvector of for the eigenvalueMDLD/eigenvalue if and only if the coordinate tupleMDLD/coordinate tuple

of with respect to the basis is an eigenvector of for the eigenvalue .

This follows directly from fact, using the diagram



Let be an -matrixMDLD/matrix over a field , and let denote an invertibleMDLD/invertible (matrix) -matrix. Let . Then an -tuple is an eigenvectorMDLD/eigenvector of for the eigenvalue if and only if

is an eigenvector of the matrix for the eigenvalue . In particular, and

have the same eigenvalues.

This follows from fact.


<< | Linear algebra (Osnabrück 2024-2025)/Part I | >>
PDF-version of this lecture
Exercise sheet for this lecture (PDF)