Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 24/refcontrol



The Theorem of Cayley-Hamilton
Arthur Cayley (1821-1895)
William Hamilton (1805-1865)


One highlight of the linear algebra is the Theorem of Cayley-Hamilton. In order to formulate this theorem, recall that we can plug in a square matrix into a polynomial, see the 20th lecture. Here, the variable is everywhere replaced by the matrix , the powers are the -th matrix product of with itself, and the addition is the (componentwise) addition of matrices. A scalar has to be interpreted as the -fold of the identity matrix. For the polynomial

and the matrix

we get

For a fixed matrix , we have the substitution mapping

This is (like the substitution mapping for an element ), a ring homomorphism,MDLD/ring homomorphism that is, the relations (see also Lemma 20.3 )

hold. The Theorem of Cayley-Hamilton answers the question of what happens when we insert a matrix in its characteristic polynomial.


TheoremTheorem 24.1 change

Let be a field,MDLD/field and let be an -matrix.MDLD/matrix Let

denote the characteristic polynomialMDLD/characteristic polynomial of . Then

This means that the matrix annihilates the characteristic polynomial.

We consider the matrix as a matrix whose entries are in the fieldMDLD/field . The adjugate matrixMDLD/adjugate matrix

belongs also to . The entries of the adjugate matrix are by definition the determinantsMDLD/determinants of -submatrices of . In the entries of this matrix, the variable occurs at most in its first power, so that, in the entries of the adjugate matrix, the variable occurs at most in its -th power. We write

with matrices

that is, we write the entries as polynomials, and we collect all coefficients referring to into a matrix. Because of Theorem 17.9 , we have

We can write the matrix on the left according to the powers of and we get

Since these polynomials coincide, their coefficients coincide. That is, we have a system of equations

We multiply these equations from the left from top down with , yielding the system of equations

If we add the left-hand side of this system, then we just get . If we add the right-hand side, then we get , because every partial summand occurs once positively and once negatively. Hence, we have .



TheoremTheorem 24.2 change

Let be a finite-dimensionalMDLD/finite-dimensional vector spaceMDLD/vector space over a fieldMDLD/field , and let

denote a linear mapping.MDLD/linear mapping Then the characteristic polynomialMDLD/characteristic polynomial of fulfills the relation

This follows immediately from Theorem 24.1 .



Minimal polynomial and characteristic polynomial

CorollaryCorollary 24.3 change

Let be a finite-dimensionalMDLD/finite-dimensional vector spaceMDLD/vector space over a fieldMDLD/field , and let

be a linear mapping.MDLD/linear mapping Then the characteristic polynomialMDLD/characteristic polynomial is a multiple of the minimal polynomialMDLD/minimal polynomial (endomorphism)

of .

This follows directly from Theorem 24.2 and Corollary 20.12 .


In particular, the degree of the minimal polynomial of is bounded by the dimension of the vector space . The minimal polynomial and the characteristic polynomial are related in several respects, for example, they have the same zeroes.


Let be a finite-dimensionalMDLD/finite-dimensional vector spaceMDLD/vector space over a fieldMDLD/field , and let

be a linear mapping.MDLD/linear mapping Let be an eigenvectorMDLD/eigenvector of with eigenvalueMDLD/eigenvalue , and let denote a polynomial. Then

In particular, is an eigenvector of with eigenvalue . The vector

belongs to the kernel of if and only if is a zero of .

We have

This implies the statement, since the assignment is compatible with addition and scalar multiplication.



LemmaLemma 24.5 change

Let be a finite-dimensionalMDLD/finite-dimensional vector spaceMDLD/vector space over a fieldMDLD/field , and let

be a linear mapping.MDLD/linear mapping Then the characteristic polynomialMDLD/characteristic polynomial and the minimal polynomialMDLD/minimal polynomial (endomorphism)

have the same zeroes.

It follows directly from Cayley-Hamilton that the zeroes of the minimal polynomial are also zeroes of the characteristic polynomial.

To prove the other implication, let be a zero of the characteristic polynomial, and let denote an eigenvectorMDLD/eigenvector of with eigenvalueMDLD/eigenvalue , its existence is guaranteed by Theorem 23.2 . We write the minimal polynomial as

where has no zero. Then

We apply this mapping to . Because of Fact *****, the factors send the vector to or to , respectively. Altogether, is sent to

As the composed mapping is the zero mapping and , we must have for some .



Further examples

For the moment, we will apply the following concept only for invertible matrices.


Let be a groupMDLD/group and an element. Then we call the smallest positive number with the order of . For this, we write . If all positive powers of are different from the neutral element, then we set

.

We consider linear mappings

with the property that some power of it is the identity, say

that is, has finite order. Typical examples are rotations around an angle of the form degree. The polynomial annihilates this endomorphism, and is, therefore, a multiple of the minimal polynomial.


Let be a field,MDLD/field and A zero of the polynomialMDLD/polynomial (1)

in is called an -th root of unity

in .

Let . The zeroes of the polynomials over are

In , we have the factorization

The proof uses some basic facts about the complex exponential function.MDLD/complex exponential function We have

Hence, the given complex numbers are indeed zeroes of the polynomial . These zeroes are all different, because

with implies, by considering the fraction, that

holds. Therefore, there exist explicit zeroes and these are all the zeroes of the polynomial. The explicit description in coordinates follows from the Euler's formula.



For a permutationMDLD/permutation on , the -matrixMDLD/matrix

where

and all other entries are , is called a

permutation matrix.

We want to determine the characteristic polynomial of a permutation matrix. Here, we use that a permutation is a product of cycles. For a cycle of the form , the corresponding permutation matrix is

Every cycle can be brought (by renumbering) into this form.


The characteristic polynomialMDLD/characteristic polynomial of a permutation matrixMDLD/permutation matrix for a cycle of orderMDLD/order (cycle) is

We may assume that the cycle has the form . The corresponding permutation matrix looks with respect to like the identity matrix and has, with respect to the first standard vectors, the form

The determinant of is multiplied with the determinant of

The expansion with respect to the first row yields



LemmaLemma 24.11 change

For a permutation matrixMDLD/permutation matrix over for a cycleMDLD/cycle (permutation) with and a -th root of unityMDLD/root of unity , the vectors

are eigenvectorsMDLD/eigenvectors of for the eigenvalueMDLD/eigenvalue . In particular, a permutation matrix of a cycle over is

diagonalizable.MDLD/diagonalizable

We have

Since there are different -th roots of unity in , these vectors are linearly independentMDLD/linearly independent due to Lemma 22.3 , and they generate a -dimensionalMDLD/dimensional (vs) linear subspaceMDLD/linear subspace of . In fact, we have

Since the vectors , , are fixed vectors,MDLD/fixed vectors the together with the , , form a basis consisting of eigenvectors of . Hence, is diagonalizable.



Proof


<< | Linear algebra (Osnabrück 2024-2025)/Part I | >>
PDF-version of this lecture
Exercise sheet for this lecture (PDF)