Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 12
- Invertible matrices
Let denote a field. For an invertible matrix , the matrix fulfilling
is called the inverse matrix of . It is denoted by
The product of invertible matrices is again invertible. Due to Lemma 9.5 , the matrix describing a base change is invertible, and the matrix of the reversed base change is its inverse matrix.
For a field and , the set of all invertible -matrices with entries in is called the general linear group
over . It is denoted by .Two square matrices are called similar, if there exists an invertible matrix with
.For a linear mapping , the describing matrices with respect to two bases are similar to each other, due to Corollary 11.12 .
- Properties of linear mappings
Let be a field, and let and be vector spaces over of dimensions and . Let
be a linear map, described by the matrix
with respect to two bases. Then the following properties hold.- is injective if and only if the columns of the matrix are linearly independent.
- is surjective if and only if the columns of the matrix form a generating system of .
- Let . Then is bijective if and only if the columns of the matrix form a basis of , and this holds if and only if is invertible.
Let and denote the bases of and respectively, and let denote the column vectors of . (1). The mapping has the property
where is the -th entry of the -th column vector . Therefore,
This is if and only if for all , and this is equivalent with
For this vector equation, there exists a nontrivial tuple , if and only if the columns are linearly dependent, and this holds if and only if the
kernel
of is not trivial. Due to
Lemma 11.4
,
this is equivalent with not being injective.
(2). See
Exercise 12.7
.
(3). Let
.
The first equivalence follows from (1) and (2). If is bijective, then there exists a
(linear)
inverse mapping
with
Let denote the matrix for , and the matrix for . The matrix for the identity is the identity matrix. Because of Lemma 11.10 , we have
and therefore is invertible. The reverse implication is proved similarly.
- Elementary matrices
We want develop effective methods to determine the rank of a matrix, and to decide whether a matrix is invertible. For this, elementary matrices are helpful.
In detail, these elementary matrices look as follows.
Elementary matrices are invertible, see Exercise 12.1 .
Let be a field and a -matrix with entries in . Then the multiplication by elementary matrices from the left with has the following effects.
- exchange of the -th and the -th row of .
- multiplication of the -th row of by .
- addition of -times the -th row of to the -th row ().
Proof
Elementary row operations do not change the solution space of a homogeneous linear system, as shown in
Lemma 5.3
.
Let be a field, and let denote an -matrix over . Then there exist elementary row operations, and a (new) numbering of the columns
and an such that, in the new matrix, the columns have the form
and
By further elementary row operations, and by swapping of columns, the matrix can be brought to the form
with
.
Let be a field, and let denote an invertible -matrix over . Then there exist elementary row operations such that, after these manipulations, a matrix of the form
with arises. By further elementary row operation, one can also obtain the
identity matrix.This rests on the manipulations of the elimination procedure, and on the fact that elementary row manipulations are achieved, due to Lemma 12.18 , by multiplications with elementary matrices from the left. In doing this, it can not happen that a zero-column or a zero-row arises, because the elementary matrices are invertible, and, in each step, invertibility is preserved. If we have an invertible upper triangular matrix, then the diagonal entries are not by Exercise 12.6 , and, by multiplication with a scalar, we can normalize them to . With this, we can further achieve, in every column, that all entries above the diagonal entry are .
In particular, for an invertible matrix , there are elementary matrices such that
is the identity matrix.
- Finding the inverse matrix
Let denote a square matrix. How can we decide whether the matrix is invertible, and how can we find the inverse matrix ?
For this we write down a table, on the left-hand side we write down the matrix , and on the right-hand side we write down the identity matrix (of the right size). Now we apply on both sides step by step the same elementary row manipulations. The goal is to produce in the left-hand column, starting with the matrix, in the end the identity matrix. This is possible if and only if the matrix is invertible. We claim that we produce, by this method, in the right column the matrix in the end. This rests on the following invariance principle. Every elementary row manipulation can be realized, according to Lemma 12.18 , as a matrix multiplication with some elementary matrix from the left. If in the table we have somewhere the pair
after the next step (in the next line) we have
If we multiply the inverse of the second matrix (which we do not know yet; however, we do know its existence, in case the matrix is invertible) with the first matrix, then we get
This means that this expression is not changed in each single step. In the beginning, this expression equals , hence in the end, the pair must fulfil
We want to find for the matrix its inverse matrix , following Method 12.11 .
- Rank of matrices
Let be a field, and let denote an -matrix over . Then the dimension of the linear subspace of , generated by the columns, is called the column rank of the matrix, written
Let denote a field, and let and denote -vector spaces of dimensions and . Let
be a linear mapping, which is described by the matrix , with respect to bases of the spaces. Then
Proof
To formulate the next statement, we introduce row rank of an -matrix to be the dimension of the linear subspace of generated by the rows.
Let be a field, and let denote an -matrix over . Then the column rank coincides with the row rank. The rank equals the number from
Theorem 12.9 .In an elementary row manipulation, the linear subspace generated by the rows is not changed, therefore the row rank is not changed. The row rank of equals the row rank of the matrix in echelon form obtained in Theorem 12.9 . This matrix has row rank , since the first rows are linearly independent, and, apart from this, this, there are only zero rows. It has also column rank , since the columns, where there is a new step, are linearly independent, and the other columns are linear combinations of these columns. By Exercise 12.20 , the column rank is preserved by elementary row manipulations.
Both ranks coincide, so we only talk about the rank of a matrix.
Let be a field, and let denote an -matrix
over . Then the following statements are equivalent.- is invertible.
- The rank of is .
- The rows of are linearly independent.
- The columns of are linearly independent.
<< | Linear algebra (Osnabrück 2024-2025)/Part I | >> PDF-version of this lecture Exercise sheet for this lecture (PDF) |
---|