Mathematics for Applied Sciences (Osnabrück 2023-2024)/Part I/Lecture 24
- Base change
We know, due to Theorem 23.15 , that in a finite-dimensional vector space, any two bases have the same length, the same number of vectors. Every vector has, with respect to every basis, unique coordinates (the coefficient tuple). How do these coordinates behave when we change the bases? This is answered by the following statement.
Let be a field, and let be a -vector space of dimension . Let and denote bases of . Suppose that
with coefficients , which we collect into the -matrix
The matrix , which describes the base change from to , is called the transformation matrix. In the -th column of the transformation matrix, there are the coordinates of with respect to the basis . When we denote, for a vector
and a basis , the corresponding coordinate tuple by , then the transformation can be quickly written as
We consider in the standard basis,
and the basis
The basis vectors of can be expressed directly with the standard basis, namely
Therefore, we get immediately
For example, the vector which has with respect to the coordinates , has the coordinates
with respect to the standard basis . The transformation matrix is more difficult to compute: We have to write the standard vectors as linear combinations of and . A direct computation (solving two linear systems) yields
and
Hence,
- Linear mappings
Let be a field, and let and be -vector spaces. A mapping
is called a linear mapping, if the following two properties are fulfilled.
- for all .
- for all and .
Here, the first property is called additivity and the second property is called compatibility with scaling. When we want to stress the base field, then we say -linearity. The identity , the null mapping and the inclusion of a linear subspace are the simplest examples for linear mappings. For a linear mapping, the compatibility with arbitrary linear combination holds, that is,
Let denote a field, and let be the -dimensional standard space. Then the -th projection, this is the mapping
is a -linear mapping. This follows immediately from componentwise addition and scalar multiplication on the standard space. The -th projection is also called the -th coordinate function.
Let denote a field, and let denote vector spaces over . Suppose that
are linear mappings. Then also the composition
Proof
Let be a field, and let and be -vector spaces. Let
be a bijective linear map. Then also the inverse mapping
Proof
- Determination on a basis
Behind the following statement (the determination theorem), there is the important principle that in linear algebra (of finite-dimensional vector spaces), the objects are determined by finitely many data.
Let be a field, and let and be -vector spaces. Let , , denote a basis of , and let , , denote elements in . Then there exists a unique linear mapping
with
Proof
The easiest linear mappings are (beside the null mapping) the linear maps from to . Such a linear mapping
is determined (by Theorem 24.7 , but this is also directly clear) by , or by the value for a single element , . In particular, , with a uniquely determined . In the context of physics, for , and if there is a linear relation between two measurable quantities, we talk about proportionality, and is called the proportionality factor. In school, such a linear relation occurs as "rule of three“.
- Linear mappings and matrices
Due to Theorem 24.7 , a linear mapping
is determined by the images , , of the standard vectors. Every is a linear combination
and therefore the linear mapping is determined by the elements . So, such a linear map is determined by the elements , , , from the field. We can write such a data set as a matrix. Because of the determination theorem, this holds for linear maps in general, as soon as in both vector spaces bases are fixed.
Let denote a field, and let be an -dimensional vector space with a basis , and let be an -dimensional vector space with a basis .
For a linear mapping
the matrix
where is the -th coordinate of with respect to the basis , is called the describing matrix for with respect to the bases.
For a matrix , the linear mapping determined by
in the sense of Theorem 24.7 ,
is called the linear mapping determined by the matrix .For a linear mapping , we always assume that everything is with respect to the standard bases, unless otherwise stated. For a linear mapping from a vector space in itself (what is called an endomorphism), one usually takes the same bases on both sides. The identity on a vector space of dimension is described by the identity matrix, with respect to every basis.
Let be a field, and let be an -dimensional vector space with a basis , and let be an -dimensional vector space with a basis . Then the mappings
defined in definition, are inverse
to each other.Proof
is usually described by the matrix with respect to the standard bases on the left and on the right. The result of the matrix multiplication
can be interpreted directly as a point in . The -th column of is the image of the -th standard vector .
- Rotations
A rotation of the real plane around the origin, given the angle counterclockwise, maps to and to . Therefore, plane rotations are described in the following way.
which is given by a rotation matrix (with some )with respect to the standard basis is called
rotation.A space rotation is a linear mapping of the space in itself around a rotation axis (a line through the origin) with an certain angle . If the vector defines the axis, and and are orthogonal to and to each other, and all have length , then the rotation is described by the matrix
with respect to the basis .
- The kernel of a linear mapping
Let denote a field, let and denote -vector spaces, and let
denote a -linear mapping. Then
The kernel is a linear subspace of .
The following criterion for injectivity is important.
Let denote a field, let and denote -vector spaces, and let
denote a -linear mapping. Then is injective if and only if
holds.If the mapping is injective, then there can exist, apart from
,
no other vector
with
.
Hence,
.
So suppose that
,
and let
be given with
.
Then, due to linearity,
Therefore,
,
and so
.
<< | Mathematics for Applied Sciences (Osnabrück 2023-2024)/Part I | >> PDF-version of this lecture Exercise sheet for this lecture (PDF) |
---|