Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 10
- Linear mappings
We are interested in mappings between two vector spaces that respect the structures, that is, they are compatible with addition and with scalar multiplication.
Let be a field, and let and be -vector spaces. A mapping
is called a linear mapping if the following two properties are fulfilled.
- for all .
- for all and .
Here, the first property is called additivity and the second property is called compatibility with scaling. When we want to stress the base field, then we say -linear. The identity , the null mapping , and the inclusion of a linear subspace are the simplest examples of a linear mapping. For a linear mapping, the compatibility with arbitrary linear combination holds, that is,
see Exercise 10.2 . Instead of linear mapping, we also say homomorphism.
The easiest linear mappings are (beside the null mapping) the linear maps from to . Such a linear mapping
is determined (by Theorem 10.10 , but this is also directly clear) by , or by the value for a single element , . In particular, , with a uniquely determined . In the context of physics, for , and if there is a linear relation between two measurable quantities, we talk about proportionality, and is called the proportionality factor. In school, such a linear relation occurs as "rule of three“.
Many important functions, in particular from to , are not linear. For example, the squaring , the square root, the trigonometric functions, the exponential function, the logarithm is not linear. But also for such more complicated functions there are, in the framework of differential calculus, linear approximations, which help to understand these functions.
Let denote a field, and let be the -dimensional standard space. Then the -th projection, this is the mapping
is a -linear mapping. This follows immediately from componentwise addition and scalar multiplication on the standard space. The -th projection is also called the -th coordinate function.
In a shop, there are different products to buy, and the price of the -th product (with respect to a certain unit) is . A purchase is described by the -tuple
where is the amount of the -th product bought. The price for the purchase is hence . The price mapping
is linear. This means, for example, that if we do first the purchase and then, a week later, the purchase , then the price of the two purchases together is the same as the price of the purchase .
Let be a vector space over a field . For , the linear mapping
is called homothety (or dilation)
with scaling factor .For a homothety, the domain space and the target space are the same. The number is called scaling factor. For , we get the identity, for , we get the point reflection at the origin.
Let denote the space of continuous functions from to , and let denote the space of continuously differentiable functions. Then the mapping
which assigns to a function its derivative, is linear. In analysis, we prove that
holds for and another function .
Let denote a field, and let denote vector spaces over . Suppose that
are linear mappings. Then also the composition
Proof
Let be a field, and let and be -vector spaces. Let
be a bijective linear map. Then also the inverse mapping
Proof
- Determination on a basis
Behind the following statement (the determination theorem) is the important principle, that in linear algebra (of finite dimensional vector spaces), the objects are determined by finitely many data.
Let be a field, and let and be -vector spaces. Let , , denote a basis of , and let , , denote elements in . Then there exists a unique linear mapping
with
Since we want , and since a linear mapping respects all linear combinations, that is [1]
holds, and since every vector
is such a linear combination, there can exist at most one such linear mapping.
We define now a
mapping
in the following way: we write every vector with the given basis as
(where for almost all ) and define
Since the representation of as such a
linear combination
is unique, this mapping is well-defined. Also,
is clear.
Linearity. For two vectors
and ,
we have
The compatibility with scalar multiplication is shown in a similar way, see
Exercise 10.21
.
In particular, a linear mapping
is uniquely determined by .
In many situations, a certain object (like a cube) in space shall be drawn in the plane . One possibility is to work a projection. This is a linear mapping
which is given (with respect to the standard bases and ) by
where the coefficients are usually chosen in the range . Linearity has the effect that parallel lines are mapped to parallel lines (unless they are mapped to a point). The point is mapped to . The image of the object under such a linear mapping is called a projection image.
- Linear mappings and matrices
A linear mapping
is determined uniquely by the images , , of the standard vectors, and every is a linear combination
and hence determined by the elements . This means all together that such a linear mapping is given by the elements , , . Such a set of data can be written as a matrix. Due to Theorem 10.10 , this observation holds for all finite-dimensional vector spaces, as long as bases are fixed on the source space and on the target space of the linear mapping.
Let denote a field, and let be an -dimensional vector space with a basis , and let be an -dimensional vector space with a basis .
For a linear mapping
the matrix
where is the -th coordinate of with respect to the basis , is called the describing matrix for with respect to the bases.
For a matrix , the linear mapping determined by
in the sense of Theorem 10.10 ,
is called the linear mapping determined by the matrix .For a linear mapping , we always assume that everything is with respect to the standard bases, unless otherwise stated. For a linear mapping from a vector space in itself (what is called an endomorphism), one usually takes the same bases on both sides. The identity on a vector space of dimension is described by the identity matrix, with respect to every basis.
If , then we are usually interested in the describing matrix with respect to one basis of .
Let denote a vector space with bases and . If we consider the identity
with respect to the basis on the source and the basis on the target, we get, because of
directly
This means that the describing matrix of the identical linear mapping is the transformation matrix for the base change from to .
Let denote a field and let denote an -dimensional vector space with a basis . Let be an -dimensional vector space with a basis , and let
and
be the corresponding mappings. Let
denote a linear mapping with describing matrix . Then
hold, that is, the diagram
commutes. For a vector ,
we can compute by determining the coefficient tuple of with respect to the basis , applying the matrix and determining for the resulting -tuple the corresponding vector with respect to .Proof
Let be a field, and let be an -dimensional vector space with a basis , and let be an -dimensional vector space with a basis . Then the mappings
defined in definition, are inverse
to each other.We show that both compositions are the identity. We start with a matrix and consider the matrix
Two matrices are equal, when the entries coincide for every index pair . We have
Now, let be a linear mapping, we consider
Two linear mappings coincide, due to Theorem 10.10 , when they have the same values on the basis . We have
Due to the definition, the coefficient is the -th coordinate of with respect to the basis . Hence, this sum equals .
We denote the set of all linear mappings from to by .
Theorem 10.15
means that the mapping
is bijective with the given inverse mapping. A linear mapping
is called an endomorphism. The set of all endomorphisms on is denoted by .
- Isomorphic vector spaces
Let denote a field and let and denote -vector spaces. A bijective linear mapping
An isomorphism from to is called automorphism.
Let denote a field. Two -vector spaces and are called isomorphic, if there exists an isomorphism
from to .Let denote a field and let and denote finite-dimensional -vector spaces. Then and are isomorphic to each other if and only if their dimension
coincides. In particular, an -dimensional -vector space is isomorphic to .Proof
An isomorphism between an -dimensional vector space and the standard space is essentially equivalent with the choice of a basis of . For a basis
we associate the linear mapping
which maps from the standard space to the vector space by sending the -th standard vector to the -th basis vector of the given basis. This defines a unique linear mapping due to Theorem 10.10 . Due to Exercise 10.23 , this mapping is bijective. It is just the mapping
The inverse mapping
is also linear, and it is called the coordinate mapping for this basis. The -th component of this map, that is, the composed mapping
is called the -th coordinate function. It is denoted by . It assigns to vector with the unique representation
the coordinate . Note that the linear mapping depends on the basis, not only on the vector .
If an isomorphism
is given, then the images
form a basis of .
- Footnotes
- ↑ If is an infinite index set, then, in all sums considered here, only finitely many coefficients are not .
<< | Linear algebra (Osnabrück 2024-2025)/Part I | >> PDF-version of this lecture Exercise sheet for this lecture (PDF) |
---|