Linear subspaces/Sum/Section
For a -vector space and a family of linear subspaces , we define the sum of these linear subspaces by
For this, we also write . The sum is again a linear subspace. In case
we say that is the sum of the linear subspaces . The following theorem describes an important relation between the dimension of the sum of two linear subspaces and the dimension of their intersection.
Let denote a field, and let denote a -vector space of finite dimension. Let denote linear subspaces. Then
Let be a basis of . On one hand, we can extend this basis, according to fact, to a basis of , on the other hand, we can extend it to a basis of . Then
is a generating system of . We claim that it is even a basis. To see this, let
This implies that the element
belongs to . From this, we get directly for , and for . From the equation before, we can then infer that also holds for all . Hence, we have linear independence. This gives altogether
The intersection of two planes
(through the origin)
in is "usually“ a line; it is the plane itself if the same plane is taken twice, but it is never just a point. This observation is generalized in the following statement.
Let be a field, and let be a -vector space of dimension . Let denote linear subspaces of dimensions and . Then
Due to fact, we have
Recall that, for a linear subspace
,
the difference is called the codimension of in . With this concept, we can paraphrase the statement above by saying that the codimension of an intersection of linear subspaces equals at most the sum of their codimensions.
Let a homogeneous system of linear equations with equations in variables be given. Then the dimension
of the solution space of the system is at least .The solution space of one linear equation in variables has dimension or . The solution space of the system is the intersection of the solution spaces of the individual equations. Therefore, the statement follows by applying fact to the individual solution spaces.