Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 9
- Base change
We know, due to Theorem 8.4 , that in a finite-dimensional vector space, any two bases have the same length, the same number of vectors. Every vector has, with respect to every basis, unique coordinates (the coefficient tuple). How do these coordinates behave when we change the bases? This is answered by the following statement.
Let be a field, and let be a -vector space of dimension . Let and denote bases of . Suppose that
with coefficients , which we collect into the -matrix
If for a basis , we consider the corresponding bijective mapping
(see
Remark 7.12
)
Let denote a field, and let denote a -vector space of dimension . Let and denote two bases of . Let
with coefficients . Then the -matrix
The -th column of a transformation matrix consists of the coordinates of with respect to the basis . The vector has the coordinate tuple with respect to the basis , and when we apply the matrix to , we get the -th column of the matrix, and this is just the coordinate tuple of with respect to the basis .
For a one-dimensional space and
we have , where the fraction is well-defined. This might help in memorizing the order of the bases in this notation.
Another important relation is
Note that here, the matrix is not applied to an -tuple of but to an -tuple of , yielding a new -tuple of . This equation might be an argument to define the transformation matrix the other way around; however, we consider the behavior in Lemma 9.1 as decisive.
In case
if is the standard basis, and some further basis, we obtain the transformation matrix of the base change from to by expressing each as a linear combination of the basis vectors , and writing down the corresponding tuples as columns. The inverse transformation matrix, , consists simply in , written as columns.
We consider in the standard basis,
and the basis
The basis vectors of can be expressed directly with the standard basis, namely
Therefore, we get immediately
For example, the vector that has the coordinates with respect to , has the coordinates
with respect to the standard basis . The transformation matrix is more difficult to compute. We have to write the standard vectors as linear combinations of and . A direct computation (solving two linear systems) yields
and
Hence,
Let be a field, and let be a -vector space of dimension . Let and denote bases of . Then the three transformation matrices fulfill the relation
Proof
- Sum of linear subspaces
For a -vector space and a family of linear subspaces , we define the sum of these linear subspaces by
For this, we also write . The sum is again a linear subspace. In case
we say that is the sum of the linear subspaces . The following theorem describes an important relation between the dimension of the sum of two linear subspaces and the dimension of their intersection.
Let denote a field, and let denote a -vector space of finite dimension. Let denote linear subspaces. Then
Let be a basis of . On one hand, we can extend this basis, according to Theorem 8.10 , to a basis of , on the other hand, we can extend it to a basis of . Then
is a generating system of . We claim that it is even a basis. To see this, let
This implies that the element
belongs to . From this, we get directly for , and for . From the equation before, we can then infer that also holds for all . Hence, we have linear independence. This gives altogether
The intersection of two planes
(through the origin)
in is "usually“ a line; it is the plane itself if the same plane is taken twice, but it is never just a point. This observation is generalized in the following statement.
Let be a field, and let be a -vector space of dimension . Let denote linear subspaces of dimensions and . Then
Due to Theorem 9.7 , we have
Recall that, for a linear subspace
,
the difference is called the codimension of in . With this concept, we can paraphrase the statement above by saying that the codimension of an intersection of linear subspaces equals at most the sum of their codimensions.
Let a homogeneous system of linear equations with equations in variables be given. Then the dimension
of the solution space of the system is at least .The solution space of one linear equation in variables has dimension or . The solution space of the system is the intersection of the solution spaces of the individual equations. Therefore, the statement follows by applying Corollary 9.8 to the individual solution spaces.
- Direct sum
Let denote a field, and let denote a -vector space. Let be a family of linear subspaces of . We say that is the direct sum of the if the following conditions are fulfilled.
- Every vector
has a representation
where .
- for all .
If the sum of the is direct, then we also write instead of . For two linear subspaces
the second condition just means .
Let denote a finite-dimensional -vector space together with a basis . Let
be a partition of the index set. Let
be the linear subspaces generated by the subfamilies. Then
The extreme case yields the direct sum
with one-dimensional linear subspaces.
Let be a finite-dimensional -vector space, and let be a linear subspace. Then there exists a linear subspace such that we have the direct sum decomposition
Let denote a basis of . We can extend this basis, according to Theorem 8.10 , to a basis of . Then
fulfills all the properties of a direct sum.
In the preceding statement, the linear subspace is called a direct complement for
(in ).
In general, there are many different direct complements.
- Direct sum and product
Recall that, for a family , , of sets , the product set is defined. If all are -vector spaces over a field , then this is, using componentwise addition and scalar multiplication, again a -vector space. This is called the direct product of vector spaces. If it is always the same space, say , then we also write . This is just the mapping space .
Each vector space is a linear subspace inside the direct product, namely as the set of all tuples
The set of all these tuples that are only (at most) at one place different from generates a linear subspace of the direct product. For infinite, it is not the direct product.
Let denote a set, and let denote a field. Suppose that, for every , a -vector space is given. Then the set
We have the linear subspace relation
If we always have the same vector space, then we write for this direct sum. In particular,
is a linear subspace. For finite, there is no difference, but for an infinite index set, this inclusion is strict. For example, is the space of all real sequences, but consists only of those sequences satisfying the property that only finitely many members are different from . The polynomial ring is the direct sum of the vector spaces . Every -vector space with a basis , , is "isomorphic“ to the direct sum .
- Footnotes
- ↑ The commutativity of such a diagram of arrows and mappings means that all composed mappings coincide as long as their domain and codomain coincide. In this case, it simply means that holds.
<< | Linear algebra (Osnabrück 2024-2025)/Part I | >> PDF-version of this lecture Exercise sheet for this lecture (PDF) |
---|