Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 4
In linear algebra, everything is worked out over a field , and the reader might think about the real numbers . But, at the moment, only the algebraic properties of are relevant, so instead one can think about the rational numbers . Starting with the theory of eigenvalues, also more specific properties of the field (like the existence of roots) are important.
The "mother of all systems of linear equations“ is just one linear equation in one variable of the form
with given elements from a field and wanted . We have three possibilities how the solution behavior might look like. For , we can multiply the equation with the inverse of , yielding the unique solution
Computationally, one can find the solution, as long as one can find the inverse element and can perform the multiplication in the field. For , the solution behavior depends on . If , then every is a solution; if , then there is no solution.
- Linear systems
Firstly, we give three further introductory examples, one from every day's life, one from geometry, and one from physics. They all lead to systems of linear equations.
At a booth on the Christmas market, there are three different pots of mulled wine. All three contain the ingredients cinnamon, cloves, red wine, and sugar, but the compositions differ. The mixtures of the mulled wines are
Every mulled wine is represented by a four-tuple, where the entries represent the respective shares of the ingredients. The set of all (possible) mulled wines forms a vector space (we will introduce this concept in the next lecture) and the three concrete mulled wines are vectors in this space.
Now suppose that none of the three mulled wines meets exactly our taste; in fact, the wanted mulled wine has the mixture
Is there a possibility to get the wanted mulled wine by pouring together the given mulled wines in some way? Are there numbers[1] such that
holds? This vector equation can be expressed by four equations in the "variables“ , where the equations come from the rows. When does there exist a solution, when none, when many? These are typical questions of linear algebra.
Suppose that two planes are given in ,[2]
and
How can we describe the intersecting line ? A point belongs to the intersection line if and only if it satisfies both plane equations. Therefore, both equations,
must hold. We multiply the first equation by , and subtract from that four times the second equation, and get
If we set , then and must hold. This means that the point belongs to . In the same way, setting , we find the point . Therefore, the intersecting line is the line connecting these points, so
An electrical network consists of several connected wires, which we call the edges of the network in this context. In every edge , there is a certain (depending on the material and the length of the edge) resistance . The points , where the edges meet, are called the vertices of the network. If we put to some edges of the network a certain electric tension (voltage), then we will have in every edge a certain current . The goal is to determine the currents from the data of the network and the voltages.
It is helpful to assign to each edge a fixed direction in order to distinguish the direction of the current in this edge (if the current is in the opposite direction, it gets a minus sign). We call these directed edges. In every vertex of the network, the currents of the adjacent edges come together; therefore, their sum must be . In an edge , there is a voltage drop , determined by Ohm's law to be
We call a closed, directed alignment of edges in a network a mesh. For such a mesh, the sum of voltages is , unless a certain voltage is enforced from "outside“.
We list these Kirchhoff's laws again.
- In every vertex, the sum of the currents equals .
- In every mesh, the sum of the voltages equals .
- If in a mesh, a voltage is enforced, then the sum of the voltages equals .
Due to "physical reasons“, we expect that, given voltages in every edge, there should be a well-defined current in every edge. In fact, these currents can be computed if we translate the stated laws into a system of linear equations and solve this system.
In the example given by the picture, suppose that the edges (with the resistances ) are directed from left to right and that the connecting edge from to (where the voltage is applied) is directed upwards. The four vertices and the three meshes and yield the system of linear equations
Here the and are given numbers, and the are the unknowns we are looking for.
We give now the definition of a homogeneous and of an inhomogeneous system of linear equations over a field for a given set of variables.
Let denote a field, and let for and . We call
a (homogeneous) system of linear equations in the variables . A tuple is called a solution of the linear system if holds for all .
If is given,[3] then
is called an inhomogeneous system of linear equations. A tuple is called a solution to the inhomogeneous linear system if
holds for all .
The set of all solutions of the system is called the solution set. In the homogeneous case, this is also called the solution space, as it is indeed, by Lemma 6.11 , a vector space.
A homogeneous system of linear equations always has the so-called trivial solution . An inhomogeneous system does not necessarily have a solution. For a given inhomogeneous linear system of equations, the homogeneous system that arises when we replace the tuple on the right-hand side by the null vector is called the corresponding homogeneous system.
The following situation describes a more abstract version of Example 4.1 .
Let denote a field, and . Suppose that in , there are vectors (or -tuples)
given. Let
be another vector. We want to know whether can be written as a linear combination of the . Thus, we are dealing with the question whether there are elements such that
holds. This equality of vectors means identity in every component, so that this condition yields a system of linear equations
It might happen that a system of linear equations is given in such a way that there are variables on both sides of the equations, like in
In this case, one first transforms this system to the standard form by simple additions and processing the coefficients in each equation.
- Matrices
A system of linear equations can easily be written with a matrix. This allows us to make the manipulations that lead to the solution of such a system without writing down the variables. Matrices are quite simple objects; however, they can represent quite different mathematical objects (e.g., a family of column vectors, a family of row vectors, a linear mapping, a table of physical interactions, a relation, a linear vector field, etc.), which one has to keep in mind in order to prevent wrong conclusions.
We will usually restrict to this last situation.
For every , the family , , is called the -th row of the matrix, which is usually written as a row tuple (or row vector)
For every , the family , , is called the -th column of the matrix, usually written as a column tuple (or column vector)
The elements are called the entries of the matrix. For , the number is called the row index, and is called the column index of the entry. The position of the entry is where the -th row meets the -th column. A matrix with is called a square matrix. An -matrix is simply a column tuple (or column vector) of length , and an -matrix is simply a row tuple (or row vector) of length . The set of all matrices with rows and columns (and with entries in ) is denoted by ; in case we also write .
Two matrices are added by adding corresponding entries. The multiplication of a matrix with an element (a scalar) is also defined entrywise, so
and
The multiplication of matrices is defined in the following way:
Let denote a field, and let denote an -matrix and an -matrix over . Then the matrix product
is the -matrix, whose entries are given by
A matrix multiplication is only possible when the number of columns of the left-hand matrix equals the number of rows of the right-hand matrix. Just think of the scheme
the result is an -Matrix. In particular, one can multiply an -matrix with a column vector of length (the vector on the right), and the result is a column vector of length . The two matrices can also be multiplied with roles interchanged,
The -matrix
The identity matrix has the property , for an arbitrary -matrix . Hence, the identity matrix is the neutral element with respect to matrix multiplication.
If we multiply an -matrix with a column vector , then we get
Hence, an inhomogeneous system of linear equations with disturbance vector can be written briefly as
Then, the manipulations on the equations that do not change the solution set, can be replaced by corresponding manipulations on the rows of the matrix. It is not necessary to write down the variables.
An -matrix of the form
The transposed matrix arises by interchanging the roles of the rows and the columns. For example, we have
- Footnotes
- ↑ In this example, only positive numbers have a practical interpretation. In linear algebra, everything is over a field, so we also allow negative numbers.
- ↑ Right here, we do not discuss that such equations define a plane. The solution sets are "shifted linear subspaces of dimension two“.
- ↑ Such a vector is sometimes called a disturbance vector of the system.
<< | Linear algebra (Osnabrück 2024-2025)/Part I | >> PDF-version of this lecture Exercise sheet for this lecture (PDF) |
---|