Jump to content

Linear algebra (Osnabrück 2024-2025)/Part I/Lecture 4

From Wikiversity

In linear algebra, everything is worked out over a field , and the reader might think about the real numbers . But in the moment, only the algebraic properties of are relevant, so instead one can think about the rational numbers . Starting with the theory of eigenvalues, also more specific properties of the field (like the existence of roots) are important.

The "Mother of all system of linear equations“ is just one linear equation in one variable of the form

with given elements from a field , and wanted . We have three possibilities, how the solution behavior might look like. For , we can multiply the equation with the inverse of , yielding the unique solution

Numerically, one can find the solution, as long as one can find the inverse element and can perform the multiplication. For , the solution behavior depends on . If , then every is a solution, if , then there is no solution.



Linear systems

Firstly, we give three further introductory examples, one from every day's life, one from geometry and one from physics. They all lead to systems of linear equations.


At a booth on the Christmas market, there are three different pots of mulled wine. All three contain the ingredients cinnamon, cloves, red wine and sugar, but the compositions differ. The mixtures of the mulled wines are

Every mulled wine is represented by a four-tuple, where the entries represent the respective shares of the ingredients. The set of all (possible) mulled wines form a vector space  (we will introduce this concept in the next lecture) and the three concrete mulled wines are vectors in this space.

Now suppose that none of the three mulled wines meets exactly our taste, in fact the wanted mulled wine has the mixture

Is there a possibility to get the wanted mulled wine, by pouring together the given mulled wines in some way? Are there numbers[1] such that

holds? This vector-equation can be expressed by four equations in the "variables“ , where the equations come from the rows. When does there exist a solution, when none, when many? These are typical questions of linear algebra.


Two planes in space, intersecting in a line.

Suppose that two planes are given in ,[2]

and

How can we describe the intersecting line ? A point belongs to the intersection line if and only if it satisfies both plane equations. Therefore, both equations,

must hold. We multiply the first equation by , and subtract from that four times the second equation, and get

If we set , then and must hold. This means that the point belongs to . In the same way, setting , we find the point . Therefore, the intersecting line is the line connecting these points, so


An electrical network consists of several connected wires, which we call the edges of the network in this context. In every edge , there is a certain (depending on the material and the length of the edge) resistance . The points , where the edges meet, are called the vertices of the network. If we put to some edges of the network a certain electric tension (voltage), then we will have in every edge a certain current . It is helpful to assign to each edge a fixed direction, in order to distinguish the direction of the current in this edge (if the current is in the opposite direction, it gets a minus sign). We call these directed edges. In every vertex of the network, the currents of the adjacent edges come together, their sum must be . In an edge , there is a voltage drop , determined by Ohm's law to be

We call a closed directed alignment of edges in a network a mesh. For such a mesh, the sum of voltages is , unless from "outside“ a certain voltage is enforced.

We list these Kirchhoff's laws again.

  1. In every vertex, the sum of the currents equal .
  2. In every mesh, the sum of the voltages equals .
  3. If in a mesh, a voltage is enforced, then the sum of the voltages equals .

Due to "physical reasons“, we expect that, given voltages in every edge, there should be a well-defined current in every edge. In fact, these currents can be computed, if we translate the stated laws into a system of linear equations and solve this system.

In the example given by the picture, suppose that the edges (with the resistances ) are directed from left to right, and that the connecting edge from to (where the voltage is applied) is directed upwards. The four vertices and the three meshes and yield the system of linear equations

Here the and are given numbers, and the are the unknowns we are looking for.

We give now the definition of a homogeneous and of an inhomogeneous system of linear equations over a field, for a given set of variables.


Let denote a field, and let for and . We call

a (homogeneous) system of linear equations in the variables . A tuple is called a solution of the linear system, if holds for all .

If is given,[3] then

is called an inhomogeneous system of linear equations. A tuple is called a solution to the inhomogeneous linear system, if

holds for all .

The set of all solutions of the system is called the solution set. In the homogeneous case, this is also called the solution space, as it is indeed by fact a vector space.

A homogeneous system of linear equations has always the so-called trivial solution . An inhomogeneous system does not necessarily have a solution. For a given inhomogeneous linear system of equations, the homogeneous system which arises when we replace the vector on the right-hand side by the null vector , is called the corresponding homogeneous system.

The following situation describes a more abstract version of example.


Let denote a field and . Suppose that in , there are vectors (or -tuples)

given. Let

be another vector. We want to know whether can be written as a linear combination of the . Thus, we are dealing with the question whether there are elements , such that

holds. This equality of vectors means identity in every component, so that this condition yields a system of linear equations


It might happen that a system of linear equations is given in such a way that there are variables on both sides of the equations, like in

In this case, one first transforms this system to the standard form by simple additions and processing the coefficients in each equation.



Matrices

A system of linear equations can easily be written with a matrix. This allows us to make the manipulations which lead to the solution of such a system, without writing down the variables. Matrices are quite simple objects; however, they can represent quite different mathematical objects (e.g., a family of column vectors, a family of row vectors, a linear mapping, a table of physical interactions, a relation, a linear vector field, etc.), which one has to keep in mind in order to prevent wrong conclusions.


Let denote a field, and let and denote index sets. An -matrix is a mapping

If and , then we talk about an -matrix. In this case, the matrix is usually written as

We will usually restrict to this last situation.


For every , the family  , , is called the -th row of the matrix, which is usually written as a row tuple (or row vector)

For every , the family  , , is called the -th column of the matrix, usually written as a column tuple (or column vector)

The elements are called the entries of the matrix. For , the number is called the row index, and is called the column index of the entry. The position of the entry is where the -th row meets the -th column. A matrix with is called a square matrix. An -matrix is simply a column tuple (or column vector) of length , and an -matrix is simply a row tuple (or row vector) of length . The set of all matrices with rows and columns (and with entries in ) is denoted by , in case we also write .


Two matrices are added by adding entries with corresponding entries. The multiplication of a matrix with an element (a scalar) is also defined entrywise, so

and

The multiplication of matrices is defined in the following way.


Let denote a field, and let denote an -matrix and an -matrix over . Then the matrix product

is the -matrix, whose entries are given by


A matrix multiplication is only possible when the number of columns of the left-hand matrix equals the number of rows of the right-hand matrix. Just think of the scheme

the result is an -Matrix. In particular, one can multiply an -matrix with a column vector of length (the vector on the right), and the result is a column vector of length . The two matrices can also be multiplied with roles interchanged,


The -matrix

is called identity matrix.

The identity matrix has the property , for an arbitrary -matrix . Hence, the identity matrix is the neutral element with respect to matrix multiplication.


If we multiply an -matrix with an column vector , then we get

Hence, an inhomogeneous system of linear equations with disturbance vector , can be written briefly as

Then, the manipulations on the equations, which do not change the solution set, can be replaced by corresponding manipulations on the rows of the matrix. It is not necessary to write down the variables.


An -matrix of the form

is called a diagonal matrix.

Let be a field, and let be an -matrix over . Then the -matrix

is called the transposed matrix for .

The transposed matrix arises by interchanging the role of the rows and the columns. For example, we have



Footnotes
  1. In this example, only positive numbers have a practical interpretation. In linear algebra, everything is over a field, so we allow also negative numbers.
  2. Right here, we do not discuss that such equations define a plane. The solution sets are "shifted linear subspaces of dimension two“.
  3. Such a vector is sometimes called a disturbance vector of the system.


<< | Linear algebra (Osnabrück 2024-2025)/Part I | >>
PDF-version of this lecture
Exercise sheet for this lecture (PDF)