Jump to content

Matrix/Stochastic/Powers/Section

From Wikiversity

We investigate now the powers of a stochastic matrix with the help of the sum norm.


For an arbitrary vector , we have

Iterative application of this observation shows that fact  (2) is satisfied.



Let

be a real square matrix with non-negative entries. Then is column stochastic if and only if is isometric with respect to the sum norm for vectors with non-negative entries, that is,

holds for all

.

Let be a column stochastic matrix, and

be a vector with non-negative entries. Then

holds. If the described isometric property holds, then it holds in particular for the images of the standard vectors. That means that their sum norm equals . These images are the corresponding columns of the matrix; therefore, all column sums are .



Let be a

column stochastic matrix. Then the following statements hold.
  1. There exists eigenvectors to the eigenvalue .
  2. If there exists a row such that all its entries are positive, then for every vector that has a positive and also a negative entry, the estimate

    holds.

  3. If there exists a row such that all its entries are positive, then the eigenspace of the eigenvalue is one-dimensional. There exists an eigenvector where all entries are no-negative; in particular, there is a uniquely determined stationary distribution.
  1. The transposed matrix is row stochastic; therefore, it has an eigenvector to the eigenvalue . Due to fact, the characteristic polynomial of the transposed matrix has a zero at . Because of exercise, this also holds for the characteristic polynomial of the matrix we have started with. Hence, has an eigenvector to the eigenvalue .
  2. We now assume also that all entries of the -th row are positive, and let denote a vector with (at least) a positive and a negative entry. Then
    holds.
  3. As in the proof of (2), let all entries of the -th row be positive. For any eigenvector to the eigenvalue , according to (2), either all entries are non-negative, or non-positive. Hence, for such a vector, because of , its -th entry is not . Let be such eigenvectors. Then belongs to the fixed space. However, the -th component of this vector equals ; therefore, it is the zero vector. This means that and are linearly dependent. Therefore, this eigenspace is one-dimensional. Because of (2), there exists an eigenvector to the eigenvalue with non-negative entries. By normalizing, we get a stationary distribution.



We consider the column stochastic -matrix

here, all entries of the first row are positive. Due to fact, there exists a unique eigendistribution. In order to determine this distribution, we compute the kernel of

This kernel is generated by , and the stationary distribution is


For the column stochastic -matrix

the eigenspace to the eigenvalue equals ;, it is two-dimensional. This shows that the conclusion of fact does not hold even when there exists a column (but not a row) with strictly positive entries.


Let be a column stochastic matrix, fulfilling the property that there exists a row in which all entries are positive. Then for every distribution vector with , the sequence converges to the uniquely determined stationary distribution

of .

Let be the stationary distribution, which is uniquely determined because of fact  (3). We set

This is a linear subspace of of dimension . Due to fact  (2), has only non-negative entries; therefore, it does not belong to . Because of

is invariant under the matrix . Hence,

is a direct sum decomposition into invariant linear subspaces. For every with , we have

due to fact  (2). The sphere of radius is compact with respect to every norm; therefore, the induced maximum norm of is smaller than . Because of fact and fact, the sequence converges for every to the zero vector.

Let now be a distribution vector; because of

we can write

with . Because of

and the reasoning before, this sequence converges to .



In the situation of fact, we can find the eigendistribution by solving a system of linear equations. If we are dealing with a huge matrix (think about vertices), then such a computation time-consuming. Often, it is not necessary to know the eigendistribution precisely; it is enough to know a good approximation. For this, we may start with an arbitrary distribution, and we compute finitely many iterations. Because of fact, we know that this method gives arbitrarily good approximations of the eigendistribution. For example, a search engine for the web generates for search item an ordered list of web pages where this search item occurs. How does this ordering arise? The true answer is, at least for the first entries, that it depends on how much someone has paid. Despite of this, it is a natural approach, and this is also the basis for the Page ranks, to consider the numerical ordering in the eigendistribution. The first entry is the one where most people would "finally“ end up when they follow with the same probability any possible link. This movement is modelled[1] by the stochastic matrix described in example.

The numerical difference between finding an exact solution of a system of linear equations to determine an eigenvector, and the power method, can be grasped in the following way. Let an -matrix be given. The elimination process needs in order to eliminate the first variable in equations, multiplications (here, we do not consider the easier additions); therefore, the seize of the multiplications in the complete elimination process is

For the evaluation of the matrix at a vector, multiplications are necessary. If we want to compute iterations, we need operations. Hence, if is substantially smaller than , then thetotal expenditure is substantially smaller.

  1. Modelling means in (in particular, applied) mathematics the process to understand phenomena of the real world with mathematical tools. We model physical processes, weather phenomena, transactions in finance, etc.