Irreducible Markov chains and regular Markov chains
Definition.
A Markov chain is called an **irreducible chain** if it is possible to go from every state to every state (not necessarily in one move).
Definition.
A Markov chain is called a **regular chain** if some power of the transition matrix has only positive elements.
Example.
Let the transition matrix of a Markov chain be defined by
If we start from state 1, then next step we jump to state 2. If we start from state 2, then we jump to state 1. In other words, we are flipping between two states after each step. So the chain is irreducible. But
So the chain is not regular.
We shall discuss two important theorems relating to regular chains. But before, we will use the following two lemmas.
Lemma.
\label{lemma:8.1}
Suppose is a random variable with finite number of outcomes, with the greatest being and the least being . Then we have
Proof ▸
Assume has outcomes, and let , then we have
◼
Lemma.
\label{lemma:8.2}
Suppose a transition matrix only has positive entries and the smallest entry is . Consider a vector . Denote its maximum value as and its minimum value as . Now consider the product . Denote its maximum value as and its minimum value as . Then we have
Proof ▸
Note that is a transition matrix. So each row of sums up to . From lemma , we note that each entry in the vector is a weighted average of the entries in , therefore, we have
The largest weighted average that could be obtained in the present case would occur if all but one of the entries of have value and one entry has value , and this one small entry is weightedby the smallest possible weight, namely . In this case, the weighted average would equal .
%
Similarly, the smallest possible weighted average equals .
%
Therefore, we have
%Since is true for any , we can take some particular form of to find the best and worst outcome of . First let such that only one entry being and such entry captures the minimum entry of . Then we can show that
%Similarly, by letting such that only one entry being and such entry captures the minimum entry of , we can obtain
- yields
◼
Now we give two important theorems.
Theorem. [Fundamental limit theorem for regular chains]
\label{thm:8.1}
Let be the transition matrix for a regular chain. Then, as , the powers approach a limiting matrix with all rows the **same** vector . The vector is a strictly positive probability vector (i.e.., the components are all **positive** and they sum to one).
Proof ▸
We consider an arbitrary vector , and denote the maximum entry of as and the minimum entry of as . From lemma 1, we must have
Writing all these inequality from 1 to gives
Therefore is a decreasing sequence bounded below and is an increasing sequence bounded above. From sequence theory we know that and must converge. Let
From Lemma we have
On the other hand, since is the minimum entry of , we must have , which gives . Thus from we have
We choose a particular form of such that with th entry being 1 and others being 0. Then gives the th column of . Using , we have
which shows the th column has the same entries.
Thus, by letting , , we conclude that each column of has the same entries. This finishes the proof.
◼
The th entry of , , is the probability that the process will be in state after steps if it starts in state . If we denote the common row of by , then Theorem states that the probability of being in in the long run is approximately , the th entry of , and is independent of the starting state.
Theorem.
\label{thm:8.2}
Let be a regular transition matrix, let
and let be the common row of . Then
-
, which means is a left eigenvector of with eigenvalue 1, and any row vector such that is a constant multiple of .
-
, which means is an eigenvector of with eigenvalue 1, and any column vector such that is a multiple of .
Proof ▸
To prove part (a), we note that from Theorem ,
Thus
But , and so and .
Let be any vector with . Then , and passing to the limit, . Let be the sum of the components of , i.e., . Then it is easily checked that . So, .
To prove part (b), assume that . Then , and again passing to the limit, . Since all rows of are the same, the components of are all equal, so is a multiple of .
◼
Note that an immediate consequence of Theorem is the fact that there is only one probability vector such that , which is . The probability vector is the vector such that .
Example.
Let be a transition matrix such that
Compute .
This is equivalent to compute . Note that is a probability vector, so .
Therefore , and .
Fixed vectors
Definition.
A row vector with the property is called a **fixed row vector** for . Similarly, a column vector such that is called a **fixed column vector** for .
Thus, the common row of is the unique vector which is both a fixed row vector for and a probability vector. Theorem shows that any fixed row vector for is a multiple of and any fixed column vector for is a constant vector. One can also state the above definition in terms of eigenvalues and eigenvectors. A fixed row vector is a left eigenvector of the matrix corresponding to the eigenvalue 1. A similar statement can be made about fixed column vectors.
The following theorem generalizes Theorem to the case where the starting state is itself
determined by a probability vector.
Theorem 8.3.
Let be the transition matrix for a regular chain and an arbitrary probability vector. Then
where is the unique fixed probability vector for .
Proof ▸
By Theorem ,
Hence
But the entries in sum to 1, and each row of equals . From these statements, it is easy to check that
◼
Does the result holds for irreducible chains?
NOT true for irreducible chains. The following is a counterexample.
Example.
Let
If we calculate the left eigenvector, we can obtain
But if we start from or , we will never goes to this equilibrium.
Definition.
The period of a state is defined by , the greatest common divisor of the epochs at which return is possible. We call state periodic if and aperiodic if .