Markov Chain Properties
In the previous post, I gave the definition of Markov chain and as you can see in the page that I introduced in the previous post, a Markov chain can be represented by a directed graph. Each node represents a state and each edge of the graph represents a probability from one state to another state.
Imagine that there are various directed graphs. It means that there are various Markov chains. To understand Markov chain more deeply, you want to categorise Markov chain by statistic characteristics.
One of the most important properties that represents Markov chain statistic characteristics is ergodicity. Before understanding ergodicity, there are some other properties that you need to understand.
Reducibility
A Markov chain is said to be irreducible if it is possible to get to any state from any state.
Transience and recurrence
When you are given a state of a Markov chain, if there is a non-zero probability that you will never return to the state, the state is said to be transient.
If a state is not transient, it’s said to be recurrent.
The state A is transient even though it has a self loop. There is a probability that you will return to A when you start from A. But you will never return to A once you go to the state B. B, C, D are recurrent.
If the expected return time of a state is finite, the state is said to be positive recurrent.
To understand an expected return time, we need to define the first return time of a state i.
And we define the probability of the first return time after steps as follows:
Then we can define the expected return time as follows:
Periodicity
When you are given a state of a Markov chain, if you return to the same state after a multiples of k steps, k is said to be the period of the state.
If a state has 1 period, the state is said to be aperiodic.
Ergodicity
If a state is aperiodic and positive recurrent, the state is said to be ergodic. If all states in an irreducible Markov chain are ergodic, the Markov chain is said to be ergodic.
Stationarity
One of the remarkable characteristics of a ergodic Markov chain is that a ergodic Markov chain has a stationary distribution.
It means that regardless of the initial state, the proportion of time the chain spends in a state is going to be converge to a finite value.
Let’s give a formal definition of a stationary distribution of Markov chain.
Let is a probability of a transition from a state to another state .
A Markov chain can be represented by a transition matrix.
\begin{equation} \begin{bmatrix} p_{11} & p_{12} & \cdots & p_{1n} \\ p_{21} & p_{22} & \cdots & p_{2n} \\ \vdots & \vdots & p_{ij} & \vdots \\ p_{m1} & p_{m2} & \cdots & p_{mn} \end{bmatrix} \end{equation}
If a Markov chain is ergodic, there is a vector such as:
is said to be the stationary distribution of the Markov chain.
, which is the proportion of time the chain spends in a state i, is related to the expected return time .
( is the normalizing constant.)