Markov Chains Limiting Probabilities

Limiting Probabilities

Definition 1: Suppose that the limit on the left side of the following equation exists for all r, s in S and suppose further that

Stable limits

for any r and u in S. Then we can define the limiting probability at s to be

Limiting probability

Property 1: The limiting probabilities ps are the unique solution to the equations (for all s in S)

Limiting probabilities property

Proof: By Property 1 of Markov Chains Basic Concepts

Proof 1

Proof 2

Finally, we need to show that the system of linear equations has a unique solution. Let qs for all s in S be a solution to the equations and let Q be the n × 1 vector Q = [qs], where n = the number of states in S, and let P be the n × n matrix P = [prs]. Since

q_s formula

we see that Q = QP. It then follows that QP = QP2 and hence Q = QP2. In a similar fashion it follows that Q = QPk. This means that

new q_s formula

From this and the assumption that

Assumption

it follows that

Completion of the proof

This completes the proof.

Property 2: For any states r and s, the following limit always exists

Cesaro limit

Proof: In general, for any sequence a1, a2, …, the Cesàro limit

Cesàro limit

exists and is equal to the limit of an as n → ∞ if this ordinary limit exists.

Periodic chains

Definition 2: A Markov chain is periodic if there is a non-empty partition of S = S1S2 ⋃ … ⋃ Sm with m > 1 and SiSj = ∅ for all ij such that if rSi with i < m then

Periodic characteristic

(i.e. prs = 0 for any s not in Si+1) and for any rSm then

Another periodic characteristic

(i.e. prs = 0 for any s not in S1).

For example, S = {1, 2} where p12 = p21 = 1 (and so p11 = p22 = 0) is a trivial example of a periodic Markov chain.

Stationary chains

Definition 3: A probability distribution f(s) = qs on S is an equilibrium (aka stationary) distribution of a Markov chain if

Equilibrium distribution

The first set of equations are called the equilibrium equations and the last equation is called the normalizing equation. If S has n elements, then these equations can be re-expressed as n-1 equation in n-1 unknowns.

Let Q be the n × 1 vector Q =[qs] and let P be the n × n matrix P = [prs]. Then, as we saw in the proof of Property 3, the equilibrium equations can be expressed as Q = QP.

The reason for calling qs a stationary distribution is explained in the following property.

Property 3: If qs is an equilibrium distribution and P(x0 = s) = qs for all s, then for all i and all s, P(xi = s) = qs.

Proof: The proof is by induction on i. The case i = 0 is obvious. Assume the property for i. Now

Stationary property

Observation: If the limiting probabilities exist for all states s, then by Property 1, f(s) = ps is the unique equilibrium distribution. Unfortunately, these limiting probabilities don’t always exist; e.g. if the set of states is periodic.

Definition 4: A non-empty set of states C is closed provided there is no transition out of C, i.e. for all r in C and s not in C, prs = 0.

Property 4: A (finite-state) Markov chain x0, x1, … which does not contain two or more disjoint closed sets has a unique equilibrium distribution qs such that for any states s

q_s as a limit

where qs is independent of the starting state r.

Furthermore, if the Markov chain is not periodic, then the ps exist, and so qs = ps.

Observations

Note that the Markov chain in Example 2 of Markov Chain Examples is not periodic and does not contain two or more disjoint closed sets. Thus the ps exist and satisfy the equilibrium equations of Definition 3. This explains why the  arising from high powers of Pk (corresponding to the limit of the pkrs as k → ∞) were the same as the solutions to the simultaneous linear equations (as in Definition 3).

Note too that if the Markov chain has more than one disjoint closed sets, then each such closed set would satisfy Property 4.

References

Fewster, R. (2014) Markov chains
https://www.stat.auckland.ac.nz/~fewster/325/notes/ch8.pdf

Pishro-Nik, H. (2014) Discrete-time Markov chains
https://www.probabilitycourse.com/chapter11/11_2_1_introduction.php

Norris, J. (2004) Discrete-time Markov chains
https://www.statslab.cam.ac.uk/~james/Markov/

Lalley, S. (2016) Markov chains
https://galton.uchicago.edu/~lalley/Courses/312/MarkovChains.pdf

Konstantopoulos, T. (2009) Markov chains and random walks
https://www2.math.uu.se/~takis/L/McRw/mcrw.pdf

Wikipedia (2025) Cesàro summation
https://en.wikipedia.org/wiki/Ces%C3%A0ro_summation

Leave a Comment