Lecture-16: Invariant Distribution
1 Transient and recurrent states
1.1 Hitting and return times
Definition 1.1. For a homogeneous Markov chainX:Ω→XZ+, we can definefirst hitting timeto state x∈X, as
τx+≜inf{n∈N:Xn=x}. IfX0=x, thenτx+is called thefirst return timeto statex.
Lemma 1.2. For an irreducible Markov chain X:Ω→XZ+ on finite state spaceX, we haveExτy+<∞for all states x,y∈X.
Proof. From the definition of irreducibility, for each pair of statesz,w∈X, we have a positive integer nzw∈Nsuch thatpnzwzw>ϵzw>0. Since the state spaceXis finite, We define
ϵ≜ inf
z,w∈Xϵzw>0, r≜ sup
z,w∈Xnzw∈N.
Hence, there exists a positive integerr∈Nand a realϵ>0 such that p(n)zw >ϵfor somen⩽rand all statesz,w∈X. It follows thatP(∪n∈[r]{Xn=y})>ϵor Pz
n
τy+>ro
⩽1−ϵfor any initial condition X0=z∈X. Therefore, we can write fork∈N
Px
n
τy+>kro
=Px
n
τy+>(k−1)ro
P(nτy+>kro
|nτy+>(k−1)r,X0=xo
)⩽(1−ϵ)Px
n
τy+>(k−1)ro . By induction, we havePx
n
τy+>kro
⩽(1−ϵ)k. SincePx
n
τy+>no
is decreasing inn, we can write Exτy+=
∑
k∈Z+ r−1
∑
i=0
Px{τy+>kr+i}⩽
∑
k∈Z+
rPx{τy+>kr}⩽ r ϵ<∞.
Corollary 1.3. For an irreducible Markov chain X:Ω→XZ+on finite state spaceX, we have Px
n
τy+<∞o=1 for all states x,y∈X.
Proof. This follows from the fact thatτy+ is a positive random variable with finite mean for all states y∈Xand any initial statex∈X.
1.2 Recurrence and transience
Definition 1.4. Let fxy(n)denote the probability that starting from statex, the first transition into statey happens at timen. Then, fxy(n)=Px
n
τy+=no
. Then we denote the probability of eventually entering stateygiven that we start at statex, asfxy=∑∞n=1fxy(n)=Px
n
τy+<∞o. The stateyis said to betransient if fyy<1 andrecurrentif fyy=1.
Definition 1.5. For a discrete time processX:Ω→XZ+, the total number of visits to a statey∈Xin first nsteps is denoted by Ny(n)≜∑ni=11{Xi=y}. Total number of visits to statey∈Xis denoted by Ny≜Ny(∞).
Remark 1. From the linearity of expectations and monotone convergence theorem, we get EyNy =
∑n∈Np(n)yy.
1
Lemma 1.6. Consider a homogeneous Markov chain X:Ω→XZ+. For each m∈Z+ and state x,y∈X, we have
Px
Ny=m =
(1− fxy m=0, fxyfyym−1(1−fyy) m∈N.
Proof. For eachk∈N, the time τy+(k)of the kth visit to the statey is a stopping time. From strong Markov property, the next return to statey is independent of the past. That is, (τy+(k+1)−τy+(k): k∈N)is ani.i.d. sequence, distributed identically toτy+ starting from an initial stateX0=y. When X0=x̸=y, thenτy+is independent of sequence(τy+(k+1)−τy+(k):k∈N)and distributed differently.
We observe that
Ny=m =nτy+(m)<∞,τy+(m+1) =∞o=∩mk=1nτy+(k)−τy+(k−1)<∞o∩nτy+(m+1)−τy+(m) =∞o. It follows from the strong Markov property for processX, that
Px
Ny=m =Px
n
τy+(1)<∞o m
k=2
∏
Py
n
τy+(k)−τy+(k−1)<∞o Py
n
τy+(m+1) =∞o .
Corollary 1.7. For a homogeneous Markov chain X:Ω→XZ+, we have Py
Ny<∞ =1{fyy<1}.
Proof. We can write the event
Ny<∞ as the disjoint union of events
Ny=m form∈Z+, and the result follows from additivity of probability over disjoint events, and the expression for the conditional probability mass functionPy
Ny=m in Lemma 1.6.
Remark2. In particular, this corollary implies the following.
1. A transient state is visited a finite amount of times almost surely.
2. A recurrent state is visited infinitely often almost surely.
3. Since∑y∈XNy=∞, it follows that all states can be transient in a finite state Markov chain.
Proposition 1.8. A state y∈Xis recurrent iff∑k∈Np(k)yy =∞.
Proof. For any statey∈X, we can write
p(k)yy =Px{Xk=y}=Ex1{X
k=y}.
Using monotone convergence theorem to exchange expectation and summation, we obtain
k∈
∑
Np(k)yy =Ey
∑
k∈N
1{Xk=y}=EyNy.
Thus,∑k∈Np(k)yy represents the expected number of returnsEyNyto a stateystarting from statey, which we know to be finite if the state is transient and infinite if the state is recurrent.
Proposition 1.9. Transience and recurrence are class properties.
Proof. Let us start with proving recurrence is a class property. Letxbe a recurrent state and letx↔y.
Then, we will show thatyis a recurrent state. From the reachability, there exist somem,n>0, such that p(m)xy >0 andp(n)yx >0. As a consequence of the recurrence,∑s∈Z+p(s)xx =∞. It follows thatyis recurrent by observing
k∈
∑
Z+p(k)yy ⩾
∑
s∈Z+
p(m+n+s)yy ⩾
∑
s∈Z+
p(n)yx p(s)xxp(m)xy =∞.
Now, ifxwere transient instead, we conclude thatyis also transient by the following observation
s∈
∑
Z+p(s)yy ⩽∑s∈Z+p
(m+n+s) xx
p(n)yx p(m)xy
<∞.
Corollary 1.10. If y is recurrent, then for any state x such that x↔y, fxy=1.
2
2 Invariant distribution
Definition 2.1. For a time-homogeneous Markov chainX:Ω→XZ+ with transition matrixP, a distri- butionπ∈ M(X)is calledinvariantif it is a left eigenvector of the probability transition matrixPwith eigenvalue unity, or
π=πP.
Remark 3. Recall that ν(n)∈ M(X) where νx(n) =P{Xn=x}for all x∈X, denotes the probability distribution of the Markov chainXbeing in one of the states at stepn∈N. Then, if ν(0) =π, then ν(n) =ν(0)Pn=πfor all time-stepsn∈N.
Definition 2.2. For a time-homogeneous Markov chain X:Ω →XZ+ with transition matrix P, the stationary distribution is defined asν(∞)≜limn→∞ν(n).
Remark 4. For a Markov chain with initial distribution being invariant, the stationary distribution is invariant distribution.
Example 2.3 (Simple random walk on a directed graph). LetG= (V,E)be a finite directed graph.
We define a simple random walk on this graph as a Markov chain with state spaceVand transition matrixP:V×V→[0, 1]wherepxy≜deg1
out(x)1{(x,y)∈E}. We observe that vector(degout(x):x∈X) is a left eigenvector of the transition matrixPwith unit eigenvalue. Indeed we can very that
x∈
∑
Xdegout(x)pxy=
∑
x∈X1{(x,y)∈E}=degout(y).
Since∑x∈Xdegout(x) =2|E|, it follows thatπ:X→[0, 1]defined byπx≜ deg2|E|out(x)for eachx∈V, is the equilibrium distribution of this simple random walk.
2.1 Existence of an invariant distribution
Proposition 2.4. Consider an irreducible and aperiodic homogeneous DTMC X:Ω→XZ+ with transition matrix P and starting from initial state X0=x. Let the positive vectorπ˜x:X→R+defined as
˜
πx(y)≜Ex τx+ n=1
∑
1{Xn=y}=Ex
∑
n∈N1{n⩽τx+}1{Xn=y}, y∈X.
Thenπ˜x=π˜xP if Px{τx+<∞}=1, andπ≜Eπ˜x
xτx+ is a stationary distribution ifExτx+<∞.
Proof. We will first show thatπis a distribution on state spaceX. We first observe that
y∈
∑
X˜
πx(y) =
∑
y∈X τx+ n=1
∑
1{Xn=y}=
τx+ n=1
∑
1{Xn∈X}=Exτx+.
Thus ˜πx(y) =Ex∑τn=1x+ 1{Xn=y}⩽Exτx+for all statesy∈X. IfExτx+<∞, then ˜πx(y)<∞for eachy∈X.
Further, we have ˜πx(x) =1. Since ˜πx(y)⩾0, it follows that Eπ˜x
xτx+ is a distribution on the state spaceX.
We next show that ˜πxis an invariant distribution of DTMCX. Using the monotone convergence theorem, we can write
w∈
∑
X˜
πx(w)pwz=
∑
n∈N
∑
w∈X
Px
τx+⩾n,Xn=w P({Xn+1=z} | {Xn=w}).
We first focus on the termw=x. We see that{Xn=x,τx+⩾n}={τx+=n}. Hence, from the strong Markov property, we havePx{Xn=x,Xn+1=z,τx+⩾n}=Px{τx+=n}pxz. Summing over alln∈N, we get
˜
πx(x)pxz=
∑
n∈N
Px
Xn=x,Xn+1=z,τx+⩾n =pxz
∑
n∈N
Px
τx+=n =pxz.
3
We next focus on the termsw̸=x, such that{Xn=w,τx+⩾n}={Xn=w,τx+⩾n+1} ∈Fn. Hence, from the Markov property ofX, we can write
Px
τx+⩾n+1,Xn=w,Xn+1=z =Px
τx+⩾n,Xn=w P({Xn+1=z} |Xn=w,τx+⩾n,X0=x )
=Px
τx+⩾n,Xn=w pwz. Summing both sides overn∈Nandw̸=x, we get
w̸=x
∑
˜
πx(w)pwz=
∑
n∈N
∑
w̸=x
Px
τx+⩾n+1,Xn=w,Xn+1=z =
∑
n⩾2
Px
τx+⩾n,Xn=z
=π˜x(z)−Px
τx+⩾1,X1=z =π˜x(z)−pxz. The result follows from summing both the cases.
2.2 Uniqueness of stationary distribution
Recall that distributionsπon state spaceXsuch thatπP=πis called a stationary distribution. Similarly, a functionh:X→Ris calledharmonic atxif
h(x) =
∑
y∈X
pxyh(y).
A function isharmonic on a subsetD⊂Xif it is harmonic at every statex∈D. That is, Ph=hfor a function harmonic on the entire state spaceX.
Lemma 2.5. For a finite irreducible Markov chain, a function f that is harmonic on all states inXis a constant.
Proof. Supposehis not a constant, then there exists a statex0∈X, such thath(x0)⩾h(y)for all states y∈X. Since the Markov chain is irreducible, there exists a statez∈Xsuch thatpx0,z>0. Let’s assume h(z)<h(x0), then
h(x0) =px0,zh(z) +
∑
y̸=z
px0,yh(y)<h(x0).
This implies that h(x0) =h(z)for all states zsuch thatpx0,z >0. By induction, this implies that any h(x0) =h(y)for any states y reachable from state x0. Since all states are reachable from state x0by irreducibility, this implieshis a constant on the state spaceX.
Corollary 2.6. For any irreducible and aperiodic finite Markov chain, there exists a unique stationary distribu- tionπ.
Proof. For an aperiodic and irreducible DTMCX:Ω→XZ+with finite state spaceX, we havePx
n
τy+<∞o= 1 andExτy+<∞for all statesx,y∈X. Therefore, we have seen the existence of a positive stationary distributionπfor an irreducible and aperiodic finite Markov chain. Further, from previous Lemma we have that the dimension of null-space of(P−I)is unity. Hence, the rank ofP−Iis|X| −1. Therefore, all vectors satisfyingν=νPare scalar multiples ofπ.
2.3 Stationary distribution for irreducible and aperiodic finite DTMC
For a finite state irreducible and aperiodic DTMCX:Ω→XZ+, we haveExτy+<∞andPx
τy<∞ =1 for allx,y∈X. That is, the return times are finite almost surely, and hence we can apply strong Markov property at these stopping times to obtain that DTMCXis a regenerative process with delayed renewal sequenceτy+:Ω→NN, whereτy+(0)≜0, andτy+(n)≜infn
m>τy+(n−1):Xm=yo .
Theorem 2.7. For a finite state irreducible and aperiodic Markov chain X:Ω→XZ+, its invariant distribution is same as its stationary distribution.
Proof. We can create an on-off alternating renewal function on this DTMCX, which is ON when in state y. Then, from the limiting ON probability of alternating renewal function, we know that
π(y)≜ lim
k→∞Px{Xk=y}= lim
n→∞
1 n
∑
n k=11{Xk=y}= 1 Eyτy+. We observe thatπ(y) =π˜y(y)
Eyτy+ for each statey∈X. From the uniqueness of invariant distribution, it follows thatπis the unique invariant distribution of the DTMCX. We observe thatπ(x)is the long- term average of the amount of time spent in statexand from renewal reward theoremπ(x) =E1
xτx+. 4