Professional Documents
Culture Documents
Biman Chakraborty
5 Period of a state
6 Stationary Distribution
Definition
A process is deterministic if its future is completely determined by its present and past.
Example
Growth of some biological system is governed by X(t) = X0 ert .
Definition
A stochastic process is a random process evolving in time.
Even if you have full knowledge of the state of the system and its entire past, you can
not be sure of its value at future times with certainty.
Example
Snake and ladder game
Biman Chakraborty (Aliah University) Stochastic Processes May 3, 2020 3 / 43
Definition
A stochastic process, or often random process, is a collection of random variables
{Xt , t ∈ T }, representing the evolution of some system of random values over time.
Definition
The set of all possible values of Xt is called State Space.
The elements of this set are called State. The State Space may be discrete as well as
continuous.
If the State Space is discrete (continuous) then the stochastic process is called dis-
crete (continuous) state space stochastic process.
According to state space and time stochastic process can be classified into four
categories.
1 Discrete time and discrete state space (DTDS) stochastic processes:
Example : Random walk on the set of integers
2 Discrete time and continuous state space (DTCS) stochastic processes:
Example: Autoregressive processes
3 Continuous time and discrete state space (CTDS) stochastic processes:
Example: Population size of any species over time.
4 Continuous time and Continuous state space (CTCS) stochastic processes:
Example: Brownion Motion
Example
1 Random Walk on Integers:
2 snakes and ladders game:
Definition
A stochastic matrix (also termed probability matrix, transition matrix, substitution matrix, or
Markov matrix) is a matrix used to describe the transitions of a Markov chain. The (i, j)th
element of one step transition probability matrix ( denoted by P ) is pi,j i.e.
P [X1 = j|X0 = i]
.
(n)
The (i, j)th element of n-step transition probability matrix ( denoted by P (n) ) is pi,j i.e.
P [Xn = j|X0 = i].
Time-dependent distribution: defines the probability that Xn takes a value in a particular subset of S at
a given time n. Note that we can calculate this distribution using initial distribution (π (0) ) and transition
probability matrix P . For a state j ∈ S
X X (n)
P (Xn = j) = P (Xn = j, X0 = i) = P (X0 = i)pij = π (0) P (n) [, j] = π (0) P n [, j]
i∈S i∈S
Stationary distribution: defines the probability that Xt takes a value in a particular subset of S as t → ∞
(assuming the limit exists)
Hitting probability: the probability that a given state is S will ever be entered
First passage time: the instant at which the stochastic process first time enters a given state or set of
states starting from a given initial state
P (E1 E2 E3 ...En ) = P (E1 )P (E2 |E1 )P (E3 |E1 E2 )...P (En |E1 E2 ...En−1 )
For a set of n events such that P (Ei |E1 E2 ...Ei−1 ) = P (Ei |Ei−1 ), we have
P (E1 E2 E3 ...En ) = P (E1 )P (E2 |E1 )P (E3 |E2 )...P (En |En−1 )
Probability of a Path
P (X0 = i0 , X1 = i1 , ...Xn = in ) = P (X0 = i0 )P (X1 = i1 |X0 = i0 )...P (Xn = in |Xn−1 = in−1 )
Definition
The relation ↔ forms an equivalent relation (How?). The set of equivalence classes in a
DTMC are called the communication classes or, more simply, the classes of the Markov
chain. If every state in the Markov chain can be reached from every other state, then there
is only one communication class (all the states are in the same class).
Definition
A Markov chain is said to be irreducible if there is only one equivalence class, i.e., if all
states communicate with each other.
Biman Chakraborty (Aliah University) Stochastic Processes May 3, 2020 12 / 43
Example
Consider a Markov chain with state space {1, 2, 3, 4} and transition probability matrix P ,
where
1 1
2 2 0 0
1 1 0 0
P = 2 2
1 1 1 1 .
4 4 4 4
0 0 0 1
Find the equivalence classes.
Whether all the states will be in a single communicating class? In that case all states will
communicate with 1. So we check it in the following.
Lemma
(n)
A state i ∈ S is said to be absorbing if and only if pii = 1, ∀n ≥ 0.
If a state i ∈ S is absorbing then it does not communicate to any other state (How?).
Definition
(n)
If state i is recurrent, then the set {fii }i≥1 defines a probability distribution for a random
variable (Tii ) which defines the first return time.
1 The mean of the distribution of Tii is referred to as the mean recurrence time to state
i.
∞
(n)
X
µii = nfii
n=1
2 A recurrent state i is positive recurrent if µii < ∞, otherwise null recurrent.
(1) 1
f11 = P (X1 = 1|X0 = 1) = p11 =
2
(2) 1
f11 = P (X2 = 1, X1 6= 1|X0 = 1) = P (X2 = 1, X1 = 2|X0 = 1) = p12 p21 =
6
(3)
f11 = P (X3 = 1, X2 6= 1, X1 6= 1|X0 = 1) = P (X3 = 1, X2 = 2, X1 = 2|X0 = 1)
1 2
= p12 p22 p21 = ×
6 3
(4)
f11 = P (X4 = 1, X3 6= 1, X2 6= 1, X1 6= 1|X0 = 1) = P (X4 = 1, X3 = 2, X2 = 2, X1 = 2|X0 = 1)
2
1 2
= p12 p222 p21 = ×
6 3
Therefore the state 1 is recurrent. The mean recurrent time for the state 1 is
∞ 2
X (n) 1 1 1 2 1 2
µ11 = =1× +2× +3× × +4× ×
nf11 + ...
n=1
2 6 6 3 6 3
2 ! 2 !
1 1 2 2 1 2 2
= +2× 1+ + + ... + + + ... + ...
2 6 3 3 6 3 3
2 !
1 2 2 5
= 1+ 1+ + + ... =
2 3 3 2
5
Since µ11 = 2
< ∞ state 1 is positive recurrent.
Lemma
For i, j ∈ S
∞ ∞ ∞
P (n) P (n) P (n)
1 pij = fij pjj = fij 1 + pjj
n=1 n=0 n=1
∞
P (n)
pij
n=1
2 fij = ∞
P (n)
1+ pjj
n=1
∞
(n) P (n)
3 sup pij ≤ fij ≤ pij
n≥1 n=1
Two states i and j communicate (i ↔ j) if and only if fij > 0 and fji > 0.
∞
(n)
X
pii = ∞ (< ∞)
n=0
Theorem
If i ↔ j then
1 i is recurrent ⇐⇒ j is recurrent.
2 i is transient ⇐⇒ j is transient.
Theorem
∞
P (n) (n)
If j ∈ S is transient, then prove that ∀ i ∈ S, pij < ∞ and lim pij = 0.
n=1 n→∞
Theorem
In a finite and irreducible Markov chain all states are recurrent.
Exercise
Consider a homogeneous Markov chain {Xn : n ≥0}. Determine
which states are tran-
0 0 12 21
1 0 0 0
sient and which are recurrent? S = {1, 2, 3, 4}; P =
0
1 0 0
0 1 0 0
Theorem
i ∈ S recurrent ⇐⇒ E[number of returns to i |X0 = i] = ∞
Example
1 {0, 1, 2, . . . , N}-finite, 2 boundaries at 0 and N .
2 {0,1,2, . . . } - semi infinite, 1 boundary at 0.
Example
1 Absorbing boundary: An absorbing boundary at x = 0 assumes the one step
transition probability is p00 = 1.
2 Reflecting boundary: A reflecting boundary at x = 0 assumes the transition
probabilities are p11 = 1 − p and p12 = p.
3 Elastic boundary: An elastic boundary at x = 0 assumes the transitions probabilities
are p12 = p, p11 = sq, p10 = (1 − s)q, p + q = 1, p00 = 1, for 0 < p, s < 1.
Xn : Gambler’s fortune at time n, then the process {Xn , n = 0, 1, 2, ...} is a Markov chain
1 0 0 ... 0 Communication Classes:
p 0 q ... 0
{0}, {1, 2, ..., N − 1}, {N }
P =
.. .. .. . . ..
. . . . .
Recurrent State: {0}, {N }
0 0 q 0 p
0 0 0 0 1 Transient States: {1, 2, ..., N − 1}
(n)
The state i has period d if pii = 0 unless n = vd, is a multiple of d.
Theorem
Let Xn be a Markov chain with state space S. If i, j ∈ S are in the same communication
class, then d(i) = d(j). That is they have the same period.
Definition
An aperiodic positive recurrent state is called ergodic. If the Marcov chain is irreducible
and all states are ergodic then the Markov chain is called ergodic Markov chain.
0.5 0.25 0.25
P 2 = 0.25 0.5 0.25 0.25 0.375 0.375
0.25 0.25 0.5 P 3 = 0.375 0.25 0.375
0.375 0.375 0.25
Hence the period of the state 1 is g.c.d.(2, 3, ..) = 1. Similarly period of each state is 1.
Definition
A set C of states is said to be closed if once the process enters it, it can not get out of it.
(n)
i.e., if fij = 0 ∀i ∈ C, j ∈ C c . If C is closed then pij = 0 ∀n ≥ 1, ∀ i ∈ C, j ∈ C c .
(nt) t
lim pjj = , where t is period of state j
n→∞ µjj
.
(nt)
2 If j is recurrent null (whether periodic or aperiodic) then limn→∞ pjj = 0.
(n)
3 If k is recurrent null then for any j ∈ S, limn→∞ pjk = 0.
(n) fjk
lim pjk = .
n→∞ µkk
Definition
k k 35
However, this eigenvector will be a probability distribution if x + y = 1 =⇒ 5
+ 7
= 1 =⇒ k = 12
.
7 5
Therefore x = 12 , y = 12 .
7 5
Note that lim P (Xn = 1) = 12
and lim P (Xn = 2) = 12
.
n→∞ n→∞
7 5
Again, we can calculate lim P (n) = 12
7
12
5 .
n→∞
12 12
Proof.
Let aj = P [X0 = j], j = 1, 2, ...
First suppose that the distribution of Xn does not depend on n. Then
∞
X ∞
X
aj = P [X1 = j] = P [X1 = j|X0 = k]P [X0 = k] = pkj ak
k=1 k=1
0 51 35 1
5
1 1 1 1
P = 4 4 4 4 .
1 0 0 0
0 21 12 0
Theorem
Suppose a Markov chain is irreducible and that a stationary distribution π exists:
X
π 0 = π 0 P, pij = 1, πj > 0.
j∈S
It is easy to verify that the Markov chain is not irreducible and its communication classes are
{{1, 2}, {3, 4, 5}}.
We can find that
0.5294 0.4706 0.0000 0.0000 0.0000
0.5294 0.4706 0.0000 0.0000 0.0000
(n)
lim P = 0.0000 0.0000 0.2424 0.1212 0.6364 .
n→∞
0.0000 0.0000 0.2424 0.1212 0.6364
Does limiting
0.0000 0.0000 0.2424 0.1212 0.6364
distribution exist?
If not, Why?