Professional Documents
Culture Documents
Markov processes
Summer 2022
3 Feller processes 26
3.1 Feller processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.2 From the Feller process to its infinitesimal description . . . . . . . . . . . . 30
3.3 From the infinitesimal description to the Feller process . . . . . . . . . . . 32
5 Spin systems 46
5.1 Spin systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
5.2 Ergodicity of spin systems . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
5.3 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 52
5.4 Attractive spin systems and coupling . . . . . . . . . . . . . . . . . . . . . 53
1
Chapter 1
Definition 1.1.
Consider the smallest σ-field F on Ω such that ω 7→ ω(t) is measurable for all t ≥ 0.
(ii) a right-continuous filtration (Ft )t≥0 on Ω, Ft ⊂ F for all t ≥ 0 such that (Xt )t≥0 is
adapted to (Ft )t≥0 and Px (X0 = x) = 1 for all x ∈ S, and
and all bounded measurable functions g : Ω → R. (1.1) is called the Markov prop-
erty. (We write Ex for the expectation with respect to Px ).
2
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 3
Px (Xt1 = x1 , ..., Xtm = xm ) = pt1 (x, x1 )pt2 −t1 (x1 , x2 ) · · · ptm −tm−1 (xm−1 , xm )
q(x, y) ≥ 0, ∀x 6= y
X
q(x, y) = 0, ∀x ∈ S.
y∈S
Example 1.4 (Poisson process). (Xt )t≥0 has values in Z. After an exponentially dis-
tributed waiting time, the process jumps by 1, i.e. it goes from its present state x to x + 1.
Example 1.5. Take a discrete-time MC given by the transition matrix P = (p(x, y))x,y∈S ,
where P is a stochastic matrix, i.e.
X
p(x, y) ≥ 0, ∀x, y ∈ S, p(x, z) = 1 ∀x ∈ S.
z∈S
Then we define (Xt )t≥0 as follows: take i.i.d. exp(1) distributed random variables. Then
at x wait for one of these and jump according to P .
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 4
Remark. With Ft = σ(Xu , u ≤ t), (1.1) holds true due to the loss of memory property
of the exponential distribution.
Proof.
∞
−t 0 −t
X tk
pt (x, y) = e P (x, y) + e P k (x, y)
k=1
k!
∞ ∞
d X tk X tk−1
pt (x, y) = −e−t P 0 (x, y) − e −t k
P (x, y) + e −t
P k (x, y)
dt k=1
k! k=1
(k − 1)!
∞ k−1 k
X t t
= −e−t P 0 (x, y) + e−t − P k (x, y)
k=1
(k − 1)! k!
d
pt (x, y) = −1 + P 1 (x, y).
dt t=0
Claim. For all β > 0, δ > 0, pt (·, ·) belongs to the following MC (Xt )t≥0 : The MC stays,
when in 0, for an exp(β)-distributed waiting time in 0 and then goes to 1. When in 1, it
stays a exp(δ)-distributed time in 1 and then it goes to 0, etc. (Note that all waiting times
are independent.)
Proof. Exercise.
Lemma 1.7. Let S be finite and Q a Q-matrix. Then Pt = (pt (x, y))x,y∈S given by
∞
tQ
X tn
Pt = e := Qn
n=0
n!
lim pt (x, x) = 1
t→0
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 6
d tQ
e = QetQ .
dt
Example 1.8. A birth and death chain is a MC with state space N ∪ {0} and the
following Q-matrix
q(k, k + 1) = ρk , k ≥ 0
q(k, k − 1) = λk , k ≥ 1
q(k, k) = −ρk − λk , k ≥ 0.
Interpretation. (Xt )t≥0 is the size of a population at time t. Each particle has two inde-
pendent clocks, which ring after exponentially distributed times T and Te with parameters
ρ and λ. If time T < Te, the particle has at time T an offspring, if Te ≤ T the particle dies
at time Te. The offspring particles are doing the same, and all particles behave indepen-
dently. This means that in state Xk = k the process goes with rate ρk to k + 1 and with
rate λk to k − 1. Note that if T1 , T2 , ..., Tk are i.i.d. with law exp(ρ) then min(T1 , ..., Tk )
has the distribution exp(ρk). Note that c(k) = (ρ + λ)k is not bounded in this example.
Theorem 1.9. Let (Xt )t≥0 be a continuous time MC. Let pt (x, y) = Px (Xt = y) for t ≥ 0
and x, y ∈ S. Then
Proof. For (1.2) it remains to show that limt↓0 pt (x, x) = p0 (x, x) = 1. Due to the right-
continuity of the paths, we have
Since pt (x, x) ≥ Px (T > t) the claim follows. The Chapman-Kolmogorov equations follow
from the Markov property: (1.1), with g = 1{Xt =y} tells us that
Px (Xs+t = y|Fs ) = PXs (Xt = y) = pt (Xs , y) Px -a.s. for all x ∈ S.
Take expectation w.r.t. Px to get
X
pt+s (x, y) = Px (Xs+t = y) = ps (x, z)pt (z, y),
z∈S
We get (1.7) (formally, since one would need to justify that t → pt (x, y) is continuously
differentiable!) by deriving (1.3) w.r.t. s and taking s = 0.
X
ps+t (x, y) = ps (x, z)pt (z, y)
z∈S
d d X d
pt (x, y) = ps+t (x, y)
= ps (x, z) pt (z, y)
dt ds s=0 z∈S
ds s=0
X
= q(x, z)pt (z, y).
z∈S
In the same way one could derive w.r.t. t and take t = 0 to get
d d X
ps (x, y) = ps+t (x, y) = ps (x, z)q(z, y). (1.8)
ds dt t=0 z∈S
If S is finite (1.7) is a finite system of linear differential equations, which has the unique
solution
∞
X tk k
Pt = etQ = Q .
k=0
k!
If S is infinite, uniqueness of the solution is in general not satisfied, but true under
additional assumptions.
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 8
We start with some important facts on transition functions, which we will need later.
Theorem 1.10. Let (pt )t≥0 be a transition function. Then the following hold:
implies that pt (x, x) > 0 for t small enough. Furthermore, due to (1.3), we have
Note that
|pt (x, y)(ps (x, x) − 1)| ≤ |ps (x, x) − 1| = 1 − ps (x, x),
and X X
0≤ ps (x, z)pt (z, y) ≤ ps (x, z) = 1 − ps (x, x).
z6=x z6=x
Thus, we obtain
|pt+s (x, y) − pt (x, y)| ≤ 2(1 − ps (x, x)).
We deduce that
(iii) The proof relies on the following lemma, which is often called subadditivity lemma:
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 9
then
lim g(t)
t
= sup g(t)
t
t&0 t>0
For a proof we refer to [Liggett, Theorem A.59]. Due to (i) and (ii), the function
lim f (s)
s
= sup f (s)
s
s&0 s>0
exists in [0, ∞]. Using the chain rule and limt&0 pt (x, x) = 1, we see that the limit
on the l.h.s. equals u(x) and the inequality (1.9) follows also readily.
The proof is complete.
Now, we explain informally how to come from the transition function to a Q-matrix.
This direction requires some extra assumptions in certain places. Let (pt )t≥0 be a transition
function and let x ∈ S be such that u(x) < ∞. Then, one can show that for all x 6= y the
right-hand derivative
d
u(x, y) := pt (x, y)
dt t=0
If this is true for all x ∈ S, then (u(x, y))x,y∈S is a Q-matrix corresponding to the transition
function (pt )t≥0 . For proofs we refer to [Liggett, Theorem 2.14].
Definition 1.12. Let (pt )t≥0 be a transition function. The state x ∈ S is called absorbing
if pt (x, x) = 1 for all t ≥ 0. The state x ∈ S is called instantaneous if u(x) = ∞, see
Theorem 1.10.
If S is countable, there are examples (see Section 1.3 below) of transition functions
where all x ∈ S are instantaneous and there is no MC with this transition function.
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 10
Z is called the embedded discrete time MC. Let τ0 , τ1 , ... be random variables whose
conditional distribution given (Z0 , Z1 , ...) is the following: Given (Z0 , Z1 , ...) the τk are
independent and have laws
exp(c(Zk )) k = 0, 1, 2, ... with τk = ∞ if c(Zk ) = 0.
The finite dimensional marginals of ((Z0 , Z1 , ....), (τ0 , τ1 , ...)) are given as follows:
Let
( P∞
min{m ≥ 0 : τ0 + τ1 + ... + τm ≥ t}, if k=1 τk > t
N (t) =
+∞, otherwise.
Hence N (t) = 0 for a time interval of length τ0 , then N (t) = 1 for a time interval of length
et = ZN (t) on {N (t) < ∞}: X
τ1 , etc. Finally, let X e has right-continuous paths (if defined)
and waits, when at x, an exponentially distributed time with parameter c(x), and then
jumps to state y with probability p(x, y). Everything is fine, except that X e could jump
infinitely often in a finite time interval.
Theorem 1.13. The following statements are equivalent:
(a) (1.7) has a unique solution, which is a transition function
[e) (X
et )t≥0 is an MC in continuous time.
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 11
Recalling that countable unions of countable sets are countable, we conclude that S is
countable. Now, set
Y
pt (x, y) := P(Xt = y|X0 = x) = Pxk (Xtk = yk ), x, y ∈ S. (1.11)
k∈N
Next, we give a criterion on the parameters βk and δk such that (pt )t≥0 is a transition
function. Afterwards, we will also see a criterion such that all states are instantaneous.
Theorem 1.15. Suppose that
∞
X βk
< ∞. (1.12)
k=1
βk + δk
Then,
P(Xt ∈ S|X0 ∈ S) = 1 for all t ≥ 0, (1.13)
and (pt )t≥0 is a transition function.
CHAPTER 1. MARKOV CHAINS IN CONTINUOUS TIME 12
Proof. The function (pt )t≥0 is a transition function if we show (1.2) and the Chapman-
Kolmogorov equations (1.3). First, pt (x, y) ≥ 0 is obvious. Second, we have
X
pt (x, y) = P(Xt ∈ S|X0 = x).
y∈S
We show that
P(Xt ∈ S|X0 = x) = 1 (1.14)
for all x ∈ S. Fix x ∈ S. Then, there exists a nx < ∞ such that xi = 0 for all i ≥ nx . Note
that Xt ∈ S whenever Xtn = 1 for only finitely many n ≥ nx . We show that this holds
with probability one due to the Lemma of Borel-Cantelli. Indeed, recalling (1.6), we see
that for any n ≥ nx
−t(βn +δn )
P(Xtn = 1|X = x) = δnβ+β ≤ δnβ+β
n
n
1 − e n
n
.
Thus, recalling our hypothesis (1.12), the Borel-Cantelli lemma yields that
P(Xtn = 1 for only finitely many n ≥ nx |X0 = x) = 1,
i.e. (1.14) holds. Moreover, we have
P
P(X0 =x)
X
P(Xt ∈ S|X0 ∈ S) = P(Xt ∈ S|X0 = x) P(X0 =x)
P(X0 ∈S)
= x∈S
P(X0 ∈S)
= 1,
x∈S
which proves (1.13). Next, we prove the last part in (1.2). Using once again (1.6), we see
that for any n ≥ nx
n−1 ∞
Y Y δi
pt (x, x) ≥ Pxk (Xtk = xk ) . (1.15)
k=1
β + δi
i=n i
Let us shortly
Q argue that the infinite product makes sense. It is well-known
P that an infinite
product k ak converges to a non-zero limit iff the infinite sum k log(ak ) converges. We
have (using the estimate log x ≥ 2(x − 1) for x ∈ (1/2, 1])
∞ ∞ ∞
X X X βk
log βkδ+δ
δk
k
≤ 2 − 1 = <∞
βk +δk
βk + δk
k
k=1 k=1 k=1
by hypothesis. Thus, the infinite product in (1.15) is well-defined. Now, in (1.15), letting
first t & 0 and then n % ∞, we see that
lim pt (x, x) = 1.
t&0
X n−1
Y
= Pxk (Xtk = zk )Pzk (Xsk = yk )Px1 (Xt+s
1
= y1 )
z2 ,...,zn−1 k=2
n−1
Y
k
= ··· = Pxk (Xt+s = yk ),
k=1
where we use the Chapman-Kolmogorov equations for the MCs (Xtk )t≥0 . We conclude
that
X ∞
Y n−1
Y
k k j
pt (x, z)ps (z, y) = pt (0, 0)ps (0, 0) Pxj (Xt+s = yj ).
z∈An k=n j=1
Letting n % ∞, we obtained the Chapman-Kolmogorov equations for (pt )t≥0 and the
proof is complete.
Remark. We stress that X is always a stochastic process with values in {0, 1}∞ . The
condition (1.12) guarantees that we can consider X also as a stochastic process with
values in S.
So far, we have seen a condition on the sequences (βk )k∈N and (δk )k∈N such that (1.11)
defines a transition function. In the following theorem, we also give a condition which
implies that all states of (1.11) are instantaneous. For an intuitive explanation of the
condition, see the remark below.
Theorem 1.16. Suppose that (1.12) holds and that
∞
X
βk = ∞. (1.16)
k=1
Finally, letting n → ∞ and using our hypothesis yields that u(x) = ∞. This finishes the
proof.
Remark. Let us informally understand the condition (1.16). Roughly speaking, a state is
instantaneous if the probability of remaining in it in an interval is zero. The probability
that X k remains in 0 in an interval of length h is e−βk h . The probability that the entries
of (X n , X n+1 , ...) remain in 0 is consequently
∞
Y
e−βk h ,
k=n
using the independence of the entries. This product, however, is zero if (1.16) is satisfied.
Chapter 2
1
Note that this is not necessarily a probability measure.
15
CHAPTER 2. INVARIANT MEASURES, RECURRENCE AND TRANSIENCE 16
P
(ii) Assume now π(x)q(x, y) = 0. Then
x∈S
d X (1.7) X X
π(x)pt (x, y) = π(x) q(x, z)pt (z, y)
dt x∈S x∈S z∈S
X X
Fubini (to verify) = pt (z, y) π(x)q(x, y) = 0.
z∈S x∈S
Hence
X X
π(x)pt (x, y) = π(x)p0 (x, y) = π(y).
x∈S x∈S
Definition 2.3. A measure π is reversible for the MC with transition function pt (x, y)
if
A reversible measure is also invariant, since (2.3) implies (2.1). However there are
invariant measures that are not reversible.
Example. Fix λ > 0. Let (Xt )t≥0 be a Poisson process with intensity λ, i.e.
X X ∞
X
pt (x, y) = pt (x, y) = pt (0, z) = 1.
x∈Z x≤y z=0
On the other hand, (2.3) can not hold since for x < y, pt (x, y) > 0 but pt (y, x) = 0.
Example. Let
−β β
Q= .
δ −δ
δ β
π(0) = , π(1) = ,
β+δ β+δ
Definition 2.5. A stochastic process (Xt )t≥0 is strictly stationary, if for all n ≥ 1 and
t1 < t2 < · · · < tn , the law of (Xt1 +s , ..., Xtn +s ) does not depend on s.
Lemma 2.6. Let π be a probability measure on S and (Xt )t≥0 a MC with transition
function pt (·, ·) and starting distribution π (i.e. X0 has law π). Then (Xt )t≥0 is strictly
stationary if and only if π is invariant.
Proof of Theorem 2.9. Due to our construction of (X et )t≥0 , the length of the time intervals
spent in x are i.i.d. with law exp(c(x)). If c(x) = 0 we have pt (x, x) = 1 for all t ≥ 0 and
x is recurrent. Hence assume c(x) > 0. Due to the strong Markov property, the number G e
of visits of (Zn )n∈N to x has a geometric law (where we admit the case G e = ∞). Further,
G
e is independent of the waiting times (τ0 , τ1 , ...). Hence either
Z Z G
e
X
1{Xt =x} dt = ∞ or 1{Xt =x} dt = τ ik ,
R+ R+ k=0
e = ∞ ⇒ G(x, x) = ∞
G
and x is recurrent or
G
X
e
1
G(x, x) = Ex τ ik = ,
k=0
c(x)e
p
Example 2.10. Linear birth and death chain (cf. Example 1.8). Fix ρ, λ > 0 and consider
the MC with state space S = {0, 1, 2, . . . } and Q-matrix Q = q(k, l)k,l∈S given by
ρk
l =k+1
q(k, l)k,l∈S = λk l =k−1
−(λ + ρ)k l = k
2. Which states are transient? Which states are recurrent? 0 is clearly a recurrent state.
Claim: All other states are transient.
Consider the discrete embedded MC (Zn ). Intuitively, if x > 0 is recurrent, then it is
visited by (Zn ) infinitely often. But on each visit, we have a probability ≥ ε to visit
0 before coming back to x. Since these trials are iid, one of them will be successful
and the MC will be absorbed at 0: contradiction!
Proof of the claim. Let x > 0 and let T1 := min{n > 0 : Zn = x} and Ti+1 :=
min{n > Ti : Zn = x}. For H0 := min{n > 0 : Zn = 0} we have
x
λ
Px (H0 < T1 ) ≥ =ε
λ+ρ
(which is the probability that the first x steps are going to the left). Now by the
strong Markov property
Thus,
≤ (1 − ε)i .
Therefore, ∀k ∈ N
=⇒ x is transient.
The arguments of 1. and 2. go through for the general birth and death chain with q(0, 0) = 0
and, for k ≥ 1
ρk
l =k+1
q(k, l)k,l∈S = λk l =k−1
−(λk + ρk ) l = k
ρ 1 ρ
If ρ+λ
= 2
or ρ+λ
< 12 then
P1 lim Zn = 0 = P1 Z̃k = 0 for some k ≥ 1 = 1
n→∞
ρ ρ
since Z̃n is either recurrent (if ρ+λ = 12 ) or transient to the left (i.e. if ρ+λ
< 1
2
we
have P1 (limn→∞ Z̃n = −∞) = 1). Hence
P1 lim Xt = ∞ > 0 ⇐⇒ ρ > λ.
t→∞
Proposition 2.11. For an irreducible MC, either all states are recurrent or all states are
transient. In the transient case, we have
hence G(x, x) ≥ pt (x, y)G(y, y)pt (y, x). Hence if the MC is irreducible, G(x, x) < ∞
implies G(y, y) < ∞. To show the second statement, let Hy := inf{s ≥ 0 : Xs = y} the
first hitting time of y. Then
Z
G(x, y) = Ex 1{Xt =y} dt1{Hy <∞}
R+
(strong Markov property) = Px (Hy < ∞)G(y, y) ≤ G(y, y).
Definition 2.12. An irreducible MC is called recurrent if all states are recurrent and
transient if all states are transient.
Remark. The chain (Xt )t≥0 is irreducible if and only if the embedded discrete time MC
(Zn )n∈N is irreducible. In this case (Xt )t≥0 is recurrent if and only if (Zn )n∈N is recurrent.
Definition 2.13. A function f : S → R is superharmonic for (Xt )t≥0 if, for all t ≥ 0
and x ∈ S
Conversely assume that (Xt )t≥0 is recurrent and that f is bounded and superharmonic.
Then (f (Xt ))t≥0 is a bounded Px -supermartingale, which converges Px -a.s. Since (Xt )t≥0
is recurrent, each x ∈ S is visited at arbitrary large times and hence f is constant.
The last argument also shows the following.
Lemma 2.16. For an irreducible, recurrent MC (Xt )t≥0 each non-negative superharmonic
function is constant.
Lemma 2.17. If an irreducible MC has an invariant distribution, it is recurrent.
Proof. Let π be an invariant distribution. Then for all t ≥ 0
Z
X 1
π(y) = π(x) ps (x, y)ds. (2.6)
x∈S
t [0,t]
which is a contradiction.
Lemma 2.18. An irreducible and recurrent MC has an invariant measure π. Moreover
π(x) > 0, ∀x ∈ S.
CHAPTER 2. INVARIANT MEASURES, RECURRENCE AND TRANSIENCE 22
Proof. Fix z ∈ S.
1
Then π(z) = c(z) and in particular π(z) > 0 and π(z) < ∞.
We show that π is invariant. Due to the strong Markov property
Z Z
Ez 1{Xt =x} dt = Ez 1{Xt =x} dt , s > 0.
[0,s] [Rz ,Rz +s]
Therefore
Z
π(x) = Ez 1{Xt =x} dt
[0,Rz ]
Z Z
= Ez 1{Xt =x} dt − Ez 1{Xt =x} dt
[0,Rz +s] [Rz ,Rz +s]
Z
= Ez 1{Xt =x} dt
[s,Rz +s]
Z
= Ez 1{Xt+s =x} dt
[0,Rz ]
Z X
= Pz (Xt = y, Rz > t)ps (y, x)dt
R+ y∈S
X
= π(y)ps (y, x),
y∈S
R R
since π(y) = Ez [0,Rz ] 1{Xt =y} dt = R+ Pz (Xt = y, Rz > t)dt. In particular, π(z) =
P
y∈S π(y)ps (y, z) and since we know that π(z) < ∞ and ps (y, z) > 0 for all y ∈ S
and all t > 0 due to irreducibility, we see that π(y) < ∞ for all y ∈ S. It remains
to showPthat π(x) > 0 for all x ∈ S. Since pt (z, x) > 0 for all t > 0, x ∈ S and
π(x) = y∈S π(y)pt (y, x) ≥ π(z)pt (z, x) > 0, we have π(x) > 0 for all x ∈ S.
Lemma 2.19. There is (up to a multiplication with a constant) exactly one invariant
measure for an irreducible and recurrent MC.
CHAPTER 2. INVARIANT MEASURES, RECURRENCE AND TRANSIENCE 23
Proof. Assume π1 , π2 are invariant measures for (Xt )t≥0 . Note that π1 (x) > 0 and π2 (x) >
0 for all x ∈ S, due to the same argument as in the proof of Lemma 2.18. Define another
transition function
π1 (y)pt (y, x)
pt (x, y) = .
π1 (x)
Due to Theorem 2.9 the MC (X t )t≥0 with transition function p is recurrent as well. Set
α(x) = ππ12 (x)
(x)
and note that
X
pt (x, y)α(y) = α(x),
y∈S
In particular, (Xt )t≥0 is transient, since there are two truly different invariant measures.
In this example, more is known: show that limt→∞ Xtt > 0 (exercise).
Definition 2.21. An irreducible MC which has an invariant distribution is positive
recurrent. An irreducible recurrent MC which does not have an invariant distribution is
null recurrent.
Remark 2.22. An irreducible recurrent MC with invariant measure π is positive recurrent
if and only if the expected return times are finite, i.e.
Ez [Rz ] < ∞, ∀z ∈ S.
Proof. The proof is based on a coupling argument, which goes as follows: Let (Xt )t≥0
and (Yt )t≥0 be two independent MC with the same transition function pt (·, ·). (We will
fix the starting distribution later.) Define Zt := (Xt , Yt ) for t ≥ 0. Then (Zt )t≥0 is an
irreducible MC on S × S with transition function
It is easy to check that (Zt )t≥0 has the invariant distribution π e given by πe(x1 , x2 ) =
π(x1 )π(x2 ), x1 , x2 ∈ S. Hence (Zt )t≥0 is recurrent due to Lemma 2.17. In particular, for
any starting points x1 , x2
P(x1 ,x2 ) (Zt ∈ {(i, i) ∈ S × S|i ∈ S} for some t ≥ 0) = P(x1 ,x2 ) (Xt = Yt some t ≥ 0) = 1
τ := inf{t ≥ 0 : Xt = Yt }
(Wt )t≥0 is a MC with transition function pt (·, ·). We choose now the starting distribution
d d
such that X0 = x, Y0 = π. Then Yt = π for all t ≥ 0, since π is invariant. But we also
d
have Wt = π for all t ≥ 0.
d
since τ is finite P-a.s. Here P(x,π) is the law of (Zt )t≥0 , where X0 = x, Y0 = π.
(b) If (Xt )t≥0 is irreducible and recurrent each non-negative harmonic function is con-
stant.
Proof. Exercise.
CHAPTER 2. INVARIANT MEASURES, RECURRENCE AND TRANSIENCE 25
Remark 2.27. If (Xt )t≥0 is irreducible and recurrent, Theorem 2.15 implies that each
bounded harmonic function is constant. Consider a graph with finite degrees. It has the
Liouville property if each bounded harmonic function (for simple random walk on the
graph) is constant. Simple random walk in continuous time on a graph with finite degrees
can be defined with the Q-matrix
1
deg(x) , x 6= y, x ∼ y
q(x, y) = 0, x 6= y, x 6∼ y
−1, x = y,
where we write x ∼ y if x and y are neighbours, i.e. there is an edge connecting x and y,
and “finite degrees”means that each x has only finitely many neighbours. It is known that
Zd has the Liouville property for d ≥ 1. In particular there are transient graphs with the
Liouville property.
Chapter 3
Feller processes
Here, “vanishing at infinity” means that for each ε > 0, there is a compact subset Kε of
S such that |f (x)| ≤ ε for all x ∈
/ Kε . Let kf k := supx∈S |f (x)|. Then C(S) is a separable
Banach space. The functions in C(S) are uniformly continuous.
The σ-field F on Ω is the smallest σ-field, such that all projections Xt are measurable
w.r.t. F.
Definition 3.1. A Feller process with values in S consists of
(a) a collection of probability measures (Px )x∈S on (Ω, F)
(b) a right-continuous filtration (Ft )t≥0 on Ω, such that (Xt )t≥0 is adapted to (Ft )t≥0
with the following properties:
Px (X0 = x) = 1, ∀x ∈ S (3.1)
The mapping
and
and all g : Ω → R bounded and measurable. (3.3) is called the Markov property and
(3.2) the Feller property.
26
CHAPTER 3. FELLER PROCESSES 27
kTt f k ≤ kf k, ∀f ∈ C(S)
i.e. that the Tt are contractions. More precisely: let S be compact, then
g = kf k − f ≥ 0, g ∈ C(S)
and
P(X0 = 0) = 1.
Rest: exercise.
Exercise. Assume S is countable, pt (x, y) a transition function and
X
(Tt f )(x) = pt (x, y)f (y), x ∈ S, t ≥ 0.
y∈S
Then
CHAPTER 3. FELLER PROCESSES 29
The last inequality follows from property (b) applied to −f . Therefore f 7→ f − λLf is
injective, and together with (c) we conclude that (1 − λL)−1 is defined everywhere for
λ > 0 small enough. Moreover it is a contraction that maps non-negative function to
non-negative functions.
Example 3.9. Let S = R, D(L) = {f ∈ C(S) : f 0 ∈ C(S)} and Lf = f 0 . Then L is a
generator.
Proof. See exercises.
CHAPTER 3. FELLER PROCESSES 30
is a generator
Proof. See Exercises.
We will see that Feller processes, transition semigroups and generators are
in one-to-one correspondence.
t 7→ Tt f
Remark. (b) says ”Tt = etL ”. We know that this is true if S is finite:
X
Lf (x) = q(x, y)f (y), Pt = Tt = etQ .
y∈S
It turns out that the third way is the best for generalization to unbounded operators.
Remark. Often generators are (second order) differential operators. Then, one can con-
sider the partial differential equation (PDE)
d
u = Lu, (3.9)
dt
where u = u(t, x) is a function of space and time. L only acts on the space variable. Under
mild conditions the solution of this PDE with initial condition u(0, x) = f (x) is given by
Example 3.13 (Linear motion). Let s = R and ω (x) (t) = x + t and Px = δω(x) . Then,
the corresponding process (Xt )t≥0 is a Feller process.
Example 3.14 (Brownian motion). Let S = R and let L be the following operator
1
Lf = f 00 ,
2
where
D := {f ∈ C(R) : f twice differentiable with f 00 ∈ C(R)}. (3.11)
(a) One could check directly that L is a generator in the sense of Definition 3.8.
Since f 00 is uniformly continuous, the second term converges to zero for t → 0, uniformly
in x, and the first term becomes arbitrarily small for K large.
Remark. In fact, D = D(L). The proof above does not yet show this since the limit of
Tt f (x)−f (x)
t
for t → 0 could exist as well for functions which are not in D.
Remark. In the proof above, there is no requirement on the first derivative of f . However,
it is easy to see that in fact, D = D e = {f ∈ C(R) : f twice differentiable with f 0 , f 00 ∈
C(R)}. Apply the Taylor formula to get, for a function f ∈ D, f (x + 1) − f (x) = f 0 (x) +
1 00
2
f (x + ax ) where ax ∈ (0, 1). But, for each ε > 0, we can choose K large enough such
/ [−K, K] and |f 00 (x + ax )| ≤ ε for x ∈
that |f (x + 1) − f (x)| ≤ ε for x ∈ / [−K, K], and we
0
conclude that f ∈ C(R) as well and hence f ∈ D. e
Proof.
According to (c) in Definition 3.8: For ε > 0 small enough and g ∈ C(S) there is
f ∈ D(L) such that f − εLf = g. Further, with f, g as above, recalling (??)
f −g kf k + kgk 2
kL(ε) gk = kLf k = k k≤ ≤ kgk.
ε ε ε
(ε)
Hence L(ε) is a bounded operator. Hence we can define Tt by
∞
(ε) tL(ε)
X tk (ε) k
Tt =e = L .
k=0
k!
Lemma 3.15. If A is a bounded operator, then A satisfies (c) in Definition 3.8, i.e.
Proof. (a)
g − εLg = h ⇒ g = (1 − εL)−1 h
(a) from Definition 3.8: D(L(ε) ) = C(S). True since R(1 − εL) = C(S) for ε small
enough.
CHAPTER 3. FELLER PROCESSES 34
hence
λ ε
inf f (x) ≥ inf f (x) + inf g(x).
x∈S ε + λ x∈S ε + λ x∈S
Therefore inf x∈S f (x) ≥ inf x∈S g(x).
(c) from Definition 3.8: This is clear since L(ε) is bounded, see Lemma 3.15.
(d) from Definition 3.8: S compact: L(ε) 1 = L(1 − εL)−1 1, but (1 − εL)−1 1 = 1,
because f − εLf = 1 is solved uniquely by f = 1, L1 = 0. Therefore L(ε) 1 = 0.
For the locally compact case see [Liggett].
(ε)
It remains to show that (Tt )t≥0 is a semigroup with generator L(ε) . But, more
generally: If a bounded (linear) operator A on C(S) is a generator, then (Tt )t≥0
defined by Tt = etA , t ≥ 0 (defined as power series) is a transition semigroup with
generator A. To prove this check Definition 3.3.
exists in C(S) and the convergence is uniform on compact intervals. Further (Tt )t≥0 is a
transition semigroup with generator L.
Proof. See [Liggett] Theorem 3.24.
Now we know that for each generator, there is a semigroup (Tt )t≥0 . Next we convince
ourselves that for each transition semigroup (Tt )t≥0 , there is a corresponding Feller process.
CHAPTER 3. FELLER PROCESSES 35
Theorem 3.18. Let (Tt )t≥0 be a transition semigroup. Then there is a Feller process
(Xt )t≥0 such that
Ex [f (Xs )g(Xt )] = Ex [f (Xs )EXs [g(Xt−s )]] = Ex [f (Xs )(Tt−s g)(Xs )] = (Ts h)(x),
where h(z) = f (z)(Tt−s g)(z) (this corresponds to the law of (Xs , Xt )). One now has to
show that (Xt )t≥0 can be chosen such that t 7→ Xt (ω) is in D[0, ∞), for this we refer to
[Liggett] Theorem 3.26.
Example 3.20. For a ∈ R, let (Bt )t≥0 be a BM and let Xt := at + Bt be BM with drift.
Then (Xt )t≥0 is again a Feller process with semigroup
Z Z √
Tt f (x) = Ex [f (at + Bt )] = ϕx+at,t (y)f (y)dy = ϕ0,1 (y)f (x + at + ty)dy .
By the same arguments as before, for f ∈ D, the first term converges to 12 f 00 (x), uniformly
in x, while the second term converges to af 0 (x). Moreover the convergence of the second
term is uniform in x (see Exercises). Therefore the generator L of (Xt )t≥0 is given by
1
Lf (x) = f 00 + af 0
2
and D ⊆ D(L).
Chapter 4
Example. Let S = [0, 1], D(A) = {f ∈ C(S) : f 0 (0) exists} and define A by Af (x) =
f 0 (0) for all x ∈ S. Note that S is compact, and that A is not a probability generator
(why not?). Then the closure of {(f, Af ), f ∈ D(A)} is not the graph of a linear operator:
∃(fn )n∈N ∈ C(S) such that
but g 6= Af . Take for instance fn (x) := n−1 sin(nx), fn0 (x) = cos(nx), hence fn0 (0) = 1 but
fn → 0 in C(S).
Lemma 4.2. (a) Assume A satisfies (a) and (b) in Definition 3.8, then A exists and
A satisfies (a) and (b) in Definition 3.8 as well.
(b) If A satisfies (a), (b) and (c) in Definition 3.8, then A is closed.
(d) If A is closed and satisfies (b) in Definition 3.8, then R(1 − λA) is a closed subset
of C(S).
36
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 37
and for f ∈ D(A) define Af := limn→∞ Afn . We have to show that this is well-defined,
that is fn ∈ D(A), fn → 0, Afn → h implies h = 0.
Choose g ∈ D(A). Due to (3.5) we have
kg − h − λAgk ≥ kgk.
Letting first λ → 0 and then approximating h by g (using (a) in Definition 3.8) we get
h = 0.
The closure A of A satisfies clearly (a) in Definition 3.8 since it extends A. To show
that A satisfies (b), take f ∈ D(A), λ ≥ 0 and f − λAf = g, then there exists a sequence
(fn )n∈N in D(A) such that fn → f and Afn → Af , and since A satisfies (b) in Definition
3.8, we have
where gn = fn − λAfn . Now letting n → ∞ we get inf x∈S f (x) ≥ inf x∈S g(x).
Part (b): Let A be the closure of A. We want to show that A = A. Let f ∈ D(A) and
λ > 0 small enough. Due to (c) in Definition 3.8 there is h ∈ D(A) such that
Definition 4.3. A collection D ⊂ D(L) is called core for L, if L is the closure of L|D .
In particular L is determined by its values on D.
Remark. We will see that in general a generator L is not determined by its restriction
to an arbitrary dense subset of D(L).
Xtrefl := |Bt |, t ≥ 0.
(Xtrefl )t≥0 is a Feller process, hence there is a semigroup (Ttrefl )t≥0 and a generator Lrefl .
For f ∈ C[0, ∞) let the even extension of f be
then we have
(Ttrefl f )(x) = Ex [f (|Bt |)] = Ex [fe (Bt )],
for x ≥ 0. Hence
D(Lrefl ) = f ∈ C[0, ∞) : fe ∈ D(L) .
But since fe0 (x) = f 0 (x)1{x>0} − f 0 (−x)1{x≤0} we need f 0 (0) = 0 to have that fe0 is contin-
uous. Hence (see Example 3.14 and the two remarks following it)
1
Lrefl f = f 00 , D(Lrefl ) = f ∈ C[0, ∞) : f 0 , f 00 ∈ C[0, ∞), f 0 (0) = 0 .
2
Example 4.5 (Brownian motion with absorption). Let τ = inf{s ≥ 0 : Bs = 0}.
Define (with S = R+ )
Xtabs = Bt 1{t<τ } .
Xtabs is a Feller process, hence there is a corresponding transition semigroup (Ttabs )t≥0 and
generator Labs . For f ∈ C[0, ∞) let fo be the odd extension of f in R, i.e.
fo (x) = f (x)1{x≥0} + 2f (0) − f (−x) 1{x<0} .
Using the strong Markov property, we have
1
Ex [fo (Bt )1{τ ≤t} ] = Ex [fo (−Bt )1{τ ≤t} ] = Ex fo (Bt ) + fo (−Bt ) 1{τ ≤t} = f (0)Px (τ ≤ t).
2
Therefore for x ≥ 0
(Ttabs f )(x) = Ex f (Xtabs )1{τ >t} + Ex f (Xtabs )1{τ ≤t} ]
generator, although Dr is dense in C[0, ∞). Namely Labs and Lrefl are true extensions of
A. But a generator cannot truly extend another generator.
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 39
(ii) For s < t we can use the Markov property to write Ex [Mt |Fs ] in the following way
Z Z
Ex [Mt |Fs ] = Ex [f (Xt−s ◦ θs )|Fs ] − Lf (Xu )du − Ex Lf (Xu ◦ θs )duFs
[0,s] [0,t−s]
Z Z
= EXs [f (Xt−s )] − Lf (Xu )du − EXs Lf (Xu )du
[0,s] [0,t−s]
Z
= EXs [Mt−s ] − Lf (Xu )du
[0,s]
Z
= f (Xs ) − Lf (Xu )du Px -a.s. for all x ∈ S.
[0,s]
Hence
Z Z
f d(µTt ) = (Tt f )dµ. (4.2)
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 40
Definition 4.8. µ ∈ M1 (S) is invariant for the Feller process (Xt )t≥0 with transition
semigroup (Tt )t≥0 if µTt = µ for all t ≥ 0, i.e.
Z Z
(Tt f )dµ = f dµ, ∀f ∈ C(S), t ≥ 0. (4.3)
We write I for the collection of invariant distributions. Due to (4.3), I is a convex set.
We write Ie for the set of extremal elements1 of I.
w
Exercise 4.9. Let µ ∈ M1 (S) and assume that µTt −−−→ ν. Then ν is invariant.
t→∞
Theorem 4.10. Let D be a coreR for L. Then µ ∈ M1 (S) is invariant for the corresponding
Feller process if and only if Lf dµ = 0 for all f ∈ D.
(ii) Assume
Z
Lf dµ = 0, ∀f ∈ D.
Since D is a core, there is, for each f ∈ D(L), a sequence (fn )n∈N , fn ∈ D for all n
such that fn → f and Lfn → Lf . Therefore
Z
Lf dµ = 0, ∀f ∈ D(L).
R R
If f ∈ D(L), f − λLf = g, then f dµ = gdµ. Iteration yields
Z Z
−n
(1 − λL) gdµ = gdµ.
The following theorem gives a sufficient condition for the existence of an in-
variant distribution.
1
An element v in a convex set C is extremal if
Hence
Z Z Z Z
f dνn − (Tt f )dνn = f dνn − f d(νn Tt )
S S S S
Z Z Z Z
1
= Tr f dµdr − Tr f dµdr
n [0,n] S [t,n+t] S (4.4)
Z Z Z Z
1
= Tr f dµdr − Tr f dµdr
n [0,t] S [n,t+n] S
n→∞
−−−→ 0.
Since S is compact, M1 (S) is compact as well (Prohorov’s theorem, see [Liggett] Theorem
A.21) and therefore there exists a subsequence (nk )k∈N such that
w
νnk −−−→ ν,
k→∞
for some ν ∈ M1 (S). Since Tt f ∈ C(S), we can take the limit in (4.4) along the subse-
quence (nk )k∈N and we get
Z Z
f dν = Tt f dν.
4.4 Examples
Example 4.13. Had seen: If (Xt )t≥0 is BM, then the corresponding generator is Lf = 12 f 00
with
D(L) = {f ∈ C(R) : f 0 , f 00 ∈ C(R)}.
Consider Yt = Xct , c > 0, t ≥ 0, then (Yt )t≥0 has the generator
Ex [f (Yt )] − f (x) Ex [f (Xct )] − f (x) c
lim = c lim = f 00 (x), f ∈ D(L).
t↓0 t t↓0 ct 2
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 42
N2 + 2N3
((1 − x)2 , 2x(1 − x), x2 ), where x = .
2N
In other words,
N 2n N − n 1
P(N
e1 = n1 , N
e2 = n2 , N
e3 = n3 ) = (1 − x) 1
(2x(1 − x))n2 x2n3
n1 n2
if n1 + n2 + n3 = N and zero otherwise. Therefore N e2 + 2Ne3 has the law Bin(2N, x). Let
Xn denote the proportion of allele A in the n-th generation. (Xn )n∈N is a Markov chain
1 2
in discrete time with state space {0, 2N , 2N , ..., 1}. Heuristics:
2N
X 2N k 2N −k k
LN f (x) = Ex [f (X1 ) − f (x)] = x (1 − x) f − f (x) . (4.5)
k=0
k 2N
Theorem 4.15. Let S = [0, 1] and Lf (x) = 21 x(1 − x)f 00 (x), f polynomial. Then
and
Z
Px (Xτ = 1) = x, Ex Xt (1 − Xt )dt = x(1 − x).
R+
(b) Let f be a polynomial, f −λLf = g for some λ ≥ 0. f has a minimum in x0 ∈ [0, 1]. If
x0 ∈ (0, 1) then f 00 (x0 ) ≥ 0. If x0 ∈ {0, 1} then Lf (x0 ) = 0. In any case Lf (x0 ) ≥ 0,
hence
f − λLf = g. (4.7)
Pn k
Assume f (x) = k=0 bk x , then (4.7) becomes
1
bk − λ k(k + 1)bk+1 − (k − 1)kbk = ak , k ∈ {1, . . . n}, a0 = b0
2
with bn+1 = 0. Solve this recursively, starting with bn and ending with b1 . Hence
R(1−λL) contains all polynomials and is dense in C[0, 1]. Hence R(1−λL) = C[0, 1]
due to Lemma 4.2 (d).
lim Xt =: X∞ exists P-a.s. and Ex [Xt ] = x for all x ∈ [0, 1]. (4.8)
t→∞
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 44
for some γ, β, C > 0. Then the paths t 7→ Xt (ω), 0 ≤ t ≤ T are continuous for P-a.a. ω.
Proof. See [Liggett] Theorem 3.27.
To check the hypothesis, fix y ∈ [0, 1], f (x) = (x − y)2 . Then Lf (x) = x(1 − x) and
therefore
Z
2
(Xt − y) − Xs (1 − Xs )ds, t ≥ 0
[0,t]
is a martingale. Hence
Z
2 1
Ey [(Xt − y) ] = Ey [Xs (1 − Xs )]ds ≤ t, (4.10)
[0,t] 4
since x(1 − x) ≤ 41 . But this does not suffice to apply Theorem 4.16! Take f (x) = (x − y)4 ,
then Lf (x) = 6x(1 − x)(y − x)2 . Therefore also
Z
4
(Xt − y) − 6 Xs (1 − Xs )(Xs − y)2 ds, t ≥ 0
[0,t]
is a martingale and
Z
4
Ey [(Xt − y) ] = 6 Ey [Xs (1 − Xs )(Xs − y)2 ]ds
[0,t]
Z
3
≤ Ey [(Xs − y)2 ]ds
2 [0,t]
(4.10)
Z
3 1 3
≤ tds = t2 .
2 [0,t] 4 16
CHAPTER 4. GENERATORS, MARTINGALES, INVARIANT DISTRIBUTIONS 45
where f (x) = 2x log(x) + 2(1 − x) log(1 − x). Note that f 00 (x) = x2 + 1−x 2
, x ∈ (0, 1) and
1 00 2
hence 2 x(1 − x)f (x) = 1 for all x ∈ (0, 1). However f 6∈ C [0, 1] and f 6∈ D(L) because
otherwise f (Xt ) − t would be a martingale, but f (Xt ) − t ≤ 0 and f (Xt ) − t → −∞, which
yields a contradiction to the martingale convergence theorem. Therefore take fε ∈ C 2 [0, 1]
with fε = f in [ε, 1 − ε]
then
Z
fε (Xt ) − Lfε (Xs )ds
[0,t]
is a martingale and hence f (Xτε ∧t ) − (τε ∧ t) is a Px -martingale for ε < x < 1 − ε. Hence
Exercise 4.17. Let (Bt )t≥0 be BM, x ∈ [0, 1] and τ = inf{s ≥ 0 : Xs ∈ {0, 1}}. Then
Ex [τ ] = x(1 − x).
Spin systems
We want to find conditions on c(x, η) such that the closure of L is a generator. Define
X
|||f ||| := sup |f (η (x) ) − f (η)|
η∈S
x∈V
and let
D := f ∈ C(S) : |||f ||| < ∞ .
46
CHAPTER 5. SPIN SYSTEMS 47
Proof. If f ∈ D, let
i→∞
where ξi ∈ S is a sequence with ξ0 = η, f (ξi ) −−−→ f (ξ). Then
X
|f (η) − f (ξ)| ≤ αf (x) = dαf (ξ, η),
x∈I
(ii) |||f ||| < ∞ alone does not imply the continuity of f . Proof: exercise.
Proof. Let V = {x1 , x2 , ...}. For all ε > 0 there exists Nε such that for η ∈ S
X X
c(xi , η)|f (η (xi ) ) − f (η)| ≤ Cαf (xi ) < ε.
i≥Nε i≥Nε
Now the claim follows since f ∈ C(S) and c(xi , ·) is continuous for every i ∈ N.
Interpretation:
• inf η∈S [c(u, η) + c(u, η (u) )] is the amount of flipping at u that occurs independently
of the rest of the configuration.
• So ε is the minimal amount of flipping that occurs without interaction between sites.
• a(x, u) describes how much the flip rate at site x depends on the configuration at u
• and M gives the maximal amount of dependence of the flip rate at a site on the rest
of the configuration.
Let `1 (V ) be the Banach space of functions α : V → R that satisfies
X
kαk := |α(x)| < ∞.
x∈V
Note that if M < ∞, then Γ is well-defined and bounded, and we have kΓk = M . More
precisely, note that
!
X X X X X X
kΓαk = |Γα(u)| ≤ |α(x)a(x, u)| = a(x, u) |α(x)| ≤ M |α(x)|
u∈V u∈V x:x6=u x∈V u:u6=x x∈V
which gives kΓk ≤ M and we leave it as an exercise to show equality. For f ∈ C(S) and
x ∈ V , let
Remark. Note that if λM < 1 + λε, then the operator in (5.5) is well-defined, since
∞ k
(1 + λε)1 − λΓ]−1
≤
1 X λ
kΓkk .
1 + λε k=0 1 + λε
Theorem 5.2. Let M < ∞ and define L as in (5.1) for f ∈ D. Then L is a generator
and the corresponding semigroup (Tt )t≥0 satisfies
for all f ∈ D(L), where the inequality is again pointwise. In particular, if f ∈ D then
Tt f ∈ D and we have
Here
(
0 |u − x| =
6 1
a(x, u) = 1
2d
|u − x| = 1.
and
X
M = sup a(x, u) = 1.
x∈V
u:u6=x
We have
(
(u) 0 |u − x| =
6 1
a(x, u) = sup |c(x, η) − c(x, η )| =
η∈S λ |u − x| = 1.
Ln is a generator for the spin system, where the spins {η(x), x 6∈ Vn } are frozen. Let us
write Γn for the quantity from (5.2) corresponding to Ln , and note that Γn ≤ Γ. That is,
if a ∈ `1 (V ) is non-negative then
In particular kΓn k ≤ kΓk. Since Ln is a bounded operator (the sum in (5.7) is finite), we
have
due to Lemma 3.15 and Lemma 4.2 (c) for all λ ≥ 0. Hence for a given g ∈ D, there is
fn ∈ C(S) such that
fn − λLn fn = g.
Now let λ be small enough for λM < 1 + λε. Since Ln satisfies (5.3) we must have fn ∈ D
for all n ∈ N, due to the remark after Lemma 5.1. Hence we can define
The first inequality is from Lemma 5.1, and the last from (5.8) and the definition of the
inverse. Since ∆g ∈ `1 (S) the r.h.s. of (5.9) goes to zero as n → ∞, hence gn → g. It
follows that g ∈ R(1 − λL), so
D ⊂ R(1 − λL).
We conclude that R(1 − λL) is dense in C(S) as well. Due to Lemma 4.2 (d), R(1 − λL)
is a closed subspace of C(S). Therefore
Hence, L has the properties (a) - (d) in Definition 3.8 and is a generator.
Definition 5.5. The spin system (ηt )t≥0 (the Feller process with generator L from The-
orem 5.2) is called ergodic if there exists a unique invariant distribution µ and
Z
lim (Tt f )(η) = f dµ, ∀η ∈ S, f ∈ C(S).
t→∞
In other words (ηt )t≥0 is ergodic if there exists µ ∈ M1 (S) such that
w
(νTt ) −−−→ µ, ∀ν ∈ M1 (S).
t→∞
Example (Voter model (continued)). Recall Example 5.3. The process (ηt )t≥0 is not
ergodic, since there are at least two invariant distributions: δ0 (the Dirac measure on
the configuration η(x) = 0, ∀x ∈ V ) and δ1 (the Dirac measure on the configuration
η(x) = 1, ∀x ∈ V ) are invariant.
Therefore
X
sup |f (η) − f (ξ)| ≤ ∆f (x) = |||f |||.
η,ξ
x∈V
and hence
sup |Tt f (η) − Tt f (ξ)| ≤ |||Tt f ||| ≤ e(M −ε)t |||f |||.
η,ξ
5.3 Examples
Example 5.7. Voter model (more general). Let V be a vertex set and let (p(x, y))x,y∈V
be an irreducible stochastic matrix with p(x, x) = 0 for all x ∈ V . Let
X
c(x, η) = p(x, y).
y:η(x)6=η(y)
Interpretation: With rate p(x, y), the voter in x takes the opinion of the voter in y.
Have
X
M = sup p(x, u) = 1.
x∈V
u:u6=x
Example 5.8. Contact process (more general) Let (V, E) be a graph with vertices
V and edges E and assume that the graph has bounded degree. Write x ∼ y if x, y are
neighbors, i.e. if (x, y) ∈ E. Fix λ > 0. Let
(
1 η(x) = 1
c(x, η) =
λ#{y : y ∼ x, η(y) = 1} η(x) = 0.
Interpretation: η(x) = 1 means that the particle at x is infected, η(x) = 0 means that
the particle at x is healthy. Infected particles recover with rate 1, healthy particles are
infected by their neighbors with rate λ. The Dirac measure δ0 on “everybody healthy” is
invariant.
Question: Are there other invariant measures? Theorem 5.6 implies that I = {δ0 } if
1
λ< .
maxx∈V deg(x)
Example 5.9. Noisy voter model. Let (p(x, y))x,y∈V be an irreducible stochastic matrix
with p(x, x) = 0 for all x ∈ V . Fix β, δ ≥ 0 and let
X
c(x, η) = p(x, y) + δ1{η(x)=1} + β1{η(x)=0} .
η:η(x)6=η(y)
We have ε = 2 and M = 2de2dβ (1 − e−2β ). Hence (ηt )t≥0 is well-defined, and ergodic
if β is small enough. In fact, as we will see soon:
(i) (ηt )t≥0 is always ergodic if d = 1.
(ii) (ηt )t≥0 is not ergodic for β large enough if d ≥ 2.
Theorem 5.12. Let (ηt )t≥0 be a spin system with rates c1 (x, η) and (ξt )t≥0 be a spin
system with rates c2 (x, ξ). If η ≤ ξ implies that
then there is a coupling (ηt , ξt )t≥0 for starting values η ≤ ξ such that
P(η,ξ) (ηt ≤ ξt , ∀t ≥ 0) = 1.
(ηt , ξt )t≥0 is a Feller process with S = {(0, 0), (0, 1), (1, 1)}V and its generator is given in
terms of the rates as in (5.1). The construction of the process is analogous to the proof of
Theorem 5.2. The assumptions guarantee that all rates are non-negative. Finally we have
to check that (ηt )t≥0 and (ξt )t≥0 are Feller processes with rates c1 (x, η) and c2 (x, η). For
instance
and
Note that (ηt )t≥0 and (ξt )t≥0 are not independent. In the particular case c1 (x, η) =
c2 (x, ξ) in (5.10), this leads to the definition of an attractive spin system.
Definition 5.13 (Attractive spin system). The spin system with rates c(x, η) is at-
tractive if for all η ≤ ξ,
Example. The contact process, the noisy voter model and the Ising model are
attractive.
Example. The anti-voter model on Zd , given by
1 X
c(x, η) = 1{η(y)=η(x)}
2d y:x∼y
is not attractive.
Corollary 5.14. For an attractive spin system, there is a coupling (ηt , ξt )t≥0 of two copies
of the process, such that for η ≤ ξ we have
P(η,ξ) (ηt ≤ ξt , ∀t ≥ 0) = 1.
Definition 5.15.
(i) A function f ∈ C(S) is increasing if
η ≤ ξ ⇒ f (η) ≤ f (ξ).
Lemma 5.16. Let (ηt )t≥0 be an attractive spin system. Then the transition semigroup
(Tt )t≥0 satisfies
(i) f ∈ C ↑ ⇒ Tt f ∈ C ↑ , ∀t ≥ 0.
(ii) µ ν ⇒ µTt νTt .
Proof. (i) Let η ≤ ξ. Then
(Tt f )(η) = Eη [f (ηt )]
= E(η,ξ) [f1 (ηt , ξt )], f1 (η, ξ) := f (η)
Corollary 5.14 ≤ E(η,ξ) [f2 (ηt , ξt )], f2 (η, ξ) := f (ξ)
= Eξ [f (ξt )]
= (Tt f )(ξ).
Theorem 5.17. Let (ηt )t≥0 be an attractive spin system with semigroup (Tt )t≥0 . Then we
have
ν and ν are called upper invariant measure and lower invariant measure.
(ii) For each sequence (tk ) with tk → ∞, δ0 Ttk has a limit point (due to compactness) -
but due to (i), all these limit points agree. (Exercise: make this argument precise!).
Theorem 5.19. The stochastic Ising model (see Example 5.10) with d = 1 is ergodic for
all β ≥ 0.
and
Let m ∈ N,
c(x, η), |x| < m
cm (x, η) = K, |x| ≥ m, η(x) = 0
0, |x| ≥ m, η(x) = 1.
1
the limits are w.r.t. weak convergence of probability measures on S
CHAPTER 5. SPIN SYSTEMS 57
Due to Theorem 5.12 one can couple the spin systems (ηt )t≥0 and (ξt )t≥0 with starting
configurations η0 ≡ 1, ξ0 ≡ 1 and rates c(x, η) and cm (x, ξ) such that ηt ≤ ξt for all t ≥ 0.
The process ξt (x), −m+1 ≤ x ≤ m−1, is an irreducible MC in continuous time with state
space {0, 1}2m−1 . Therefore the law of ξt converges weakly towards the unique invariant
distribution µm for t → ∞. Extend µm on {0, 1}Z by setting, for ζ ∈ {0, 1}Z ,
Then ν µm . Let us compute µm . Let µ ∈ M1 ({0, 1}Z ) the law of a (two-sided) time-
discrete MC with state space {0, 1} and transition matrix
e−β
β
1 e
P = β
e + e−β e−β eβ
1
and with µ(ζ(0) = 0) = 2
= µ(ζ(0) = 1). The finite-dimensional marginals of µ are given
by
l−1
1Y
µ({ζ : ζ(x) = z(x) for k ≤ x ≤ l}) = p(z(j), z(j + 1)) (5.11)
2 j=k
for z ∈ {0, 1}Z . Claim The conditional law of µ(·|ζ(x) = 1, ∀xwith |x| ≥ m) is reversible
for the (time-continuous) MC (ξt )t≥0 , hence it is invariant. This implies that
Proof. Exercise.
Now we show that µ = limm→∞ µm . For −m < k < l < m
where P (i) (·, ·) are the entries of the i-th power of P . The convergence theorem for MC
in discrete time gives us that
1
lim P (n) (u, v) = , ∀u, v ∈ {0, 1}.
n→∞ 2
Hence the term in (5.12) converges to the r.h.s. of (5.11). Since ν µm and limm→∞ µm =
µ, we conclude that ν µ. The same game, exchanging 0 and 1 yields µ ν and hence
ν = ν, which implies that (ηt )t≥0 is ergodic with unique invariant distribution µ.
Corollary 5.20. The unique invariant distribution for the stochastic Ising model with
d = 1 and β ≥ 0 is the probability measure µ with finite dimensional marginals given by
(5.11).
CHAPTER 5. SPIN SYSTEMS 58
d
Consider the stochastic Ising model with S = {0, 1}Z . For simplicity of notation we
d
can pass to S = {−1, +1}Z and take
X
c(x, η) = exp − β η(x)η(y) . (5.13)
y:|y−x|=1
Theorem 5.21. Assume that d ≥ 2. Then there exists βc = βc (d) such that for all
β > βc , the stochastic Ising model has at least two invariant distributions and is therefore
not ergodic.
Remark. We know by Theorem 5.6 that the stochastic Ising model is ergodic if β is small
enough. In the stochastic Ising model we have
ε = 2, M = 2de2dβ (1 − e−2β ),
Proof of Theorem 5.21. Assume first that d = 2. We have to show that for β large enough
ν 6= ν. Let Λ = {−n, . . . , n}×{−n, . . . , n} and let µΛ,+ be the (unique) invariant distribu-
tion for the (finite state) MC on S = {−1, 1}Λ given by the rates in (5.13) and η(x) = 1,
∀x ∈ Λc . (Exercise: what is the Q-matrix of this Markov chain?) We will show that for β
large enough
1
lim µΛ,+ (η(0) = −1) < .
|Λ|→∞ 2
Lemma 5.22.
1 X
µΛ,+ (η|Λ ) = exp β η(x)η(y) .
Z x∈Λ,y∈Λ∪∂Λ
|y−x|=1
Peierl’s contour argument: For each configuration, draw edges between spins of oppo-
site signs. Then, these edges form contours and any pair of configurations (obtained from
each other by flipping all signs) corresponds to a unique contour configuration. Contours
are cycles in the dual graph given by {(x, y) : (x + 1/2, y + 1/2) ∈ Z2 }, with edges between
x and y if |x − y| = 1.
- + + + +
- + + + +
+ + - - +
+ + - - +
+ + + + +
To estimate the probability that η(0) = −1, observe that if η(0) = −1, the origin must
lie inside at least one contour. Let B(η) be the set of all contours of η. We estimate, for
γ a contour surrounding 0, µΛ,+ (γ ∈ B(η)). Let
|B(η)| = #{edges in B(η)}.
Then, due to Lemma 5.22
−2β|B(η)|
P
η:γ∈B(η) e
µΛ,+ (η : γ ∈ B(η)) = P −2β|B(η)|
.
η∈S e
Define ηe by
(
−η(x), if γ surrounds x
ηe(x) =
η(x), otherwise.
Then B(e
η ) is B(η) without the contour γ. Hence
−2β|B(eη )|
P
−2β|γ| η:γ∈B(η) e
µΛ,+ (η : γ ∈ B(γ)) = e P −2β|B(η)|
,
η∈S e
| {z }
≤1
We saw already that on any graph with bounded degree there is a Feller process - the
contact process - corresponding to these rates, and that the contact process is ergodic if
1
λ< .
max deg(x)
x∈V
Our goal is now to show that on Zd the contact process is not ergodic if λ is large enough.
t
δ
δ δ
δ δ
-2 -1 0 1 2
60
CHAPTER 6. THE CONTACT PROCESS 61
Let (x, s), (y, t) ∈ V × [0, ∞), s ≤ t. We define a (directed) path from (x, s) to (y, t) to be
a sequence
We write (x, s) → (y, t) if there exists such a directed path. Note that (x, s) → (y, t)
means that y is infected at time t if x is infected at time s. Let η0 ∈ S = {0, 1}V and
define ηt ∈ S for t ∈ [0, ∞), by taking ηt (y) = 1 if and only if there exists x ∈ V such
that η0 (x) = 1 and (x, 0) → (y, t). Then (ηt )t≥0 is a contact process with parameter λ.
1. It gives a coupling of all the contact processes with parameter λ and any starting
configuration η0 .
PA B
λ (ηt = 0 on B) = Pλ (ηt = 0 on A),
Proof. The event on the l.h.s. of (6.1) is the union over a ∈ A and b ∈ B of the events
(a, 0) → (b, t). If we reverse the direction of time and the direction of the arrows of
infection, the probability of the event does not change and is now the probability of the
event on the r.h.s. of (6.1).
CHAPTER 6. THE CONTACT PROCESS 62
By Theorem 6.1,
{0} d
Pλ (ηt ∩ Zd 6= ∅) = Pλ (ηtZ (0) = 1)
and by the weak convergence of δ1 Tt to ν,
d t→∞
Pλ (ηtZ (0) = 1) −−−→ ν({γ ∈ S : γ(0) = 1}).
Since ν is shift-invariant, the claim follows.
We define the critical value λc of the contact process by
λc = λc (d) = sup{λ : θ(λ) = 0} ∈ [0, ∞].
λ 7→ θ(λ) is non-decreasing. Hence
(
= 0 λ < λc
θ(λ)
> 0 λ > λc .
Theorem 6.3. For d ≥ 1, we have
1
≤ λc (d) < ∞.
2d
Remark. The bounds can be improved.
Proof of Theorem 6.3. We know already that
1
λc (d) ≥ ,
2d
1
since for the contact process ε = 1 and M = 2dλ, hence M < ε if λ < 2d
. Since Z is a
subgraph of Zd , we have
λc (d) ≤ λc (1).
Hence it suffices to show λc (1) < ∞. Fix ∆ > 0 and let m, n ∈ Z such that m + n are
even. We define independent random variables Xm,n , taking values 0 or 1. Let Xm,n = 1
and call (m, n) open, if in the graphical representation, the following two events occur:
CHAPTER 6. THE CONTACT PROCESS 63
λ2
2
2+ λ
> psite
c
(1 + λ)
and the last inequality is satisfied for λ large enough (we have p(∆) → 1 for λ → ∞).
Hence λc (1) < ∞.
Proof of Lemma 6.4. Peierl’s argument: A finite cluster of open sites has to be sur-
rounded by a closed contour. Consider contours surrounding the connected component of
sites which are reachable from (0, 0).
Pp (∃ closed contour of length n) ≤ 3n (1 − p)n ,
since there are at most 3n possible contours of length n and each is closed with probability
(1 − p)n . Now, for p close to 1,
X
3n (1 − p)n < 1
n∈N
Theorem 6.5. Consider the contact process on Zd , d ≥ 1. Then there is λc ∈ (0, ∞) such
that
I = {ν} = {δ0 } ⇐⇒ λ ≤ λc .
Remark. We showed that there is λc ∈ (0, ∞) such that λ < λc ⇒ {ν} = {δ0 } and
λ > λc ⇒ Ie = {δ0 , ν} and ν 6= δ0 . The critical case λ = λc is delicate, but it has been
proved that θ(λc ) = 0 for d ≥ 1.
Chapter 7
Let V be a countable
P vertex set and let q(x, y) ≥ 0, for all x 6= y, x, y ∈ V . Assume that
M = supx∈V u:u6=x q(x, u) < ∞. Then we know that the spin system (ηt )t≥0 with rates
X
c(x, η) = q(x, y), x ∈ V, η ∈ {0, 1}V
y:η(y)6=η(x)
is well-defined and a Feller process. The voter model is never ergodic, since δ0 and δ1
are invariant.
Question. Are there other extremal invariant distributions, i.e. Ie = {δ0 , δ1 }?
65
CHAPTER 7. THE VOTER MODEL 66
(1)
Check that in this case, Tt and L2 commute for any t ≥ 0. Let
(1)
u(t, z1 , z2 ) = Ez1 [H(Zt1 , z2 )] = Tt H(·, z2 )(z1 ). (7.2)
(1)
where we use the assumption in the second equality and the fact that Tt and L2 commute
in the third equality. For each z1 ∈ S1 , the unique solution of this differential equation
with initial condition u(0, z1 , z2 ) = H(z1 , z2 ) is given by
X
pt (z2 , y)H(z1 , y) = Ez2 [H(z1 , Zt2 )], (7.3)
y∈S2
see [Liggett] Theorem 2.39. (7.1) follows from (7.2) and (7.3).
For the voter model we consider
Y
H(η, A) = η(x) = 1{η(x)=1,∀x∈A} ,
x∈A
where η ∈ {0, 1}V , A ⊆ V, |A| < ∞. The dual process (At )t≥0 is a system of coalescing
MC with Q-matrix q(x, y). More precisely, the points in A move independently according
to this MC, and coalesce when they meet.
A0 At1
(At )t≥0 is a MC in continuous time. The state space consists of all finite subsets of V and
the Q-matrix is given by
Remark. Due to our assumption M < ∞, there is a continuous time MC with this
Q-matrix.
Theorem 7.3. The processes (ηt )t≥0 and (At )t≥0 are dual. w.r.t.
Y
H(η, A) = η(x).
x∈A
Interpretation. The Markov chain describes the back-tracking of opinions. Let t > 0.
Follow the development which lead to ηt (x). Possibly, there is t1 < t, t1 > 0 and x1 ∈ V
such that the voter in x took the opinion ηt1 (x1 ), otherwise ηt (x) = η0 (x). We continue
going back in time: 0 < tm < tm−1 < ... < t0 = t and ηt (x) = η0 (xm ). The path going
at time ti , from xi−1 to xi is the path of a continuous time MC Yx (t) starting at x, with
Q-matrix q(x, y). For different x, the MC are not independent, but they coalesce when
they meet. With this construction ηt (x) = η0 (Yx (t)).
CHAPTER 7. THE VOTER MODEL 68
Remark. Lets assume the starting law is the product measure νρ with finite-dimensional
marginals
νρ ({η : η(x) = 1, ∀x ∈ A}) = ρ|A| , ρ ∈ (0, 1).
Then we have
P(ηt (x) = 1) = P(η0 (Yx (t)) = 1) = ρ.
P(ηt (x) = ηt (y) = 1) = P(η0 (Yx (t)) = η0 (Yy (t)) = 1)
= ρP(Yx (t) = Yy (t)) + ρ2 P(Yx (t) 6= Yy (t))
Since (ηt )t≥0 has the same law as (1 − ηt )t≥0 we also have
Pη (ηt (x1 ) = ηt (x2 ) = 0) =E1−η (ηt (x1 ) = ηt (x2 ) = 1)
(1) (2)
=E{x1 ,x2 } 1 − η(At ) = 1 − η(At ) = 1
(1) (2)
=P{x1 ,x2 } η(At ) = η(At ) = 0
In the same way,
P(ηt (x) = 0) = P(η0 (Yx (t)) = 0) = 1 − ρ.
P(ηt (x) = ηt (y) = 0) = P(η0 (Yx (t)) = η0 (Yy (t)) = 0)
= (1 − ρ)P(Yx (t) = Yy (t)) + (1 − ρ)2 P(Yx (t) 6= Yy (t)) .
Letting t → ∞,we see that the probability that ηt (x) and ηt (y) have the same value goes
to 1 if and only if the two MC, (Yx ) and (Yy ) coalesce almost surely.
(1) (2)
Let (Xt )t≥0 and (Xt )t≥0 be independent MC with rates q(·, ·) and starting points
(1) (2)
x1 , x2 ∈ V . Take now V = Zd and assume (Xt )t≥0 and (Xt )t≥0 are irreducible. Let
(1) (2)
Zt = Xt − Xt , t ≥ 0.
(ii) Ie = {δ0 , δ1 }.
(iii) We have
lim µTt = αδ1 + (1 − α)δ0
t→∞
if and only if
X
lim pt (x, y)µ({η : η(y) = 1}) = α, ∀x ∈ V (7.4)
t→∞
y∈V
(1)
where pt (·, ·) is the transition function of (Xt )t≥0 .
CHAPTER 7. THE VOTER MODEL 69
Pη (ηt (x1 ) = ηt (x2 ) = 1) =Eη (H(ηt , {x1 , x2 })) = E{x1 ,x2 } (H(η, At ))
(1) (2)
=P{x1 ,x2 } η(At ) = η(At ) = 1 .
Using the bijection between (ηt )t≥0 and (1 − ηt )t≥0 we also have
due to (i). Hence η is µ-a.s. constant and therefore µ = αδ1 + (1 − α)δ0 for some
α ∈ [0, 1].
therefore (µTt ) converges for t → ∞ (see the exercise below), hence limt→∞ µTt is
invariant. Hence
due to (ii).
for any finite A ⊆ V . Then µTt converges weakly to an element of M1 (S) for t → ∞.
Then
Z
(νρ Tt )({η : η(x) = 1, ∀x ∈ A}) = Pη (ηt (x) = 1, ∀x ∈ A)νρ (dη)
Z
= PA (η(x) = 1, ∀x ∈ At )νρ (dη)
X (7.5)
= PA (At = B)νρ ({η : η(x) = 1, ∀x ∈ B})
B:B⊆V
finite
= EA [ρ|At | ]
Since |At | is non-increasing, we know a∞ = limt→∞ |At | exists (note that a∞ is a random
variable). Let t → ∞ in (7.5) and conclude that
Hence µρ is invariant. Now take V = Zd , assume that q(x, y) = q(0, |y − x|) and that
(1) (2) (1) (2)
(Zt )t≥0 is transient. Recall that Zt = Xt − Xt , t ≥ 0,P
where Xt , Xt are independent
MC with Q-matrix Q = (q(x, y)) x,y∈Zd and q(x, x) = − u6=x q(x, u).
x6=y
In the third step, we also used the reversibility: Px (Zt = z) = Pz (Zt = x) since
q(x, y) = q(0, |y − x|) = q(y, x) and this implies that the measure π(x) = 1, ∀x is a
reversible measure for the MC (Zt ). Hence we have
Px (Zs = y for some s > t) ≤ P0 (Zs = 0 for some s > t). (7.8)
More precisely
Z Z
Px (Zs = y)ds = Ex 1{Zs =y} ds = Px (Zs = y for some s > t) G(y, y).
[t,∞) [t,∞) | {z }
=G(0,0)
(7.9)
Taking x = y = 0, we get
Z Z
P0 (Zs = 0)ds = E0 1{Zs =0} ds = P0 (Zs = 0 for some s > t)G(0, 0).
[t,∞) [t,∞)
(7.10)
and in particular,
XZ
P0 (Zτx +s = x, τx ≤ t)ds ≤ t + 1
x∈V [0,1]
Hence
X M (t + 1)
P0 (τx < t) ≤
x∈V
1 − e−M
µ ◦ θx−1 = µ, ∀x ∈ Zd .
d
A translation invariant probability measure µ on S = {0, 1}Z is mixing if
(ii) The probability measures µρ are translation invariant and mixing for ρ ∈ [0, 1].
(iii) µρ ({η : η(x) = 1}) = ρ.
(iv) The covariances of µρ are given by
G(x, y)
Covµρ (η(x), η(y)) = ρ(1 − ρ) . (7.13)
G(0, 0)
Proof. (i)
µρ ({η : η(x) = 1, ∀x ∈ A}) − ρ|A| = EA [ρA∞ − ρ|A| ]
= EA [(ρa∞ − ρ|A| )1{a∞ <|A|} ]
≤ g(A).
where c∞ = limt→∞ |Ct |, b∞ = limt→∞ |Bt |. Replace B with B + x and apply Lemma
7.6 (i), then the last term goes to zero as |x| → ∞.
(iii) and (iv)] This follows from (7.5). Note that, with A = {x, y},
Covµρ (η(x), η(y)) = µρ ({η : η(x) = η(y) = 1}) − ρ2
= E{x,y} [ρa∞ − ρ2 ]
= ρ(1 − ρ)P{x,y} (a∞ = 1)
= ρ(1 − ρ)Px−y (Zt = 0 for some t > 0)
= ρ(1 − ρ)Px (Zt = y for some t > 0)
G(x, y)
= ρ(1 − ρ) ,
G(0, 0)
where the last equality follows from (7.9) with t = 0 and
Z
G(x, y) = Px (Zs = y)ds = Px (Zs = y for some s > 0)G(0, 0).
R+
CHAPTER 7. THE VOTER MODEL 75
Further
Z
Covµρ (η(x), η(y)) = Covµρe (η(x), η(y))γ(deρ)
Z
G(x, y)
Theorem 7.8 (iv) = ρe(1 − ρe)γ(de
ρ)
G(0, 0)
Z Z
G(x, y)
Jensen ≤ ρ) (1 − ρe)γ(de
ρe γ(de ρ)
G(0, 0)
G(x, y)
= ρ(1 − ρ).
G(0, 0)
Hence it follows that γ = δρ , since Jensen’s inequality is strict if γ is not a Dirac measure.
This shows that {µρ : ρ ∈ [0, 1]} ⊆ Ie . For the rest of the proof, we refer to [Liggett]
Theorem 4.43.
.
Bibliography
76