Professional Documents
Culture Documents
e
H(xxx)
, (9.1)
where H(xxx) denotes the Hamilton function of the system, Z
=
_
e
H(xxx)
dxxx (9.2)
ensures proper normalization 1 = 1 and yields, e.g., the free energy F() =
k
B
T lnZ
. The main idea of the Monte Carlo method is to treat this problem as
a problem for numerical integration, where, in view of the usually high dimension of
the integral (dimension is proportional to the degrees of freedom N with the num-
ber of particles N for a many-particle system), an equidistant or regular grid in phase
space cannot be used to compute the integral (cf. Table 9.1), but grid coordinates xxx
i
have to be statistically selected, to obtain
A =
1
M
M
i=1
A(xxx
i
) p(xxx
i
) (9.3)
in the limit M. Since the distribution function usually varies over several orders
of magnitude H(xxx) is an extensive variable, H(xxx) N , a regular phase space grid
is often inefcient.
Table 9.1. Convergence behaviors for N function evaluations in d dimensions
Method Scaling Behavior
Trapezoidal N
2/d
regular grid in coordinate space
Simpson N
4/d
regular grid in coordinate space
Standard pseudorandom Monte Carlo N
1/2
(independent of d)
Quasi Monte Carlo N
1
log
D
N (for some D, realized for N > e
d
)
9.1 Expectation Values, Metropolis Monte Carlo
The Metropolis algorithm aims to calculate A without the need to calculate Z
. It
is based on the idea to prefer phase space coordinates located in the relevant part of
the phase space, which depends on temperature and further independent thermody-
namic variables: importance sampling. This is achieved by recursively creating a
Markov chain (one-step memory) of states xxx
i
: xxx
i
xxx
i+1
xxx
i+2
. . . using appro-
priate transition probabilities w(xxx xxx
. These probabilities
must be chosen such that after a large number of steps the states xxx are distributed
according to the given probability distribution p(xxx). It is achieved, for example, by
fullling the so called condition of detailed balance
9.1 Expectation Values, Metropolis Monte Carlo 147
p(xxx)w(xxx xxx
) = p(xxx
)w(xxx
xxx) , (9.4)
where it is important to realize that Z
) = min(1, e
H
), H H(xxx
) H(xxx) . (9.5)
For M the phase space coordinates xxx
i
obtained using a jump probability fulll-
ing (9.4) are distributed according to p(xxx), and the expectation value A is evaluated
as arithmetic average A =
1
M
A(xxx
i
). This approach usually does not yield infor-
mation about the partition function Z or free energy F, even though the partition
function can be expressed itself as an expectation value Z
= 1/
e
H
_
, because, for
the evaluation of
e
H
_
states with high energy H are relevant, but rarely visited.
On the other hand, rare visits of high energy states are the key for understanding
the efciency of the Metropolis scheme for calculating conventional expectation
values such as the mean energy, pressure, or the end-to-end distance of polymers. A
positive aspect of the Monte Carlo method is the exibility in choosing the step size
for a move xxx xxx
. The optimum step, and step size depends on the system under
study. For dense liquids, due to excluded volume, random congurations are usually
very unlikely to be statistically relevant, while small displacements of a statistically
relevant conguration will be often relevant, too. For the Ising model, a single step
may be the reorientation of a single spin, or the reorientation of a cluster of spins, and
the conventional game is to search for most efcient implementations. It is important
to realize that the Monte Carlo trajectory through phase space can be interpreted as
resulting from an underlying dynamics obeying a master equation (loss- and gain-
equation) for the (time dependent) probability distribution. The equation of change
for p(xxx, t) reads
d
dt
p(xxx, t) =
xxx
w(xxx xxx
)p(xxx, t) +
xxx
w(xxx
xxx)p(xxx
, t) . (9.6)
It is obvious that p(xxx), which fullls the condition of detailed balance (9.4) is the
stationary solution of the master equation (9.6). Of course, the meaning of time is
not a physical one, a time unit corresponds to a single Monte Carlo step, and we
have freedom to choose the Monte Carlo step. Dynamic correlations, however, are
important to quantify the precision of the obtained results, and eventually have a
physical meaning, if the conguration space xxx is a subspace of the space of slow
variables, which are coupled to fast degrees of freedom (acting as a heat bath).
We have seen that the Metropolis scheme is not useful for calculating Z
. While
expectation values often contain sufcient information to validate models, the above-
mentioned approach still has strong deciencies. First of all, a new simulation has
to be performed for each temperature. Second, low energy states are visited more
often than potentially needed to extract the relevant information about their effective
contribution to expectation values. Third, close to phase transitions (close to criti-
cal temperatures), the Metropolis scheme seriously slows down, if not sophisticated
moves are introduced which overcome the problem of long-range correlations.
148 9 Equilibrium Statistics: Monte Carlo Methods
9.2 Normalization Constants, Partition Function
In this section we shortly summarize methods to directly compute the partition func-
tion Z
_
q(xxx, ) dxxx . (9.7)
The normalized density, fullling
_
p dxxx = 1, explicitly reads
p(xxx, ) =
q(xxx, )
Z
. (9.8)
A particular q is the so-called geometric path between functions q
0
and q
1
:
q(xxx, ) = q
0
(xxx)
1
q
1
(xxx)
. (9.9)
We see, that the canonical distribution can be considered as a special geometric (tem-
perature) path if we choose
q
0
(xxx)
can
= 1, q
1
(xxx)
can
= e
H(xxx)
,
Z
can
=
_
e
H(xxx)
dxxx, Z
0
can
=
_
1 dxxx =V . (9.10)
Here, we introduce the notation
can
= whenever the particular choice of a canonical
(can) distribution for q
1
, q
0
of (9.10) is made to simplify a more general expression,
and V denotes volume. Of course it is then identical calculating Z
1
for all , or, for
all , Z
at xed = 1. From Z
lnZ
, H =
d lnZ
d
, (9.11)
as well as entropy S = (H F), heat capacity C
V
=
2
(
H
2
_
H
2
) etc.. The
canonical distribution maximizes S =
_
pln pdxxx = ln p under the constraint
H =constant and
_
pdxxx = 1. Notice, that appears to be inverse temperature, but
will be eventually used below as a pure, unphysical, interpolating parameter where
0 1 holds. For the same reason, we will be only interested in calculating Z
1
.
Once we have a method to compute Z
1
for given hamiltonian H(xxx), we also know
how to compute Z
=
_
q(xxx, )
P(xxx)
P(xxx) dxxx =I
P
with I =
q(xxx, )
P(xxx)
. (9.12)
The error for the calculated Z
is estimated on the y as
2
=
1
M
(
I
2
_
P
I
2
P
) . (9.13)
This formula is derived making use of the law of large numbers and offers the reason
why the standard deviation decreases with M
1/2
independent of dimension d.
Two special choices for P, denoted as uniform and canonical, are
A) P uniform (constant)
(use uniform pseudo or quasi random generator to approximately realize P)
P =
1
_
1 dxxx
=V
1
Z
=V q
P
can
= V
_
e
H
_
uniform
; (9.14)
B) P q (canonical)
(use Metropolis, rejection method, inversion etc. to approximately realize P)
P =
q
Z
=
V
q
1
P
can
=
V
e
H
_
canonical
. (9.15)
The standard Monte Carlo method still requires one run for each . For the uniform
case (A) it is straightforward to obtain realizations. Using a build-in pseudo random
generator is upon the approximate solutions. For case (A), we miss the relevant
low energy regions, which is bad for above-mentioned expectation values, but good
for calculating Z
to
desired accuracy. Just for the one-dimensional case (d =1) we would like to mention
the variable transformation method to obtain realizations y distributed according to
P(y) using uniformly distributed random numbers with p(x) = V
1
, x [0,V]. By
noticing the identity
_
p(x)dx = 1 =
_
P(y)dy =
_
p(y(x))x
(y)dy (9.16)
we need to solve the differential equation x
p
0
q
0
p
1
can
=
e
H
_
uniform
canonical
. (9.17)
Special cases:
A) = 1/q
0
can
= 1 (uniform)
B) = 1/(q
0
q
1
)
can
= e
H
(canonical)
C) = min(1/q
0
, 1/q
1
) (acceptance ratio method)
D) = q1
2
/(q
0
q
1
) with q1
2
in between q
0
and q
1
(bridge sampling, umbrella sampling, thermodynamic integration)
Z
1
Z
0
=
_
q1
2
/q
0
_
p
0
_
q1
2
/q
1
_
p
1
can
=
_
q1
2
_
uniform
_
q1
2
e
H
_
canonical
(9.18)
Functions q1
2
and q
0
, q
1
should possess overlap. Otherwise rene by using several
spans, or perform iterations with several q1
2
. Optimal path (weighted harmonic
mean):
q1
2
(xxx) =
p
0
p
1
s
0
p
0
(xxx) +s
1
p
1
(xxx)
, s
t
= n
t
/(n
0
+n
1
) . (9.19)
This is not directly usable since it depends on Z
1
/Z
0
. Iterative schemes have
been proposed, e.g., in [407], which can reduce the simulation error by orders
of magnitude when compared to the conventional importance sampling method.
Considering innitely many spans q
, q
2
, . . . q
1
, bridge sampling equals path
sampling.
9.2.3 Path Sampling
Path sampling is based on basic identity
d
d
lnZ
=
_
1
Z
d
d
q(xxx, ) dxxx =
_
d
d
lnq(xxx, )
_
p
can
= H(xxx)
p
, (9.20)
where denition (9.8) was used. Integrating (9.20) over from 0 to 1 yields, for the
physical case,
ln
Z
1
Z
0
can
=
_
H(xxx)p(xxx, ) dxxx d =
_
H(xxx)
P()
p(xxx; ) dxxx d , (9.21)
This expression involves a joint distribution p(xxx; ) = p(xxx, )P(), and P() which
may be, for example, uniform. In practise, the goal is to nd the optimum path P().
9.3 Density of States Monte Carlo (DSMC) 151
9.3 Density of States Monte Carlo (DSMC)
Density of states Monte Carlo takes an orthogonal view to the problem of high-
dimensional integration. It is based on the identity, with Q(E) e
E
Z
Z
0
=
_
q(xxx, ) dxxx
Z
0
can
=
_
Q(H(xxx))
dxxx
Z
0
=
_
n(E)e
E
dE =
_
Q
(E)
_
n(E)
, (9.22)
where n(E), due to (9.22), is dened as a normalized density of states,
_
n(E)dE =1.
The appeal of this method, introduced in [408], lies in the fact, that n(E) is indepen-
dent of . Thus Z
), t)/p(H(xxx), t)).
Start (i = 1):
p(E, t) = f
h
i
(E,t)
(h
i
is the histogram recorded in round i up to time t) with
large f = e to rapidly explore the whole E range, and converge towards a at
h
i
(E, t
max
) (achieved at time t
max
). Set p
i
(E) p(E, t
max
).
Iteration step (i +1):
p(E, t) = p
i
(E) f
h
i+1
(E)/(i1)
to obtain for large i a steady, polished, p(E)
and to approximately fulll detailed balance in the corresponding Metropolis
scheme.
152 9 Equilibrium Statistics: Monte Carlo Methods
The method is easy to implement, and it is possible to extend the method by us-
ing parallel runs with different, overlapping E windows. A drawback of density of
states Monte Carlo seems to lie in the limited range of accessible expectation values
A with A=A(E). Histogram recording becomes memory consuming if generalized
canonical distributions are considered (involving more than a single Lagrange para-
meter). However, one can record realizations A and obtain A in a postprocessing
step, or further explore the numerical precision of the following, trivial, identity
_
A
n
_
=
,
Z
_
e
H(xxx)
dxxx,
H(xxx) H(xxx) nlnA(xxx) , (9.23)
which requires performing two independent simulations, with two different hamil-
tonians H and
H (but otherwise identical) to obtain
A
n
_
for all and a single
n. To obtain A for several , which may the most classical task, we would need
to run several simulations with xed n = 1/, which must be compared with the
conventional Metropolis scheme of Sect. 9.1.
Further Extensions
The density of states Monte Carlo method had been recently extended to compute
phase diagrams of dense LennardJones uids and binary glasses with N particles
in a volume V, characterized by a two-body interaction potential U(xxx) in [410, 411].
Here, use is made of the identity (xed N,V)
(E) = k
B
S
E
|
V
=
ln(N,V, E)
E
, (9.24)
where denotes the microcanonical partition function. Integration over E yields
ln(E) =
_
E
E
0
(E
)dE
. (9.25)
Further, an expression for the congurational temperature [412, 413],
(E) =
i
FFF
i
_
i
|FFF
i
|
2
_ =
a(E)
b(E)
(9.26)
is employed, which is obtained using a phase point transformation, assuming that at
constant E, the phase space is uniformly occupied, while U is smooth. Histograms
a, b, h, are evaluated and is used as guide for a walker in energy space. Further,
the known probability of observing a conguration xxx having total potential energy
U(xxx) is used as guide for a walker in conguration space. Finally, one obtains a
smooth, since integrated, distribution from a and b. In all other respects, the above
listed algorithm is adapted for this application.
9.4 Quasi Monte Carlo 153
9.4 Quasi Monte Carlo
The strikingly simple idea of quasi Monte Carlo is to replace pseudo random num-
bers by a quasi random sequence which is known to be uniform in high dimensional
xxx space (exhibiting low discrepancy). Quasi Monte Carlo methods use determin-
istic samples at points that belong to low discrepancy sequences and approximate
the integrals by the arithmetic average of N function evaluations. According to the
KoksmaHlawka inequality their worst case error is of order (1/N) ln
d
N; where d
denotes the dimension. Since this term becomes huge when N is xed and d is large,
as sometimes happens in practice, traditionally, there has been a certain degree of
concern about Quasi Monte Carlo [414] presents sufcient conditions for fast Quasi
Monte Carlo convergence which apply to isotropic and non-isotropic problems. It is
shown that the convergence rate of Quasi Monte Carlo is of order N
1+p/lnN
1/2
with
p 0. Compared to the expected rate N
(
1/2) of Monte Carlo it shows the superi-
ority of Quasi Monte Carlo. To understand the success of Quasi Monte Carlo in some
applications, also the notion of effective dimension has been introduced, cf. [415].
One example is the Richtmeyer sequence xxx
(n)
[0, 1]
d
with x
(n)
= n
_
P
mod
1 with prime number P. Other often used sequences are the so called Faure and
Niedermeyer sequences. There is no exhaustive knowledge about the overall ef-
ciency of quasi Monte Carlo methods for computing high dimensional integrals.
Quasi Monte Carlo is certainly good for low-dimensional integration, but supersedes
pseudo Monte Carlo only if N > e
d
(huge for large d). It is yet an empirical observa-
tion that quasi random numbers should not be used for solving stochastic differential
equations (Langevin equations) [416].