Professional Documents
Culture Documents
NE StatPhys Greffetv2
NE StatPhys Greffetv2
Introduction to Nonequilibrium
Statistical Physics
Master QLMN
Jean-Jacques Greffet
2
Contents
3
4 CONTENTS
1.1 Introduction
From a historical perspective, statistical physics is the consequence of an attempt to un-
derstand the physical properties of gases from a mechanical modelling of the movement of
individual molecules. The goal is therefore to recover concepts of thermodynamics such
as pressure, temperature, entropy from mechanical concepts such as momentum or kinetic
energy. It is clearly an attempt to go from a microscopic description to a macroscopic de-
scription.
From a microscopic description to a macroscopic description. In practice, a me-
chanical description of a gas based on the computation of the position and velocity of all
the molecules is impossible. Numerical simulations can be done on thousands of atoms
but certainly not with a mole of atoms. Furthermore, it is impossible to know with enough
accuracy the initial conditions of a large system to predict its evolution. Hence, we see that
a statistical approach has to rely on some average procedure for practical reasons. At this
point, a statistical approach seems to be necessary to simplify the equations. While this is
true, it is only one of the features of statistical physics.
Complex systems. The statistical approach introduced by Boltzmann and Gibbs to
study gases has been extended to many other physical systems and gases are certainly not
the best example of the potential of a statistical approach. For instance, it is not possible to
understand the behaviour of a neutron star or the properties of a magnetic bit in a hard disk
without using statistical physics. Beyond physical phenomena, the procedure that allows
to move from a microscopic description to a macroscopic description is widely used in
different sciences. Information theory due to Shannon borrowed the concept of entropy.
Many concepts used in macroeconomy are based on similar concepts: here the problem is
to move from the behaviour of individuals to the economic quantities relevant at the scale
of a country. Another example is the study of trafic jams. Here, the issue is to understand
the dynamics of flow of cars from the description of the behaviour of individual cars. We
see that the statistical approach enables to study a large class of systems called complex
systems. We now introduce the idea that a large system is more than the mere addition of
its parts.
Emergence of new concepts and new phenomena. One of the most fascinating fea-
tures of statistical physics is its ability to introduce purely macroscopic concepts. To illus-
7
8 CHAPTER 1. INTRODUCTION TO STATISTICAL PHYSICS. BASIC CONCEPTS.
trate this point, let us consider the example of a gas consisting in N atoms with mass m. If
all velocities and positions are known at a given instant, the classical dynamical variables
are well characterized and we have all the relevant information. From thermodynamics,
we know that a gas can be described using both entropy and temperature. These concepts
do not appear in the microscopic description of the system. They are purely macroscopic
quantities. Hence, we see that statistical physics is not a mere filter that would keep the
essential information and remove the superfluous data. It is a different way to look at a
system. More importantly, statistical physics enable to study phenomena that cannot be
captured by a complete microscopic description such as phase transition. The solid-liquid
phase transition of a gas or the paramagnetic-ferromagnetic phase transition of a metal are
examples where the microscopic description does not help understanding the observed phe-
nomena. These phenomena are truly collective effects resulting from the interaction of its
elements.
Fluctuations. Finally, it is important to stress that statistical physics allows to go be-
yond classical thermodynamics. Indeed, it is possible to compute the fluctuations of the
physical quantities, something that is beyond reach for classical thermodynamics. This is
all the more important as it will be shown that the knowledge of fluctuations allows deriv-
ing the non-equilibrium properties of physical systems. For example, knowing the voltage
fluctuations of an electric circuit allows deriving its impedance and therefore its response to
an external excitation such as an applied voltage.
Summary of equilibrium staistical physics. The purpose of the first chapter is to in-
troduce the key concepts required to describe a system with a statistical approach. We will
first define what is a state of a physical system. We will then introduce the probability that
the system is in a particular state. With that, we will be able to compute mean physical
quantities. Hence, it follows that finding this probability is the fundamental task. Once
this is achieved, the rest of the lectures will consist in learning how to use in practice this
knowledge to compute physical quantities of interest such as a state equation for instance.
In order to illustrate the power of the statistical physics formalism, we will show that we
can recover the basic laws of thermodynamics. We will also show that many laws of clas-
sical thermodynamics such as the perfect gas state equation can be derived. When doing
so, we will realize that one of the most remarkable features of statistical physics is that
an oversimplified model of the microscopic systems often produces a remarkably accurate
macroscopic law.
Outline The non-equilibrium phenomena are the subject of chapters 4-9. We will first
introduce the general properties of linear systems. We will then establish a very general
relation connecting fluctuations and linear response known as the Fluctuation-Dissipation
theorem. We will then introduce the Langevin model which is an alternative technique to
model fluctuations. The next step will be to introduce the Boltzmann equation which allows
to discuss transport phenomena. Finally, we will discuss the basic ideas of irreversible
thermodynamics, a theory introduced to extend the capabilities of classical thermodynamics
to non-equilibrium situations.
the measurement may produce different results. The probability of obtaining a given result
Oi corresponding to the eigenstate φi is given by the square modulus |ci |2 . In summary,
quantum physics introduces the concept of mean value of the outcome of a measurement
for a given physical quantity and a given physical state. What is fundamentally quantum, is
the fact that although the state of the system is fully known and denoted |r >, the outcome
of a measurement may not be fully determined. On the other hand, the mean value of the
observable is given by
D E
r|Ô|r .
A mean value in statistical physics has a completely different origin. It is due to the lack
of knowledge of the state occupied by the system. For instance, when dealing with a gas, we
do not know what is the position and momentum of the 3N atoms. In other words, we need
to assign a probability that the system is in state r. The key problem of statistical physics is
to find this probability. Once this probability is known, a mean value can be computed and
therefore the physical quantities can be computed.
10 CHAPTER 1. INTRODUCTION TO STATISTICAL PHYSICS. BASIC CONCEPTS.
Let us denote Pr the probability that the system occupies the quantum state |ri. Then,
we know that Dif the system
E is in this state, the mean outcome of a measurement of the ob-
servable Ô is r|Ô|r . The statistical average O of the measurement, given the uncertainty
on which state is actually occupied, is thus given by
X D E
O= Pr r|Ô|r .
r
We use a bar to denote the statistical average to emphasize the difference with the quan-
tum average. Let us now make three important remarks:
i) If we know the probability law, it is possible to derive the mean values. For instance,
we can derive the mean energy of a gas of N non-interacting atoms by adding the kinetic
energy of the N atoms and averaging over all possible states. For a system of N spins with
unknown orientations, it will be possible to derive the mean magnetization, etc.
ii) A very important feature of a statistical approach is that if the probabilities are known,
then one can also compute the fluctuations of the physical quantities. This is a very impor-
tant property that goes well beyond the concept of average value of a quantity at equilibrium.
This is a major contribution of statistical physics.
iii) The key concept is the probability that a system occupies a given state. It is very
important to note that such a probability can be assigned to a system independently of its
size. In other words, there is nothing that prevents to use statistical physics when dealing
with a system that contains three atoms. The specificity of a very small system is that the
fluctuations will be very large. However, all the concepts that we introduce can be applied.
Remark : Although the definition of the statistical average was introduced above using
a quantum description of the system, we emphasize that the concept of statistical average
can also be used in a classical framework. The bottom line is considering a system that can
be in different microscopic (classical or quantum) states with different probabilities.
function, it is clearly seen that the probability density can be defined for systems with either
large or small number of atoms N . What is needed, is a large number of realizations of the
system in the (fictitious) ensemble of realizations.
In summary, there is no need to have a large number of particles to use a statistical
physics approach. The statistical character of the approach comes from the fact that the
microscopic state of the system is unknown. The reason for a statistical model is not related
to the number of particles in the system. It will be seen later that the number of particles
does
√ play a significant role regarding the fluctuations: the relative fluctuations will decay as
1 N . Hence, systems with a small number of particles will display large fluctuations so
that a statistical approach becomes essential.
1.2.5 Ergodicity
In order to define the probability, we have introduced the concept of ensemble of micro-
scopic realizations of the same macroscopic state. An alternative approach consists in
studying a single system as a function of time. One could think of measuring all the states
occupied by the system as a function of time and derive from this procedure the probability
density. This would be a time average.
A system is ergodic if the time average is equal to the ensemble average. In other
words, a system is ergodic if the time spent by a given system at each point in phase space
is proportional to the probability density.
The question of ergodicity is often emphasized because it allows to provide an answer
to the following question. In practice, one is interested in a given system. For example, if
we study a gas in a volume V with N atoms, we make measurements on this single system.
Hence, the question arises: why an average over a collection (an ensemble) of microscopic
realizations gives a relevant model of the experimental mesurements performed on a single
system ? Assuming that the experimental measurement yields a time average of the result
and assuming that the system is ergodic, we can assume an ensemble average to be a good
model of the experimental measurements. While this argument seems plausible, it cannot be
used in the vast majority of experiments as the time required to explore all the phase space
(and therefore perform a meaningful time average) is usually many orders of magnitude
larger than the actual measurements.
So it has to be concluded that if the ensemble average provides a good agreement with
the measurements, it is because the macroscopic quantities which are measured are not
12 CHAPTER 1. INTRODUCTION TO STATISTICAL PHYSICS. BASIC CONCEPTS.
sensitive on the details of the microscopic realizations. This argument can be supported by
the direct calculation of the fluctuations over different realizations. We will perform these
calculations and show that indeed, for large systems, the fluctuations are often negligible.
∆E kB T.
However, some care has to be taken when using an integral over phase space. The
sum over quantum numbers has no dimensions whereas an integral in phase space is not
dimensionless. For example, the volume of the phase space associated with the position
and momentum of one atom has the dimension of m6 kg 3 s−3 . Furthermore, if a state is
associated with a single point, there is an infinite number of states. We need to introduce
the elementary volume associated to a single state and this elementary volume must have a
dimension. It can be shown that this elementary volume is given by hN df where Ndf is the
number of degree of freedom. We call degree of freedom, a pair of conjugated variables x
and px in the phase space. For instance, in the case of a single atom, Ndf = 3 corresponding
to the three axis x, y and z. For N atoms, we have Ndf = 3N . As a practical way of
remembering the elementary volume associated to a single state, the Heisenberg uncertainty
principle can be used. It states that ∆x∆px ≈ h. Hence, a single state cannot be a point
(perfect knowledge of position and momentum). An area h is consistent with the Heisenberg
principle. This argument is by no means a proof but can be used as a practical mnemonic
argument.
More generally, the number of states dN in a system with Ndf degrees of freedom is
given by the ratio of the volume element in phase space dΓ divided by the volume element
hNdf :
dΓ
dN ≈ .
hNdf
Here, we give this relation without a proof. However, its validity can be checked easily
using simple systems. Let us consider the case of a simple harmonic oscillator. In what
follows, we derive the number of states using either the quantum approach or the classical
approach. From a classical point of view, we know that there are two variables x and px .
The classical energy of the harmonic oscillator is given by
1 p2
E = mω 2 x2 + x
2 2m
1.3. SUMMING OVER ALL POSSIBLE STATES. THE CLASSICAL APPROXIMATION.13
where ω is the eigenfrequency of the harmonic oscillator. The volume of phase space Γ
2 2
with energy lower than E is the area of the ellipse given by the equation xa2 + pb2x = 1 where
a2 = 2E/mω 2 and b2 = 2mE. This area is given by
2πE
Γ = πab = .
ω
We now estimate the number of states with energy lower than E using quantum me-
chanics. It is known that the energy spectrum of a harmonic oscillator is given by En =
(n + 1/2)~ω. We now assume that E ~ω. It follows that the number of states with
energy between 0 and E is given by
N = E/~ω.
Upon comparison of the two results, it is seen that N = Γ/h. Note that the quantum
estimate of N is accurate only if N is much larger than 1. This means that the energy of the
system (which is typically given by kB T ) is much larger than the quantum of energy ~ω. In
other words, the energy spectrum can be considered to be continuous.
In summary, if the quantum of energy is much smaller than the actual energy of the
system, it is possible to move from a quantum description to a classical description of the
states of the system. This is called the classical approximation. From a practical point of
view, one has to compute sums over all states weighted by the statistical probability. The
classical approximation enables to replace sum over discrete states by an integral over the
phase space volume with the proper normalization:
Z
X dΓ
−→ N
.
r Γ h
In the case of N atoms with no other degrees of freedom than their translational degrees
of freedom, we have:
N R R
X Y dxi dyi dzi dpxi dpyi dpzi
−→ .
r
h3
i=1
It is often practical to perform the sum by grouping all states r with same energy En . We
then need to introduce the degeneracy gn of the energy level En . The sum can then be cast
in the form: X X
−→ gn
r n
where r is a label for a state and n is a label for an energy level. If the classical approx-
imation is valid, then a similar procedure can be used when performing the integral over
the phase space. We then replace the integral over dΓ by an integral over the energy. We
also need to introduce the number of states with energy in the interval [E, E + dE] which
is given by g(E)dE. Here, we have introduced the function g(E) called density of states.
Finally, we can write:
X X −−−−−−−−−−−−−−−−−→ Z
−→ gn classical approximation g(E)dE.
r n
It will be seen later on that g(E) is not a mere technical definition but instead a quantity
that plays a fundamental role in the physical description of the system.
14 CHAPTER 1. INTRODUCTION TO STATISTICAL PHYSICS. BASIC CONCEPTS.
We now address the central question of statistical physics. How can we assign a probability
to a given state ? We first consider a system in equilibrium and isolated. Hence, the energy
of the system takes a given value E. It follows that we have to consider only states with
energy E. The states which are compatible with the information that we have about the
system (e.g. energy, number of particles, volume) are called accessible states. Having
restricted the number of possible states, we now need to assign a probability to each of
them. As we do not have any additional information allowing to suppose that one state
is more likely than another one, we have to assign the same probability to all the states.
Any other choice would require some additional information to be justified. Given this
consideration, we take as a postulate that all states have equal probability at equilibrium.
This statement is the fundamental principle of statistical physics:
For an isolated system at equilibrium, all accessible states have the same probability.
This choice is justified a posteriori by the fact that the theory allows to describe the
experimental observations. The ensemble corresponding to isolated systems is called mi-
crocanonical ensemble. The corresponding probability is called microcanonical probability.
Discussion
In this paragraph, we want to show that the fundamental principle does play a role analogous
to the second principle in classical thermodynamics. We know that the second principle is
the principle that allows to find the equilibrium state of the system among all possible states.
At first glance, the fundamental principle does not enable to choose as it postulates that
all states are equivalent. In order to clarify this issue, we emphasize that the fundamen-
tal principle states that all microscopic states have the same probability. The question of
the choice of the equilibrium state in classical thermodynamics is about the choice of a
macroscopic state. So the actual question is : How many microscopic states correspond to
a given macroscopic state ? The equilibrium macroscopic state is the one corresponding to
the largest number of microscopic states.
Let us take an example. We consider a simple model of paramagnetism. The system
consists of N spins 1/2 without interactions. No magnetic field is applied so that each spin
has a probability 1/2 to be in the state + (sz = 1/2) and a probability 1/2 to be in the
state − (sz = −1/2). We expect such a system to have a zero magnetic moment2 . We
assume that the spins are localized in sites in a crystal so that they are distinguishable. A
given state is specified by the list of the orientation of the N spins (+, +, −, +, −, −, .....).
According to the fundamental principle, the state (+, +, +, +, +, ....+, +) where all the
spins are in the state + has the same probability that a state with N/2 + and N/2 −.
This seems counterintuitive as we expect the system to be in a state with no magnetization.
To resolve this apparent paradox, we clarify the difference between a microscopic state
(+, +, −, −, −) and a macroscopic state given by the number n+ of spins +.
2
We remind the reader that the magnetic moment is proportional to the spin
1.4. FUNDAMENTAL PRINCIPLE OF STATISTICAL PHYSICS 15
We do expect the state n+ = N/2 to be more likely than the state n+ = N . However,
the microscopic state (+, +, +, ...+, +) has indeed the same probability that any other state.
The key point is that there is only one microscopic state corresponding to n+ = N whereas
there are N !/[n+ !(N −n+ )!] different microscopic states corresponding to the macroscopic
state n+ . The observed macroscopic state is the state corresponding to the largest number
of microscopic states. It can be checked using ln N ! ≈ N ln N − N that it corresponds to
n+ = N/2.
The key feature that allows deciding which ”macroscopic” state is actually observed is
therefore the number of microscopic states corresponding to a given macroscopic situation.
Let us denote by Ω this number. The entropy introduced by L. von Boltzman is S =
kB ln Ω so that maximizing the entropy and maximizing the number of microscopic states
corresponding to a given macroscopic situation is one and the same.
where Pr is the probability that the system is in a microscopic state r. Boltzmann had
two goals: to find the velocity distribution derived by Maxwell and to establish a form of
the entropy that would allow showing that entropy can only increase. Note that this form
of the statistical entropy allows to recover the simple Boltzmann formula in the case of an
isolated system. Indeed, if Pr = 1/Ω, then S = kB ln Ω.
the entropy will provide a positive result S > 0. With this first example, we indeed see
that the entropy increases as the amount of missing information increases. In other words,
the missing information is the fact that we do not know in which microscopic state is the
system. If the number of possible microscopic states increases, then the amount of missing
information increases.
Let us now take a step back and think about the link between information and proba-
bility. We toss the dices 600 times and we find that the outcome is always 1. This is an
extremely unlikely result. It is so unlikely that we will probably think that the dices are
loaded and we will ask to take a close look at the dice. By contrast, if the outcome is 100
times the value 1, 98 times the value 2, etc, we will not be surprised and will not pay much
attention. In the first case, the probability of the event is very unlikely and the event get our
attention. It contains a lot of information: the dice is loaded! In the second case, the event
is very likely and does not contain much information.
Hence, information increases as probability decreases. If we introduce a function H(Pr )
measuring the amount of information, it has to be a decreasing function of Pr . We now go
a step forward and consider the particular case of the event consisting of tossing the dices
four times. Obviously, this event A can be split in two independent events B and C : tossing
twice the dices and repeating it. As the tosses are independent, the probability is given by
P (A) = P (B and C) = P (B)P (C). It follows that : H[P (A)] = H[P (B)P (C)]. We
now look for a measurement of information as an additive law. It seems natural to define H
as a function where the amount of information contained in B and in C can be added. We
then require H[P (B)P (C)] = H[P (B)] + H[P (C)]. This entails that H(P ) = −K ln(P )
where K is an arbitrary positive constant.
We have thus found that the measurement of missing information must be given by
−K ln P in order to have a law that satisfies the two basic requirements : decaying function
of P and additive function. We now introduce the average value of this missing information
over all possible events and call it entropy:
X
S = −K Pr ln Pr .
r
2.1 Introduction
In this chapter, we study systems in contact with a thermostat. A thermostat can exchange
energy and has a fixed temperature. It can be viewed as a system with an infinite heat ca-
pacity. The first purpose of this chapter is to introduce a technique to derive the probability
for a system to be in a given state from the fundamental principle. The second purpose is to
study how to extract physical quantities (mean value and fluctuations) from the knowledge
of these probabilities. We will see that a quantity plays a key role in this context: the parti-
tion function. The general technique of study of a system in contact with a thermostat will
be illustrated using the example of a perfect gas. Finally, we will compare the formalism of
statistical physics with classical thermodynamics thereby revisiting concepts such as work
and heat in the statistical physics framework.
∂S
= 0, (2.1)
∂Pr
for all states r. Yet, the probabilities are not independent. Indeed, the normalization condi-
tion yields:
17
18CHAPTER 2. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE
X
Pr = 1. (2.2)
r
Hence, we need to deal withP the problem of finding the values Pr that maximize the
entropy given the constraint r Pr = 1. This problem of finding an extremum in the
presence of constraints of the form f (P1 , ...Pr , ..) = 0 can be solved using the Lagrangian
multipliers technique. Here, we only show how to use this technique without giving any
proof.
where a parameter λi called Lagrangian multiplier has been introduced for each con-
straint. Here, we have introduced two constraint but the procedure can be reproduced with
any number of constraints. We now search the maximum of this new function considering
that the variables x1 , x2 , ..., xN are independent. The final result depends on the unknown
Lagrangian multipliers. Their value can be found using the constraints but this step can be
avoided in most cases as we will see.
As an example, you can quickly check the efficiency of the technique with the following
example. What are the values of the sides a and b of a rectangle with perimeter P that
maximizes the area of the rectangle ? Here, the variables are a, b, the function is f (a, b) =
ab and the constraint is g(a, b) = 2(a + b) − P = 0. You should find that the solution is a
square so that a = b = P/4. We can further check that the extremum is indeed a maximum
in this case.
Hence, Pr0 depends on λ but does not depend on the state P label r so that the probability
is the same for all states. Accounting for the constraint r Pr = 1, we find immediately
Pr = 1/Ω where Ω is the number of accessible states in agreement with the first formulation
of the fundamental principle.
2.3. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE 19
X
Pr = 1 (2.6)
r
X
Pr Er = E (2.7)
r
where we have introduced two Lagrangian multipliers λ1 , λ2 . Writing that the derivative
with respect to Pr0 is zero yields:
−kB ln Pr0 − k + λ1 + λ2 Er0 = 0. (2.9)
It is seen that Pr0 depends on the energy Er0 and can be cast in the form:
exp(−βEr0 )
Pr0 = . (2.10)
Z
In the previous equation, instead of solving in terms of the Lagrangian multipliers, we
have introduced a parameter β called statistical temperature. We have also introduced a
normalization factor Z called partition function. By definition, it is given by:
X
Z= exp(−βEr ). (2.11)
r
Hence, we have been able to derive an explicit form of the probability Pr that the system
is in a given state r from the principle of maximum entropy. We now show how to use this
probability law to derive physical quantities.
We now note that the numerator can be derived from the partition function by taking a
derivative with respect to β:
∂Z X
=− Er exp(−βEr ), (2.13)
∂β r
20CHAPTER 2. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE
∂ ln Z
E=− . (2.14)
∂β
We stress that this expression reveals another aspect of the partition function Z. While it
has been first introduced as a normalization factor, it appears now that it plays the role of a
potential which contains information on the system. This is the first time that we observe
this feature. It will be seen in what follows that the partition function can be related to the
potentials introduced in classical thermodynamics. It is a fundamental quantity in statistical
physics.
X exp(−βEr ) 1 ∂2Z
E2 = Er2 = (2.16)
r
Z Z ∂β 2
!2
1 ∂Z 2
X exp(−βEr )
2
E = Er = (2.17)
r
Z Z ∂β
Hence, we find:
2
1 ∂2Z ∂ 2 ln Z
1 ∂Z ∂ 1 ∂Z
(E − E)2 = − = = (2.18)
Z ∂β 2 Z ∂β ∂β Z ∂β ∂β 2
Again, we note that this quantity can de derived from the knowledge of the partition
function.
2.3.4 Entropy. Link between the statistical temperature and the temperature
in classical thermodynamics
The calculation of the entropy is straightforward. It suffices to insert the form of the proba-
bility in the definition.
X X
S = −kB Pr ln Pr = −kB Pr [−βEr − ln Z] = kB ln Z + kB βE. (2.19)
r r
Note that we can compute the entropy and not only an entropy variation during a trans-
formation, in marked contrast with classical thermodynamics. We now proceed to identify
the temperature. We use the thermodynamic identity:
1 ∂S
= , (2.20)
T ∂U
2.3. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE 21
where U denotes the internal energy of the system. Here, we identify U with E and we note
that ln Z is an explicit function of β but does not depend on E. Indeed, from its definition,
the partition function depends on β and the list of energy levels of the system. We find:
∂S
= kB β. (2.21)
∂E
By comparing with the definition of the temperature in classical thermodynamics, we find:
1
β= . (2.22)
kB T
We can thus recover the free energy:
F = E − T S = −kB T ln Z, (2.23)
and remark that it is proportional to the partition function. This is the first example where
we can identify −kB T ln Z with a thermodynamic potential.
∂E ∂E ∂β 1 ∂ 2 ln Z (E − E)2
cv = = = 2 2
= . (2.24)
∂T ∂β ∂T kB T ∂β kB T 2
We find that the heat capacity, which is an example of a linear response coefficient, is
proportional to the rms deviation of the energy.
1
An interaction energy term between two particles such V (r1 , r2 ) could not be assigned to the energy of
particle 1 or particle 2
2.4. EXAMPLE: PERFECT GAS 23
p2n
where Er = N
P
n=1 2m . As the classical approximation is valid, we can replace the sum
over discrete states by an integral over phase space.
N Z
Y dxn dyn dzn dpx,n dpy,n dpz,n p2n
Z= exp[−β ] = ζN . (2.27)
h3 2m
n=1
where
4πp2 dp −βp2 /2m
Z
ζ=V e (2.28)
h3
Note that the factorization is made possible by the fact that the energy of the system is
the sum of the energies of the independent particles with no interaction terms. Hence,
the problem is simply the computation
R∞ function ζ corresponding to a single
of a partitionp
particle. Using the identity: 0 t2 exp(−at2 )dt = 41 aπ3 , we get
3/2
2πm
ζ=V . (2.29)
βh2
∂ ln Z 3N 3N kB T
E=− = = . (2.32)
∂β 2β 2
Let us now study the energy fluctuations using Eq. (2.18).
2 ∂ 2 ln Z 3N
E2 − E = 2
= . (2.33)
∂β 2β 2
24CHAPTER 2. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE
We are thus led to examine if these two terms can be identified with the elementary heat and
work variations.given by Eq.(2.35). This is indeed the case for some particular cases. To
proceed, we start with the calculation of the elementary variation of the entropy.
Entropy variation
Since we know the explicit form of the entropy as a function of the probabilities Pr , we can
derive the entropy variation:
X
dS = −kB d[Pr ln(Pr )]. (2.37)
r
P
We note that the sum of probabilities is equal to 1 so that r dPr = 0. We thus get:
X
dS = −kB ln(Pr )dPr . (2.38)
r
Work
P
We now consider the second term r Pr dEr . Let us again consider a particular transfor-
mation. We consider an adiabatic (δQ = 0) tranformation
P so that dU = δW . We also
consider that the transformation is isentropic (dS = r Er dPr = 0 as was shown above.
In that case, the energy variation dU = dE yields:
X
δW = Pr dEr . (2.41)
r
It is seen that the elementary work in a reversible and adiabaitic transformation is as-
sociated to a modification of the energy of the states of the system. Let us illustrate this
idea with the perfect gas.The work received by the system is given by −P dV in classical
thermodynamics. We now revisit this issue using a microcopic point of view. We consider
a vessel with volume V = Lx Ly Lz containing atoms. If the length Lx is modified using a
~2 4π 2
moving piston, the kinetic energy levels of the particles given by n2x 2m L2x
are also mod-
ified. If this modification is slow enough, it does not induce quantum transitions between
states.
In summary, providing energy to the system through work amounts to change slowly
the energy of the states of the system.
26CHAPTER 2. SYSTEM IN CONTACT WITH A THERMOSTAT. CANONICAL ENSEMBLE
X
Er dPr = T dS
r
X
Pr dEr = δWrev . (2.46)
r
The parameter l can be the volume of the system, a magnetic field, an electric field, etc.
P tehn derive the corresponding generalized force F through the relation δW = F dl =
We
r Pr dEr :
δW = F dl (2.49)
X ∂Er
= Pr dl (2.50)
r
∂l
X exp(−βEr ) ∂Er
= dl (2.51)
r
Z ∂l
1 ∂ ln Z
= − dl. (2.52)
β ∂l
The final result can be cast in the form:
1 ∂ ln Z
F =− (2.53)
β ∂l
This relation connects the generalized force with its conjugate variable. Hence, the relation
is the state equation connecting F and l.
1 ∂ ln Z
P = . (2.54)
β ∂V
Inserting the partition function derived in Eq.(2.30), we get
N N kB T
P = = . (2.55)
βV V
Chapter 3
3.1 Introduction
The goal of this chapter is to extend the canonical approach to systems which can exchange
energy with their environment and which have an additional constraint. We start with an
open system that can exchange energy and particles with a reservoir. In that case, the
number of particles N in the system can fluctuate but is fixed on average. Hence, a new
constraint is given and the probability law is modified. Other examples of systems in equi-
librium include systems with fluctuating quantities such as electric charge, volume, magne-
tization or polarization which are fixed on average upon interaction with the environment.
The first goal of this chapter is to introduce a general technique to deal with these systems.
The ensemble corresponding to systems exchanging particles is called grand canonical and
the ensemble of systems with a constraint on the quantity X will be called generalized grand
canonical.
Another goal of this chapter is to show that the relevant thermodynamic potential will
be given by −kB T ln ZGC where the subscript stands for grand canonical. We will also
introduce the thermodynamic limit and discuss the equilibrium conditions between subsys-
tems.
29
30CHAPTER 3. SYSTEM IN CONTACT WITH A RESERVOIR. GRAND CANONICAL ENSEMBLE
where Nr is the number of particles when the system is in state r and N is the mean number
of particles. Applying the same method than in the previous chapter, we find:
exp(−βEr + αNr )
Pr = (3.4)
ZGC
dU = T dS + µdN, (3.9)
µ = kB T α. (3.11)
Hence, it is seen that the chemical potential is the intensive variable that controls the ex-
change of particles between the system and the reservoir in the same way than the temper-
ature controls the exchange of energy between the system and the thermostat. We can now
cast the probability law in the form:
exp(−βEr + βµNr )
Pr = . (3.12)
ZGC
Finally, we can introduce the thermodynamic potential:
exp(−βEr + αXr )
Pr = . (3.17)
ZGCX
3.3.2 Entropy
Inserting the probability in the definition of the statistical entropy, we obtain:
X X
S = −kB Pr ln Pr = −k Pr [−βEr +αXr −ln ZGCX ] = kB βE−kB αX+kB ln ZGCX .
r r
(3.18)
We note that the Lagrangian multiplier is given by:
1 ∂S
α=− . (3.19)
kB ∂X E
dU = T dS + xdX. (3.20)
so that α = x/kB T .
By expressing the Lagragian multiplier in terms of the conjugated parameter x, we find:
exp(−βEr + βxXr )
Pr = . (3.22)
ZGCX
32CHAPTER 3. SYSTEM IN CONTACT WITH A RESERVOIR. GRAND CANONICAL ENSEMBLE
2 ∂ 2 ln ZGCX
X2 − X = . (3.24)
∂α2
∂X
χ= , (3.25)
∂x
so that ∆X = χ∆x.
We now insert the explicit form of the mean value X in the definition of the suscepti-
bility:
2
∂ 2 ln ZGCX X2 − X
χ = kB T 2
= . (3.26)
∂x kB T
Hence, it is seen that for any quantity X, the corresponding susceptibility is given by its
fluctuations. This relation has been derived for a static perturbation. This general relation
will be generalized to a time-dependent perturbation. A frequency dependent susceptibility
will then be defined. Its connection with the time-dependent fluctuations characterized by
the correlation function < X(t)X(t0 ) > is called the fluctuation-dissipation theorem.
∂ ln ZGCX
X = kB T .
∂x
This equation relates x and X, it is the state equation of the system. Here, it is given in
terms of the grand canonical partition function. It is also possible to derive the state equation
when using a canonical approach. Indeed, the perfect gas state equation is valid for both a
closed and open system. We now derive this alternative form of the state equation. Instead
of computing the average value of X, we can compute the value Xmax of X that maximizes
the probability P (X). The two values, X and Xmax can be considered to be equal if and
only if the probability P (X) has a very narrow distribution. This assumption
√ is valid for
large systems as we have seen that the relative fluctuations decay as 1/ N .
3.4. THERMODYNAMIC LIMIT 33
Let us first write the probability P (X). It is simply the sum of the probabilities Pr over
all the states such that Xr = X:
X exp(−βEr ) ZC (β, X)
P (X) = exp(αX) = exp(βxX) (3.27)
ZGCX ZGCX
r/Xr =X
We now remark that if X is fixed, the system is described by a canonical distribution and
its partition function is given by ZC (X). It is now a simple matter to find the maximum of
the probability:
∂P (X)
= 0. (3.28)
∂X
We find:
∂Zc
βxZC + = 0, (3.29)
∂X
so that:
∂ ln ZC ∂ ln ZC
x = −kB T = −kB T (3.30)
∂X Xmax ∂X X
This equation yields an alternative form of the state equation connecting x and X in
terms of the canonical partition function.
2 1 ∂ 2 ln ZGC
N2 − N = . (3.31)
β2 ∂µ2
34CHAPTER 3. SYSTEM IN CONTACT WITH A RESERVOIR. GRAND CANONICAL ENSEMBLE
so that the average number of atoms and their fluctuations are given by:
1 ∂ ln ZGC 2
N= = ζ exp(βµ) ; N 2 − N = N . (3.34)
β ∂µ
√
We see that the relative fluctuations vary as 1/ N . If N is on the order of the Avogadro
number, this is on the order of 10−12 . This conclusion suggests that using a canonical model
to describe the system considering that the number of atoms is fixed and takes the value N
should provide accurate results.
We repeat the analysis for a perfect gas and compute the fluctuations of the energy for a
perfect gas in contact with a thermostat. The energy fluctuations were shown to be:
∂ 2 ln Z
(∆E)2 = . (3.35)
∂β 2
Using the canonical partition function, we find:
ζN
Z = (3.36)
N!
2πm 3/2
ζ = V (3.37)
βh2
3N
(∆E)2 = (3.38)
2 β2
∆E 1
≈ √ (3.39)
E N
Here again, we see that the relative energy fluctuations are negligible if 1 N .
system. This limit is called the thermodynamic limit. It is valid for macroscopic systems
with a number of particles on the order of the Avogadro number.
This limit paves the way to simplify the calculations by choosing the most convenient
ensemble. In practice, we need to know how to compute a temperature in an isolated system,
how to compute a chemical potential in a system which is not in contact with a reservoir,
etc.
1 ∂S
= . (3.40)
T ∂U V
∂ZC (N )
µ = −kB T . (3.41)
∂N
We have seen that for a canonical ensemble, it coincides with the free energy. It is worth
pointing out two features of the grand potential:
i) All the relevant information on the properties of the system can be derived from it. Indeed,
we have seen that the partition function contains all the information.
ii) By definition of the partition function, the grand potential is a natural function of β and
x. By contrast, it is not a function of the mean energy and of X.
A = E − T S − xX. (3.44)
so that TR S−U +xR X is maximum. Thus, the grand potential U −TR S−xR X is minimum
at equilibrium. This form of the equilibrium condition will be very useful to discuss phase
transitions.
dU = T dS − P dV (3.46)
dH = T dS − P dV + d(P V ) = T dS + V dP. (3.47)
Similarly, systems are either isolated or can exchange heat with their environment so
that there is a variation of entropy. When dealing with systems in contact with a thermostat
that controls the temperature, the natural potential is F = U − T S whose elementary
variation is given by:
dF = −P dV − SdT (3.48)
3.7. EXAMPLES 37
For open systems exchanging particles with a reservoir, the particle number can fluc-
tuate and the chemical potential is imposed by the reservoir. The relevant potential is then
F − µN = U − T S − µN .
In statistical physics, we have introduced the potential as −kB T ln Z where Z is the
partition function corresponding to the relevant ensemble. Relevant ensemble means that
the constraints associated with the fact that a variable X is fixed or fluctuating has already
been taken into account. In the general case, the grand potential can be cast in the form
A = E − T S − xX. (3.49)
Let us emphasize that the grand potential A = −kT ln ZGCX is a function of T and x as
can be seen directly on the explicit form of the partition function:
X
ZGCX = exp(−Er /kT + xXr /kT ). (3.50)
r
3.7 Examples
3.7.1 Exercise: Grand canonical ensemble P-T
Problem text
1. Use the Lagrange multiplier technique to derive the probability of a state with energy
Er and volume Vr . Show that it can be cast in the form exp(−βEr + αVr )/ZGCT P .
2. Derive the mean volume and the volume fluctuations.
3. Compute the entropy. Derive the grand potential −kB T ln ZGCT P . What is the corre-
sponding thermodynamic potential?
4. Derive the form of α as a function of the pressure P.
5. Derive the mean energy E:
∂ ln Z
E = −P V − .
∂β
Solution
1.1 Probability P
We maximize the entropy S = − r Pr ln(Pr ) accounting for the constraints:
X
Pr =1 (3.51)
r
X
Pr Er = E
r
X
Pr V r = V
r
using the Lagrangian multiplier technique. We then search the zero of the function:
38CHAPTER 3. SYSTEM IN CONTACT WITH A RESERVOIR. GRAND CANONICAL ENSEMBLE
" #
∂ X X X X
Pr ln Pr + α1 ( Pr Er − E) + α2 ( Pr Vr − V ) + α3 ( Pr − 1)] .
∂Pr0 r r r r
(3.52)
We obtain:
− ln Pr0 = 1 + α3 + α1 Er0 + α2 Vr0 , (3.53)
so that the probability can be cast in the form:
exp[−βEr0 + αVr0 ]
Pr0 = . (3.54)
ZGCT P
Rms fluctuations :
X 1 ∂ 2 ZGCT P
Vr2 = Pr Vr2 = , (3.56)
r
ZGCT P ∂α2
∂ZGCT P 2
2 1
Vr = 2 ,
ZGCT P ∂α
∂ 2 ZGCT P ∂ZGCT P 2
2 1 1
Vr2 − Vr = − ,
ZGCT P ∂α ZGCT P ∂α
∂ 2 ln ZGCT P
∂ 1 ∂ZGCT P
= = .
∂α ZGCT P ∂α ∂α2
1.3 Entropy
X X
S = −k Pr ln Pr = −k Pr [− ln Z − βEr + αVr ] = k ln Z + kβE − kαV . (3.57)
r r
so that
∂ ln ZGCT P
−E − P V = . (3.60)
∂β
Chapter 4
4.1 Introduction
The concept of linear response applies to a large number of phenomena where it is possi-
ble to identify a cause (or excitation) and a consequence (or response) related by a linear
relation. In these lectures, we are primarily interested in coefficients introduced for de-
scribing transport phenomena. Let us mention for instance Fourier’s law φcd = −k∇T for
heat transport, Ohm’s law for charge transport j = σE and Fick’s law jN = −D∇N for
particles transport. Here, the fluxes are the consequences and the gradients are the causes.
Yet, the scope of linear response theory is not restricted to transport phenomena. Linear
responses can also be found in different contexts: the polarisation induced by an electric
field P = 0 χE, the magnetization induced by a magnetic field M = µ0 (µr − 1)H, the
length variation of a spring when a force is applied x = −F/k, the charge in a capacitor
when applying a voltage Q = CU , etc
It is worth pointing out that the linear coefficients can be introduced phenomenolog-
ically from an experimental approach. There are tables where the electrical or thermal
conductivities can be found for different materials. From a theoretical point of view, these
coefficients can be derived by assuming that the systems are in a situation close to equi-
librium and considering the presence of an excitation (the cause) as a perturbation. The
linearity of the response is a direct consequence of the fact that it is possible to seek a so-
lution using a first order perturbation approximation. This is the route that we will follow
in the next chapter devoted to linear response theory and fluctuation-dissipation theorem.
Here, we assume that the linear relation is an experimental fact and we do not attempt to
derive it from first principles. However, this linear response must be consistent with general
principles of physics. For instance, a linear response relation must:
Accounting for all these conditions entails some specific properties of the linear re-
sponse coefficients. It is the purpose of this chapter to establish these properties. Let us
note that we will only make the assumption of a linear response. Hence, the properties that
39
40 CHAPTER 4. LINEAR RESPONSE COEFFICIENTS. GENERAL PROPERTIES
where we have introduced a linear response function χ(t, t0 ). Let us now consider a station-
ary system. Let us consider the example of a spring whose force constant does not change
in time. It follows that the linear response function is only a function of t − t0 so that if
the same experiment is repeated at two different times, it will produce the same result be-
cause the system does not depend on time t. Only the interval t − t0 between excitation
and response matters. Hence, for a stationary system, the linear response can be cast in the
form:
Z ∞
X(t) = χ(t − t0 )F (t0 )dt0 . (4.2)
−∞
This form suggests that the response at time t may depend on forces at times t0 > t.
This would violate the causality principle. Indeed, the cause must precede the consequence.
Hence, the linear response function needs to satisfy the condition:
Finally, we obtain the following general form for a linear response satisfying stationarity
and causality:
Z t
X(t) = χ(t − t0 )F (t0 )dt0 . (4.4)
−∞
We can now discuss some properties of the linear response function. First of all, let us
assume that we excite the system with a pulse F0 δ(t − t0 ). We obtain a response X(t) =
χ(t − t0 )F0 . It is seen that χ is the response of the system to a pulse. Hence, this response
function is also called impulse response function.
It is important to note that the response function introduces a time scale. Indeed, the
response function decays with a typical time denoted τ . This time scale is typically the time
needed by the system to return to equilibrium. Let us consider for example a circuit with a
capacitor and a resistance. The typical decay time is given by τ = RC. In other words, this
typical time scale characterizes the memory of the system. Note that it is a general property
that the response function tends to zero for long times. This follows from the existence of
dissipation processes.
4.3. SUSCEPTIBILITY 41
Finally, we consider the particular case of a force varying very slowly as compared to
the typical system time scale τ . It is then possible to approximate the linear relation:
Z t Z t
0 0 0
X(t) = χ(t − t )F (t )dt ≈ F (t) χ(t − t0 )dt0 = χ0 F (t), (4.5)
−∞ −∞
Rt R∞
where we have introduced χ0 = −∞ χ(t − t0 )dt0 = 0 χ(u)du. We have obtained a
linear relation connecting the force and the response at the same time. This approximation
amounts to consider that the response can be simplified using χ0 δ(t−t0 ). This clearly means
that the response is considered to be instantaneous. It is very important to realize that this
is strictly equivalent to assume that dispersion can be neglected. This will be discussed in
the following section.
4.3 Susceptibility
The goal of this section is to introduce a spectral analysis of the linear response. We will
establish that stationarity, causality and the second principle impose severe constraints on
the Fourier transform of the linear response function.
4.3.1 Definition
We first note that for a stationary system, the linear response relation is a convolution prod-
uct: Z ∞
X(t) = χ(t − t0 )F (t0 )dt0 , where χ(t) = 0 if t < 0. (4.6)
−∞
The Fourier transform of the linear response function χ(ω) is called susceptibility or
complex admittance. Given that the linear response function is a real function, we have the
property:
χ(−ω) = χ∗ (ω). (4.9)
Let us emphasize that it is equivalent to know the linear response or the susceptibility
for all the spectrum. Knowing the frequency dependence of the susceptibility amounts to
know the dispersive behaviour of the system. For the particular case of an instantaneous
response χ(t) = χ0 δ(t), we find that χ(ω) = χ0 does not depend on frequency: this is a
non-dispersive system.
To summarize, when a system is excited by by a generalized force with slow time vari-
ations as compared to the time scale of the system, the system ”follows” the excitation
instantaneously. In frequency domain, this corresponds to a non-dispersive behaviour.
42 CHAPTER 4. LINEAR RESPONSE COEFFICIENTS. GENERAL PROPERTIES
4.3.2 Dissipation
The goal of this section is to establish that the dissipation rate of the system energy is
proportional to the imaginary part of the susceptibility χ00 = =(χ). We will also prove
that at equilibrium, the imaginary part of the susceptibility χ00 = =(χ) is positive so that
”negative” losses (i.e. amplification) are impossible. This property follows from the first
and second principle of thermodynamics. To proceed, let us first consider a system that
receives work done by the generalized force and exchanges heat with its environment which
is considered to be a thermostat at temperature T0 . Our starting point is energy conservation
expressed by the first principle in the form:
dU = δW + δQ, (4.10)
where δW stands for the work received by the system and δQ is the heat received by
the system. The work done by F on the system is given by δW = F dX = F vdt
where we have introduced the notation v = Ẋ. Let us now consider a harmonic force
F0 cos(2πt/T ) = F0 cos(ωt) = Re[F0 exp(−iωt)]. The response is also monochromatic
and given by X(t) = Re[X0 exp(−iωt] = Re[χ(ω)F0 exp(−iωt]. During a period, the
internal energy changes and comes back to its initial value so that its variation is null. It
follows that: Z Z t+T Z t+T t+T
dU = 0 = δW + δQ. (4.11)
t t t
Heat and work exchanged during a cycle are therefore related by:
Z t+T Z t+T Z t+T
δW = F (t0 ) v(t0 ) dt0 = − δQ. (4.12)
t t t
This energy conservation equation shows that the energy received by the system is released
as heat by the system to the environment. The work done during a cycle can be computed
directly:
t+T
1 t+T F0 e−iωt + F0∗ eiωt −χ(ω)F0 e−iωt + χ∗ (ω)F0∗ eiωt
Z Z
1 0
F v dt = iω
T t T t 2 2
ω
= =[χ(ω)]|F (ω)|2 . (4.13)
2
This formula shows that the average power transferred to the system is given by ω2 |F (ω)|2 =(χ).
Finally, the second principle imposes:
δQ
dS ≥ , (4.14)
T0
where T0 is the temperature of the thermostat in contact with the system. We know that
entropy is a state function so that its variation during a cycle is null. It follows that
Z t+T Z t+T
ωT
T0 dS = 0 ≥ δQ = − |F (ω)|2 Im(χ). (4.15)
t t 2
Note that the final result depends Ron the conventional choice of the sign in the Fourier
transform. If we had used χ(t) = χ(ω) exp(iωt)dω/2π, we would have obtained a
velocity v = iωX instead of −iωX so that we would have obtained χ00 (ω) ≤ 0.
Let us finally note that equation (4.15) shows that the work done by the generalized force
on the system is positive whereas the heat received by the system is negative. This work
corresponds to the energy transferred to the thermostat as heat. In other words, the system
transforms work in heat: this is the absorption process. This transformation of work into
heat coincides with a production of entropy transferred to the thermostat. The imaginary
part of the susceptibility is therefore the quantity that characterizes the absorption by the
system.
Let us finally emphasize that we have assumed that the system is at equilibrium at all
times because we have used the first and second principle. A system out of equilibrium can
have a susceptibility with the opposite sign. This is for instance the case of the amplify-
ing medium in a laser where absorption is replaced by amplification. Needless to say, the
amplifying medium is not in equilibrium.
It follows that the inversion formula of Fourier transforms Eq.(4.8) can be cast in the
form: Z ∞
χ(ω) = χ(t) exp(iωt)dt. (4.18)
0
Here, causality appears because we integrate only over positive values of time t. In what
follows, we need to define χ(ω) for complex values of ω (ω = ω 0 + iω 00 ). Equation (4.18)
allows computing χ(ω) for any complex value of ω. This is called an analytic continuation.
We now examine the consequences of causality on the properties of χ(ω). Causality implies
the following identity:
Since the Fourier transform of a product is the convolution product of the Fourier trans-
forms, we find, using the Fourier transform of the Heaviside function, the identity:
Z ∞
dω 0
0 0 i
χ(ω) = χ(ω ) πδ(ω − ω ) + pv , (4.21)
−∞ 2π ω − ω0
44 CHAPTER 4. LINEAR RESPONSE COEFFICIENTS. GENERAL PROPERTIES
∞ ∞
χ00 (ω 0 ) 0 −1 χ0 (ω 0 ) 0
Z Z
0 1 00
χ (ω) = pv dω χ (ω) = pv dω (4.23)
π −∞ ω0 − ω π −∞ ω0 − ω
a time constant τ = RC. This is an example of relaxation, i.e. return to equilibrium after
suppressing a constraint.
We now introduce a formal definition of the relaxation function. Let us consider a con-
stant force F0 applied to the system. The response of the system is a displacement X0 . At
time t = 0, the force is suppressed. The system response is then given by F0 Ψ(t) where we
have introduced the relaxation function Ψ(t).
- Since the system is linear, the relaxation function must be related to the response function.
- Relaxation is fundamentally driven by losses mechanisms. In the first example, the energy
given to the system by polarizing it is dissipated in water through collisions. In the second
example, the energy given by the voltage supplier to the capacitor is dissipated by Joule
effect in the resistor.
- Relaxation is a measurement of the memory of the system. The decay of the relaxation
function tells how long it takes until the system forgets its initial non-equilibrium state.
These remarks give a hint on the fundamental properties underlying the linear response
theory based on fluctuation-dissipation theorem. We can guess that there must be a link
between losses, memory time and linear response.
where we have introduced the new variable u = t − t0 . It is seen that the response of the
system can be cast in the form
X(t) = F0 Ψ(t)
where Z ∞
Ψ(t) = χ(u) du. (4.29)
t
Hence, we can derive the response function from the relaxation function. For t < 0, causal-
ity imposes that the response function is null. For t > 0, it can be derived from the relaxation
function so that we finally obtain:
dΨ(t)
χ(t) = −H(t) , (4.30)
dt
4.5 Example
Here, we apply the formalism introduced in this chapter to a well-known example. In order
to illustrate the last remark, we will derive the complex impedance Z(ω) of an RC circuit
starting from the knowledge of its relaxation function. This is certainly not the easiest way
to find the complex impedance of a RC circuit. Yet, it nicely illustrates how the spectral
information can be recovered from the relaxation function. The relaxation function Ψ(t)
is identified from the time decay of the capacitor charge Q(t) = Q0 exp(−t/RC). The
generalized force is the voltage V such that Q = CV where C is the capacitance. We have
Q0 = CV0 so that V0 is the generalized force. Hence, Q(t) = V0 Ψ(t) so that
Ψ(t) = C exp(−t/RC).
dΨ(t) H(t)
χ(t) = −H(t) = exp(−t/RC). (4.31)
dt R
The susceptibility is the Fourier transform of the response function:
Z ∞
1 C
χ(ω) = exp(−t/RC) exp(iωt)dt = (4.32)
0 R 1 − iωτ
where we have used the notation τ = RC. In order to find the impedance, we need to
identify the coefficient relating the intensity to the voltage:
−iωC
I(ω) = −iωQ(ω) = −iωχ(ω)U (ω) = U (ω). (4.33)
1 − iωτ
It follows that the impedance is given by:
U (ω) 1
Z(ω) = =R− (4.34)
I(ω) iωC
Chapter 5
Linear response.
Fluctuation-Dissipation theorem
5.1 Introduction
The goal of this chapter is to derive a general form of the coefficients introduced in linear
response theory from a statistical approach. There are two basic ideas that will be developed
here. First, we study the relaxation of the system. As discussed in the preceding chapter,
relaxation contains all the information on the linear response of the system. The second idea
is that the generalized force applied to the system can be considered to be a perturbation.
Hence, a first order perturbative solution can be found leading to a linear response. The
main result that will be established here is a simple relation between the relaxation function
and the time correlation function of the fluctuations at equilibrium.
To be more explicit, let us consider the two examples used in the introduction of the
concept of relaxation in the previous chapter. We had considered water molecules oriented
along an electrostatic field thereby producing a macroscopic polarization. When turning off
the electric field, the polarization returns to zero, its equilibrium value, with a dynamic be-
haviour given by the relaxation function Ψ(t). Similarly, the charge of a capacitor returns to
zero exponentially when the applied voltage is replaced by a short circuit. At equilibrium,
in the absence of external electric field, the mean polarization is null but its instantaneous
value is non zero. At equilibrium, without applied voltage, the charge of the capacitor
also fluctuates in time. In both cases, it is possible to compute the correlation function
< P (t)P (t + τ ) > or < Q(t)Q(t + τ ) >. These equilibrium fluctuations will be shown
to be related to the relaxation function. In other words, the time dependence of relaxation
and fluctuation is the same. This may be surprising at first glance as the relaxation function
deals with the mean value whereas the correlation function deals with fluctuations. This can
be understood qualitatively by realizing that both the relaxation function and the correlation
function are a measurement of the memory of the system. The relaxation function mea-
sures the memory of an ordered system while the fluctuation measures the memory of one
particular random state. In both cases, this memory is governed by the same microscopic
mechanisms.
The chapter is organized as follows. We first derive the relaxation function. The main
result is to show that it is proportional to the correlation function. The second section deals
with the susceptibility χ. We will then be able to establish a relation between the imaginary
part of the susceptibility which is related to dissipation and the fluctuations. In the last
47
48 CHAPTER 5. LINEAR RESPONSE. FLUCTUATION-DISSIPATION THEOREM
section, we show some applications of this result to the study of fluctuations of systems at
equilibrium.
We now assume that βx0 Xr0 << 1. This assumption is valid if the interaction energy
x0 Xr0 is much smaller than the typical thermal energy kB T . A first order expansion yields:
X exp(−βEr )[1 + β x0 Xr (0)]
0 0
< X > (t) = Xr0 (t). (5.4)
r
Z GCX
0
and ZGC0 : X
ZGC0 = exp(−βEr0 ). (5.7)
r0
X exp(−βEr0 )
< X > (t) = [1 + β x0 Xr0 (0)] Xr0 (t)
r0
ZGC0 [1 + β x0 X0 ]
X0 (t) + β x0 < X(0)X(t) >
= . (5.8)
1 + β x0 X0
We observe that X0 (t) is the mean value of X without applied external force x0 . Hence,
this is the mean value at equilibrium X0 without external force and therefore it is time
independent so that X0 = X0 (t) = X0 (0).
Note that the mean value < X(0)X(t) > is taken for a zero value of the generalized
force x0 . This is the fluctuation at equilibrium. Finally, the mean value at time t can be cast
in the form:
< X > (t) − X0 = β x0 [< X(0)X(t) > − < X > (t)X0 ] (5.9)
The structure that we obtain explicitly shows a linear link between the response < X >
(t) − X0 and the cause x0 . Our derivation allows identifying the origin of this linearity: the
first order expansion of the probability. Hence, one may expect a linear response if x0 X <<
kB T . This inequality gives a precise meaning to the condition ”close to equilibrium” often
invoked as a validity condition for linear response. We now consider the particular case
where X is null without external applied force (i.e. X0 = 0 if x0 = 0):
where CXX (t) =< X(0)X(t) > is the correlation function at equiilibrium. It follows that
the relaxation function is given by:
dΨ dCXX (t)
χ(t) = −H(t) = −βH(t) . (5.12)
dt dt
Let us emphasize a remarkable property of this equality. The term on the left-hand side
describes a non-equilibrium property: linear response to an external force. The term on the
right hand side describes time fluctuations at equilibrium with no external force applied. In
other words, the study of the fluctuations at equilibrium contains the information on the lin-
ear response to an external force out of equilibrium. We have already indicated the physical
origin of this property. Both phenomena are driven by the microscopic mechanisms at work
when the system loses the memory of its initial state, be it an ordered state (relaxation) or
a particular microscopic state (time fluctuation). In both cases, the relevant time scale is
given by the collision time.
50 CHAPTER 5. LINEAR RESPONSE. FLUCTUATION-DISSIPATION THEOREM
Let us also stress a second remarkable property. As discussed previously, the relaxation
function contains all the information on the susceptibility and therefore on the absorption
spectrum. In other words, the left-hand side of the equality deals with dissipation while
the right hand-side deals with fluctuations. This is the origin of the name ”Fluctuation-
Dissipation” theorem. We will analyse in more detail this aspect of the fundamental result
obtained in the subsequent sections.
5.3 Susceptibility
We have introduced the susceptibility as the Fourier transform of the linear response func-
tion. Since we have established that the linear response is given by the correlation function
of a stationary random process, we need to deal with the spectral analysis of a stationary
processes. This is the purpose of the Wiener-Kinchine theorem.
There is no limit for fT as T tends to infinity. However, it is possible to define the limit:
1 ˜
lim |fT (ω)|2 , (5.15)
T →∞ T
for a class of random processes. This is typically the case of white noise for instance.
This quantity is called power spectral density and denoted If (ω). Here, the word power
refers to the fact that the square of the function is associated with energy for electric quan-
tities. Dividing energy by time yields power. In other words, as T goes to infinity, the
energy diverges but the power remains constant. A major result of spectral analysis of sta-
tionary random processes is the Wiener-Khinchine theorem which states that the correlation
5.3. SUSCEPTIBILITY 51
Coming back to light, it is seen that the spectrum can be derived from the knowledge
of the time correlation function of the field. This is the basis of the Fourier transform
spectroscopy technique which is based on measuring the time correlation function using a
Michelson interferometer.
CXX (t) =< X(t)X(0) >=< X(0)X(t) >=< X(−t)X(0) >= CXX (−t). (5.20)
It is important to emphasize that the commutation is not always valid for operators in quan-
tum mechanics. Hence, a quantum derivation of the fluctuation dissipation theorem yields
a different result. This is left as a problem. The final classical result is the second form of
the fluctuation-dissipation theorem:
2kB T
IXX (ω) = χ”(ω) (5.21)
ω
The term on the left-hand side corresponds to fluctuations while the term on the right-hand
side is proportional to the losses spectrum. A quantum derivation produces a different result:
52 CHAPTER 5. LINEAR RESPONSE. FLUCTUATION-DISSIPATION THEOREM
2χ”(ω)
IXX (ω) = Θ(ω, T ) (5.22)
ω
where Θ(ω, T ) is the mean energy of a quantum harmonic oscillator given by:
1 1
Θ(ω, T ) = ~ω + . (5.23)
2 exp(~ω/kB T ) − 1
We have derived the fluctuations of the response X. It is also possible to derive the
fluctuations of the generalized force x(t). Indeed, we have a deterministic relation between
both quantities given by X(ω) = χ(ω)x(ω). It follows that:
Finally, the thermal fluctuations of the generalized force are given by:
Z ∞
2 2kB T χ”(ω) dω
< x (0) >= Cxx (0) = . (5.27)
π 0 |χ(ω)|2 ω
Z ∞
2 χ”(ω) dω
< x2 (0) >= Cxx (0) = Θ(ω, T ) . (5.28)
π 0 |χ(ω)|2 ω
5.4 Applications
In this section, we will illustrate how the fluctuation-dissipation theorem can be used to de-
rive the thermal fluctuations. We first consider the voltage fluctuations across an impedance
Z.
5.4. APPLICATIONS 53
i
χ(ω) =
Z(ω)ω
R(ω)
χ”(ω) = . (5.30)
ω 2 |Z|2
We have introduced the usual notation R for the real part of the impedance Z. Equation
(5.27) yields the fluctuation of voltage across the impedance:
2kB T ∞
Z
2
< V >= R(ω) dω. (5.31)
π 0
χ” ω3
= (5.35)
|χ|2 6π0 c3
It is now a simple matter to derive the energy density of the blackbody radiation. First, we
derive the energy associated with the n-component of the electric field by multiplying by
0 /2. We need to multiply by a factor 3 to account for the three components of the electric
field. We also multiply by 2 to account for the magnetic energy 1 . Hence, we obtain:
∞
ω2
Z
~ω
U= dω. (5.37)
0 exp(~ω/kB T ) − 1 π 2 c3
Here, we have not included the additional term ~ω/2 per mode due to the vacuum fluctua-
tion. It is responsible for the Casimir force between two parallel mirrors.
Solution
1. Linear response
We introduce the linear relation between E0 and the dipole moment
The linear response function can be computed using χ(t) = −Θ(t) dΨ(t) χ0
dt = Θ(t) τ exp(−t/τ ).
2. Computing the susceptibility amounts to take the Fourier transform:
Z ∞
χ0 χ0
χ(ω) = exp(−t/τ + iωt)dt = . (5.40)
0 τ 1 − iωτ
3. The relation between relaxation function and correlation function yields < pi (t)pi (0) >=
kB T Ψ(t) = kB T χ0 exp(−t/τ ).
4. A direct application of the FD theorem yields :
2χ”(ω) 2χ0 τ kB T
Ipp (ω) = kB T = (5.41)
ω 1 + ω2τ 2
56 CHAPTER 5. LINEAR RESPONSE. FLUCTUATION-DISSIPATION THEOREM
Chapter 6
6.1 Introduction
This chapter is an introduction to a technique developped by Langevin to model fluctua-
tions of linear systems. This technique allows to go beyond statistical equilibrium because
the time dynamics of fluctuations can be modelled. It has been introduced historically in
order to explain Brownian motion. Hereafter, we summarize the key features of Brownian
motion. We then introduce Langevin model in the context of 1D Brownian motion. In the
following section, we use the Langevin model to recover Fluctuation-Dissipation theorem
in the context of 1D Brownian motion. Finally, we derive the Einstein relation existing
between diffusion coefficient and mobility.
57
58CHAPTER 6. LANGEVIN MODEL OF BROWNIAN MOTION. APPLICATION: FLUCTUATIONS OF LIN
dv
m = F(t) =< F(t) > +R(t), (6.1)
dt
where < F(t) > stands for the mean value of the force F(t) produced by the molecules
on the particle. R(t) is the random component of that force. The mean value is the drag
force. It can be cast in the phenomenological form −γmv(t). This form is only valid at
low velocities. 1
Of course, when using this differential equation, one needs to specify explicitly what
is this random force. The first idea is to try to derive a microscopic model of that random
force by analysing the collisions with molecules. However, we will show that it is possible
to derive the required information without need to perform such a microscopic analysis.
Indeed, in the framework of Langevin model, the random force is the source of fluctuations
at equibrium. The later are known. Hence, extracting the required information on the
random force from the mere knowledge of the fluctuations seems possible.
The power spectral density of the velocity fluctuations depends on two factors: i) the
exciting force spectrum and ii) the dynamics of the linear system. The later is given by the
dynamical equations describing the system. In the case of Brownian motion, it is possible to
simplify the problem. We assume that the random force spectrum is a white noise spectrum.
In other words, the time correlation function of the random force is given by:
1
More precisely, for a spherical particle, the Reynolds number has to be lower than 0.5. This is a condition
more stringent than having a laminar flow.
6.3. FLUCTUATION-DISSIPATION THEOREM. SECOND FORM. 59
0 b
τcoll = . (6.3)
v
This time can be compared with the typical time between two collisions τcoll . It can
be estimated using a hard sphere model. Let us assume that a molecule with a velocity v
interacts with molecules located in a cylinder with radius 2b et axis parallel to the velocity.
If the cylinder length is vτ , it contains one molecule on average. By denoting n the number
of molecules per unit volume, we get
It is convenient to introduce the distance d such that nd3 = 1. This length is the side of
a cube that contains one molecule. We can cast the collision time as:
3
0 b
τcoll ≈ τcoll 4π . (6.5)
d
In normal conditions, d takes the value 3.3 nm. Hence, b/d 1 so that the collision
duration is negligible compared to all other time scales. Hence, we can consider that the
collision is instantaneous so that the correlation function is proportional to δ(t − t0 ) and the
spectrum is a white noise spectrum.
where fT (ω) is the Fourier transform of the restriction of the function f (t) on the interval
[0, T ]. We get:
RT (ω) 1
vT (ω) = , (6.7)
m γ − iω
60CHAPTER 6. LANGEVIN MODEL OF BROWNIAN MOTION. APPLICATION: FLUCTUATIONS OF LIN
so that
|RT |2 1
|vT |2 = . (6.8)
m γ + ω2
2 2
IR = 2γmkB T . (6.13)
This equation allows identifying the amplitude of the force fluctuations. It establishes
a link between the drag force coefficient γ and the power spectral density of the velocity
fluctuations. Again, we see that there is a fundamental link between dissipation and fluctu-
ation. This relation is often called the second form of the fluctuation-dissipation theorem.
However, in the previous chapter, we have seen that the fluctation-dissipation theorem not
only provides a link between fluctuation and dissipation but also yields the full susceptibil-
ity (not only its imaginary part) as first shown by Kubo. This more complete form of the
theorem is called first form of the Fluctuation-Dissipation theorem. We now show that it
can be derived in the framework of Langevin model.
6.4.1 Susceptibility
Here, we repeat the study of the motion in harmonic regime. Yet, we now assume that there
is an external deterministic force denoted Fext . Newton’s equation applied to the functions
restricted to a finite time interval T are:
It follows that the velocity and the external force are linearly related:
1
< vT >= Fext,T . (6.16)
m(γ − iω)
Thus, we can define a susceptibility:
1
χv = (6.17)
m(γ − iω)
Note that this linear response has been defined as a coefficient between velocity and
force. This is what is commonly used in the framework of Brownian motion but this is
not the standard approach we used when introducing the Fluctuation-Dissipation theorem.
Indeed, the product of the two quantities v and F does not have the dimension of an energy.
which is the form of the first Fluctuation-dissipation theorem, namely the link between the
susceptibility and correlation function.
domain. The result for the velocity correlation function will be the same. Only the technical
approach changes. In the next section, we will use this time dependent approach to derive
the rms mean position of a particle. This is an introduction to the diffusion processes in the
framework of Langevin model. The dynamic equations governing the particle velocity in
time domain can be written as:
dv R(t)
= −γv + . (6.21)
dt m
The solution of the homogeneous equation is v(t) = v(0) exp(−γt). A standard technique
to find a solution of a first order differential equation is the so-called constant variation
method. We seek a solution assuming the form v(t) = f (t) exp(−γt). By inserting this
form in the differential equation, we get:
df R(t)
= exp(γt). (6.22)
dt m
so that:
t
R(t0 )
Z
f (t) = f (0) + exp(γt0 )dt0 . (6.23)
0 m
It follows that:
t
R(t0 )
Z
v(t) = v(0) exp(−γt) + exp[−γ(t − t0 )]dt0 . (6.24)
0 m
Finally, upon multiplying the equation (6.24) by v(0) and averaging, we obtain:
v(0)
< x(t) >= x(0) + [1 − exp(−γt)], (6.28)
γ
so that: Z t Z t1
R(t2 )
x(t)− < x(t) >= dt1 dt2 exp[γ(t2 − t1 )]. (6.29)
0 0 m
6.5. EINSTEIN RELATION. BROWNIAN MOTION 63
Instead of deriving the average of the square of this quantity it turns out to be simpler to
study (x− < x >)(v− < v >) = 1/2(d(x− < x >)2 /dt) and then integrate this quantity:
Z t Z t1 Z t
R(t2 ) R(t3 )
(x− < x >)(v− < v >) = dt1 dt2 exp[γ(t2 −t1 )] dt3 exp[γ(t3 −t)].
0 0 m 0 m
(6.30)
Using the equation < R(t2 )R(t3 ) >= IR δ(t2 − t3 ), we find:
IR t
Z Z t1
(x− < x >)(v− < v >) = dt1 dt2 exp[γ(t2 − t1 )] exp[γ(t2 − t)]
m2 0 0
Z t
IR [exp(2γt1 ) − 1]
= 2
exp(−γt) dt1 exp(−γt1 )
m 0 2γ
IR 1 − exp(−γt) exp(−2γt) − exp(−γt)
= +
m2 2γ 2 2γ 2
IR
= [1 − exp(−γt)]2 (6.31)
2m2 γ 2
It follows that v = qE/γm. By definition, the mobility µ is given v = µE. Upon identifi-
cation, we find:
µ = q/γm. (6.36)
The ratio µ/d can now be derived. This is Einstein relation:
µ q
= . (6.37)
D kB T
64CHAPTER 6. LANGEVIN MODEL OF BROWNIAN MOTION. APPLICATION: FLUCTUATIONS OF LIN
IV = 2RkB T. (6.43)
Finally, we get:
< q(t)q(0) >= CkB T exp(−t/RC). (6.44)
Solution
1. Dynamical equation
density is thus:
IR
Ix = 2 (6.49)
m (ω − ω1 )(ω − ω1∗ )(ω − ω2 )(ω − ω2∗ )
For positive times t, we evaluate the integral using a contour closed by a semicircle in
the half space =(ω) < 0 so that the exponential decay of exp(−iωt) guarantees that the
contribution of the semicircle vanishes. We get:
IR cos(ω00 τ + φ) exp(−γτ /2)
< x(t)x(t + τ ) >= (6.50)
2m2 γω0 ω00
where tan(φ) = γ/2ω00 . Note that cos(φ) = 1 − γ 2 /4ω02 . At t = 0, we have < x2 >=
p
1. Write the differential equation that describes the exponential decay of the dipole mo-
ment (external field switched off).
66CHAPTER 6. LANGEVIN MODEL OF BROWNIAN MOTION. APPLICATION: FLUCTUATIONS OF LIN
2. Write the same equation including a noise term f that causes fluctuations. We use the
assumption that < fi (t)fj (t + τ ) >= If δij δ(τ ). Explain this approximation.
3. Write the expression of the power spectral density for the fluctuations of dipole
moment. Use this expression to derive the correlation function < pi (t)pi (0) > as a function
of If . At equilibrium, the correlation is < pi (0)pi (0) >= kB T χ0 . What is the value of If
? Derive the expression of the power spectral density of the dipole moment fluctuations.
Solution
1. The time dependent dipole moment is ruled by the differential equation
dp
+ γp = 0.
dt
dp
+ γp = f (t).
dt
The time correlation function of the random force is modelled by δ(t), owing to the short
time correlation function as compared to the decay time τ . The term δij accounts for the
isotropy of the system resulting in the absence of correlation between different components.
3. In frequency domain, we get
p(ω)[−iω + γ] = f (ω).
If (ω)
Ipp (ω) = . (6.53)
ω2 + γ 2
We now use Wiener-Kinchine theorem to get the time correlation function. Using the form
of the power spectral density of the noise and integrating over frequencies in the complex
plane we get :
Z ∞
If e−iωt dω e−γt
< pi (t)pi (0) >= = If . (6.54)
−∞ (ω + iγ)(ω − iγ) 2π 2γ
2kB T χ0 τ
Ipp (ω) = . (6.55)
1 + ω2τ 2
Chapter 7
7.1 Introduction
When an intensive parameter such as temperature, chemical potential, voltage, etc. is in-
homogeneous, fluxes of heat, particles, charges, are generated in order to restore the ho-
mogeneity. We will study the fluxes that appear in the systems for situations ”close” to
equilibrium. The concept of distance to equilibrium will be given a precise meaning later
when introducing the dimensionless Knudsen number.
Let us give a few examples of the type of systems that will be considered. Phenomeno-
logical laws have been introduced to account for these transport effects:
Kinetic theory.
The basic idea is to consider that the transport of some quantity X is mediated by the trans-
port of particles. The physical quantity X could be the charge, the mass, the energy, the spin
or the momentum of the particle for instance. Obviously, as a particle moves, it transports
all these quantities. A kinetic approach is based on computing particles fluxes and deriving
from them the fluxes of X. In order to achieve this program, it is necessary to know the
velocity distribution of particles.
67
68CHAPTER 7. INTRODUCTION TO THE KINETIC THEORY OF TRANSPORT PHENOMENA.
As opposed to kinetic theory, this approach is not based on a microscopic point of view.
It is an extension of the macroscopic thermodynamics to non-equilibrium phenomena. The
approach is based on the concept of local flux density. It allows computing the local entropy
generation rate. It also allows deriving very general reciprocity laws connecting coupled
transport phenomena such as a charge current due a temperature gradient.
In this chapter, we will give a sketch of the kinetic theory using elementary fluxes es-
timations. In the next chapter, we will introduce a more systematic technique to evaluate
the velocity distribution out of equilibrium based on the so-called Boltzmann equation. We
will show how this equation can be solved using a first order perturbation solution in the
framework of the relaxation time approximation. Finally, we will apply this approach to
both gases and solids.
Particles crossing the plane through an area Σ during dt in the positive direction (pos-
itive velocity) are in a cylinder with volume Σv dt. To count how many particles cross,
we simply multiply this volume by the number of particles per unit volume. Among all
particles in the volume, only half of them have a positive velocity so that we divide by a
factor of 2. The number of particles per unit volume is not homogeneous. We note that we
consider particles that will not be scattered before reaching the plane. Hence, they must be
close enough to the plane x0 . On average, particles are coming from a distance given by l
so that the number of molecules that we use is given by n(x0 − l)/2. Finally, the flux per
unit area is n(x0 − l)v/2. 1
1
The choice of l may seem arbitrary. Hereafter, we repeat the derivation with a more acurate justification
of the choice of the mean free path as a typical length. We first revisit the concept of mean free path and try to
offer a more general view. Let us introduce the probability density that a particle has a collision between x and
x+dx. This probability can be written as P (x)dx where P (x) is a probability density. We now assume that the
probability is homogeneous in x so that P (x) is a constant function. As P (x)dx is a probability, P (x) has the
dimension of the inverse of a length. Let us denote l this length so that we have P (x)dx = dx/l. We can now
derive the probability Π(x) that a particle has no collisions between 0 and x. In order to derive this probability,
we need to make a further assumption. We assume that the collisions are a markovian random process. In other
words, we assume that collisions are uncorrelated. Hence the probabilities of having a collision between 0 and
7.2. ELEMENTARY INTRODUCTION TO KINETIC THEORY OF TRANSPORT PHENOMENA69
The total flux through the interface results from a balance between particles moving in
both directions. We obtain:
n(x0 − l) n(x0 + l) dn
jN = v− v ≈ −lv (x0 ). (7.4)
2 2 dx
The result depends on the gradient of the number of particles per unit volume in agreement
with Fick’s law. The diffusion coefficient D is the product of the mean free path l and the
velocity v.
This simple model can be generalized to the three-dimensional case by dividing the
number of particles by a factor of 6 instead of 2. This rough estimate is based on the fact
that the particles can be distributed along six directions, 1/6 along positive axis Ox, 1/6
along the negative Ox axis, 1/6 along the positive Oy axis, etc. This leads to the relation
D = lv/3. More elaborate theories show that this relation is exact.
n n n dv
j = −e[ v(x0 − l) − v(x0 + l)] ≈ e l. (7.8)
6 6 3 dx
Noting l = vτ where τ is the mean time between two collisions and using mv 2 /2 =
3kB T /2, we get:
eτ dmv 2 eτ nkB dT
j≈ n ≈ . (7.9)
6m dx 2m dx
The current density is zero:
eτ nkB dT
j = 0 = σE + , (7.10)
2m dx
where σ = ne2 τ /m so that the field induced by the temperature gradient can be cast in the
form:
kB dT
E=− (7.11)
2e dx
For the sake of illustration,we consider the case of a gas of particles with mass m. At
equilibrium, the velocity distribution function f (0) (r, v) is given by a Maxwell-Boltzmann
distribution: 3/2
mv2
(0) m
f (r, v) = n0 exp − . (8.2)
2πkB T 2kB T
It is useful to note that the Maxwell-Boltzmann distribution function can be factorized
as the product of the number of particles per unit volume n0 and a velocity probability
density function P (v), where:
3/2
mv2
m
P (v) = exp − . (8.3)
2πkB T 2kB T
Flux density
In order to derive the flux density, we first consider the flux of the quantity X through a
plane normal to the unit vector n due to particles with a velocity v. The particles crossing
the plane through an area Σ during dt are contained in a volume v · ndtΣ. In this volume,
71
72CHAPTER 8. INTRODUCTION TO KINETIC THEORY OF TRANSPORT PHENOMENA. BOLTZMANN
there are f (r, v, t)d3 v(v · n)dtΣ particules so that there are f (r, v, t)d3 v(v · n)dt particles
per unit area. Each particle transports the quantity X so that the contribution to the flux
density is given by :
djX = [X]f (r, v, t) d3 v (v · n). (8.4)
To obtain the total flux, we need to add the contributions of all particles with all possible
velocities. Thus, we get:
Z
djX = [X]f (r, v, t) (v · n) d3 v. (8.5)
We note that in this equation, the scalar product v · n changes sign with the velocity sign
along the normal n so that particles with a positive velocity give a positive contribution
to the flux whereas molecules with negative velocity give a negative contribution. We can
define a flux density j in such a way that the flux through an elementary surface dΣn is
ΦX = jX · (dΣn):
Z
jX = [X]f (r, v, t) v d3 v. (8.6)
From this expression, it is clearly seen that for an isotropic distribution of velocity, the
flux of a quantity X independent of the velocity is null. At equilibrium, the Maxwell-
Boltzmann law is symmetric so that the formula yields a zero flux as expected1 . The same
behaviour is obtained for electrons in degenerate systems governed by Fermi-Dirac law. 2
To evaluate a flux, it is thus necessary to derive the velocity distribution out of equilibrium.
The non equilibrium may be due to a perturbation such as an electric field, a temperature
gradient, a concentration gradient, etc. To proceed, it is necessary to derive an equation for
the velocity distribution. This equation is known as Boltzmann equation and will be derived
in the next section.
neighbors on the boundary so that the particle stays on the boundary. Since particle stays in
a volume element and the volume does not change, we can therefore state that the number
of particles per unit phase space volume is a constant:
df (r, v, t)
= 0. (8.7)
dt
The derivative can be expanded as follows:
∂f (r, v, t)
+ v · ∇r f (r, v, t) + γ · ∇v f (r, v, t) = 0, (8.8)
∂t
where we have introduced the acceleration γ = dv/dt.
∂f (r, v, t)
+ v · ∇r f (r, v, t) + γ · ∇v f (r, v, t) = Γcoll (8.9)
∂t
The derivation of the collision term requires to carefully study the interaction between all
particles. While this can be done, we will skip a detailed treatment and instead introduce a
simple model.
f − f (0)
Γcoll = − (8.10)
τ
where we have introduced the relaxation time τ . This characteristic time usually depends
on the particle energy. In what follows, we will neglect the dependence of the relaxation
time on the energy. However, we stress that this is a rough approximation. The reader needs
to keep in mind that the relaxation time is not the same for particles with different energies
as the losses channels may strongly depend on the energy of the particles.
In what follows, we will assume that the system is close to equilibrium. Hence, we look
for a correction of the velocity distribution using a perturbative technique. Thus, we cast
the velocity distribution f in the form:
The term on the left hand side is of first order in ϕ. Hence, we only keep first order terms
on the right hand side. So far, we have not identified what is the perturbative parameter. We
have only formally postulated that a perturbative expansion is possible. To proceed, it proves
useful to make a dimension analysis of the equation. Let us introduce a dimensionless time
t = t+ θ where θ is a characteristic time of the system. Boltzmann equation can be cast in
the form:
We see that the right hand side term is proportional to a dimensionless number called Knud-
sen number given by Kn = τ /θ. This number determines the order of magnitue of ϕ. It
provides a quantitative measurement of the degree of non equilibrium of the system. A low
Knudsen number as compared to 1 corresponds to a system close to equilibrium. In that
case, a perturbative solution is justified. This situation corresponds to a system that changes
slowly in time as compared to τ . This is called a collisional regime as many collisions may
take place while the system gradually evolves. These collisions result in the possibility of
maintaining a local equilibrium at any time. Let us now examine in more detail the different
terms of the right hand side member in order to identify the first order terms in Kn .
Before analysing an example, it is important to make a technical remark on the analysis
of order of magnitude of different terms in an equation when derivatives are involved. Let us
consider a quantity with the form cos(2πx/d) where 1 and a function cos(2πx/D).
It is clearly seen that the order of magnitude of these two functions are given respectively
by and 1 respectively so that the first is much smaller than the second. Let us now suppose
that we compare their derivatives with respect to x. We find −(2π/d) sin(2πx/d) and
−(2π/D) sin(2πx/D). Now, the comparison depends on d/D so that the first term may
be the largest one. While this example is trivial, when comparing different terms in an
equation, this effect may be hidden. It can become obvious by introducing dimensionless
variables. Let us introduce here the variable x+ = 2πx/d. Hence, the derivative d/dx
becomes (d/2π)d/dx+ for the first term. A similar procedure for the second term yields
(D/2π)d/dx+ . It is seen that the typical length scale appears naturally when introducing
dimensionless variables. Indeed, the function cos(x+ ) has the same order of magnitude
than its derivative − sin(x+ ). When dealing with functions with dimensionless variables,
8.5. EXAMPLES 75
the functions and their derivatives have the same order of magnitude. Hence, the difficulty
to compare terms has been removed.
To summarize, when comparing different terms of an equation containing both functions
and their derivatives, it is necessary to first introduce dimensionless variables. By doing so,
the natural length scales of the problem appear explicitly in the equations and the resulting
terms can be compared directly.
8.5 Examples
8.5.1 Electrical conduction in an ionic solution
Here, we study the electrical conduction in a liquid containing n ions per unit volume. We
assume that the system is in stationary regime. The only cause of nonequilibrium is the
presence of the electric field E. Our goal is to derive the current density and therefore to
recover Ohm’s law. The procedure is always the same:
i) General form of the flux density;
ii) Derivation of the velocity distribution by seeking a perturbative solution of Boltzmann
equation in the relaxation time approximation;
iii) Derivation of the transport coefficient.
Here, the quantity X transported is the charge e of each ion. We only consider the
contribution of the ions with positive charge e and mass m. A similar term should be
included to account for the contribution of the negative ions. The flux can be cast in the
form: Z
j = e f (r, v)vd3 v (8.14)
Boltzmann equation in the relaxation time approximation can be cast in the form:
τ ∂f (r, v, t)
ϕ(r, v, t) = − + v · ∇r f (r, v, t) + γ · ∇v f (r, v, t) (8.15)
f (0) ∂t
Now, we need to analyse this equation in detail in order to identify the first order terms
that give a contribution to the velocity distribution pertrubation. It is obviously expected that
the electric field plays a key role. Let us examine where it appears. The first term is zero
because the system is in stationary regime so that time derivatives are null. The second term
is also null (∇r f = 0) because the system is homogeneous. The third term introduces the
derivative with respect to velocity. Since f (0) depends explicitly on velocity, this derivative
is non zero. The acceleration term is equally non zero because there is an electric force such
that mγ = eE. We finally obtain:
τ eE
ϕ(r, v, t) = − · ∇v f (r, v, t). (8.16)
f (0) m
It is easy to identify the Knudsen number from this equation:
eEτ
Kn = (8.17)
mv0
p
where v0 is the order of magnitude of the velocity given by kB T /m. This dimension
analysis provides the order of magnitude of ϕ. Hence, the perturbative expansion is valid
76CHAPTER 8. INTRODUCTION TO KINETIC THEORY OF TRANSPORT PHENOMENA. BOLTZMANN
if the velocity eEτ /m due to the work done by the field between two collisions is much
smaller than the mean thermal velocity v0 . It can also be viewed as a comparison between
two times: the collision time τ and the time θ = mv0 /eE required to accelerate the ion to
a velocity v0 . The perturbation solution is valid for τ θ.
To compare ∇v φf (0) with ∇v f (0) we note that the first term is of order one in Kn while
the second is of order two in Kn . By keeping only first order terms, we finally obtain:
e2 τ E ∂f (0)
Z
jx = − vx dvx dvy dvz
m ∂vx
jy = 0
jz = 0 (8.20)
obtain:
e2 τ E ne2 τ
Z
(0) ∞ (0) 3
jx = − [vx f ]−∞ − f d v = E. (8.21)
m m
This result has the structure of Ohm’s law. We now have an explicit form of the conduc-
tivity as a function of the characteristic of the system. This result can be applied to either
ionic solutions or plasmas. It cannot be applied to degenerate electronic systems such as
metals. We will address this issue in the next chapter and we will show that a similar form
is obtained for the conductivity.
Solution
1. Momentum flux density px :
Z
jpx ,z = d3 v[px ]f (r, v)vz . (8.24)
∂ f − f (0) ϕf (0)
vz f (r, v) = − =− . (8.25)
∂z τ τ
To first order, we get:
∂f (0) (r, v) ∂V
ϕf (0) = τ vz . (8.26)
∂vx ∂z
3. Fluid viscosity.
Inserting the velocity distribution function in Eq.(8.24), we get:
∂V
jpx ,z = −µ (8.27)
∂z
where the viscosity is given by:
Z
∂ (0)
µ=− d3 vτ vz2 mvx f (r, v). (8.28)
∂vx
To compute the integral, it is useful to note that the velocity distribution at equilibrium can
be cast in the form of the product of the particle density n and a velocity probability density
function P (v). Thus, we have:
Z
2 ∂P (vx )
µ = −nmτ < vz > vx = nkB T τ (8.29)
∂vx
Introduction to irreversible
thermodynamics.
9.1 Introduction
79
80 CHAPTER 9. INTRODUCTION TO IRREVERSIBLE THERMODYNAMICS.
The second postulate is that entropy is an additive and increasing function of energy.
Hence, when the system is divided into two subsystems, the entropy of the system is the
sum of the entropy of the two subsystems.
The third postulate fixes the entropy at zero for zero temperature.
section is to introduce in a systematic way a prescription to identify what is the force and
what is the response. For instance, what should be the ”force” at work for heat fluxes : ∇T
or ∇1/T ?
In order to introduce heuristically the relevant quantities for a formulation of a theory
of non-equilibrium phenomena, we revisit the equilibrium case. At equilibrium, we have a
well-established principle: the entropy is maximum. In order to introduce non-equilibrium,
we consider an isolated system Σ0 which can be splitted in two subsystems:
Σ0 = Σ ∪ Σ0 .
The extensive quantities in each subsystem are denoted by (X1 , X2 , ...XN ) and (X10 , X20 , ...XN
0 )
Xk0 = Xk + Xk0 ,
so that
dXk0 = dXk + dXk0 = 0.
According to the second principle, the entropy is maximum at equilibrium so that:
∂S 0 ∂S ∂S 0 ∂S ∂S 0
= + = − = 0. (9.2)
∂Xk ∂Xk ∂Xk ∂Xk ∂Xk0
This equation shows that the equilibrium is characterized by the equality of the partial
derivatives of the entropy with respect to its natural variables (i.e. extensive variables).
In the following, we denote Fk and Fk0 the partial derivatives ∂S/∂Xk and ∂S 0 /∂Xk0 . This
suggests that if they are not equal, then the difference Fk − Fk0 plays the role of the driving
force for the exchange of quantity Xk . Obviously, such an approach is expected to be valid
for small differences so that this is a theory for systems close to equilibrium. In other words,
we are working in the spirit of the linear response theory approach. With this in mind, we
introduce the quantity:
∂S 0
FXk = = FXk − FX0 k , (9.3)
∂Xk
which is called affinity or generalized force. We now give a number of examples of these
affinities. Let us start by considering the internal energy U . We get:
∂S 1 1 1
X = U; FU = = ; FU = − 0.
∂U T T T
Let us now consider the volume. Our purpose here is to show how to proceed to identify
systematically the correct affinity. We start from the thermodynamic identity as it is usually
expressed in classical thermodynamics in terms of energy:
dU = T dS − P dV + µdN + xdX.
dU P µ x
dS = + dV − dN − dX.
T T T T
From this identity, we immediately identify the affinity for the volume:
82 CHAPTER 9. INTRODUCTION TO IRREVERSIBLE THERMODYNAMICS.
∂S P P P0
X =V; FV = = ; FV = − 0,
∂V T T T
for the number of particles:
∂S µ µ µ0
X = n; Fn = =− ; Fn = − + 0,
∂n T T T
and for any extensive quantity X:
∂S x x x0
X; FX = =− ; FX = − + 0.
∂X T T T
To summarize, let us stress that the main motivation to redefine new intensive variables
(e.g. P/T instead of P) is based on the different role of entropy and energy in the theoret-
ical framework of thermodynamics. While the internal energy U is merely one extensive
variables among others, the entropy plays a central role. The usual formulation of thermody-
namics gives a particular role to internal energy for historical reasons. From a fundamental
point of view, it is the entropy that has to play a central role.
dS 0 X ∂S 0 ∂Xk X ∂S ∂S 0 ∂Xk
= = − , (9.4)
dt ∂Xk ∂t ∂Xk ∂Xk0 ∂t
k k
which can be cast in the form:
dS 0 X ∂Xk
= Fk . (9.5)
dt ∂t
k
The physical content of this equation is very important. It clearly shows that there is
a production of entropy associated to the flux of the quantity Xk if the affinity (i.e. the
discontinuity in Fk ) is non zero. Furthermore, the rate of entropy production appears as the
product of two terms: the affinity Fk and the flux ∂Xk /∂t.
We now move to systems with a continuous gradient of the internal variables such as
temperature gradient of density gradient.
We have seen previously in the framework of Boltzmann equation that the validity condi-
tion for local thermodynamic equilibrium was essentially a low Knudsen number. In other
words, the system may change over time scales which are larger than the collision time and
it may change in space with length scales which are larger than the mean free path. We now
establish the thermodynamic identity in its local version. We start from the identity:
X ∂S
dS = dXk , (9.6)
∂Xk
k
which is well defined for a system with volume V. So far, we have considered a system in
equilibrium. We can always consider the system to be an ensemble of different subsystems.
∂S
Let us introduce Fk = ∂X k
and consider a subvolume δV which can be chosen arbitrarily.
Introducing the entropy density ρS as S = ρS δV and the density of Xk denoted ρk such
that Xk = ρk δV , we see that we can write for this subsystem:
X
δV dρS = Fk δV dρk , (9.7)
k
so that X
dρS = Fk dρk . (9.8)
k
∂ρk
+ divJk = 0. (9.9)
∂t
Z Z Z
dSV d ∂ρS
= ρS (ρ1 (r, t), ...ρN (r, t))dr = dr − JS · dΣ.
dt dt V V ∂t irrev Σ
where the term on the right is the entropy production rate due to irreversible processes. We
will derive its explicit form by studying the left hand side terms. Let us first use the local
thermodynamic identity (9.8) to introduce the time rate of entropy change:
∂ρS X ∂ρk
= Fk , (9.10)
∂t ∂t
k
In the previous equation, we have used the fact that at local thermodynamic equilibrium,
the entropy has no explicit dependence on time other than mediated by the time dependence
of xk . The divergence of the current density can be cast in the form:
X X X
divJs = div Fk Jk = Fk divJk + ∇Fk · Jk
k k k
X ∂xk X
= − Fk + ∇Fk · Jk , (9.12)
∂t
k k
By analogy with the previous section, we have introduced the affinity defined by ∇Fk .
The above result shows that the source of entropy and therefore of irreversibility is propor-
tional to the gradient of the intensive quantities. More precisely, the entropy production rate
is the product of two terms: the affinity and the corresponding flux. This result provides a
clear justification of the concept of quasistatic transformation. It is seen that as the affinities
go to zero, the entropy production vanishes. It justifies the use of quasistatic tranformations
to suppress entropy generation. We now provide two examples to illustrate these concepts.
9.4.4 Examples
Heat transfer
We start by considering a cylinder of uniform section Σ and length L connecting two ther-
mostats at temperature T1 and T2 with T1 > T2 . We consider a transformation such that
9.4. FLUX AND AFFINITY IN A HETEROGENEOUS SYSTEM. 85
δQ
∆S2 = .
T2
Since entropy is additive, the total variation of entropy in the system is:
1 1
∆S = ∆S1 + ∆S2 = δQ − .
T2 T1
This formula provides the entropy generated during the heat transfer. As already pointed
out, it increases as the discontinuity T1 − T2 increases. Let us study now the local rate of
entropy production. We first start from the thermodynamic identity to identify the proper
affinity: dS = dU/T + Fk dXk . The affinity is thus given by ∇(1/T ) and the flux is the
internal energy flux which is the heat flux in this case. It is given by Fourier law:
∂ρS X 1
= ∇Fk · Jk = ∇ · [−λ∇T ].
∂t irrev T
k
This equation yields the local entropy production rate. It is interesting to observe that
the entropy is only created where the gradient is non zero and therefore, only in the wire
connecting the thermostats. In order to obtain the total entropy produced, it suffices to
sum over the whole volume of the wire with section Σ and length L. By noting that the
temperature field is one dimensional and denoting by x the variable along the wire, we get:
Z L 2
λ ∂T
dx 2 Σ.
0 T ∂x
We have T (0) = T1 and T (L) = T2 . We choose the flux orientation along the x axis so
that the flux is positive if T1 > T2 . Noting that Φ = −λ ∂T
∂x Σ is constant along the wire in
stationary regime, we can write the integral in the form:
L L ∂ T1
Z Z
1 ∂T 1 1
− dxΦ 2 =Φ dx = Φ − .
0 T ∂x 0 ∂x T2 T1
This result is clearly in agreement with the previous approach if we consider that during δt,
there was a heat flow δQ = Φ∆t.
86 CHAPTER 9. INTRODUCTION TO IRREVERSIBLE THERMODYNAMICS.
dU = T dS + V dq + µdn.
Here, we have included a term that depends on the chemical potential as an exchange of
charges implies an exchange of particles. By noting that n and q are not independent as
they are related by q = −ne in the case of electrons, we obtain the thermodynamic identity:
µ
dU = T dS + V − dq.
e
It is seen that the the electrochemical potential Vec = V − µe appears naturally. We can now
cast the thermodynamic identity in the form:
dU Vec
dS = − dq.
T T
From this identity, we can now derive the form of the affinity given by the gradient − VTec .
In the particular case of the Joule effect, the metal is isothermal and the chemical potential
is uniform so that the affinity is simply given by: −∇V /T = E/T . It follows that:
σE2
∂ρS E
= · σE = .
∂t irrev T T
The entropy production rate is thus given by the ratio of the electromagnetic power
dissipated per unit volume divided by the temperature. We know consider that Joule effect
takes place in a simple cylindrical resistor with length L and uniform section Σ. We can
introduce the voltage across the resistance U such that the electric field is given by E =
U/L. Integrating the entropy production rate over the whole volume, we get:
σE2 Σ U2 U2 1
Z
dS
= dx Σ = =
dt V T ρL T R T
where the resistance is R = ρL/Σ = L/σΣ.
X
Jk = Lkl ∇Fl . (9.14)
l
It is seen that the most general form of a flux involves all the affinities so that crossed
terms are included. Obviously, such a linear response formulation is an approximation
which is valid for systems close to equilibrium. In the framework of irreversible thermody-
namics, the existence of these fluxes is an experimental fact. The theory is a phenomeno-
logical theory. Yet, it is possible to derive general properties that must be fulfilled by the
coefficients. It is the purpose of this section to introduce these properties.
Let us first stress that the formulation of general properties can take place only if the
systematic definition of coefficients is used. We stress that the value of the coefficients are
not the same if one consider that the heat flux is proportional to ∇T as it is usually done
or if one uses the affinity ∇(µ/T ). Similarly, the charge current density is the response to
−∇Vec /T and not to E = −∇V . Hence, the linear response coefficient is not the usual
electrical conductivity.
It is worth mentioning that the linear relations established here are valid for affinities
that vary slowly in time as compared to relaxation times of the system so that the linear
coefficients are non dispersive. If that condition is satisfied, the systems are said to be
markovian. Indeed, if the system is considered to have a response that is non dispersive in
frequency domain, its response in time domain is considered to be instantaneous. In other
words, its response does not depend on the value of the excitation at time preceding the
observation of the effect.
Lα,β
X
Jαk = kl ∂β Fl , (9.15)
l,β
where exponents α, β denote the cartesian components of the vectors. The more general
form of the transport coefficient is given by a tensor of rank 2. We now assume that the
system is invariant under any rotation. This is the case for instance of thermal conductivity
of a fluid. It follows from Curie principle that the tensor Lα,β kl must be invariant under
any rotation. There are only two tensors of rank 2 which are rotationally invariant. They
are given by δ α,β and uα uβ . It follows that the linear coefficient can be cast in the form
Aδ α,β + Buα uβ . This reduces the number of parameters from 9 to 2 !
88 CHAPTER 9. INTRODUCTION TO IRREVERSIBLE THERMODYNAMICS.
We have also expressed the fluxes using linear response coefficients9.14. By inserting
this form in the previous expression, the entropy production rate appears to be a bilinear
form:
∂ρS X
= ∇Fk · Lkl ∇Fl . (9.17)
∂t irrev
k,l
This condition imposes many constraints on the linear coefficients. For instance, it
follows that thermal conductivity and electrical conductivity are positive.