Professional Documents
Culture Documents
Fall 2009
The crisis of Classical Mechanics
At the end of the XIX century classical physics consisted of two main areas:
1. Classical Mechanics: Developed by Galileo and Newton, it predicted the motion of celestial objects with
astonishing accuracy. It was also applicable (accounting for always present frictional effects) to the motion of
projectiles.
2. Classical Electromagnetism: Starting from the experiments of Coulomb, Ampère, Faraday, Lenz, etc., it
culminated with the mathematical formulation by Maxwell and the prediction that electromagnetic waves are
light (Hertz).
Both areas were highly successful in most cases in explaining experimental observations. Yet, at the threshold of
the XX century a few problems began to emerge, leading to the transition from ‘classical physics’ to ‘modern (or
quantum) physics’. The following were the major ‘puzzles’ which classical physics could not explain:
All objects emit ‘thermal radiation’, dependent on their temperature. Think of a wood log glowing red-to-white
in the fireplace, or of molten iron, etc. Roughly speaking, the composition of the object does not matter, only
its temperature: The hotter the body, the ‘bluer’ the body will appear.
Many objects do not reflect light (think of a chunk of coal) or reflect it poorly. Empirically it is found that the
light emitted by this class of bodies is almost completely independent of the details of their composition.
Thus, the concept of ‘black body’ idealizes this class of objects. It is defined as a body which absorbs light
(i.e., electromagnetic radiation), but does not reflect any of it, instead re-emitting it after having ‘thermalized’
it. The ideal model would be a black box in which radiation enters through a tiny hole. Electromagnetic wave
hitting the internal walls of the box reach equilibrium with the environment. Its spectrum (i.e., number of waves
– or ‘modes’ – n(ν)dν in an interval ν -ν + dν of frequencies ν ) is measured as light exits from the tiny hole.
0
0 1 2 3 4
ν (1014 Hz)
– Hitting a solid with light, electrons should be emitted with energy proportional to the intensity of the light
(since it’s proportional to |E|2 )...
– Instead, the electron energy grows with decreasing wavelength of the light and no electrons are emitted for
ν < ν0 , where ν0 is a function of the solid.
The light-intensity determines only the number of emitted electrons (Millikan, 1914)
– Einstein explained the experimental results with Planck’s ‘quanta’ (1905, Nobel prize in 1921). These ‘quanta’
stuff seems to be for real...
3
glass enclosure Millikan’s experiment (1914)
Ia a
LIGHT 1
electrons
ν0
Ib b
stopping potential V0 0
0 2 4 6 8 10 12
applied V V ν (1014/sec)
• Structure and spectra of atoms: Why don’t electrons fall into the nucleus?
– After Rutherford’s experiments (1913), atoms were viewed as electrons orbiting the nucleus. [Rutherford shot
radiation known as ‘α particles’ (now known to be He nuclei) at a very thin gold foil. Most α particles went through essentially
undeflected, but those who were deflected were scattered almost backwards. This showed that atoms are essentially ‘empty’ but
have a massive ‘core’ (or ‘nucleus’).] But Maxwell equations predicted that the electrons should radiate their energy
and fall into the nucleus.
h
p = mv =
λ
nh
2πr = , (6)
mωr
2 e2
ω = . (7)
4π0 mr3
e2 n2h̄2 e2 n2h̄2
= → = ,
4π0 mr3 m2 r 4 4π0 mr
or:
4π0 n2h̄2
rn = , (9)
me2
where the subscript n has been added to show that only discrete values of the orbital radius are allowed. This was de Broglie’s main
goal.
b. Squaring this equation and inserting it into the first of the equations (8):
me4
ωn = , (10)
(4π0 )2 n3h̄3
where again the subscript n shows that only discrete angular velocities are allowed. The total energy will be:
1 2 r2 − e
En = mωn n .
2 4π0 rn
Multiplying Eq. (5) by r, the second term in the right-hand-side (rhs) of this equation becomes mω 2 r2 , so:
1 2 2
En = − mωn rn .
2
In words, the total energy is one half of the potential energy. This is a particular example of a general theorem of classical mechanics
n2h̄2
En = − .
2mrn2
The factor in red is the Rydberg constant R, so, comparing Eq. (11) with Eqns. (1)-(3) above, we explain the Lyman, Balmer, and
Paschen series as the discrete wavelength of the light emitted by an electron falling from an orbit n to the orbit m=1 (Lyman), m=2
(Balmer), or m = 3 (Paschen).
Start from the classical energy (often called ‘Hamiltonian’ H ), kinetic plus potential:
p2 e2
H = − ,
2m 4π0 r
h̄2 2
− ∇ ψ(r) + V (r)ψ(r) = Eψ(r)
2m
After years of struggle and confusion, a group of physicists working at the Institute of Theoretical Physics in
Copenhagen (founded by Niels Bohr) proposed the following interpretation.
|Ψ(r, t)|2 dr = ρ(r, t) is the probability of finding an electron in a volume dr around r at time t. Of course:
ρ(r, t) dr = |Ψ(r, t)|2 dr = 1
∂ρ
+ ∇·S = 0,
∂t
h̄2 d2 ψ(z)
− = (E − V0 )ψ(z) .
2m dz 2
Put ψ(z) = eikz and find k = ±[2m(E − V0 )]1/2 /h̄ or E = h̄2 k2 /(2m), so that h̄k is the z-component
of the electron momentum.
• Another simple example: Particle in a box.
Potential of a ‘quantum well’:
∞ for z ≤ 0
V (z) = 0 for 0 < z < L .
∞ for z ≥ L
Boundary conditions:
h̄2 d2 ψ(z)
− = Eψ(z) for − L/2 < z < L/2 ,
2m dz 2
2 2 2
with En = n h̄ π2 .
2mL
The energy depends on the boundary conditions at L, like the note emitted by a vibrating violin string depends
on where one places the finger.
• A tunneling problem: Potential of a ’barrier’:
0 for z ≤ 0
V (z) = V >0 for 0 < z < L .
0 for z ≥ L
General solution for electron of energy E < V coming from the left, partially reflected back and partially
transmitted:
Aeikz + Be−ikz for z ≤ 0
ψ(z) = Ceκz + De−κz for 0 < z < L ,
F eikz for z ≥ L
with k = (2mE)1/2/h̄, κ = [2m(V − E)]1/2 /h̄.
Using continuity of ψ(z) and dψ(z)/dz at z = 0 and z = L one can find the coefficients A, B , C , D, and
F.
The ratio T = |F |2 /|A|2 is the ‘transmission coefficient’, the ratio R = |B|2 /|A|2 is the ‘reflection
coefficient’. Note that, classically, one would expect T = 0.
This is the first surprise of Quantum Mechanics: Particles can populate regions which are classically forbidden!
1. Quantum Mechanics has an intrinsic ‘probabilistic’ nature, different from all ‘classical’ physical theories. Just
consider the interpretation of the squared wavefunction |Ψ(r, t)|2 as the ‘probability’ of finding the particle
at position r at time t. So, we cannot say with precision where the particle will be at any given time, but we
can only define an ‘average’ position r:
r = Ψ∗(r, t) r Ψ(r, t) dr , (12)
V ol
(recall that the wavefunction is assumed to be normalized, that is, V ol Ψ∗(r, t) Ψ∗ (r, t) dr = 1), which
is called the ‘expectation value of the position r on the state Ψ. Similarly, we can define the expectation value
of the momentum,
∗
p = Ψ (r, t) p Ψ(r, t) dr = −ih̄ Ψ∗ (r, t) ∇ Ψ(r, t) dr . (13)
V ol V ol
In some cases, the state Ψ of the particle will be such that one variable is precisely defined (such as the
momentum h̄k for a free particle), but the position will be undefined.
• Atoms: Building up the Periodic Table
We see in this wection how the Schrödinger equation – together with a few additional concepts – can allow us
to build (conceptually) the periodic table of the elements. In turn, this will give us some ideas about atomic
orbitals, which constitute the basic concepts underlying the formation of atomic bonds in molecules and crystals
(which is what we are after).
First, we need the concept of quantum number. The example of a particle in a three-dimensional (3D) box will
illustrate this.
– A little bit of math: A free 3D particle.
We have seen before how we can solve the Schrödinger equation in a couple of simple one-dimensional cases.
h̄2 ∇2
− ψ(r) + V (r) ψ(r) = E ψ(r) . (19)
2m
Let’s assume that the potential is equal to a constant, V (r) = V0 , and let’s express the Laplacian and the
spatial coordinate r in terms of their components:
2 2 2 2
h̄ ∂ ∂ ∂
− ψ(x, y, z) + ψ(x, y, z) + ψ(x, y, z) + V0 ψ(x, y, z) = E ψ(x, y, z) .
2m ∂x2 ∂y 2 ∂z 2
(20)
A standard way to solve a partial differential equation of this type is to separate the variables, that is, to look
for a solution of the form:
ψ(x, y, z) = ψx (x) ψy (y) ψz (z) , (21)
where ψx , ψy , and ψz are three different functions in one variable only. Let’s now set V0 = 0 (which we can
always do, since the zero of the potential energy is arbitrary) and insert this into Eq. (20):
2 2 2 2
h̄ d ψx (x) d ψy (y) d ψz (z)
− ψy (y) ψz (z) + ψx (x) ψz (z) + ψx (x) ψy (y) = E ψx (x)ψy (y)ψz
2m dx2 dy 2 dz 2
(22)
Let’s now divide both sides by ψx (x)ψy (y)ψz (z):
2 2 2 2
h̄ 1 d ψx (x) 1 d ψy (y) 1 d ψz (z)
− + + = E. (23)
2m ψx (x) dx2 ψy (y) dy 2 ψz (z) dz 2
This equation tells us that the sum of 3 functions of different variables (x, y , and z ) must always be equal
to a constant (E ) for all possible values of x, y , and z . This is possible only if each function is equal to a
h̄2 d2 ψx (x)
− = Ex ψx (x) , (27)
2m dx2
which has a general solution
ik x
ψx (x) = Ax e x , (28)
where Ax is an arbitrary multiplicative constant (which we shall fix later by some ‘normalization’ procedure)
and
1
kx = 2mEx . (29)
h̄
Similarly for the y and z components, so that the full wavefunction has the form
having defined the wavevector k as the vector with components (kx , ky , kz ) and A is an arbitrary constant.
Equation (30) describes the particle as a ‘plane wave’ with wavevector k. Note that its momentum p will be
h̄k (as we saw in 1D), and the energy of the particle is E = h̄2 k2 /(2m).
– A 3D particle in a box.
Let’s now go back to Eq. (19) or (20), but now let’s assume that the particle is confined in 3D cubic box of
where V = L3 is the volume of the box and the energy of the particle will be
Note that the normalization constant A = (8/V )1/2 in this case has been fixed by requiring that there is
unit probability of finding the electron inside the box:
dr |ψ(r)|2 = 1 .
V
Note that in 1D the particle has only one degree of freedom and we need only a single ‘quantum number’ n to
label the energy levels. In 3D the particle has 3 degrees of freedom and we now need 3 quantum numbers, nx ,
ny , and nz to label the possible states (that is, the possible ‘types’ of wavefunction) of the particle.
• The hydrogen atom.
The case of the H atom is conceptually similar to the case of the particle in a cubic box. Unfortunately, it is quite
more complicated from a mathematical perspective, because the potential confining the particle (the Coulomb
potential energy is −e2 /(4π0 r)) is not ‘flat’, but it depends on the distance r from the nucleus. Therefore,
we shall not go into a detailed solution of the Schrödinger equation, but we shall consider only its qualitative
The ‘radial function’ Rn,l (r) describes the way the wavefunction spreads away from the nucleus. The ‘angular
functions’ Θl,m (θ) and Φm (φ) describe how the wavefunction is distributed as a function of the polar and
azimuthal angles.
As we saw before in the case of the 3D particle in the cubic box, we need 3 quantum numbers to label the
possible energy levels: The ‘principal’ quantum number, n, can take the values:
n = 1, 2, 3, ... ∞ ,
and is related to the number we saw when dealing with the simple semiclassical model (pages 5-7). Usually
(that is, in the absence of magnetic fields and ignoring smaller effects) the electron energy depends only on this
number via an expression identical to what we saw on page 7. The angular quantum numbers l and m are
related to the angular momentum (h̄l is related to its magnitude and h̄m to its component along the polar
aixs) of the ‘orbiting’ electron and can take the values:
Let’s start with the simplest element, H. Its nucleus has a single positively charge proton and it is surrounded by
a single electron. The lowest possible orbital the electron can occupy is the ‘ground state’ (n=1, l=0, m=0).
We use the notation 1s1 , meaning that in the orbital 1 (n=1) which is an s-wave (l=0) we have one electron
(the superscript ‘1’). Note that the electron is bound rather weakly to the nucleus (we saw that the ionization
energy is about 13.6 eV). Therefore, H loses its electron quite easily (that is, it is easlily ionized). This trait is
common to all elements with only one electron in the outermost shell. These elements are very reactive (think
of hydrogen mixing with oxygen!)
The third element, Li, has three protons in its nucleus (Z =3), so it must have 3 electrons orbiting it in order to
be charge neutral. The third electron, as we just saw, must go into the n=2 shell, so Li will have the electronic
configuration 1s2 2s1 . It is a very reactive metal, for the reasons we discussed dealing with H: It is very easy to
strip away the lone outer electrons, leaving behind a positive Li ion.
We now keep populating the n=2 shell, adding another electron in the s-orbital (berillium, Be, 1s2 2s2 with
Z =4), then 6 more into the 3 p orbitals, adding electrons with spin up and spin down in sequence. As we
move from the reactive metals (Li, Be) we build elements with 3 ‘outer’ electrons (boron, B, 1s2 2s2 2p1), 4
outer electrons (carbon, C, 1s2 2s22p2 ), 5 (nitrogen, N, 1s22s2 2p3 ), 6 (oxygen, O, 1s2 2s2 2p4 ), and 7 outer
electrons (fluorine, F, 1s2 2s2 2p5 ). As we do so, we move from elements which like to lose electrons, to those
who like to gain them. Eventually, with neon (Ne, Z =10, 1s2 2s2 2p6 ) we complete the n=2 shell and we hit
another inert gas, like He unlikely to react chemically. The 8 slots we have filled while populating the n=2 shell
are the 8 columns of the periodic table. The chemical properties of any two elements having the same number
of electrons in the outer shell are very similar.
So, as we fill the n=3 shell, we start from sodium (column I, Na, Z =11, 1s2 2s2 2p63s1 ) which, having
only one electron in the outer shell, behaves like Li. And we end up with argon (column VIII, Ar, Z =18,
Things get a little more complicated now, because the energy of the 4s shell is lower than that of the 3d shell.
So, with potassium (K) and calcium (Ca) we fill the 4s shell, but then we go back filling the 3d shell. The fact
remains that – up until further complications caused by the f states of transiton metals – each element belongs
to its own column which determines the chemical properties vua the number of electrons in the outer shell.
Of interest to us are silicon (Si, Z =14, 1s22s2 2p6 3s2 3p2 ) and germanium (Ge, Z =32,
1s22s2 2p6 3s2 3p6 3d104s2 4p2 ) which, like C, belong to the IV column. Their 4 outer electrons are
s2 p2 and can easily ‘hybridize’ into 4 sp3 orbitals, forming the tetrahedal structure required to bond atoms in
their cubic crystal form. Elements of column III (B, Al, Ga, In) have outer electrons arranged as s2 p1 , that is,
one fewer electron than in Si and Ge. Elements of column V (N, P, As, Sb) have outer electrons arranged as
s2 p3 , that is, one more electron than in Si and Ge.
Finally, electrons in fully occupied shells are called ‘core electrons’. They do not contribute to the chemical
activity of the element. Electrons in the outer shell are called ‘valence electrons’.
• Molecules and Bonds.
If we bring several atoms together, several things may happen, depending on the electronic populations of the
outer shell.
1. Nothing. For example, bringing 2 Ne atoms together, both atoms retain their electronic configuration
unaltered, as they are both in a configuration of a completely filled outer shell with 8 electrons (2 in s orbitals,
6 in p orbitals). This is why elements of the VIII column are called ‘inert’ elements (or gases).
2. Formation of an ionic bond. Let’s consider Na and Cl. Na has 1 electrons in the outer s shell, Cl 7 valence
electrons (2 in the s shell, 5 in the p shell). Energetically it is advantageous for Na to release its electron
(weakly bound, as the core electrons shield the nuclear charge), while Cl ‘loves’ to acquire that electron and
complete the filling of its outer shell. Thus the sodium atom becomes positively ionized (Na+ ) and the clorine
atom becomes negatively ionized (Cl− ). Now we have a positive (Na+ ) and negative (Cl− ) particle which are
attracted by the strong Coulomb force. The two ions are said to have formed an ‘ionic bond’ and stick together
in the NaCl molecule (regular kitchen salt). Ionic bonds are usually quite strong with ‘binding energies’ of
Cl– O –
Finally, it is important to realize that during the formation of a bond the electronic orbitals change their structure.
For example, during the formation of the O2 molecule the electrons ‘shared’ by the two oxygen atoms will ‘orbit’
both cores (nuclei+core electrons), forming new orbitals by a ‘linear superposition’ of the wavefunctions of the
2p orbitals of each atom. As shown in Fig. 3.2 of the text, out of these two ‘single atom’ orbitals we form two
molecular orbitals: A lower-energy ‘bonding orbital’ and a higher-energy ‘anti-bonding orbital’, respectively at a
slightly lower and higher energy then the original single-atom orbitals. The bonding orbital is occupied by the
two shared electrons and it exhibits a larger charge density (that is, a larger |ψ|2 ) between the atoms. In other
words, there is a large probability of finding the electrons between the atoms. Thus, this orbital contributes to
keeping the atoms together. The antibonding orbital is empty and it is associated with a charge density (again,
|ψ|2 ) which is larger away from the bond.
The main purpose of this course is the study of how current (carried by electrons) flows in semiconductor
devices. Therefore, it is important to understand the properties of the electron motion in semiconductors.
Semiconductors, as we shall see later, are materials which have current-conduction properties intermediate between
insulators (glass, wood) and conductors (copper, aluminum). Their practical importance originates from two major
properties:
1. The addition of a minute amount of impurities in perfect semiconductors (at the level of a few parts per million)
can modify their current-conduction properties by huge amounts, transforming what is essentially an insulator
into a pretty good conductor. This property is extremely useful in designing ‘devices’ which use the changing
conductor/insulator properties of the material to make tiny electronic ‘switches’ (on/off, 0/1 in binary logic).
2. Their electronic structure (as we’ll see in the next section) includes the presence of an “energy gap” which
behaves like the energy step between two energy levels in an atom. This determines the ‘optical properties’ of the
material. That is, the wavelength (color) of the light which can be emitted and absorbed by the semiconductor.
This property allows the use of semiconductors to make devices interacting with (emitting/absorbing) light in
the visible and near-visible spectrum (lasers, light-emitting diodes, etc.)
Most semiconductors are made up of elements of the IV column of the periodic table (group IV elements,
such as Si and Ge, called elemental semiconductors) or are a combination of elements from the third and fifth
column (such as GaAs, InGaAs, GaN, InGaAsP called III-V compound semiconductors) or, finally, a combination
of group II and group VI elements (such as CdSe, ZnS, HgCdTe, called II-VI compound semiconductors).
The electrical conductivity of semiconductors depends on the arrangement of the atoms which constitute the
semiconductor itself. And this dependence is very strong, judging from the fact that a few impurities can modify
it dramatically. Thus, if we want to study electronic condution, we have to analyze how atoms are arranged in
• Definitions
– Crystals are regular arrangements of atoms which look identical when viewed from 2 points
r and r = r + Rl ,
where
Rl = l1 a + l2 b + l3 c
is a ‘translation operation’ with integers l1 , l2 , and l3 and with ‘fundamental translation vectors’ a, b, and
c. The vector Rl is also called a lattice vector. The length of these vectors are called lattice constants a,
b, and c. One can define infinitely many translation vector. It is convenient to consider always the ’shortest’
ones, which we have called above ‘fundamental’ translation vectors..
– A cell is the parallelepiped formed by the fundamental translation vectors. Note that the repetition of a cell
obtained by applying a translation operation fills the entire space.
– As there are infinitely many possible choices for the translation vectors, so there are infinitely many possible
choices for the cells. It is often (not always, though!) convenient to work with the primitive or Wigner-Seitz
cell, which is the cell having the smallest possible volume, given by Vcell = |a × b · c|. It can generated
by connecting all lattice points with straight lines, bisecting these lines with planes normal to the lines, and
considering the volume enclosed by these planes.
– A lattice is the set of all points r generated by translation operations. The symmetry properties of a
three-dimensional lattice allow the existence of 14 types of Bravais lattices (those generated by Wigner-Seitz
cells).
– A basis is the set of the coordinates of all atoms within a primitive cell. For example, in the next figure we
see the structure of the GaAs (or) Si crystal: Ga atoms occupy the edges and the center of the faces of the
cube (the crystal is then said to be face-centered cubic, or fcc. As atoms are displaced from the edge at the
left by a(1/4, 1/4, 1/4). The basis of the cell is thus made up of the two coordinate vectors describing the
2.0
Ev (Γ)
In this section we shall discuss the electronic properties of semiconductor crystals. These are characterized
mainly by the existence of allowed ‘energy bands’ separated by ‘energy gaps’. Electronic transport in semiconductors
depends crucially on both the characteristics of the energy bands and by the value of the gap between occupied
and empty states.
Understanding bands quantitatively requires solving the Schrödinger equation for many electrons in the
potential due to all ions in the crystal. For example, ignoring core electrons, a crystal composed of N Si atoms
can be viewed as a lattice of N Si+4 ions and 4N electrons embedded in the complicated Coulomb potential of
all of these ions.
The problem may appear daunting. However, the periodicity of the lattice allows major simplifications:
Roughly speaking, all we have to do (so to speak...) is to study the electronic properties of a single cell subject to
the periodic boundary conditions stemming from the symmetry of the lattice. Thus, we can reduce the problem to
one involving ‘only’ 8 electrons and 2 4-fold-ionized Si ions.
General theorems are available to compute the band structure of solids: Most notably, several schemes to
approximate the potential of the ions and all ’other electrons’ to a fixed pseudopotential independent of what the
other electrons do, so that we can use a single electron picture, and Bloch theorem, which tells us how to reduce
the complicated wavefunction as a product of a simple ‘free electron envelope’ wavefunction and a complicated
‘fast-wiggling’ wavefunction, the same in each cell. Thus each electronic state is labeled by something which
resembles the wavevector of a free electron, k, and can be written as:
ik·r
ψk (r) = e uk (r) , (33)
Despite these powerfull simplifications and approximations, the problem is still mathematicaly somewhat
convoluted. So, we shall ignore quantitative aspects and discuss instead the band structure of solids only from a
qualitative point of view.
Let’s start by considering what happens when we bring 2 or more Si atoms together.
• Coupling atoms.
Consider two Si atoms, widely separated, as in the figure below. This shows the Coulomb potential around
each nucleus (∝ 1/r , where r is the distance from the nuclei), the energy of the 3 occupied levels, and their
electronic confuguration. Let’s recall here the important fact that there are 8 states available in the n=3 energy
level (2 3s-states and 6 3p-states), but Si has only 4 electrons in this shell, so that the outermost (valence)
shell is only half occupied.
2s22p6 2s22p6
1s2 1s2
Now let’s bring the nuclei much closer, so that the potential around each nucleus is affected significantly by
the other atom. The ‘core’ levels (n = 1 and n = 2) do not feel much the perturbation caused by the other
nucleus, but the highest-energy occupied level n=3 do feel a dramatic change. The orbitals now spread over
both nuclei. Clearly, their energy will change, because the potential itself has changed. Qualitatively, what
happens is that, starting from the original hybrid sp3 orbitals of a single atoms, we will have a ‘distortion’ of
each of them, caused by the new potential. In addition, we may form two superposition of the orbitals, much as
we had formed the hybrid sp3 . One superposition corresponds to orbitals with a large electron probablity-density
between the nuclei. This is a ‘bonding orbital’, since these ‘shared electrons’ constitute a covalent bond. Since
the electrons in this orbital sit close to the attractive nuclei, the energy of this orbital is low. The other linear
combination of orbitals we may form corresponds to orbitals having a small electron density between the nuclei.
These orbitals are ‘antibonding’, since they do not contribute to the covalent bond between the two Si atoms.
Their energy will be higher, since the electrons, on average, sit farther away from the nuclei.
2s22p6 2s22p6
1s2 1s2
The net effect is that the 4-fold degenerate energy levels corresponding to the 2 sp3 orbitals of the two separate
atoms, each containing one (of a possible maximum of 2) electrons, have split into 2 lower-energy bonding states
(each with a slightly diffrerent energy) and 2 higher-energy (also each at a slightly different energy) antibonding
states. The electrons will occupy the lowest-energy states to form the bond. The antibonding states will remain
unoccupied.
BONDING Ψ1+Ψ2
Ψ1–Ψ2
ANTIBONDING
Let’s now imagine what happens when we bring together 3 Si atoms: They will bond together forming a set of
3 occupied bonding states (at 3 different but very similar energies) and 3 antibonding unoccupied states.
• Energy bands.
Now let’s consider a Si crystal formed by, roughly, N = 1022 atoms. What happens now is very similar
to what we have seen in the case of 2 and 3 atoms. The N atoms will contribute a total of 8N possible
sp3 orbitals, originally all at the same energy. Only 4N of them are occupied. They will split into 4N
occupied orbitals contributing to the bond and 4N empty antibonding orbitals. The energy of each of the
The valence and conduction bands are separated by an energy gap whose existence stems from considerations
similar to those which caused the presence of forbidden energies in the isolated Si atom.
Suppose that we apply an electric field to a Si crystal (by placing in contact with two metallic plates, for
example, at the opposite ends of the crystal). In principle, the electrons would ‘like’ to respond to the field and
move away from the negatively-charged plate towards the positively-charged plate. But ‘motion’ requires gaining
kinetic energy. Therefore, the electrons in the valence band should acquire a bit of extra energy in order to move.
But since the valence band is full, they may gain energy only by ‘jumping’ to the next available empty states in
the conduction band. But this jump requires a large amount of energy (about 1.1 eV at room temperature).
Therefore, the electrons cannot move so that no current will flow in the crystal, despite the electric field we have
aplied... In other words, a perfect Si crystal (we’ll see below why we stress the word ‘perfect’) behaves like an
insulator.
Glass (silicon dioxide, SiO2 ) is another insulator, like Si, but its band-gap is about one order of magnitude larger
than that of Si (for Si we have seen that Eg ≈ 1.1 eV, while for SiO2 we have Eg ≈ 9.2 eV).
The figure below shows qualitatively the difference between semiconductors, insulators, and semiconductors. The
difference betweem semiconductors and insulators is only quantitative: If the forbidden energy gap is very large,
even the presence of impurities cannot modify the insulating poperties of the material. For semiconductors,
instead, the addition of a small number of selected foreign atoms (impurities) can alter dramatically their
properties. We shall see how this process (called ‘doping’) is intentionally done.
FULL
FULL
h̄2 k2
E(k) = , (34)
2m
where m is the mass of the electron. The figure below shows this ‘parabolic’ relation.
3
2 2
k
E(k) =
2m
KINETIC ENERGY E (eV)
2
0
–10 –5 0 5 10
WAVEVECTOR k (109 m–1)
The figure below shows results for these types of calculations applied to the study of the band structure of Si.
The labels L, X , Γ, U , K , etc. label special highly simmetric points in the BZ. For example, the point Γ is
the center of all reflections, inversions, etc. for a cubic crystal. It is called the ‘center of the BZ’ (or, simply,
‘zone center’), shown in the figure above.
4
Γ2’
Γ2’
3
L3
Γ15 Γ15
2
ENERGY ( eV )
1 L1 X1
–1 Γ8
Γ7
–2 L4,5
L6
–3
X5
–4
–5
–6
X ZWQ L Λ Γ ∆ X U,K Σ Γ
Looking at the figure, we see that the valence bands reach a maximum energy at about 1.17 eV below the
minimum of the conduction band. Three valence bands reach this maximum value at the same point, the Γ
point. The lowest-energy conduction band, instead, reaches a minimum value (arbitrarily set to zero) near the
Let’s consider now the function the function En (k) for n = 5 (the first conduction band, so we also call it
ECB (k)) in the proximity of its minimum. Let’s consider, for simplicity, this function on a line going from the
BZ-center Γ to X : This line correcponds to wavevectors of the form k = (k, 0, 0). The function ECB (k))
reaches its minimum at the point k = (k0 , 0, 0), where k0 ≈ 0.852π/a. Let’s expand this function around
its minimum, k0 :
Now: Due to the fact that we have set the zero of the energy at ECB (k0 , 0, 0), the first term at the
right-hand-side (rhs) vanishes. Due to the fact that we have an extremum (a minimum) at k = (k, 0, 0), the
first derivative vanishes by definition of extremum, so that also the second term at the rhs vanishes. Thus,
Eq. (35) becomes:
1 ∂ 2 ECB (k0 , 0, 0) 2
ECB (k, 0, 0) ≈ 2
(k − k0 ) . (36)
2 ∂kx
Let’s now shift the kx by −k0 so that k → k − k0 and, finally, rewrite Eq. (36) as:
1 ∂ 2 ECB (k0 , 0, 0) 2
ECB (k, 0, 0) ≈ 2
k . (37)
2 ∂kx
Now let’s compare Eq. (37) with Eq. (34). Both are parabolic laws (i.e., the energy grows with the square of
the wavevector). We can make them look even more similar if we introduce the quantity:
1 1 ∂ 2 ECB (k0 , 0, 0)
= 2
. (38)
m∗ h̄ ∂k 2
x
h̄2 k2
E(k) ≈ , (39)
2m∗
which is just the kinetic energy of a free electron, but with its mass m replaced by an effective mass m∗ .
The figure below shows a blow-up of the area covered by the red rectangle in previous plot, showing in more detail
the extrema of the conduction and valence bands, together with parabola (dotted red curves) approximating the
‘exact’ energy-vs-k relationship E(k) (also called ‘dispersion’) with the effective-mass parabolas.
1.0
0.5
0.0 X
ENERGY ( eV )
–0.5
Eg
–1.0
Γ
–1.5
–2.0
–0.5 0.0 0.5 1.0 1.5
k–VECTOR (2π/a)
h̄2 2
kx ky2 + kz2
E(k) ≈ + . (40)
2 ml mt
We need two effective masses: A heavier ‘longitudinal’ effective mass (about 0.9 the free electron mass m0 ) as
we move along the line from Γ to X , and a much lighter ‘transverse’ mass (mt ≈ 0.19 m0 ) as we move away
from the minimum at (k0 , 0, 0) along a direction perpendicular to the kx -axis. Equi-energy surfaces are thus
‘ellipsoids of rotation’, solids generated by the rotation of an ellipse around the axis along its major (longitudinal)
axis.
The band structure of other cubic (fcc) semiconductor is quite similar. The two figures below show those of
Ge and GaAs. Note that in both semiconductors, as also in Si, the maximum of the valence bands is triply
degenerate and it occurs at the center of the BZ at the Γ-point. The minimum of the conductions band for
Ge, instead, occurs at the 8 symmetry points called L for Ge, with local minima at a slightly higher energy at
the 6 points X (and one more at Γ). For GaAs the minimum occurs at Γ. This makes GaAs a ‘direct-gap’
semiconductor, with both the VB maximum and the CB minimum at the same position in k-space. This has
important consequences on the optical properties of GaAs (and many other III-V compund semiconductors),
since electronic transitions involving photons are ‘vertical’, as we shall see later.
4 L4,5 4
L3 Γ15
L6 Γ8
3 3
Γ15
Γ8
2 Γ6 2
ENERGY ( eV )
X3
ENERGY ( eV )
1 Γ7 X5 1
L1
L6 Γ7
Γ1 X1
0 0
Ge Γ8
–1 –1 GaAs
Γ8
L4,5 Γ7
–2 –2 L4,5
L6 Γ7
–3 –3 L6
X5
X7
–4 –4
X6
–5 –5
–6 –6
X ZWQ L Λ Γ ∆ X U,K Σ Γ X ZWQ L Λ Γ ∆ X U,K Σ Γ
Despite some minor ‘clumsiness’ due to the need of an anisotropic mass, nevertheless we have obtained an
extremely powerful simplification: The whole set of complications caused by the lattice, its symmetries, the
nature of the ionic potentials, and, to some extent, the entire quantum-mechanical machinery underlying the
1. We can describe electrons as classical particles with a wavevector k. The physical momentum of electrons with
low energy (that is, with k close to the minimum of the band) is h̄k, as for free electrons. (As k moves away
from the minimum, h̄k, properly called crystal momentum, ceases to be simply related to the real momentum
of the electron.
2. The energy of the electron is simply given by the free-electron expression provided we use the effective mass
m∗ .
3. Sufficiently close to k0 , the electron velocity is given by the classical expression (for a simple isotropic mass):
dr p h̄k 1
= v = = = ∇kECB (k) . (41)
dt m∗ m∗ h̄
The second equality is obvious from the definition p = mv, having replaced m with m∗ . The third
equality comes from observation 1 above that p = h̄k. The final equality relies on the fact that this is the
expression for the group velocity exhibited by a packet of waves in a medium in which the frequency ω is a
function of wavevector, ω(k) = ECB (k)/h̄, similarly to electromagnetic waves in a medium in which the
dielectric constant depends on wavelength. These media are called ‘dispersive’. This is the origin of the name
‘dispersion’ usually given to ECB (k).
4. Suppose that, somehow, we manage to have electrons in the conduction band (either because the material is
a metal, or because we have given enough energy – with light of short-enough wavelength so that hν > Eg ,
for example – to electrons in the valence band so that some of them can ”jump” across the gap). If we apply
an electric field to the crystal, these electrons can move. It can be shown that the (crystal) momentum h̄k of
the electron obeys the equation of motion:
d(h̄k)
= −eF . (42)
dt
Replacing the (crystal) momentum h̄k with the momentum p, we recognize this equation as Newton’s second
NV B
J = −e vi − (− e vj ) = e vj , (46)
j=1
The equations of motion (41) and (42) are also valid for holes, provided we replace the charge −e with +e and
the dispersion ECB (k) (or, more simply, the effective mass m∗ in the CB) with the dispersion in the valence
band, EV B (k) (or the effective mass in the VB).
The figure on page 48 illustrates qualitatively how electrons and holes move under the action of an electric field.
It shows a “band diagram”, in which the x-axis represents the position and the vertical axis represents the total
energy, which we shall measure in eV. The two lines represent the top of the valence band (lower line, labeld as
EV ) and the bottom of the CB, labeled as EC . At the far left and far right are regions of zero electric field.
Therefore the lines, representing potential energy, are ‘flat’. In the central region, instead, the slope of the lines
is proportional to the electric field, whose x component Ex = dEC /dx is positive (that is, the field points
to the right). [The relation Ex = dEC /dx comes from the definition of the electrostatic potential V (x) as Ex = −dV /dx.
The potential energy for electrons is −eV (x), since the electron charge is negative, −e. Since we have defined the line EC as the
potential energy – measured from the bottom of the CB –, we have −eV (x) = EC (x), so Ex = −dV /dx = (1/e)dEC /dx.
Finally, now note that when measuring energy in eV, we must divide the energy (measured in joules) by the magnitude of the electron
charge. So, numerically we have Ex = dEC /dx, when EC is expressed in eV].
Electrons can be thought as ‘entering from the right’ and move to the left, driven by the force −eEx . As
they do so, they move along horizontal lines, retaining the same total energy they had originally, but converting
potential energy into kinetic energy. This can be viewed as the vertical distance between the EC -line and the
location of the green dot (the electron) at a given position x. The parabolas superimposed to the EC -line shows
the dispersion E(k) at a given position. Accelerated electrons climb along the dispersion as their wavevector
e– EC
k
EV
electron
kinetic
energy
hole
kinetic
energy
h+
POSITION
• Doping.
– So, once more: The technological importance of semiconductors stems from the fact that we can make them
decent insulators or excellent conductors by adding ‘free carriers’ (electrons in the CB, holes in VB).
– This is achieved by introducing ‘substitutional’ impurities. In Si (a group IV element, with 4 electrons in
the bonding sp3 valence orbitals) the substitution of a group III element (‘acceptors’ such B, Ga, In, having
only 3 valence electrons) causes the absence of one electron in the valence band for each acceptor atom. Si
so doped is said to be p-type, since current will be carried by holes, which we just saw that can be viewed
as positively charged ‘particles’. By ’accepting’ an extra electron (that is, giving away the ‘hole’), acceptors
become negatively charged ionic impurities. The substitution of a group V element (’donors’ such as P, As,
Sb, having 5 valence electrons) causes the presence of one extra electron in the CB for each donor impurity.
Si so doped is said to be n-type, since negatively charged electrons carry the electric current. By ’donating’
an extra electron acceptors become positively charged ionic impurities.
– Technologically doping can be performed in several ways:
1. By diffusion: Si is exposed to a gas rich of the dopant species. These diffuse into Si. Thus, we have
a depth-dependent distribution of dopants, dictated by the diffusion process: Higher concentration at the
surface, dropping rapidly with depth. This is called the ‘doping profile’. Diffusion proceeds typically (not
always, though) ‘interstitially’, that is, the dopant atoms ‘sneak in’ between Si atoms. As such, the dopants
are electrically inactive. Therefore, the wafer must be heated for long times (tens of minutes) at some high
temperature (typically 900 C to 1100 C) to allow dopant atoms to replace thermally dislodged Si atoms and
become ‘substitutional’ impurities. This is called an ’annealing step’ used for ‘dopant activation’. Historically
this was the first process adopted. However, as devices became smaller, the re-arrangement of the dopants
– It is easy to see why impurities like dopants are easily ionized. Thus, considering for example donors like
P in Si, the fifth ‘extra’ valence electron of P is easily excited into the conduction band, without remaining
attached to the dopant. We can view the 5-th-electron+impurity system as resembling the hydrogen atom:
The five-fold ionized P core is surrounded by 4 electrons in the sp3 bonding orbitals, exactly as if it were the
original Si atom, but it is singly positively charged. The extra electron orbits the core much like an electron in
the H atom. We can use exactly the same theory, but with two major modification: First, the electron has a
different (smaller, in Si and most – but not all – semiconductors) effective mass and. Second, the dielectric
constant of Si is larger (as it is in all semiconductors). Let’s see how this affects the energy levels.
The external potential caused by the substitutional ionized impurity is a ‘screened’ Coulomb potential: It is
screened by:
1. the valence electrons and by some lattice rearrangement, which is accounted for by the macroscopic dielectric
constant s of the semiconductor
2. the free carrier (holes in the VB, electrons in the CB)
e2
Vdop (r) = −
4πs r
Thanks to effective-mass approximation, we can treat this potential as an hydrogen-like problem, as we said
above. Assuming a scalar effective mass (that is: the same in all directions, which we saw is not true for Si,
but we shall ignore this complication for now), the energy levels (measured from the band extremum) will be:
m∗ e4
En = − .
32π3 n2 h̄2 2s
Note how the magnitude of these energies is reduced by a factor m∗ 20 /(m0 2s ) with respect to the hydrogen
atom. Since for most semiconductors m∗ /m0 ∼ 0.1 and 0 /s ∼ 0.1, the ionization energy about 104
times smaller than that of H, so it is of the order of 1-to-10 meV. This is smaller than the the thermal energy
kB T (≈ 25 meV at 300 K). Thus, we see that dopants are almost always ionized. ‘Freezout’ (that is, the
condition in which free carriers are trapped into impurity states and the semiconductor actually becomes an
insulator) occurs at very low temperatures. Note also that the Bohr radius of the n-th excited state,
4πn2 h̄2 s
an = − ,
m∗ e2
is much larger than that of the H atom, being typically of the order of tens of nanometers. This justifies the
use of the macroscopic dielectric constant, since the orbit of the electrons extends over so many lattice cells
that we can approximate the crystal as a homogeneous macroscopic medium.
• Density of States.
We now deal with the important concept of ‘density of states’, which is of utmost importance in studying the
way electrons (holes) populate the CB (VB). First we must consider the density of states in momentum-space
(or k-space). Then, we translate this into density of available states per unit energy. We shall then deal with
∂ 2ψ ∂ 2ψ ∂ 2ψ 2
+ + + k ψ = 0,
∂x2 ∂y 2 ∂z 2
with k2 = 2mE/h̄2 .
Let’s separate the problem: We look for a solution of the form ψ(x, y, z) = ψx (x)ψy (y)ψz (z). Substituting
this into the Schrödinger equation and diving by ψ , one gets:
1 d2 ψx 1 d2ψy 1 d2 ψz 2
+ + + k = 0.
ψx dx2 ψy dy 2 ψz dz 2
1 d2 ψx 2
= some constant = − k .
ψx dx2 x
Thus, accounting for the boundary condition (wavefunction must vanish at the wall of the cube):
2 2 2 2
kx + ky + kz = k
nx π ny π nz π
kx = , ky = , kz = ,
L L L
2 L3
4πk dk .
4π3
(Recall Jean’s law n(ν) ∝ ν 2 ? Since for photons 2πν = ck, where c is the speed of light, the expression
above is just Jean’s law for the black-body spectrum).
We now interested in expressing the number of states with magnitude of k in the shell of magnitudes
between k and k + dk in terms of states with energy between E and E + De. To do that, we use
E(k) = h̄2 k2 /(2m∗ ) and change variable from k to E : Since k = (2m∗ E)1/2/h̄ and
h̄2 k m∗
dE = dk → kdk = dE
m∗ h̄2
we have
2 (2m∗ E)1/2 m∗ m∗ (2m∗ )1/2 E 1/2
k dk = k kdk = dE = dE ,
h̄ h̄2 h̄3
so that the number of states per unit volume (recall that L3 is the volume of the crystal) with energy between
E and E + dE , ρ(E), will be:
m∗ (2m∗ E)1/2
ρ(E) dE = dE ,
h̄3 π2
md (2mdE)1/2
ρ(E) = .
h̄3 π2
Consider a large (that is, at the human, macroscopic scale) volume of gas. This volume will contain a very large
number of molecules, N ≈ 1023 . Clearly, we cannot even think of using Newton’s laws to describe the gas:
We would have to know the position and velocity of each molecule of the gas at some initial time, and solve the
huge number N of equations of motion, accounting for the forces among the particles. Even if we could solve
the all of these equations, it would be next to impossible to know all of the initial conditions required to get the
full solutions. Yet, we can describe the gas if we are willing to be content with information only about some
global, averaged quantities: We do not require the knowledge of all positions and velocities of all molecules
at all times... we would not know what to do with this knowledge anyway... Instead, we ask for information
about the volume and pressure of the gas given its temperature. This is the information provided by the laws of
Thermodynamics.
Statistical Mechanics tackles the problem differently in a different, intermediate way in an atempt to bridge
the gap between Newton’s laws and Thermodynamics: We shall still use Newton’s laws, but in an averaged,
statistical fashion.
In our example, we do not require to have the precise knowledge of all coordinates and velocities. Rather, we
shall be satisfied with some ‘statistical’ knowledge, like the average kinetic energy of a particle, or the probability
that it will have a particular velocity.
Let’s consider again our gas of N molecules. Let’s assume that is contained in a volume V in thermal contact
with a much larger system, so large that its temperature can be assumed not to change appreciably when it
exchanges energy with our gas. Let’s also assume that our gas is at some very high temperature, having been
so prepared initially. We ask how the gas will be described when it reaches thermal equilibrium with the large
reservoir. Simple (?) considerations tell us that the gas will evolve in time towards equilibrium by moving
towards the most probable distribution of molecules among all possible energy states under the constraints that
1. the number of molecules in the gas does not change, and 2. the total energy (gas + reservoir) remains
where A is some normalization (so that the total probability adds to unity, we shall not worry here about its
value), T is the (absolute) temperature of the reservoir, and kB = 1.38 × 10−23 J/K is called the Boltzmann
constant. The function ρ(E) is the density of states we found before and we shall discuss its appearance shortly.
We have given the exponential function its own symbol, fM B = exp[−E/(kB T )], for later convenience.
Now note that the temperature is a measure of how likely it will be for a molecule to have a particular value
of energy E : The higher the temperature, the more likely it will be for the molecule to be found at energy
E . Clearly, PM B (E) can also be interpreted as the ‘energy distribution’ of the molecules in the gas: When
properly normalized, N PM B (E)dE will tell us how many molecules, on average, will be found in the energy
interval between E and E + dE . At zero temperature all molecules will be at the lowest possible energy state
E = 0. As T grows, the distribution becomes ‘flatter’, more and more energy states becoming significantly
populated. Equation (48) is known as the Maxwell-Boltzmann distribution and it applies to classical particles
only, as we shall see below.
Why does the DOS ρ(E) appear in Eq. (48)? The best to way to see it is to re-express E in terms of the
velocity v and then count all particles per unit volume in the gas. We can simply look at all velocities, multiply
each entry in our tally by the probability that a particle will be there, and add them all up to obtain the particle
density n = N/V . Since we shall deal mostly with wavevectors and quantum mechanical particles in the
Let’s go to polar coordinates in k space as we did above, change integration variable from k to E , and we get:
21/2 m∗3/2 E E
n = A dE E 1/2 exp − = A dE ρ(E) exp − . (50)
π2 h̄3 kB T kB T
This shows that the probability distribution must indeed contain the DOS.
Finally, let’s compute the average energy of each particle in the gas: By definition, the average energy will be
the average of E weighted by the probability function ρ(E)fM B (E):
∞
∞
const 0 dE E ρ(E) fM B (E) 0 dE E 3/2 exp[−E/(kB T )]
E =
∞ =
∞ , (51)
const 0 dE ρ(E) fM B (E) dE E 1/2 exp[−E/(k T )]
0 B
where ‘const’ is some normalization constant which cancels out from the calculation. Let’s change integration
variable by setting x2 = E/(kB T ), so that dE = 2kB T xdx and E = x2 kB T :
∞ 4 −x2
0 dx x e (3/8)π1/2 3
E = kB T
∞ = kB T = kB T . (52)
dx x2 e−x2 (1/4)π 1/2 2
0
(The integrals can be calculated with some trickery... Ask if you want to know... don’t ask what you do not want
to know...) This is a particular instance of a general result known as the principle of ‘equipartition’: At thermal
is called ‘Fermi-Dirac’ distribution. The energy EF is called Fermi energy (or ‘Fermi level’, or often ’chemical
potential’). It is a measure of the density of particles: Suppose we are at zero T and we must fill our ‘quantum
box’ with 4 particles. Two particles will go into the first energy state, E1 , two more into the second, E2 , by
Pauli’s principle. The Fermi energy will by slightly abobe E2 , since at T = 0 FF D is a step function equal to
1 when E < EF , equal to zero otherwise.
For those interested, below we see how one can use Pauli’s principle and the indistinguishability of electrons to
derive the Fermi-Dirac distribution.
– Let’s consider how a system of Ne ‘indistinguishable’ electrons at thermal equilibrium with the environment will be distributed over
N energy levels, each with energy Ei and containing gi (i = 1, ..., N ) available states labeled by different quantum numbers, by
N
placing ni electrons in each energy level i. The total energy of the system is Etot = i=1 Ei ni .
The number gi is called the ‘degeneracy’ of the level i.
Pauli’s principle tells us that we cannot put more than one electron in each state, so ni ≤ gi .
– The number Wi of different ways in which ni electrons can be placed into the gi available states in the ith level is:
gi !
Wi = .
(gi − ni )!ni !
(Think of gi bins in which we can put a ball or leave it empty, having ni balls in total. We can put the first ball in any of the gi
bins, the second ball in any of the remaining gi − 1 bins, etc., for a total of gi ! possible ways. But, since neither the balls nor the
’voids’ can be distinguished, we must divide this number by the ni ! equivalent ways we can place the ni balls into the bins and by
the (gi − ni )! equivalent ways we can arrange the gi − ni voids.)
– Thus, the total number of possible arrangements is just
gi !
W = Wi = .
(gi − ni )!ni !
i i
– The most likely distribution should be determined by maximizing W under the constraints of energy and particle conservation. It’s
ln x! ≈ x ln x − x ,
and write
ln W ≈ [ gi ln gi − gi − (gi − ni ) ln(gi − ni ) + (gi − ni ) − ni ln ni + ni ] =
i
= [ gi ln gi − (gi − ni ) ln(gi − ni ) − ni ln ni ] .
i
Then
∂ ln W
d(ln W ) = dni ≈ [ ln(gi − ni ) + 1 − ln ni − 1 ] dni = ln(gi /ni − 1) dni .
∂ni
i i i
now comes from solving d(ln W ) = 0 subject to the conditions
Our solution i dni = 0 (the total number of particles cannot
change) and E dn
i i i = 0 (the total energy should not change). This requires the use of the ‘method of Lagrange multipliers’:
We multiply each equation describing a constraint by a constant (a ‘multiplier’) and add these equations to the original equation:
[ ln(gi /ni − 1) − α − βEi ] dni = 0 .
i
Here the Lagrange multiplier α refers to particle conservation, the multiplier β to energy conservation. Since this must be valid for
each i, a solution will be given by:
ln(gi /ni − 1) − α − βEi = 0 for all i.
Therefore, the density of electrons in the ith energy level Ei will be:
ni 1
f (Ei ) = = .
gi 1 + eα+βEi
Here we should recall that the Fermi energy (or ‘Fermi level’, or the ‘chemical potential’ of the grand-canonical
ensemble), the total electron density, and the temperature are related, as we’ll see below.
– Some useful approximations to the Fermi-Dirac distribution are the following:
θ(E − EF ) for kB T << EF
fF D (E) ≈ ,
e(E−EF )/(kB T ) for kB T >> EF (Maxwell-Boltzmann distribution)
where θ(x) is the Heavyside step function, θ(x) = 1 for x < 0, θ(x) = 0 otherwise.
• The electron (hole) density in the CB (VB).
Let’s now use the Fermi-Dirac distribution to calculate how electrons and holes are distributed energetically in
the CB and VB of semiconductors.
– The total density of free electrons in the lowest-energy CB of a semiconductor will be given by:
1 2
n = fF D [E(k)] = fF D [E(k)] dk .
V (2π)3
k
For an isotropic effective mass m∗ , E(k) = h̄2 k2 /(2m∗ ), so that k = (2m∗E)1/2 /h̄ and dE =
where kF = (2m∗ EF )1/2 /h̄ is the Fermi wavevector. Note that the Fermi energy is positive. This means
that the Fermi level is above the minimum of the CB for electrons (or below the maximum of the VB for
holes).
2. High-temperature, non-degenerate limit. On the contrary, at high temperature we have the so-called
non-degenerate limit in which the FD-distribution approaches the form of the MB distribution (so that this
is also called the ‘semiclassical limit’): fF D (E) ≈ e−(E−EF )/(kB T ) and
3/2 ∞
1 2m∗n
nhigh T ≈ eEF /(kB T ) dE E 1/2 e−E/(kB T ) =
2π2 h̄2 0
3/2 ∞
1 2m∗nkB T E /(kB T ) 1/2 −x
= e F dx x e ,
2π2 h̄2 0
where in the last step we have used explicitly the fact that we have always measured EF from the
bottom of the CB. In the case of the ‘ellipsoidal’ CB of Si we should replace the isotropic mass m∗n with
mn,d = 62/3 (mL m2T )1/3 , the factor of 6 stemming from the 6-fold degeneracy of the CB minima. For
holes we get, similarly,
3/2
1 2m∗p kB T (−EG −EF )/(kB T ) −(EF −Ev )/(kB T )
phigh T ≈ e = Nv e ,
4 πh̄2
3
1 kB T ∗ 3/2 −EG /(kB T ) 2
pn = (mn,d mp ) e = ni ,
2 πh̄2
3/2
3/2
1 2m∗n kB T 1 2m∗p kB T
NC = and NV = ,
4 πh̄2 4 πh̄2
are called effective density of states for electrons and holes, respectively. Other useful forms are:
where Ei is the Fermi level in undoped (intrinsic) Si, which we shall derive shortly.
3. General case.
The general case of temperature not too high or not too low is mathematically a little more complicated,
because there is no analytical solution to the integral. However, the result is widely available in the literature
and it is expressed in terms of what are called ‘Fermi integrals’. In general we have:
3/2 ∞ 3/2 ∞
1 2m∗ 1/2 1 2m∗kB T 2 x1/2
n = dE E f [E(k)] = √ dx ,
2π2 h̄2 0 4 πh̄2 π 0 1 + ex−η
where, as before, we have changed the integration variable to x = E/(kB T ), we have set η = EF /(kB T ),
√
and we have multiplied and divided by 2/ π to recover the prefactor Nc introduced above. As just stated,
the integral cannot be solved in closed form, but it is a well-know integral (appropriately called Fermi-Dirac
integral of order 1/2 and labeled by F1/2 (η)). So, the electron density is usually written in terms of this
integral as:
3/2
1 2m∗ kB T
n = F1/2 (η) = Nc F1/2 (η) .
4 πh̄2
– Determining the Fermi level.
Two more cases are extremely important: How to determine the Fermi level in an intrinsic (i.e., undoped)
semiconductor and in a doped material.
so
kB T Nc
EF = Ei = Emg − ln ,
2 Nv
where we have defined the ‘mid-gap’ Emg = (Ev + Ec )/2. Since NC /NV = (m∗n /m∗p )3/2 is not
too different from unity, the log term above is usually negligible and we can almost always approximate
Ei ≈ Emg .
2. Fermi level in extrinsic (doped) Si: Use again the charge neutrality condition which now reads:
+
n = p + ND ,
+
for an n-type semiconductor. Here ND is the density of ionized donors. The probability that an electron
will occupy the energy level ED associated to the H-like impurity potential will be
1
f (ED ) = ,
1 + gD e(ED −EF )/(kB T )
where gD is a degeneracy factor dependent on the microscopic structure of the impurity, typically taken as
1/2 since only one electron can occupy the impurity state while we have derived the Fermi function assuming
+
2 electrons with opposite spins per state. Therefore: ND = ND [1 − f (ED )] ≈ ND (in words: dopants
are almost always ionized, except at very low temperatures) and:
NC ND
n ≈ p + ND → EF ≈ EC − kB T ln = Ei + kB T ln .
ND ni
Let’s summarize these results. In an intrinsic semiconductor the Fermi level is essentially at midgap,
EF ≈ Ei ≈ Emg . If we dope it n-type, EF moves up towardes the minimum of the CB and its distance
from Ei – usually denoted with φB – will be:
ND
EF − Ei = φB = kB T ln .
ni
As ND grows we reach a point at which EF enters the CB. The semiconductor at this point becomes
degenerate and the MB-approximation we have employed for fF D (E) ceases to be valid. A correct analysis
should make use of the Fermi-Dirac integrals. Similarly, if we dope the semiconductor p-type, the Fermi level
moves ‘down’ towards the top of the VB. Its distance from the intrinsic Fermi level Ei will be
NA
EF − Ei = −φB = − kB T ln .
ni
The figures in the following page show the position of the Fermi level for both electrons and holes at two
temperatures in Si.
This would imply an endless acceleration. However, electrons in a crystal will suffer collisions (with the charged
ionized dopants, with lattice vibrations, with other electrons, with interfaces of the device, etc.). Without
going into details of what happens during these collisions, let’s take a statistical view and assume that on the
average the electrons are going to collide once every τ seconds. Let’s also assume that at each collision each
electron loses all of its momentum. Then, the rate at which momentum is lost by the ‘swarm’ of electrons will
be:
dPx Px
= − . (58)
dt collisions τ
[It is not really necessary to assume total loss of momentum at each collision. Rather, the equation above is valid provided 1/τ
is re-interpreted as the ‘momentum relaxation rate’ (that is, the rate at which momentum is lost by the ‘swarm’) rather than the
collision rate]
At steady state with the field, the momentum lost via collisions must equal and oppostite to the momentum
gained via the acceleration from the field. Therefore, equating Eq. (57) to Eq. (58):
dPx dPx
− = , (59)
dt f ield dt collisions
e2 τ n
Jx = −envx = Fx . (62)
m∗
The quantity µn = eτ /m∗ is called ’electron mobility’. It is measured in cm2 /Vs and it is the constant of
proportionality between the electric field and the average electron drift velocity:
vx = µn Fx . (63)
Note that Eq. (62) is simply Ohm’s law: The current is proportional to the applied bias: If we have a chunk
of semiconductor with cross-sectional area A and length L and we apply a voltage V between two ends of
the ‘chunk’, as shown in the figure below, the current I will be I = Jx A and the field along the x-direction
will be Fx = V /L. Thus, Eq. (62) can be rewritten as:
Ae2 τ n V V
I = Jx A = = A e n µ n , (64)
m∗ L L
or, defining the resistance R = L/(enµn A),
V = IR , (65)
L
d
y
W
x
– Resistance, resistivity, conductivity, etc.
Looking again at the figure illustrating our ‘chunk’ of semiconductor, we should define several quantities which
we shall need in the following.
∗ Mobility. We have already seen that the mobility is the constant of proportionality between electric field
and electron velocity. In vector form, vd = µn F, where the subscript d refers to the fact that we mean
the ‘drift’ velocity of the ‘swarm’ and we have dropped the ‘average symbols‘ ... This is an ‘intensive’
property of the material; that is, it does not depend on the shape or dimensions of the sample, but only on
its intrinsic properties.
∗ Conductivity. The conductivity σ is another intensive property and it is the proportionality constant between
electric field and current:
J = σF. (66)
1
ρ = ≈ 0.4 Ωcm . (68)
1.602 × 1019 × 1, 500 × 1016
∗ Resistance. As we have seen before, after Eq. (64), the resistance is an ‘extensive’ quantity. This means
that is depends on the shape and dimension of the sample. For the geometry of the figure,
L
R = , (69)
enµn A
so that the resistance drops as the cross-sectional area increases (as in wires of different diameters or ‘gauge’)
and increases as the length of the sample increases.
∗ Sheet resistance. Finally, the sheet resistance is the defined as the resistance of a square of sample of
thickness d. From Eq. (69), setting W = L, so that A = Ld,
L 1 ρ
Rs = = = (70)
enµn Ld enµn d d
The sheet resistance is measured in ohm/square (Ω/✷) and it depends on the thickness of the sample.
– DOS and Conductivity Effective Masses.
The effective mass m∗ which appears in the expression for the mobility, µn = eτ /m∗ , is called the
where kL is the component of the wavevector along the longitudinal principal axis of the ‘football’, and
kT 1 and kT the 2 components perpendicular to it. So, considering the field along the x direction, for the
‘footballs’ with the longitudinal axis along the x-direction the mass relevant to the mobility will be mL , but
for the other 4 footballs it will be the transverse mass mT . Since at equilibrium 1/6 of the electron occupy
each football, the inverse mass 1/m∗ averaged over all footballs (we shall call them more properly ‘ellipsoids’
from now on!) will be:
1 1 1 2
= + . (72)
mc 3 mL mT
In a way, the conductivity mass is obtained by taking an average of the inverse masses along the three principal
directions and taking the inverse of the results. From the way we have defined it, the conductivity mass
describes the inertia of the electron, and so its acceleration and velocity under an external field. On the
contrary, the DOS mass describes the density of available states at a given energy, as we saw on page 54. In
Si the two masses differ numerically: Using mL = 0.91 (in units of the free electron mass) and mT = 0.19,
we have mc ≈ 0.26 while md = (mLm2T )1/3 ≈ 0.32.
In more general situations, the conductivity mass can be related directly to the electron group velocity, and
so to the inverse of the slope of the dispersion E(k) along the direction of the field. It also often called
the ‘optical mass’, since it is the mass which appears in optical transitions (excitation of electrons to higher
energies due to absorption of light).
vd = µn F .
The physical interpretation is obvious: If the scattering time τ is constant, then applying a field twice as
strong will result in the ‘swarm’ of elecrons to drift at a velocity twice as high.
However, as the field gets larger and larger some additional effects come into play. To get an idea of what
happens, we should make two observations:
1. First, note that the scattering time is an average of the scattering times, τ (E), of all electrons at all energies
in the swarm: By the usual expression of ‘weighted averages’,
In general, the scattering ‘frequency’ (usually called the ‘scattering rate’) 1/τ (E) is a function of the
electron kinetic energy. For example, scattering with ionized impurities weakens as E grows because the
depth of the Coulomb potential appears to decrease as the electron energy increases. On the contrary, the
frequency for scattering with phonons (and many other processes) increases like ρ(E). This is a general
principle: The more ‘final states’ are available after a collision, the easier it is to scatter and so the larger is
the scattering rate.
(Think of parallel-parking your car: It takes significantly longer to park – so, a lower ‘parking frequency’ – than to exit the parking
space – and so higher ‘deparking’ frequency . This is because when parallel-parking there are a small number of ‘final spaces’ (and
so, ‘states’) available (not too close to the curb, not too close to the car in front, not too far away from the curb, not to close
to the car behind, etc.) Instead, when exiting the parking space, one van afford to be sloppier, since there are many more ‘final
spaces’ available.
So, if electrons for some reason begin to be distributed with a probability f (E) different from the equilibrium
Note that both electrons and holes contribute with the same sign: Holes are moving in the direction of positive
field and contribute with a plus-sign to the current, as they are positively charged. Electrons move in the
opposite direction of negative field, but contribute to the current with a negative sign.
Similarly,
rate right-to-left = ρ2 (E) f2 (E) ρ1 (E) [1 − f1 (E)] . (80)
At equilibrium, since there can be no transfer of energy across the discontinuity, these two rates of fluxes at a
given energy must be equal (if they were not, electrons and energy would keep flowing one direction or other,
and we would not be at equilibrium). So:
ρ1 (E) f1 (E) ρ2 (E) − ρ1 (E) f1 (E) ρ2 (E) f2 (E)] = ρ2 (E) f2 (E) ρ1 (E) − ρ2 (E) f2 (E) ρ1 (E) f1 (E) .
(82)
The second term on the rhs cancels with the second term on lhs and we have the final result at equilibrium
f1 (E) = f2 (E), which is what we had ‘guess’ before. We conclude with the most important result that at
equilibrium the Fermi level is constant everywhere. In formula in a one-dimensional system inhomogeneous in
the x direction:
dEF (x)
=0. (83)
dx
Note that the balance of fluxes at a given energy we have used in Eq. (81) is called ‘detailed balance’.
• Recombination processes
Having created e-h pairs (thermally, with light, electron or ion bombardment, etc.), they will tend to ‘recombine’:
Whenever an electron and a hole paths cross, electron can fall into the attractive potential of the hole and
’annihilate’ itself and the hole in the VB. These ‘recombination’ processes can be broadly divided into two
classes: ’Radiative’ recombination, in which the e-h energy (typically, the bandgap Eg ) is released in the form
of light, or ‘non-radiative’, in which the excess energy is transferred to the lattice in the form of extra ionic
agitation. Recombination processes can also be assisted by traps or defect: The electron first falls into a trap
with energy-level inside the gap. A passing hole can then recombine with the trapped electron.
Regardless of the nature of the recombination processes, we can describe them with a ‘recombination rate’ r :
Since at equilibrium the electron and hole concentrations n0 and p0 (the subscript ‘0‘ refers to equilibrium
values) are such that n0 p0 = n2i , recombination processes will attempt to return the ‘perturbed’ concentration
n and p to their equilibrium values, so that the recombination rate must have the form:
r = αr (np − n2i ) .
Indeed at equilibrium np = n2i so that r = 0. The ‘recombination constant’ αr depends on the microscopic
nature of process (whether recombination is radiative or not, on the nature of the trap, if trap-assisted, etc.).
Let’s assume that we have generated an excess concentration of electrons and holes, by shining light on the
sample. Let n(t) and p(t) be the concentrations of electrons and holes as function of time. If at time t = 0
we turn off the light, the excess carriers will recombine and their concentration will decrease according to the
equation:
dn(t) 2
= −r = αr ( ni − np) .
dt
Let’s assume that n(t) = n0 + δn(t), p(t) = p0 + δp(t). Since we have generated pairs, δn = δp.
dn(t) dδn 2 2
= = αr [ ni − (n0 + δn)(p0 + δp)] ≈ αr [ ni − n0 p0 − (n0 + p0 )δn] ,
dt dt
having ignored terms of the order of (δn)2 . Since n0 p0 = n2i , the first and second terms inside the square
bracket cancel. Since the sample was assumed to be p-type, p0 >> n0 and we can ignore n0 . Thus:
dδn(t) δn(t)
≈ − αr p0 δn(t) = − , (84)
dt τn
where τn = 1/(αr p0 ) is the ‘electron lifetime’. This equation has the solution δn(t) = δn(0) exp(−t/τn ):
The concentration of excess electrons decays exponentially from its initial value δn(0) with the time constant
τn .
Suppose that now we keep shining light on the sample and that gop is the rate at which we generate e-h pairs
(measured in pairs per unit volume and unit time, so in m−3 s−1 ). Then recombination process will attempt to
reduce the concentration of excess electrons with a term similar to Eq. (84). However at the same time light
will continue to generate excess with a rate gop . Thus the rate at which the electron density changes will be
determined by the competition of these two processes:
dδn(t) δn(t)
≈ − + gop . (85)
dt τn
At steady state we must reach a constant concentration of electrons, so that δn/dt = 0 and δn = gop τn .
Similarly, we will have a steady-state concentration of excess holes, δp = gop τp .
A way to describe the concentration of excess carriers using the same mathematical machinery used at equilibrium
is based on the introduction of the concept of ‘quasi-Fermi level’. We saw that at equilibrium we can use
Off-equilibrium we now define the ‘quasi-Fermi levels’ for electrons, EF n , and holes, EF p , such that:
EF n − Ei Ei − EF p
n = ni exp p = ni exp . (87)
kB T kB T
Occasionally, with noticeable lack of good taste, EF n and EF p are called ‘imref’ (‘fermi’ spelled backwards).
We can view this as a simple mathematical change of variable, but it is better to view it as a convenient way
to keep using equilibrium concepts also off equilibrium. As long as we recall that EF n and EF p are NOT real
Fermi levels and are NOT well-defined from the point of view of Statistical Mechanics, these concepts are useful.
• Diffusion
When excess carriers are excited across the gap, they are often generated in a spatially inhomogeneous way
(that is, their concentration varies depending on position). For example, when we generate carriers with light,
they will be generated in high concentration only at the spot where the light beam strikes the sample. Clearly,
such an inhomogeneous spatial distribution cannot last: The random motion of the the electrons – hitting each
other, hit by thermally agitated ions, etc. – will randomly distribute them throughtout the entire volume. Such
a process, which occurs very frequently in nature, is called ‘diffusion’. The textbook draws the nice analogy with
molecules of perfume in a room. At first very localized within the small volume around the source of perfume
(the bottle or the area were we spray the perfume), with time their scent will ‘diffuse’ thorughout the entire
room. This is due to the random nature of the collision-dominated motion of the molecules: At the boundary
between the perfume-containing volume and the rest of the room, at a given time each molecule will have, on
average – a 50% probability of being scattered outside the volume, a 50% chance of being scattered back into
the perfume-containing volume. So, on average, half of the molecules will leave the original volume. With time,
the perfume-containing volume expands while, at the same time, the concentration of perfume molecules in the
original volume decreases. The process continues untill the entire volume is filled with a uniform concentration of
n(x)
n(x) n1
n(x)
t1 n2
t2 > t 1
t3 > t 2
x x
0 x x0–L x0 x0+L
To see how we can describe this process mathematically, let’s consider a simpler one-dimensional situation
illustrated in the figures above. The first figure shows how an ensemble of particles originally (t = 0) at x = 0
spreads as time increases. Let’s now consider one of those ’bell-shaped’ curves (they are Gaussians, as we shall
see later) and divide them into the indicated rectangles as follows. Let’s assume that electrons collide with a
mean free path L and a scattering time (or ‘mean free time’) τ (as we have seen discussing the mobility).
Therefore, in any given time step τ , electrons can move with equal probability to the left or to the right by an
amount L. Let’s now divide the x-axis into segments of length L. In each segment there will be a number
nj of electrons, where j is an index labeling the segment. Let’s now consider the final figure showing a pair
of adjacent segments, labeled 1 and 2, on the two sides of the point x0 , containing n1 and n2 electrons
respectively. In a time τ the number of electrons leaving the ‘bin’ number 1 will be n1 AL, where A is the
area of the bin in the y and z directions. Half of them will move to the left, and only a half will move to bin
2. With a similar argument applied to bin number 2, the total net number of electrons going across the point
x0 will be n1 AL/2 − n2 AL/2 = (AL/2)(n1 − n2 ). In the limit in which scattering is very frequent, so
L2 dn(x0 )
Φ(x0 ) = − . (88)
2τ dx
The quantity L2 /(2τ ) is denoted by Dn and is called the electron diffusion constant. Therefore, since
electrons carry negative charge, the current density due to the diffusion process will be:
dn(x)
jn,dif f (x) = e Dn . (89)
dx
Similarly for holes:
dp(x)
jp,dif f (x) = − e Dp . (90)
dx
Note the different sign: Carriers always diffuse in the direction of negative spatial derivative of the concentration
(i.e., decreasing concentration). This ‘minus’ sign is compensated by the negative electron charge, but it is still
present when dealing with holes.
• The drift-diffusion equations
Combining Eqns. (89) and (90) with Eqns. (62) and (78), we finally have the complete expressions for the
electron and hole current densities due to both the drift and the diffusion processes:
dn(x)
jn (x) = eµn n(x)Fx (x) + eDn , (91)
dx
F(x)
n(x) jn(diff)
jn(drift)
φn(diff)
φn(drift)
• Einstein’s relation.
Let’s consider an inhomogeneous situation in which the semiconductor is doped non-uniformly. Of course, this
implies that at equilibrium there will be a gradient of the carrier concentration. From what we have seen above,
a gradient of the carrier concentration will induce a ‘diffusion current’. But since at equilibium we cannot have
current flow, the diffusion component of the current must be compensated exactly by a drift current caused
which is called ‘Einstein’s relation’. It a very important equation which allows us to calculate µ knowing D and
viceversa. That such a relation should exist is obvious from the fact that the ‘mean free time’ (or relaxation
time) τ enters both the definition of the mobility, µn = eτ /m∗ , as well as that of the diffusion constant,
Dn = L2n /(2τ ). Since the meanfree path Ln ∝ τ , then Dn ∝ τ , like the mobility, so that the two quantities
must be related.
• The continuity and diffusion equations.
In the DDEs above, Eqns. (91) and (92), we have ignored recombination effects. These matter, since they
induce a spatial variation of the carrier concentration and, thus, can affect the diffusion current.
In order to see how we can account mathematically for recombination effects, let’s consider a semiconductor
∂p 1 jp (x) − jp (x + ∆x) δp
= − . (97)
∂t e ∆x τp
(Note the use of partial derivatives because the unknown function, δp, depends both on time t and on position
x). In the limit of ∆x → 0, we can replace the first term on the right-hand side with the derivative
−∂jp (x)/∂x, so that the equation above becomes:
∂p ∂δp 1 ∂jp δp
= = − − . (98)
∂t ∂t e ∂x τp
This is called the continuity equation for holes. Similarly we have for electrons:
∂δn 1 ∂jn δn
= − . (99)
∂t e ∂x τn
d2 δp δp δp
= = , (102)
dx2 Dp τp L2p
d2 δn δn δn
= = , (103)
dx2 Dn τp L2n
√
where Ln = Dn τn and Lp = Dp τp are the diffusion lengths of electrons and holes respectively.
These represent the distance over which electrons and holes diffuse before recombining, as we shall see below.
The general solution of Eq. (102) is:
x/Ln −x/Lp
δp(x) = C1 e + C2 e , (104)
where the integration constants C1 and C2 must be fixed using the required boundary conditions. Since
infinitely away, deep in the sample, we expect vanishing density of excess carriers, C1 must be set to zero.
(Note that if this were not the case, we would have a density of holes growing without limit as we move deep
inside the sample. This is clearly a physically unacceptable solution.) If we know that ∆p is the steady-state
−x/Lp
δp(x) = ∆p e . (105)
As we said above, Lp is the average distance over which the hole will diffuse before recombining. To see this,
let’s go back once more to the concept of ‘weighted average’. We want to compute the average distance at
which holes will be found, weighted by their concentration which is controlled by recombination. Since holes
are generated only at the surface of the sample, this will tell us exactly what we want: How far, on average,
holes will diffuse from the surface inside the sample before disappearing because of recombination. So, the
average position of the holes weighted by their concentration will be:
∞
∞
∞
−y
dx x δp(x) L2 ∞
dy y e−y −ye + 0 dy e−y
0 ∞
0 ∞
0∞
p
= = Lp = Lp , (106)
dx δp(x) L dy e−y dy e −y
0 p 0 0
where we have changed dummy integration variable y = x/Lp . This is the result we were after. Finally, if
we want to calculate the steady-state diffusion current flowing ‘vertically’ through the device, we can obtained
it from Eq. (92) in the absence of electric field:
dp(x) p(x)
jp (x) = − eDp = eDp . (107)
dx Lp
vd L2
µp = = .
F V td
This reasoning applies exactly only if the hole ‘packet’ drifts while retaining its original narrow shape. However,
diffusion prevents this. Instead, the packet spreads while shifting. Ignoring for now the drift component and
recombination processes, the equation controlling the spreading of the packet is just the diffusion equation:
∂δp ∂ 2p
= Dp , (108)
∂t ∂x2
which has the solution:
δP 2
δp(x) = e−x /(4Dp t) , (109)
2 πDp t
dn(x)
jn (x) = eµn n(x)Fx (x) + eDn . (111)
dx
Now, let’s rewrite the gradient of the electron concentration as:
dn(x) d (En −Ei )/(kB T )
n(x) dEn dEi
= ni e = − . (112)
dx dx kB T dx dx
dEp dEp
jp (x) = µp p(x) = σp (x) , (116)
dx dx
where σn and σp are the position-dependent electron and hole conductivities.
• Summary of the semiconductor equations.
Let’s summarize the basic semiconductor equations which we need in order to calculate the current flowing in a
semiconductor device:
1. The Drift-Diffusion Equations for electrons and holes. They tell us how the charge carriers move under
the action of the external field and by diffusion. In their general three-dimensional version (recall that bold
characters reprent vectors and that ∇ = (∂/∂x, ∂/∂y, ∂/∂z) is the ‘gradient’ differential operator), they
are:
jn (r, t) = eµn n(r, t)F(r, t) + eDn ∇n(r, t) , (117)
and:
jp (r, t) = eµp p(r, t)F(r, t) − eDp ∇p(r, t) . (118)
∂δn(r, t) 1 δn(r, t)
= ∇ · jn (r, t) − + gn (r, t) , (119)
∂t e τn
∂δp(r, t) 1 δp(r, t)
= − ∇ · jp (r, t) − + gp (r, t) . (120)
∂t e τp
3. Poisson Equation, finally, allows us to calculate the electric field F(r) appearing in Eqns. (117) and (118)
once we know the boundary conditions (that is, the voltages applied at the edges of the sample) and the
+
spatial distribution of the electrons and holes, n(r) and p(r), and of the ionized impurities, ND (r) and
−
NA ( r ) :
+ −
− ∇ · [∇φ(r, t)] = e [ ND (r) + p(r, t) − NA (r) − n(r, t) ] , (121)
with
F(r, t) = − ∇φ(r, t) . (122)
Note that we use only Poisson equation (so, electrostatics) and not the full set of Maxwell’s equations (which
gives the electrodynamics of the system), because we are tacitly assuming that all temporal variations are very
slow.
Note that these equations are coupled: The electric field depends on the location of the charges. But the charge
carried by electrons and holes moves under the action of the field itself. Therefore, in most cases, we must solve
them self-consistently. This is often a rather complicated mathematical task. In the following, when studying
electronic devices, we shall make some approximations to simplify this problem.