Professional Documents
Culture Documents
AND
PHYSICS
02/10/2017, 14:55
Fisica Teorica – 2012/2013
N O TA
Queste note sono pensate come supporto didattico per il corso di Fisica Teorica
(Parte A) del Corso di Laurea Magistrale in Fisica dell’Università degli Studi di
Padova. E’ vietata la loro distribuzione non autorizzata.
CONTENTS
i introduction 1
1 symmetries in physics 3
1.1 Introduction 4
1.2 Symmetries and Groups 4
1.3 Examples of Symmetries 5
1.3.1 Symmetry of Geometric Objects 5
1.3.2 Spacetime Symmetries 6
1.3.3 Gauge Symmetries 7
1.3.4 Approximate Symmetries 9
1.4 Symmetry Breaking 10
2 principle of least action 13
2.1 Euler–Lagrange Equations 13
2.2 Examples 16
2.2.1 Free relativistic particle 16
2.2.2 Electromagnetic field and its coupling to a charged parti-
cle 18
2.3 Noether’s Theorem 19
ii finite groups 23
3 groups: basic elements 25
3.1 Definitions 25
3.2 Maps between groups 27
3.3 Finite groups 29
3.3.1 The Symmetric Group Sn 31
4 subgroups, quotients and products 37
4.1 Basic definitions 37
4.2 Quotients 38
4.3 Products 42
4.4 Classifying finite groups. 44
5 representations 47
5.1 Basic definitions 48
5.2 Equivalent and reducible representations 51
i
ii Contents
INTRODUCTION
1
SYMMETRIES IN PHYSICS
2012 has been a remarkable year for physics. The discovery of the Higgs boson-
like particle at the Large Hadron Collider in Geneva provides a spectacular
confirmation of our understanding of the fundamental interactions governing
our Universe, embodied in the Standard Model of elementary particles. This
discovery marks the pinnacle of the journey started with the unification of elec-
tricity and magnetism in terms of a single symmetry principle and completed
with the further unification of the weak nuclear force, via the electro-weak sym-
metry breaking mechanism.
Symmetry principles guide theoretical physics since its birth. Geometric
symmetries can explain many physical properties of materials and character-
ize diffraction patterns found in crystallography. Dynamical symmetries are
widely used in Nuclear Physics to classify the energy of rotational and vibra-
tional modes of nuclei. Symmetry properties of spacetime itself and symme-
try transformations relating different observers are at the base of Special and
General Relativity. Broken symmetries are used to explain a wide variety of
physical phenomena, from Superconductivity to electro-weak interactions.
In this lectures we want to explore the concept of symmetry in physics. In par-
ticular we want to give a theoretical description of this concept, by construct-
ing mathematical models of the symmetries appearing in physical systems. We
are mainly going to focus on the applications of symmetry principles in the
context of Quantum Mechanics, but we are also going to discuss spacetime
symmetries, which are at the base of Relativity and Quantum Field Theory,
and internal symmetries (global and local gauge symmetries), which guide our
current understanding of the physics of fundamental interactions.
The objective is to provide students with instruments and techniques that
are useful to extract information on the physics of a definite system by using
3
4 symmetries in physics
1.1 introduction
We will use a great part of these lectures to discuss group theory. We will do
so because symmetry transformations are described by groups.
Consider three different states of a physical system: A, B and C. If state A
is similar to state B and state B is similar to state C, then state A is similar
to state C. Hence, the product of two symmetry transformations is again a
1.3 examples of symmetries 5
id S1 id S2 id
A B C
S1 ◦ S2
For example, reflections with respect to a plane form a group that contains
two elements: the reflection operation and the identity. Rotations in three-
dimensional space, on the other hand, form a group that has an infinite number
of elements. Groups with a finite (or numerable) number of elements are called
discrete groups, while groups that depend on continuous parameters are called
continuous groups.
Symmetries of a system imply relations between observable quantities, which
may be obeyed with great precision, independently of the nature of the forces
acting in the system. For example, the energies of several different states of
the Hydrogen atom are exactly equal, as a consequence of the rotational invari-
ance of the system. In other instances symmetries of physical systems are only
approximately realized or broken. Still, an approximate or a broken symmetry
can teach us as many important lessons as an exact one, like in the case of
superconductivity or of the Higgs particle.
• spatial translations:
~r → ~r +~a; (1.1)
• time translations:
t → t + b; (1.2)
• constant rotations:
~r → R~r; (1.3)
1.3 examples of symmetries 7
~r → ~r − ~v t. (1.4)
The first two symmetries derive from space and time homogeneity, while the
third one is related to space isotropy. Altogether they form a 10-dimensional
continuous group called the Galilei group, which constrains newtonian mechan-
ics. In Special Relativity Galilean boosts are replaced by Lorentz boosts and
together with the other transformations form the Poincaré Group, which we
will discuss in detail in chapter 17.
This type of symmetries is extremely important because they severely restrict
the form that physical laws can take. Also, as we will soon discuss, continuous
symmetries give rise to conserved quantities via Noether’s theorem and in the
case of spacetime symmetries we can obtain energy-momentum and angular
momentum conservation laws.
Fµν = ∂µ Aν − ∂ν Aµ . (1.6)
A µ ( x ) → A µ ( x ) + ∂ µ Λ ( x ), (1.7)
but leave Fµν invariant and hence do not change the value of the electric and
magnetic fields. As we will now see, this function is actually specifying an
element of a group of symmetries of the Schrödinger equation for a charged
particle.
8 symmetries in physics
ψ( x ) → eiΛ( x) ψ( x ), (1.8)
∂ µ → Dµ = ∂ µ − i A µ . (1.9)
Dµ ψ 7→ eiΛ Dµ ψ. (1.10)
U = exp(iΛ( x )) (1.11)
x µ → ξ µ ( x ). (1.13)
1.3 examples of symmetries 9
We already discussed the fact that often approximate symmetries are useful to
constrain the theoretical description of physical systems. These are symmetries
valid in a certain regime of energy or length and may arise dynamically.
An example is given by crystals. An infinite crystal is invariant under trans-
lations for which the displacement is an integral multiple of the distance be-
tween two adjacent atoms. Real crystal, however, have a definite size and their
surface perturbs the symmetry under translations. Nevertheless, such pertur-
bations have little effects on the properties at the interior if the crystal contains
a sufficiently large number of atoms.
Another example of a symmetry that is only approximately realized is the
isospin symmetry. One of the baryons, called ∆+ , decays into a nucleon and a
pion in two ways:
∆+ → n + π + , or ∆+ → p+ + π 0 . (1.14)
Although the masses of the neutron and the proton are very close and the same
is true for the pions, so that the only distinctive element is the electric charge
distribution, the second process occurs twice as much as the first. The explana-
tion for this discrepancy comes from a symmetry argument. In fact nucleons
and pions sit in different representations of the isospin symmetry group. This
information is actually relevant for any symmetry principle. The way physical
quantities transform under symmetry transformations is not determined just
by the symmetry group, but also by its representation. For instance, we know
that we cannot sum mass and velocity as they are quantities that transform
differently under the group of rotations in 3-dimensional space. The mass is
an invariant quantity. It is a scalar under rotations and hence transforms in
the trivial representation. The velocity transforms as a vector and hence the
group of rotation is represented in its fundamental representation. Coming
back to the example above, the two decays mode differ because nucleons form
a doublet of the isospin symmetry, pions a triplet and ∆ particles a quadruplet
and hence they are constrained by the rules governing products of representa-
tions and their decomposition in irreducible components. We stress that this
argument is valid because particles inside the same representation have nearly
10 symmetries in physics
equal masses, but different electric charges. Symmetry arguments are therefore
valid in the approximation of equal masses, which we know is violated only by
few percents.
We conclude this discussion of approximate symmetries by recalling that
we have a very simple example of a discrete symmetry that is omnipresent in
physics: the reflection symmetry Z2 : x 7→ − x. Any degree of freedom can
be approximately described by the 1-dimensional harmonic oscillator when
looking at a minimum of its potential
1
V ( x ) ' V ( x0 ) + m2 δx2 + O(δx3 ) (1.15)
2
2
V ( z i ) = − µ2 Z † Z + λ Z † Z , (1.16)
z1
where λ > 0 and Z = . It is easy to check that such a potential is
z2
invariant under unitary transformations of the Z doublet
and therefore the resulting physics should keep trace of this fact. We can also
see that the potential (1.16) has two critical points, satisfying
∂i Z = −µ2 + 2λ| Z |2 zi∗ = 0, (1.18)
1.4 symmetry breaking 11
Figure 3.: The “mexican hat” potential (1.16). The maximum has U(2) symmetry, while
the family of minima have only a residual U(1).
On the other hand, the family of minima is invariant only with respect to a
subset of such matrices, forming
p a U(1) subgroup. In fact, chosen for instance
the point z1 = 0 and z2 = σ = µ2 /(2λ) we have that
a b 0 bσ
= ⇒ b = 0, d = 1. (1.20)
c d σ dσ
This means that if we study the system around the stable minima, we explic-
itly see only a U(1) invariance, but we still have constraints on the possible
12 symmetries in physics
interaction terms due to the original larger U(2) symmetry group. For instance,
expanding the potential (1.16) around the minima we see that the symmetry
breaking terms are not generic, but fixed by the original U(2)-invariant poten-
tial.
2
THE PRINCIPLE OF LEAST ACTION AND NOETHER’S
THEOREM
In this chapter we are going to derive the Euler–Lagrange equations for con-
tinuous systems and examine the relation between conservation laws and the
existence of continuous symmetries. After a brief general discussion we will
concentrate on relativistic systems.
The principle of least action is a very elegant formalism that can be used to
derive the equations of motion of a physical system, starting from a functional,
the action, which is manifestly invariant under a definite set of symmetries.
Imagine we throw an object in the Earth’s gravitational field. We know that in
the presence of gravity the trajectory will be a parabola connecting two extrema
A and B. What is the distinctive feature of such a trajectory? We could imagine
an infinite number of different paths connecting A and B, but if we sum the
difference between kinetic and potential energy at each of the points of these
paths the result will be a number that is always greater than the one obtained
for the route followed by an actual object. In fact, while K ( A), K ( B), U ( A)
and U ( B) are the same for any trajectory, their value along the path changes
and crucially depends on the chosen trajectory. If we simplify the problem
and consider simply an object in vertical motion, whose height is described by
the function y(t), the law of motion is the one associated to an extremum (not
necessarily a minimum) of the action functional
" #
dy(t) 2
Z tB
1
S[y] = m − mg y(t) dt. (2.1)
tA 2 dt
13
14 principle of least action
The result is the same as the one we would obtain by directly solving the
equation of motion
ÿ(t) = − g, (2.2)
with boundary conditions specified by y(t A ) and y(t B ). In fact, computing the
extrema of (2.1)
δS[y] S[y + cδy] − S[y]
= lim =0 (2.3)
δy c →0 c
one reproduces (2.2).
This line of reasoning works in general for any classical system. We can
recover its equations of motion as the result of a variational problem for an
action functional constructed in terms of a Lagrangian L given by the difference
of kinetic and potential energy of the system. For a generic classical physical
system with N degrees of freedom q I (t), where I = 1, . . . , N, the state at a time
t is determined by {q I (t), q̇ I (t)} and the action giving the evolution of such
system from state A at time t A to state B at time t B is
Z tB
S[q] = L(t, q I , q̇ I ) dt. (2.4)
tA
The evolution of the system is then described by the functions q I (t) extremizing
the action S. This means that if we change path
q0I (t) = q I (t) + δq I (t), (2.5)
with the assumption that the extrema are fixed δq I (t A ) = δq I (t B ) = 0, we
require that
δS = S[q0 ] − S[q] = 0. (2.6)
This leads to (assuming synchronous variations δt = 0)
Z tB Z tB
δL δL
δS = δL dt = δq I + δq̇ I dt
tA tA δq I δq̇ I
Z tB
δL δL d
= δq I + δq I dt
tA δq I δq̇ I dt
Z tB (2.7)
δL δL δL
d d
= δq I + δq I − δq I dt
tA δq I dt δq̇ I dt δq̇ I
Z tB B
δL δL δL
d
= − δq I dt + δq I .
tA δq I dt δq̇ I δq̇ I A
2.1 euler–lagrange equations 15
The last term is vanishing because extrema are fixed and, since we consider
arbitrary variations δq I , the integral vanishes if and only if
δL δL
d
− = 0. (2.8)
δq I dt δq̇ I
These are the Euler–Lagrange equations, which must be satisfied by physical
trajectories.
These equations have a vast application in physics. Although not all systems
necessarily admit a Lagrangian description, whenever we have such a descrip-
tion we can derive its equations of motion by using (2.8). In addition, even
if we restricted our discussion to systems with a finite number of degrees of
freedom, we can extend it and repeat the derivation above for continuous sys-
tems: N → ∞. Field theories are prominent examples of this type. For instance,
the electromagnetic field can have different values at different points of space
and at different times Aµ = Aµ (t,~r ). We can therefore think of it as a vari-
able whose index I runs over an infinite number of possible values labelling
different points in space:
It is also obvious that now we need to replace the sum over I by an integration
on the ~r variable
Z
∑→ d3~r (2.11)
I
with A and B two events in spacetime. We note that while t depends on the
observer A and B do not. In the following we will often omit the extrema to
consider general equations of motion.
16 principle of least action
The Euler–Lagrange equations for fields follow as before (sum over repeated
indices a understood):
Z B Z B
4 δL δL
δS = δL d x = δφa + δ∂µ φa d4 x
A A δφa δ∂µ φa
Z B
δL δL
= δφa + ∂µ δφa d4 x
A δφa δ∂µ φa
Z B (2.13)
δL δL δL 4
= δφa + ∂µ δφa − ∂µ δφa d x
A δφa δ∂µ φa δ∂µ φa
Z B B
δL δL 4 δL
= − ∂µ δφa d x + δφa .
A δφa δ∂µ φa δ∂µ φa A
Once again the last term vanishes because we assume that the extrema are fixed
and therefore δS vanishes if and only if
δL δL
− ∂µ = 0. (2.14)
δφa δ∂µ φa
2.2 examples
x µ = x µ ( τ ), (2.15)
where τ is its proper time. Physical worldlines are selected by solving the
relativistic equations of motion, which, as we will now show, can be derived
using the principle of least action. In the case of a free particle such action must
2.2 examples 17
be a scalar function constructed solely from worldline data and the simplest
possibility is to use proper time τ for such purpose. An invariant definition of
the proper time is given by the square root of the line element
3
ds2 = dx µ ⊗ dx ν ηµν = −(dx0 )2 + ∑ (dxi )2 . (2.16)
i =1
with pµ = mcuµ , which is the correct equation of motion for a free relativistic
particle.
1
Z
S1 = − d4 x Fµν F µν . (2.22)
16πc
The Euler–Lagrange equations for this case are
δL δL
− ∂ν =0 (2.23)
δAµ δ∂ν Aµ
and hence
∂ν F µν = 0. (2.25)
However, we can see that this is a total derivative and therefore it does not
contribute to the equations of motion:
We first prove that such an action is gauge invariant provided the continuity
equations ∂µ jµ = 0 are valid. Gauge transformations map Aµ → Aµ + ∂µ Λ.
Hence, the action S2 changes accordingly
1 1 1
Z Z Z
δgauge S2 = d4 xjµ ∂µ Λ = S2 + d4 x ∂ µ ( j µ Λ ) − d4 x ∂µ jµ Λ. (2.29)
c2 c2 c2
Discarding boundary terms, the action is gauge invariant if ∂µ jµ = 0. Finally,
we can put together S0 + S1 + S2 and compute both the equations of motion for
a field in the presence of charges and for the motion of the charges in this field.
1
j δAµ d4 x we get
δS1 δS2 R µ
We already computed δA µ
. If we add δA µ
= c 2
4π ν
∂µ F µν = − j . (2.30)
c
Using the explicit expression of the 4-current for a single charged particle in S2
we can also write
q
Z
S2 = dx µ Aµ (2.31)
c
and varying with respect to δx µ (up to boundary terms)
dpµ q
Z Z
δx (S0 + S2 ) = − δxµ dτ + dδx µ Aµ + dx µ δx Aµ
dτ c
dpµ q
Z Z
= − δxµ dτ + d δx µ Aµ
dτ c
(2.32)
q q
Z Z
− δx µ dAµ + dx ν ∂µ Aν δx µ
c c
dpµ q dx ν
Z Z
= − δxµ dτ + δx µ ∂µ Aν − ∂ν Aµ dτ
dτ c dτ
leading to the equations of motion
dpµ q
= F µν uν , (2.33)
dτ c
which is the relativistic form of Lorentz force.
q I → q̂ I , q̇ I → q̂˙ I , (2.34)
which leaves the Lagrangian invariant without using the equations of motion. For-
mally
Since
δL δL δL
d d
δL(q, q̇) = − δq I + δq I (2.36)
δq I dt δq̇ I dt δq̇ I
δL
Q= δq I (2.37)
δq̇ I
must be conserved along the motion, where the first term in (2.36) is vanish-
ing because of the Euler–Lagrange equations. Note that in this derivation the
assumption that the symmetry is continuous is crucial, otherwise the infinites-
imal variation of L does not make sense.
Also in Lagrangian field theory, the invariance of the action gets translated
in the existence of conserved quantities, but now this is expressed by a local
conservation of a current. In fact, following the same steps performed above
δL δL δL
δL = − ∂µ δφ + ∂µ δφ , (2.38)
δφ δ∂µ φ δ∂µ φ
In fact
d
Z Z
Q= d3 x ∂0 j0 ( x ) = − d3 x ∂i ji ( x ) = 0, (2.41)
dt
where the last equality follows from the vanishing of ji at the boundary.
exercises
1 1
L = − ∂µ Ai ∂µ Ai − ∂µ Bi ∂µ Bi − m2 ( A1 B2 − B1 A2 ),
2 2
where Ai and Bi are real fields and i = 1, 2.
2. Compute the equations of motion for the two real scalar fields φi , i = 1, 2,
following from
1 1
L = − δij Dµ φi D µ φ j + Cijk φi φ j φk ,
2 6
1
L = −∂µ Φ† ∂µ Φ + Φ† ηΦ − (Φ† Φ)2 ,
2
φ1 ( x )
φ2 ( x )
where Φ is a complex field with 4 components Φ( x ) = φ3 ( x ) and
φ4 ( x )
η = diag{−1, −1, 1, 1}. Compute also the extrema of the potential.
Part II
FINITE GROUPS
3
GROUPS: BASIC ELEMENTS
3.1 definitions
Definition 3.1. A Group (G, ◦) is a set G with a composition law, called multi-
plication,
◦ : G×G → G
( g1 , g2 ) 7 → g1 ◦ g2
satisfying the following properties:
i) associativity: a ◦ (b ◦ c) = ( a ◦ b) ◦ c, ∀ a, b, c ∈ G;
The closure property ensures that the composition law does not generate ele-
ments outside G. Associativity implies that the computation of an n-fold prod-
uct does not depend on how the elements are grouped together.
25
26 groups: basic elements
All these examples have an infinite number of elements, but there are also
groups with a finite number of elements. We will call
Definition 3.4. Order of a group G: #( G ) = number of elements in G.
Groups with a finite order are called finite groups. An example of a finite
group is
Definition 3.5. Cyclic group of order n: Zn ≡ { gk , 1 ≤ k ≤ n| gn = 1}.
A simple way to construct this group is by means of the n-th root of unity
g = e2πi/n . Let {ζ 1 , . . . , ζ n } be the n distinct roots of the equation zn = 1, so
that ζ k = gk . This set forms a group under multiplication, with composition
law ζ k ◦ ζ m = ζ m ◦ ζ k = e2πi/n(k+m) = ζ (k+m mod n) . From this composition
we can also derive an alternative definition of Zn as (Z, +) with equivalence
relation g + n ∼ g, ∀ g ∈ Z.
From the examples above we can also distinguish discrete groups like Zn or
(Z, +) and continuous groups, like (R, +).
Example 3.2. Other examples we will in the following use are the sets of real
and complex square matrices Mat(n, R) and Mat(n, C). These are semigroups
(not all their elements have an inverse) using matrix multiplication.
The sets of real or complex square invertible matrices form a group, instead.
They are denoted by GL(n, R) and GL(n, C), respectively, where GL stands for
general linear.
We also note that while Zn , (Z, +) and (R, +) are Abelian groups, GL(n, R)
and GL(n, C) are non-Abelian.
If we have two groups with the same number of elements and analogous prop-
erties, how do we distinguish them? When can we consider the two groups to
be the same? In order to answer to these questions we need to discuss maps
between groups.
As we explained above, a group is defined by a set, but also by a multiplica-
tion. The map between the elements of the two groups may or may not respect
such multiplication law.
Definition 3.6. A Homomorphism is a map φ : G1 → G2 such that the composi-
tion in G1 is mapped into the composition in G2 :
φ ( g1 ◦ 1 g2 ) = φ ( g1 ) ◦ 2 φ ( g2 ) .
28 groups: basic elements
and
However, homomorphic maps may loose some of the structure of the original
group and relate groups with different properties. For instance we will see that
homomorphisms can map non-Abelian groups to Abelian ones.
If we want to construct a map that fully respects the structure of a group we
need to require that it is also 1 to 1:
Definition 3.7. A homomorphism φ : G1 → G2 is called isomorphism if it is
surjective and invertible:
ϕ: GL(n, R) −→ R∗
(3.5)
M 7−→ det( M).
This map preserves the group multiplication thanks to the properties of the
determinant: ϕ( AB) = det( AB) = det( A) det( B) = ϕ( A) ϕ( B).
Example 3.4. The second example is a homomorphism between a continuous
and a discrete group:
ϕ: R∗ −→ {1, −1}
(3.6)
a 7−→ |aa| .
3.3 finite groups 29
ab a b
The map ϕ satisfies ϕ( ab) = | ab|
= | a| |b|
= ϕ( A) ϕ( B), as expected for a
homomorphism.
Example 3.5. Finally, we present an isomorphism between two continuous groups:
Finite groups have properties that are not shared by infinite (or continuous)
groups. For example, if an element g of a finite group is multiplied by itself
enough times, the identity of the group is recovered. In fact, multiplying any
element of a finite group by itself more than #G times must lead to a recurrence
of the product, because there are at most #G distinct elements. Take g ∈ G,
where G is a finite group with #G = n. We can easily prove that there must be
a m ≤ n such that gm = 1. Consider the set { g, g2 , . . . , gn+1 }, which has one
element more than those of G. Since each of the elements of this set is in G by
closure, at least two of them must coincide. This means that ∃ p = q + m such
that g p = gq . We then deduce that g p = gq gm = gq and hence gm = 1, together
with m ≤ n.
We then call
Definition 3.9. The order of an element g ∈ G is the smallest integer n such
that gn = 1.
We can also prove that multiplying a group by any of its element we recover
all its elements.
Theorem 3.2. Rearrangement theorem.
g G = G, ∀ g ∈ G.
Proof. g G contains #G elements and hence they are all distinct and therefore
it is identified with G, or at least two of them are equal. If G = { g1 , . . . , gn }
and g = g j , two elements of gG are equal if g j gk = g j gm , but this implies that
gk = gm and therefore gG = G.
30 groups: basic elements
◦ 1 a
1 1 a (3.8)
a a 1
Note that the table is symmetric about the diagonal, which means that the
group is Abelian.
A different and efficient way to present a discrete group is via its generators
and their relations.
The group defined by such presentation is the result of all the inequivalent
(with respect to relations in R) words constructed by using the generators in S
(and their inverse) as letters. Note that for the same group we may have several
presentations!
Example 3.6. Cyclic groups have the following presentation:
Zn = h g | g n = 1 i. (3.9)
and from such presentation we can deduce all the elements of the group. We
see from the relations in the presentation that both ρ and σ have finite order
(n and 2). We can also use the last relation in the presentation to move all
ρ’s appearing in the definition of a group element to the left of any σ. We
3.3 finite groups 31
1 2 3 2
2 3 2 1
3 1 1 3
Figure 4.: Two identical elements of the permutation group of three objects.
First of all we can see that such maps admit a group structure. The compo-
sition of two maps f , g : X → X defined as f ◦ g : x 7→ f ( g( x )) is obviously
again a map f ◦ g : X → X and hence the group is closed under multiplication
◦. This product is also clearly associative and the identity is the map 1 : x 7→ x,
∀ x ∈ X. If we restrict our attention to bijective maps then there is an inverse
map for each one of them and hence we have an inverse element for each of
the group elements.
Note that if we have two sets X and Y that have the same number of elements,
the associated permutation groups are isomorphic.
32 groups: basic elements
We can also determine the order of Sn rather easily. Since permutations are
maps σ : X → X, we just have to check how many inequivalent ways we have
to map X to itself. The first element can be mapped into each of the n elements
of X. Hence σ (1) can have n different values. At this point σ (2) can take at
most n − 1 different values, because we have to exclude σ(1), and so on. This
means that
#Sn = n! . (3.10)
There are several different ways to describe these maps, but, instead of giving
additional abstract definitions, we will introduce them by examples.
Example 3.7. Consider S3 , the group of permutations of 3 elements. This group
has order 3! = 6. One way to describe these elements is to use matrices describ-
ing how each element is mapped:
1 2 3
. (3.11)
σ (1) σ (2) σ (3)
In this notation the top line represents the initial order of the objects and the
lower line represents the final order, after the permutation. The 6 inequivalent
maps are then
1 2 3 1 2 3 1 2 3
1= , p1 = , p2 = ,
1 2 3 2 3 1 3 1 2
(3.12)
1 2 3 1 2 3 1 2 3
d1 = , d2 = , d3 = .
1 3 2 3 2 1 2 1 3
The composition is trivial and follows by applying the maps in series. For
instance:
1 → 3 → 1
(123) ◦ (13) : 2 → 2 → 3 = (23). (3.14)
3 → 1 → 2
1 → 3 → 2
(123) = (32) ◦ (31) = 2 → 2 → 3 (3.15)
3 → 1 → 1
but also (13) ◦ (12) = (123). For each element however there is a minimum
number of 2-cycles needed to describe it and one can also see that if such
number is even all the decompositions of that group element will be given
in terms of an even number of 2-cycles and if such number is odd all the
decompositions will be given in terms of an odd number of 2-cycles.
We therefore define
We can then split our symmetric group in terms of the even and odd elements
with respect to parity. It is interesting to note that the subset of elements with
even parity define also a group:
In the case of S3 we see that A3 = {1, p1 , p2 }. Note also that in this case A3
is Abelian, while S3 is not.
34 groups: basic elements
exercises
3. Write the multiplicative table for a group of order 3 and show that it is
unique.
5. The group of quaternions is the group generated by the elements {1, −1, I, J, K }
satisfying the composition laws
I 2 = J 2 = K2 = −1, I J = K, (−1)2 = 1.
Z4 , S3 ,
Z6 , D3 ,
S3 , D3 .
7. Show that the inverse of a k-cycle is the same k-cycle written in reverse
order.
8. Compute (34) ◦ (623174) ◦ (34), (12) ◦ (13)(45) ◦ (12) and (43) ◦ (13)(452) ◦
(34). What do we learn from this?
3.3 finite groups 35
Now that we have introduced the basic notions of group theory, we would like
to understand when a subset of the elements of a group form themselves a
group under the same composition law, how to take products and quotients of
groups.
For finite groups there is no need of the requirement of having the inverse,
because an = 1 for some n for any a ∈ H.
According to this definition, the unit element {1} forms a subgroup of G as
well as the full G. These are improper subgroups.
In physics subgroups will be extremely useful, especially in the context of
symmetry breaking. We often encounter systems whose Lagrangian (or Hamil-
tonian) is invariant under a certain symmetry group, which is broken by the
addition of additional terms to a subgroup H < G.
Example 4.1. Dn contains a subgroup isomorphic to Zn . This is obtained as the
result of the repeated application of the generator ρ in the presentation (3.11).
Example 4.2. Zn < Sn . If we name ρ = (12 . . . n), we see that ρn = 1 and ρk 6= 1
for any k 6= n.
37
38 subgroups, quotients and products
Example
4.3. Dn < Sn : ρ has been identified above. σ = (1 n)(2 n − 1) . . . =
1 2 ... n−1 n
.
n n−1 ... 2 1
Once again all these examples let us sense that all finite groups can be real-
ized as subgroups of Sn for n sufficiently large (which is the content of Cayley’s
theorem). If we only had finite groups we could then forget all the theoretical
constructions we introduced so far and directly study Sn . However it is still
useful to use finite groups as a laboratory to develop the necessary intuition to
study the properties of continuous groups.
A special subgroup in every group is
ZG = { a ∈ G | a ◦ g = g ◦ a ∀ g ∈ G }.
This subgroup contains all the elements of a group that commute with all the
others. By definition, the center of a group is Abelian and it is straightforward
to see that if G is Abelian than the ZG is the whole group.
Example 4.4. Consider GL(2,R), i.e. the group of 2 by 2 invertible matrices. The
center is given by all the matrices proportional to the identity: s1, with s ∈ R∗ .
4.2 quotients
1. reflexive: a ∼ a;
2. symmetric: a ∼ b ⇒ b ∼ a;
4.2 quotients 39
3. transitive: a ∼ b and b ∼ c ⇒ a ∼ c.
Definition 4.4. An equivalence class for an element a ∈ A is the set of all ele-
ments in A equivalent to a:
[ a ] ≡ { b ∈ A | b ∼ a }.
It can be seen that equivalence relations split a set into disjoint subsets given
by equivalence classes. This splitting is going to be crucial in the definition of
group quotients.
If we now want to identify elements in G that are related by elements in H,
we need to define appropriate equivalence classes by multiplying each element
of G by all the elements in H. Since for a generic group the product is not
commutative, the result will depend on the order. Given H < G we therefore
define distinct equivalence classes using left and right multiplication.
Definition 4.5. Right cosets are the sets obtained by multiplying H < G by
elements g ∈ G from the right: Hg. Analogously, Left coset ≡ gH, with H < G
and g ∈ G.
Note that in general gH and Hg are not in H unless also g ∈ H and they are
not even subgroups of G.
Example 4.5. Consider H = {1, d1 } < S3 . This group is isomorphic to Z2 . Right
cosets are obtained by multiplying H from the right with the elements of S3 .
The group structure shows that we have only 3 distinct cosets:
As expected, left and right cosets can be different and indeed in this case they
are different. We can also note that the intersection between left cosets (or
between right cosets) is empty and that their union is the original group S3 .
In fact, this is a consequence of a general theorem:
40 subgroups, quotients and products
Theorem 4.1. Two cosets of a subgroup have the same elements or they do not share
any element in common.
Proof. Let Hg1 and Hg2 be two different (right) cosets. If there is at least one
element in common, this means that for hi , h j ∈ H, we have hi g1 = h j g2 . From
this assumption we derive that g2 ( g1 )−1 = (h j )−1 hi ∈ H. Using the rearrange-
ment theorem, we then conclude that Hg2 ( g1 )−1 = H and hence Hg2 = Hg1 ,
so all their elements are in the same.
Definition 4.6. The Right quotient G/H of a group G by its subgroup H is the
set of all equivalence classes of elements of G obtained by right multiplications
with H:
g1 ∼ g2 if ∃ h ∈ H | g1 = g2 h.
This definition implies that the elements a and b give the same element when
multiplied by g, the first from the right and the second from the left. We then
see that we can use such a definition to establish a relation between left and
right cosets. Actually, we can extend this definition to whole subgroups of a
group:
gNg−1 = N, ∀ g ∈ G.
This definition implies that the conjugate of any element in N is also an ele-
ment in N for any element in G defining the conjugation: gng−1 ∈ N, ∀g ∈
4.2 quotients 41
G and ∀n ∈ N. The straightforward consequence is that right and left cosets de-
fined in terms of N coincide: gN = Ng. Since left and right cosets are the same,
also left and right quotients coincide. This eventually leads to the fact that G/N
quotients inherit the property of being a group under the multiplication of the
original group G:
Theorem 4.2. If N C G then G/N is a group.
Proof. The elements of the quotient are defined as the equivalence classes given
by the cosets. Hence all the elements in gN for a given g ∈ G have to be
identified. The product in G/N is the product in G, where again we identify
the elements which are part of the same coset. We then have to verify that all
group properties are respected for G/N. Closure follows from the composition
in G:
g1 N ◦ g2 N = g1 ( g2 N ( g2 )−1 ) g2 N = g1 g2 N ( g2 )−1 g2 N = g1 g2 N,
where we first used the fact that N is normal and then the fact that N is a
group (N ◦ N = N). The associative property is straightforward. The existence
of an inverse for each element follows from the proof of closure above. Given
an element gN, its inverse is defined by g−1 N.
Example 4.6. We first give an instance of a subgroup that is not normal. Con-
sider Z2 < Dn , defined by the σ generator in the presentation (3.11). From
the equivalence relations given in the presentation we can see that ρσρ−1 =
σρ−2 ∈/ Z2 . As expected, this implies that the product of elements in the quo-
tient using the multiplication in Dn does not close consistently. In fact, Dn /Z2
implies the identification ρk ∼ ρk σ, but if we multiply elements in the same
equivalence class, we get inconsistent results:
ρ a ρb = ρ a+b 6= (ρ a σ )(ρb σ ) = ρ a−b . (4.4)
Example 4.7. An example of a normal subgroup is given by Zn < Dn , defined
by ρ. Since we identify all the elements ρk ∼ 1 and σρk ∼ σ, the quotient is
given by two elements, which define the group Z2 = 1, σ. The multiplication
of elements in the same equivalence class now gives consistent results. For
instance, we see that
σρ a σρb = σ2 ρb− a = ρb−a ∼ 1 = σσ. (4.5)
Example 4.8. Although for n = 3 D3 ' S3 , in general Zn < Sn is not normal
for n > 3. This can be seen right away by checking that (12)(12345)(12) =
(21345) 6= (12345) a for any a.
42 subgroups, quotients and products
4.3 products
G × K = {( g, k), g ∈ G, k ∈ K },
with composition
( g1 , k 1 ) ◦ ( g2 , k 2 ) = ( g1 ◦ g2 , k 1 ◦ k 2 ) .
For finite groups we can deduce immediately the order of the product as
#( G × K ) = #G #K. (4.6)
This means that we need to take all the words constructed from the genera-
tors of G and K with their equivalence relations plus the commutation relation
between elements of G and K.
As we discussed in the introduction to this section, we may also need to
specify groups that are products where the action of one of the factors depends
4.3 products 43
on which of the elements of the other factor is taken in the product. For this
reason we can introduce a new definition, where the last relation in (4.7) is
replaced by a new one.
Definition 4.10. Given two groups G and K their semidirect product is defined
by
Note that the symbol defining the product contains an asymmetry, which is
reflected in the different ways the two groups enter in the definition: G n K =
K o G. In fact, for semidirect products elements in G × 1K and in 1G × K do
not necessarily commute.
From the definition we also see that K is a normal subgroup, i.e. K C ( G n K ),
while G is not, i.e. G < ( G n K ).
Example 4.9. Dn ' Z2 n Zn , where Z2 = hσ | σ2 = 1i, Zn = hρ | ρn = 1i and
φσ (ρ) = ρ−1 , so that σρ = ρ−1 σ. Zn is a normal subgroup, Z2 is not.
Example 4.10. The group of orthogonal matrices in N dimensions O(N) is de-
fined as the set of matrices O ∈ GL(N, R), which satisfy O T O = 1 N = OO T .
Since det (O T O) = (det O)2 = det 1N = 1, the matrices of O(N) satisfy det
O = ±1. Those with positive determinant form a subgroup, called special or-
thogonal group SO(N) < O(N), which obviously contains the identity. The
quotient with respect to this subgroup generates two classes of equivalence,
the one of matrices with positive determinant (which contains the identity)
and the set of matrices with negative determinant. This is actually a Z2 group
under the original multiplication. In fact we leave as an exercise to the reader
to check that generically O(N ) ' SO( N ) o Z2 , which, for odd N reduces to a
direct product.
Example 4.11. The Euclidean group E( N ) ' ISO( N ). It is defined as the combi-
nation of rotations SO(N) and translations in R N :
Hence
Using direct and semidirect products we can construct groups from other
groups, but obviously these are not really new, because they can be classified
as products. So, how do we distinguish genuinely new groups?
Definition 4.11. A group is called simple if it does not admit proper normal
subgroups.
exercises
4. Show that for Abelian groups left and right cosets coincide.
4.4 classifying finite groups. 45
6. Find all proper subgroups of the group of quaternions Q and discuss the
right and left quotients of Q with respect to them.
So far we discussed groups using their abstract definition. To really start ap-
plying them (especially to physics), we must be able to represent them in a way
that could be used for practical calculations. For this reason we need to map
abstract groups into groups of mathematical objects we know how to deal with,
like matrices or differential operators. We therefore introduce in this lecture the
concept of representations of a group.
If a group codifies the symmetries of a physical system, i.e. the relations
between the transformations we can apply to a system without changing the
physics, we want to be able to represent this action on the system.
Beware that the existence of a symmetry does not mean that nothing changes
when applying it. For instance the discrete group D3 rotates or reflects a trian-
gle, leaving the image fixed, but not the vertices:
σ ρ
refl. rot.
In the same way, when we discuss a physical system, its invariance under a
certain symmetry does not imply that all its quantities remain unchanged. On
the other hand, the existence of a symmetry transformation tells us how the
variables describing a certain physical state change. For instance, Special Rela-
tivity dictates invariance of physics under Lorentz transformations. This does
not mean that two inertial frames, related by a Lorentz transformation, share
47
48 representations
exactly the same value for all their physical quantities, but rather that their
transformation is such that physical laws do not change in form. For instance,
the value of pµ from an inertial system to another changes by pµ → Λµ ν pν , but
dpµ
free particles are described by the same law dσ in both systems. In fact, Λ pro-
vides a concrete representation of how the group of Lorentz symmetries acts
on the various components of the momentum, which are grouped in a multiplet
of the symmetry group. In a similar fashion, when we consider a quantum-
mechanical system with a certain symmetry, we know that there must exist a
unitary operator commuting with the Hamiltonian, which gives a representa-
tion to the action of such symmetry. The wavefunctions, however, do not share
the symmetry of the Hamiltonian and they will transform in a way determined
by such operator and, in particular, by the representation of the symmetry
group provided by the unitary operator mentioned above. This, in turn, will
provide a way to classify the eigenfunctions of the Hamiltonian, as we will
discuss shortly.
Definition 5.2. The dimension of the vector space on which the matrices D ( g)
act is the dimension of the representation.
Since we are representing an abstract group G by a group of matrices, the
mappings between the two groups may not necessarily preserve all the infor-
mation of the original group G. For this reason we distinguish between faithful
representations, obtained by means of isomorphisms, and unfaithful representa-
tions, obtained by homomorphisms. In this last case, we will have that different
elements of the group are represented in the same way. If more than one el-
ement is represented by the same matrix, the group element connecting them
must be represented by the identity. This further implies that more than one
group element must be represented by the identity matrix and therefore that
the kernel of the map is going to be non-trivial. A faithful representation then
is such if D ( g) = D ( g0 ) ⇔ g = g0 .
It is also easy to convince ourselves that the elements in the kernel of the
representation are going to be a normal subgroup of the original group G:
H = ker D C G. In fact, if D ( g) = D ( g0 ) = 1, we also have that D ( g ◦ g0 ) =
D ( g) D ( g0 ) = 1. Obviously D (1) = 1, the inverses are again represented by
the identity and finally D ( g ◦ h ◦ g−1 ) = D ( g)1D ( g)−1 = 1 for any g ∈ G and
h ∈ H. We can then conclude that if our representation is unfaithful for G,
the same matrices are providing a representation of G/H that is faithful! We
therefore conclude that:
Theorem 5.1. All non-trivial representations of a simple group are faithful.
Once we choose a basis for V, {ei } with i = 1, . . . , N, the matrix action is
defined as
D ( g)ei = e j D ji ( g), (5.2)
where sum over repeated indices is understood. This is needed to preserve the
composition rule
D ( g ◦ g̃) = D ( g) D ( g̃). (5.3)
In fact:
D ( g ◦ g̃)ei = e j D ji ( g ◦ g̃) = D ( g)[ D ( g̃)ei ] =
(5.4)
D ( g)ek Dki ( g̃) = e j D jk ( g) Dki ( g̃)
and hence
D ji ( g ◦ g̃) = D jk ( g) Dki ( g̃), (5.5)
50 representations
which provides the standard matrix composition law. Be careful with the index
position when passing to an explicit representation.
We can now ask ourselves whether we can always represent a group in such
a way and also what kind of matrix representations are really useful. A repre-
sentation that always exists is
Definition 5.3. The trivial representation is defined by
D ( g) = 1, ∀ g ∈ G.
It is easy to check that the relations in the presentation are identically satisfied:
[ D2 (ρ)]3 = [ D (σ)]2 = 12 , (5.9)
√
1 1 3
D (ρ) D (σ) = √ = D (σ)[ D (ρ)]2 . (5.10)
2 3 −1
In general, we may have an infinite number of different representations for
the same group. To our advantage, we can often classify them, and this is the
subject of the next part of this lecture. In fact, knowing all the representations
of a given group allows us to specify all its (linear) actions on a set of quantities
and it is therefore of primary importance for applications to physics.
We already identified the conditions under which two groups are the same. In
this section we would like to identify the conditions under which two represen-
tations are the same.
Definition 5.5. Two representations D1 and D2 of the same group G are equiv-
alent if we can relate them by a change of basis, i.e.
∃S | D1 ( g) = S−1 D2 ( g)S, ∀ g ∈ G.
The matrix S is called a similarity matrix and its application to V changes the
basis of the representation space.
From the definition above, we see that S preserves the trace
1 1
D (1) = 1 , D ( d1 ) = 1 ,
1 1
1 1
D ( d2 ) = 1 , D ( d3 ) = 1 , (5.15)
1 1
1 1
D ( p1 ) = 1 , D ( p2 ) = 1 .
1 1
1 ρ ρ2 σ σρ σρ2
1
√
and therefore e3 = 1/ 3 1 is invariant. Also the subspace generated by its
1
orthogonal complement
1 1
1 1
e1 = √ − 1 , e2 = √ 1 (5.16)
2 0 6 −2
is invariant. This becomes obvious if we use as basis for the vector space
on which these matrices act the set {e1 , e2 , e3 }. In this new basis the same
representation matrices become
√
−√1/2 3/2 0 −1 0 0
D (ρ = p1 ) = − 3/2 −1/2 0 , D (σ = d3 ) = 0 1 0 . (5.17)
0 0 1 0 0 1
These are block-diagonal matrices and so will be all the matrices obtained by
their multiplications. We therefore see that this representation is the direct sum
of the 2-dimensional representation D2 and of the trivial representation.
Before proceeding let us stress the importance of the field used for the vector
space V in order to decide whether a given representation is reducible or not.
If we consider SO(2), the group of rotations on a plane, a generic element is
represented by
cos θ sin θ
D (θ ) = . (5.18)
− sin θ cos θ
This representation is irreducible if we act on R2 , but becomes reducible on C2 :
eiθ
0
D̃ (θ ) = S−1 D (θ )S = , (5.19)
0 e−iθ
1 1
where S = √1 .
2 i −i
A very interesting fact comes from the following theorem, which proves that
for finite groups unitary representations are all we need.
M≡ ∑ D † ( g ) D ( g ). (5.20)
g∈ G
This satisfies
where we used the rearrangement theorem in the second line. We also have that
M = M† and hence M is a matrix with positive real eigenvalues (v† Mv > 0).
We can diagonalize it by means of a unitary matrix U
Although we proved this theorem for finite groups, we could prove an anal-
ogous one for continuous compact groups. For continuous compact groups, how-
ever, we need to sum over infinite elements. This means that we can still per-
form the same proof provided we replace the sum with an integral over the
group, with an appropriate invariant measure (called Haar measure). Only the
existence of such a measure allows us to repeat the demonstration with
Z
M≡ dµ g D † ( g) D ( g), (5.25)
This means that for finite groups and for continuous compact groups we can
always construct representations that are unitary and block diagonal. Hence we
can constrain ourselves to studying the irreducible representations composing
them.
We stress, however, that if the group is not finite nor compact, we may have
representations that are decomposable, but not reducible, like
1 x
D(x) = , (5.28)
0 1
5.4 comments
|h2|1i|2
P12 = .
|| |1i ||2 || |2i ||2
The change of state given by the action of the symmetry group G (for instance
the rotation group or Sn for n identical particles) is represented by a unitary
operator and therefore its action on the Hilbert space is such that the transition
probabilities are not affected. We see then that knowledge of group theory
implies knowledge of QM systems.
For instance, energy levels are given by eigenstates of the Hamiltonian: H |i =
E|i. If G is a group of transformations leaving H invariant, there is a uni-
tary representation of this group that commutes with H and hence UH | Ei =
EU | Ei = H (U | Ei). This means that the unitary operator U maps eigenstates
with eigenvalue E into states with the same eigenvalue and that therefore
the portion of the Hilbert space containing the eigenvectors with energy E,
H E ⊂ H, is invariant. If such a space is invariant under the action of U, this op-
erator must be in an irreducible representation of the symmetry group. Hence
the level degeneracy, which is equivalent to the dimension of H E , must also be
the dimension of an irreducible representation.
exercises
Show that if the dimensions are the same (i.e. p = r and q = s) then
( A ⊕ B)(C ⊕ D ) = AC ⊕ BD and ( A ⊗ B)(C ⊗ D ) = AC ⊗ BD.
58 representations
tr( A ⊗ B) = (trA)(trB).
γ0 = −i σ1 ⊗ 1, γ1 = σ2 ⊗ σ1 , γ2 = σ2 ⊗ σ2 , γ3 = σ2 ⊗ σ3 .
Verify that σi σj = ieijk σk + δij 1 and, using what you just learned, compute
γ5 = −iγ0 γ1 γ2 γ3 .
−1 0 0 0
0 1 0 0
0 1 0 0 0 0 1 0
D ( a) =
0 0 −1 0 ,
D (b) =
1 0 0
.
0
0 0 0 1 0 0 0 e2πi/3
D (ρ)[ψ ⊗ χ] = χ ⊗ ψ, ∀ψ, χ ∈ H.
The Hamiltonian is
1 h 2 i
Ĥ = p̂ x + p̂2y + V ( x, y) .
2m
The group D4 acts on the wavefunction Ψ( x, y) as (U ( g)Ψ)( x, y) ≡ Ψ( D ( g)( x, y)),
where g ∈ D4 and D ( g) is the 2-dimensional representation of D4 .
a) Show that Ĥ commutes with the representation U.
b) For n positive integer and k n = nπ/2 define the functions
cos(k n z) if n is odd,
f n (z) =
sin(k n z) if n is even.
Show that Ψn,m ( x, y) = f n ( x ) f m (y) inside the square (and zero out-
side) is an eigenfunction of Ĥ and compute its eigenvalue. Discuss
their degeneracy.
60 representations
D1 ( g) T = TD2 ( g), ∀ g ∈ G,
61
62 properties of irreducible representations
0 = D1 ( Tv) = T ( D2 v)
Theorem 6.2. An invertible matrix T commuting with all the matrices of an irre-
ducible representation is proportional to the identity:
[ D ( g), T ] = 0, ∀g ∈ G ⇒ T = c 1.
Let us stress that this corollary uses the fact that the vector space on which
the representation acts is V = C N . As we have seen before we can have irre-
ducible 2-dimensional representations for SO(2) over the reals.
Now that we established these technical results, we can finally prove the great
orthogonality theorem between matrix elements.
In order to introduce an orthogonality criterion, we need first to introduce
an inner product between matrix elements. This is easily identified if we real-
ize that matrix elements can be viewed as complex functions defined over the
group G:
Dij : G → C.
They are therefore part of the vector space C[ G ] of complex functions defined
over G and we can use as inner product the ordinary inner product for a com-
plex vector space. Note that this vector space has complex dimension equal to
the order of the group, because C[ G ] ' C#G , where each factor C is the set of
all possible values on which one can map each element g ∈ G. We can then
define a scalar product on this space by summing the values of the functions
f , h ∈ C[ G ] (vectors in this space) over their components, labelled by the group
elements:
1
( f , h) ≡
#G ∑ f ( g ) h ( g ). (6.1)
g∈ G
Using this scalar product we can finally introduce the concept of orthogonality
between matrix elements.
Theorem 6.4. Great Orthogonality Theorem. Given D1 and D2 irreducible repre-
sentations of dimensions N1 and N2 of a finite group G. We have that either D1 '
/ D2
and then
( Dij1 , Dkl
2
) = 0,
64 properties of irreducible representations
where kl label different matrices and ij are the row and column indices. We
then build T such that D1 T = TD2 . There are actually N1 N2 matrices that play
this role and they are given by
If, on the other hand, N1 = N2 = N, then we have that either the two represen-
tations are inequivalent, and then again T = 0, or they are equivalent D1 ' D2 ,
in which case we can assume they coincide. If D1 = D2 = D then we can repeat
the same argument, using the corollary of Schur’s lemmas to argue that
This means that Tij ( E(kl ) ) = c(kl ) δij , whose trace determines the constants c(kl ) :
We can also combine the two conditions in a unique condition on two generic
irreducible representations a and b saying that
δ ab
b
( Dija , Dkl )= δ δ . (6.9)
N ik jl
From this theorem we also get an important limit on the number of inequiv-
alent irreducible representations for a finite group. In fact we said we can con-
sider the matrix elements of a given representation as elements in the vector
space C[ G ]. Since dim C[ G ] = #G we can have at most #G mutually orthogonal
vectors in this space. On the other hand, for each representation a of dimension
Na we can have at most Na2 different matrix elements, from which we deduce
that the total of orthogonal elements for a given representation is Na2 . Summing
over al irreducible representations we obtain the inequality
We will actually see in the next lecture that such inequality should be an exact
equality.
exercises
2. Verify the Great Orthogonality Theorem for the following irreducible rep-
resentation of S3 :
√
1 −√1 3 −1 0
D (ρ) = , D (σ) = .
2 − 3 −1 0 1
Recall that if g1 = gg2 g−1 for some g ∈ G, then g1 and g2 are conjugate. Also
recall that conjugation is an equivalence relation. We can therefore split a given
group in disjoint subsets defined by equivalence classes under conjugation as
follows:
Definition 7.1. All the elements of a group G that are conjugate to g ∈ G define
the class of conjugation of g.
We can immediately see that the identity is always alone in its conjugacy
class:
1 = g ◦ 1 ◦ g −1 , ∀ g ∈ G. (7.1)
67
68 properties of irreducible representations
We can also see that all the equivalence classes in an Abelian group have a
single element:
The group elements split therefore into three different conjugacy classes
where we used the cyclic property of the trace. Because of this fact we can see
that the value of χ D is the same on all the elements of a conjugacy class. Using
the cyclic property of the trace we can also conclude that all equivalent repre-
sentations have the same character. If two representations are equivalent then
D1 = SD2 S−1 and therefore χ D1 ( g) = trD1 ( g) = tr(SD2 ( g)S−1 ) = tr( D2 ( g)) =
χ D2 ( g).
As we did for the single matrix elements of a representation, we can intro-
duce the concept of orthogonal characters, as follows from their interpretation
as elements of the vector space C[ G ], where we introduced the inner product
(6.1). Actually, it is an immediate consequence of the GOT that
δ ab δ
( Diia , Dkk
b
)= δ δ = δ ab ii = δ ab . (7.7)
N ik ik N
However we can say more. For finite groups all decomposable represen-
tations are also reducible. This means that they are always direct sums of
irreducible representations. We then see that
Theorem 7.2. Representations with the same set of characters are equivalent.
(χ D , χ a ) = c a (7.8)
Since we just saw that characters fully specify a representation we can limit
our study to the irreducible ones. The first thing we should determine then
is how many inequivalent irreducible representations we can construct for a
given group. This is determined by a theorem in the following way:
where the first equality is really just a definition of an element of C#G and
the last one follows from the definition of regular representation. The regular
representation is reducible and using previous results we can always write
D R = ∑ a c a Da , which implies
As we said before, characters are central functions and therefore they have the
same value on each element of the same conjugacy class. This means that when
we evaluate the inner product between two characters we can actually restrict
the sum to elements of different conjugacy classes:
1
(χ a , χb ) =
#G ∑ nc χa ( gc )χb ( gc ) = δab , (7.13)
c
where c runs on the different conjugacy classes and nc is the number of ele-
ments in the c conjugacy class. Using this rewriting we can now produce some
interesting definitions and relations that will completely specify irreducible rep-
resentations of a given group. The first thing we can introduce is the matrix of
characters. For any group we have the same number of conjugacy classes and
irreducible representations and this means that to fully specify all irreducible
representations of a given group we need a square matrix encoding the value
of the character of a given irreducible representation on an element of a given
conjugacy class, for each one of them.
Definition 7.4. The matrix of characters of a representation is
r
na
Uab ≡ χ ( g a ).
#G b
72 properties of irreducible representations
This matrix has some very interesting properties that restrict its possible
entries.
Proof.
r 2
nb
†
Uab Ubc = Ūba Ubc = ∑ #G
χ a ( gb )χc ( gb ) = δac . (7.14)
b
This property means that different rows or different columns are orthogonal
and therefore we can fix their entries. Using it in the opposite way we can also
say more and actually prove a theorem that allows us to fix the number and
dimensions of the irreducible representations of a finite group.
Theorem 7.5.
∑ Na2 = #G.
a
Proof. We use unitarity of the character matrix, only in a way different from
(7.14):
√
n a nc
†
Uab Ubc = Uab U cb = ∑ χb ( gc )χb ( ga ) = δac . (7.15)
b
#G
Another noteworthy relation follows from what we just proved and tells us
when a representation is irreducible:
Definition 7.5. The character table is the matrix collecting all the values of the
characters for each irreducible representations and for every conjugacy class.
C1 C2 C3
1 1 1 1
(7.17)
10 1 ω ω2
1” 1 ω2 ω
#G = 6 = 12 + 12 + 22 . (7.18)
The characters of the trivial representation are 1 and that of the dimension 2
representation can be computed from their explicit expression, we gave previ-
ously. The rest follows by orthogonality
C1 C2 C3
1 1 1 1
(7.19)
10 1 1 −1
2 2 −1 0
74 properties of irreducible representations
exercises
C1 = {1}, C2 = { a, a5 }, C3 = { a2 , a4 },
ei2π/3
0 0 1
D2 ( a) = , D2 (b) = ,
0 e−i2π/3 1 0
e−i2π/3
0 0 −1
D3 ( a) = , D3 (b) = ,
0 e2iπ/3 −1 0
cos π3 i sin π3 0 −1
D4 ( a) = , D4 (b) = .
i sin π3 cos π3 1 0
8
I R R E D U C I B L E R E P R E S E N TAT I O N S O F S n
All finite groups are subgroups of some group of permutations Sn . Also, ir-
reducible representations of any H < G can be obtained by restricting the
representations of G to the elements of H. Combining these two facts we see
that we can obtain the representations of any finite group by studying and
building the irreducible representations of an appropriate permutation group.
It is therefore extremely useful and interesting to try to provide a classification
of the irreducible representations of Sn .
In order to do so, we will use a graphical description that is very efficient:
Young Tableaux. This type of diagrams not only is useful to discuss and classify
both conjugacy classes as well as irreducible representations of permutation
groups, but, as we will see in a subsequent lecture, this graphical description
will also be very useful for the construction and classification of representations
of continuous groups.
Definition 8.1. A standard Young Tableau of order k is a diagram with k boxes,
where the ( j + 1)th row contains a number of boxes that is less or equal to that
of the jth row and the ( j + 1)th column contains a number of boxes that is less
or equal to that of the jth column.
For instance, good Young Tableaux are
, , ,
, , , .
77
78 irreducible representations of s n
In this first part we discuss the relation between Young Tableaux and conjugacy
classes. We will see that there is a 1-to-1 correspondence between regular Young
Tableaux and conjugacy classes. In order to establish such a correspondence we
first need to classify the conjugacy classes of Sn .
Recall that any permutation associated to an irreducible j-cycle can be written
as a product of 2-cycles. For instance
We also established that the inverse of a given j-cycle is a j-cycle with the same
elements written in the opposite order, i.e.
and this in turn can be written as the product of the same 2-cycles defining
the original j-cycle, but in the opposite order, (1m . . . 2) = (12)(13) . . . (1 m −
1)(1 m). This implies that we can constrain our analysis of the conjugacy classes
of Sn using 2-cycles.
It is easy to see that the conjugates of a j-cycle by 2-cycles is again a j-cycle,
where the elements of the 2-cycle in the j-cycle are swapped:
. The same thing is true for the cycle (361)(24)(5) ∈ S6 , which is an element
by the length of the cycles defining them. In terms of Young Tableaux we can
represent them as
C1 ∼ , C2 ∼ , C3 ∼ .
From this analysis we can deduce a general lesson: we can find all the conju-
gacy classes of Sn by considering all Young Tableaux with n boxes arranged in
all possible admissible ways. We also learn that the number of classes depends
on the number of ways we can partition n in terms of integer numbers. For
instance, for S3 , we split n = 3 in 3 possible ways
3 = 1 + 1 + 1 = 2 + 1 = 3, (8.2)
corresponding to its three conjugacy classes and to its 3 inequivalent Young
Tableaux.
The shape of the tableau also tells us how many elements are present in
each conjugacy class. This is determined by the number of different ways
one can arrange the n elements one is permuting in the boxes of the tableau.
For instance, the Young Tableau corresponds to a triple 1-cycle and since
(1)(2)(3) = (1)(3)(2) = (2)(1)(3) = . . . = (3)(2)(1) we see that there is only
one element in this class:
1 2 3 = 1 3 2 = 2 1 3 = ... = 3 2 1 .
The Young Tableau on the other hand has three different elements in its
class
1 3 = (12)(3), 2 3 = (23)(1), 1 3 = (13)(2). (8.3)
2 1 2
We see then that computing the number of elements in the conjugacy class
defined by the Young Tableau under consideration follows by a combinatorial
analysis. In detail, if we have k j j-cycles, the corresponding Tableau represents
n!
nr. of permutations = k
. (8.4)
∏nj=1 j j k j !
Example 8.2. The Tableau has n = 3, k1 = 3, k2 = k3 = 0 and hence it
contains 1 element. The Tableau has n = 3, k1 = k2 = 1 and k3 = 0, hence it
n!
d= , (8.5)
∏ik=1 hi
where hi is the hook number associated to each single box in the diagram. The
hook number is the sum of all the boxes to the right and below a given box
plus one (see the picture 7 for an example).
9 6 3 1
7 4 1
1 5 2
4 1
2
4 1
The proof of this formula is very long, tedious and not particularly illumi-
nating. For this reason, we will not see here a full proof of this formula, but
we will discuss a significative example that is also useful to understand how to
construct all the irreducible representations of a given permutation group.
The starting point is given by the regular representation of Sn , which is
reducible and decomposes in the sum of all the irreducible representations
of a given permutation group. For the sake of simplicity we will focus our
analysis to the group S3 , but we will mention the consequences of consider-
ing general n whenever possible. The regular representation of S3 is a map
8.2 young tableaux and irreducible representations of Sn 81
where π is the parity of the element σ ∈ Sn , defined in Def. 3.13. For in-
1 3
stance, the tableau 2 corresponds to the operators s T = D R (1) + D R (13),
a T = D R (1) − D R (12).
We will now explicitly show that for S3 the regular representation admits
invariant subspaces associated to the vector spaces V , V and V , defined
1 2 3, 1 3 2, 2 1 3, 2 3 1, 3 1 2, 3 2 1 (8.8)
and we can see that they are related by all the permutations in S3 . Conse-
quently s T is given by the sum of all the elements in S3 and a T = D R (1). The
corresponding vector, generating V , is
i = D R (1) D R (1) + D R (12) + D R (13) + D R (23)
|v
D R ( g)|v i = |v i, ∀ g ∈ S3 . (8.10)
the vector
We are left with the space V . In this case s T and a T are not uniquely
determined, but they depend on the arrangement of the elements of Sn in the
1 2
tableau. For instance, for 3 we have s T = D R (1) + D R (12) and a T = D R (1) −
1 3
D R (13), while for we have s T = D R (1) + D R (13) and a T = D R (1) − D R (12).
2
These vectors however are not invariant under the action of the elements of S3 ,
but rather generate two orthogonal invariant subspaces of dimension 2. In fact,
the action of the S3 elements on the first gives
D R (1)|v 1 2 i = | v 1 2 i,
3 3
D R (1)|v 1 3 i = | v 1 3 i,
2 2
D R = D1 + D10 + D2 + D2 , (8.17)
exercises
1. Find the character table of S4 . To do so, we can use that the charac-
ters of the symmetric group are always integers and that the character
of the product of two representations is the product of the characters,
i.e. χ Da ⊗ Db = χ a · χb .
a) Draw all Young Tableaux with 4 boxes and then determine the con-
jugacy classes, the number of their elements and the dimension of
the corresponding irreducible representations.
b) Fill the column related to χ(1) and the raw associated to the trivial
representation.
c) By using the results of this chapter, fill the raw χ1 A , where 1 A refers
to the 1-dimensional alternating representation defined by the parity
of the elements of S4 .
d) Using χ Da ⊗ Db = χ a · χb deduce that 2 ⊗ 1 A ' 2. What does this
imply for the conjugacy classes containing elements constructed by
odd numbers of 2-cycles?
e) Use orthogonality of the characters of irreducible representations to
complete χ2 .
86 irreducible representations of s n
89
90 homotopy groups
9.1 homotopy
One way to construct invariants is the use of homotopy. This is a weaker notion
than that coming from homeomorphisms. In fact, in the case of homotopic
spaces we consider continuous functions between them that need not have an
inverse.
Homotopy also creates equivalence classes of continuous maps.
Definition 9.2. Let X and Y be two topological spaces. Consider two contin-
uous maps f 0 , f 1 : X → Y, then f 0 is homotopic to f 1 ( f 0 can be continuously
deformed into f 1 ):
f0 ∼ f1 (9.1)
F : X × [0, 1] → Y (9.2)
such that
F ( x, 0) = f 0 ( x ), and F ( x, 1) = f 1 ( x ). (9.3)
f ◦ g ∼ 1Y and g ◦ f ∼ 1X . (9.4)
9.1 homotopy 91
Note that we did not require that g was the inverse of the function f , but
only that their combination is homotopically equivalent to the identity.
Example 9.1. Here is an example of homotopic spaces. Let X = Rn r {0},
Y = Sn−1 . We can show that these two spaces are homotopic, which means
that we can continuously deform one into the other. The proof is simple if one
consider the maps
~x
f : Rn r {0} 3 ~x 7→ ∈ S n −1 , (9.5)
|~x |
~x
F (~x, t) = (1 − t)~x + t . (9.7)
|~x |
S1
These maps generate homotopy classes for different α and fixed integer n. Ex-
plicitly, for a fixed n, f α0 ∼ f α1 by means of the continuous family of functions
einθ ∼ imθ
/e , if n 6= m. (9.10)
In fact, a map generalizing (9.9) and interpolating between the two would be
but this is not continuous around θ = 0, 2π for t ∈]0, 1[ and generic m and n.
We actually have a multivalued function.
We can then differentiate equivalence classes in terms of the integer defining
the maps f α . This integer is called winding number, because we can see that it
represents how many times the circle is covered when θ goes from 0 to 2π.
n=0 n=1
such that
H s, 41
H s, 12
x0
β
x0
Figure 12.: Space X with a hole. Loop α cannot be shrunk to a point. Loop β can.
The interesting point for us is that the equivalence classes defined by loop
homotopies have a group structure.
Definition 9.7. The fundamental group or first homotopy group of a topological
space X at a point x0 , Π1 ( X, x0 ) is the group of equivalence classes of homo-
topic loops on X with base point x0 .
We can check the group properties rather easily by composing loops having
the same base point. In fact we define the product of loops by
α(2s), for s ∈ [0, 1/2],
β◦α = (9.15)
β(2s − 1), for s ∈ [1/2, 1].
This means that we first follow the loop α, from x0 to x0 and then the loop β
from the same point to x0 . The result is a special loop with basepoint x0 , which
passes in x0 for s = 1/2. The inverse is trivially defined by
α −1 ( s ) = α (1 − s ), (9.16)
and the identity map is 1(s) = x0 , ∀s. It must be noted that α−1 ◦ α 6= 1, but
obviously it is homotopic to the identity map α−1 ◦ α ∼ 1.
It is very interesting that in the case of arcwise connected topological spaces,
the fundamental group does not depend on the base point x0 anymore:
Theorem 9.2. If X is an arcwise connected topological space and x0 , x1 ∈ X, then
Π1 ( X, x0 ) ' Π1 ( X, x1 ) (9.17)
α −1 α ◦ α −1
α
x0 x0
Figure 13.: A loop α times its inverse α−1 is homotopic to the identity.
α
β
x0 x1
β −1
Figure 14.: Loops with base-point x0 can be extended to loops with base-point x1 if the
space is arcwise connected.
Example 9.3. From our previous discussion we immediately see that Π1 (U(1)) '
Π1 (S1 ) = Z, because inequivalent loops are classified by n ∈ Z. Combining
loops in the equivalence classes n and m we obtain a loop in the equivalence
class defined by n + m.
Example 9.4. Π1 (Sn≥2 ) = 0. In this case all loops can be shrunk to a point.
e 1 ( G ) = G.
G/Π (9.18)
In the following lectures we will also need that Π1 (SO(n)) = Z2 , for n > 2.
We will define then the spin group
^
Spin(n) ' SO ( n ). (9.19)
exercises
10.1 introduction
ϕα : M ⊃ Uα → Rα ⊂ Rn .
97
98 differentiable manifolds
ϕ1
Rn
U1 M
ϕ1 ◦ ϕ2−1
U2 ϕ2 ◦ ϕ1−1
Rn
ϕ2
Figure 15.: Example of a manifold with arrows exemplifying its local maps ϕ1 , ϕ2 to
Rn and the maps ϕ1 ◦ ϕ2−1 , ϕ2 ◦ ϕ1−1 giving the coordinate changes in the
overlapping region.
Definition 10.2. Two charts (U1 , ϕ1 ) and (U2 , ϕ2 ) are compatible if ϕ1 ◦ ϕ2−1 ∈
C ∞ and ϕ2 ◦ ϕ1−1 ∈ C ∞ (when U1 ∩ U2 6= ∅).
Finally, we want to cover the whole topological space M with charts, so that
we can give a coordinate in Rn to all its points.
This may not be a unique procedure and in this case we will say that two
atlases are compatible if all their charts are compatible.
We are now in a position that lets us introduce the concept of differentiable
manifold.
1. M is a topological space;
10.1.2 Properties
Now that we have introduced the notion of a manifold, we can start looking at
its properties and at the possibility of deforming different manifolds onto each
other. Since we are now interested also in the differential structure and not
just to the topological structure, we better introduce a new definition of map
between topological spaces that preserves also the differentiable properties.
Definition 10.6. A map f : M1 → M2 continuous and differentiable with in-
verse continuous and differentiable is a diffeomorphism.
Two manifolds related by a diffeomorphism are called diffeomorphic.
Other properties of manifolds should be decided looking at the underlying
topological space, but, thanks to their atlases, we can often analyze them by
looking at their image in Rn (with some care). This is the case for their com-
pactness or orientability.
Definition 10.7. A manifold is compact if for any atlas {Uα }α=1,...,∞ there is a
finite sub-collection {Uα }α=1,...,n covering the whole manifold.
For instance, in R we can always have a single open subset covering the
whole manifold, R itself, but if we choose the atlas given by {Uj =] j, j + 2[} j∈Z
there is no finite sub-collection covering the whole manifold.
100 differentiable manifolds
10.1.3 Examples
We now give some examples to illustrate the concepts introduced so far. First
of all let us introduce some obvious examples of manifolds: Euclidean spaces.
Example 10.1. Rn is a manifold with an obvious atlas given by a unique chart
with homeomorphism the identity map.
On such manifolds it is easier to see why the concept of compatible charts is
important.
Example 10.2. Let M = R. We can construct charts covering the whole manifold
by using the maps ϕ1 : R → R and ϕ2 : R → R such that ϕ1 ( x ) = x and
ϕ2 ( x ) = x3 . These are not compatible, because ϕ1 ◦ ϕ2−1 = x1/3 , which is not
C ∞ at x = 0.
It is very easy to produce examples of manifolds that are not simply Rn .
For instance, the n-sphere Sn , defined as the locus of all points at some fixed
distance from the origin of Rn+1 , is a manifold. The n-torus T n , resulting from
taking an n-dimensional cube and identifying points on opposite sides, is a
manifold. And we can easily continue.
It is therefore also interesting to understand what is not a manifold and for
this reason we now give a few simple examples.
The second one does the same, but in a way that covers the point we excluded
in the previous map. This means that now we map the open set U2 = S1 \
{(−1, 0)} to ] − π, π [⊂ R, so that also the homeomorphism ϕ2 : U2 →] − π, π [
has an inverse map that has the same form as before:
y y
ϕ1
2π π
θ x x
(-1,0)
(1,0) θ
0 ϕ2 −π
With this procedure we are assigning two different coordinates sets to two
different patches of the manifold. We should then require that the two corre-
sponding charts are compatible in the overlapping regions. The points that are
not covered by one of the two charts are at y = 0 in the R2 embedding. We
102 differentiable manifolds
then have two regions of overlap: one given by all the points with y > 0 and
one with all the points with y < 0.
When y > 0 the coordinate changes are trivial because ϕ1 ◦ ϕ2−1 (θ ) = θ and
ϕ2 ◦ ϕ1−1 (θ ) = θ. This effectively means that we do not really need to change
the coordinate assigned to the points of S1 in this region when we pass from
one chart to the other. However, when y < 0 we have ϕ1 ◦ ϕ2−1 (θ ) = 2π + θ and
ϕ2 ◦ ϕ1−1 (θ ) = θ − 2π, which means that we have assigned different coordinates
to the same points and we need to perform a coordinate change when we go
from one chart to the other. It is straightforward to see that both change of
coordinates are C ∞ .
Example 10.5. The sphere Sn = {~r ∈ Rn+1 | |~r | = 1} also needs two charts,
constructed generalizing the previous example. They come from the so-called
stereographic projection. Using once again the embedding of the sphere in a
higher-dimensional space, we can define the two charts as follows. From a pole,
calling ~x ∈ Rn the local coordinates we can use
1 − |~x |2
2~x
ϕ1−1 (~x ) = , ,
1 + |~x |2 1 + |~x |2
which covers the whole Sn except the point {0, 0, . . . , 1}, and
−1 + |~x |2
2~x
ϕ2−1 (~x ) = , ,
1 + |~x |2 1 + |~x |2
Useful structures on manifolds are vector fields and tangent spaces and we
now discuss how to define and construct them.
We are already familiar with the concept of tangent vectors to curves in Eu-
clidean spaces. We know that they can be constructed by evaluating derivatives
of the functions defining such curves. When we deal with a generic topological
space M that is also a differentiable manifold, we can construct the tangent
space to one of the points of M by generalizing this approach, describing the
vectors tangent to the curves passing through this point via its image in Rn .
A curve on M is a function
γ : R ⊃ I → M. (10.3)
10.2 tangent and cotangent space 103
xi ( p)
p γ(t) xi
Figure 17.: We can define tangent vectors to a manifold by looking at tangent vectors
to curves defined on the same manifold using their image in Rn .
When its image has a non-trivial intersection with the open subset of M that
defines a chart x : M ⊃ U → Rn , we can associate to the curve an image in Rn
with coordinates xi (γ(t)), i = 1, . . . , n (see fig. 17 for a representation of this).
The vector tangent to the curve γ at the point p = γ(t∗ ) ∈ M can be defined
via the tangent vector to its image:
d i
≡ V i ( p ).
x (γ(t)) (10.4)
dt t=t∗
f : M → R. (10.5)
where V i = ddt xi (γ(t)), so that the directional derivative of the test function f
along the curve γ can be obtained by application of the differential operator V:
d
V[ f ] ≡ f (γ(t)). (10.8)
dt
Using these differential operators we can eventually give an intrinsic definition
of vectors tangent to a curve.
xj
e i ( p) ∂e xi e j
∂e
V ( p ) = V i ( p ) ∂i = V
e i ( p) e
∂i = V ∂j ⇒ V i ( p) = V ( p). (10.9)
∂xi ∂x j
We can also generalize this concept in order to obtain a differential operator
defined on the whole manifold and not just on a curve. This is called a vector
field.
V (x)
Figure 18.: Integral curves following from the definition of the vector field V on the
manifold M.
Tp M
p
M
Figure 19.: The tangent space Tp M to a point p ∈ M can be defined by the equivalence
classes of directional derivatives along all the curves passing through p.
{ γ : R → M | γ (0) = p ∈ M }
Tp M ≡ . (10.12)
{γ1 ∼ γ2 ⇔ ddt γ1 (0) = ddt γ2 (0)}
The collection of all the tangent vector spaces at each of the points of M is
then the tangent space of the manifold.
Definition 10.12. The Tangent Space to a manifold M is defined as
[
TM = Tp M. (10.13)
p∈ M
106 differentiable manifolds
Although we will not make much use of it in this course, starting from Tp M
we can define its dual space Tp∗ M, called the cotangent space.
d i
x (t, x0 ) = V i ( x (t, x0 )), and xi (0, x0 ) = x0i , (10.15)
dt
where x0 ∈ M and xi are the local coordinates, is called the integral curve or
flow generated by the vector field V, passing through x0 at the time t = 0.
Theorem 10.1. Let V ∈ TM be a vector field on the manifold M. For any point
x0 ∈ M there is an integral curve of V, a flow x : R × M → M such that x (t, x0 ) is
a solution of the differential equation (10.15).
2. there is an identity: x0 = 1;
10.3 flows and exponential map 107
= exp(tV )[ x0i ],
where in the last step we used the definition (10.15).
We give here only a simple example of the construction of the tangent space
to a manifold and the reconstruction of the curves on the manifold via the
exponential map.
Example 10.6. Take M = R. The tangent space Tx0 R can be constructed by
considering the diffeomorphisms γ( x0 , t) = x0 + t. In fact
d f (γ( x0 , t)) − f (γ( x0 , t))
f (γ( x0 , t)) = lim
dt t →0 t
(10.18)
f ( x0 + t ) − f ( x0 ) d
= lim = f ( x0 ).
t →0 t dx
This means that ddt γ( x0 , t) = 1 = V x and the vector field is V = dx
d
.
We can get back the integral curve that originated V by the exponential map
of V:
d
exp(tV )[ x0 ] = exp t [ x0 ]
dx
(10.19)
d
= 1+t + . . . x = x0 + t = γ ( x0 , t ).
dx x = x0
108 differentiable manifolds
00
W ( p) V ( p )
p0 [V, W ]
p
V ( p) p0 W ( p0 )
Figure 20.: Geometrically the Lie bracket [V, W ] describes the non-commutativity of
the flows generated by the vector fields V and W.
The last concept we need to introduce is the Lie bracket. This is an operation
that associates a new vector field to the combination of two vector fields. It
expresses how a vector field changes along the flow generated by another vector
field. In detail, we have seen that there is a direct relation between a vector field
and its integral curves by means of the exponential map. If, starting from the
same point p ∈ M, we compose the result of integral curves coming from
different vector fields, the result will depend on the order. At the linearized
level, this should be represented once again by a vector associated to another
vector field and this is represented by the Lie brackets.
Definition 10.16. The Lie brackets are the commutator of two vector fields on
M:
[V, W ] = V i ∂i W j ∂ j − W i ∂i V j ∂ j =
= V i ∂i W j − W i ∂i V j ∂ j
(10.20)
= Z j ∂ j ≡ Z.
We do not discuss further this new ingredient at this stage, but we will come
back to it when we will discuss the relation between a Lie group and its tangent
space, defined by a Lie algebra.
exercises
φi−1 : Rn → RPn
( x0 , . . . , x̂i , . . . , xn ) 7→ [ x0 : . . . : |{z}
1 : . . . : x n ],
i
where x̂i signals the fact that we remove the coordinate xi from the set.
Compute the transition functions. Recall the stereographic projection of
the spheres and build an homeomorphism between RP1 and S1 . (One
can proceed in a similar way for CP1 and S2 or for HP1 , the projective
space on quaternions, and S4 ).
LIE GROUPS
11
LIE GROUPS.
In the first part of these lectures we mainly discussed discrete and finite groups.
We now want to consider the theory of continuous groups, where the group
elements depend continuously on a set of parameters.
The theory of continuous groups is much more complicated than the the-
ory of discrete groups. In this case the elements cannot be enumerated and
in addition to algebraic concepts we need topological notions like continuity,
connectedness and compactness and this is why we reviewed in the previous
lectures some of these notions. However, we will not discuss all possible types
of continuous group, but rather focus on a particular class of continuous groups
relevant for physical applications: Lie groups.
Definition 11.1. A Lie group is a smooth differentiable manifold with a group
structure such that the following group operations are differentiable (smooth):
• the composition: ◦ : G × G → G;
• the inversion map: G → G given by g 7→ g−1 .
The dimension of a Lie group G is determined by the dimension of G as a
manifold.
A Lie group is a special kind of differentiable manifold, where each point
is an element of a group and where different points may be related by group
operations.
Remember that our goal is not to give a full mathematical treatment of Lie
groups and their associated Lie algebras, but to show the results that are most
relevant for physical applications. For this reason we will often mention theo-
rems without demonstrations, just to inform the student of their existence, and
provide examples that are instructive to understand more general mathematical
results.
113
114 lie groups
Most of the theorems we proved in the first part of this course relied on carrying
out sums over the group elements, often in conjuction with the Rearrangement
Theorem 3.2. For instance, thanks to the rearrangement theorem, we often
replaced the sum over the group elements of a function of the elements of
the group, with the sum of the values of the same function evaluated on the
elements multiplied by the same g0 from the left
∑ f ( g0 g) = ∑ f ( h = g0 g) = ∑ f ( g ), (11.1)
g∈ G h∈ G g∈ G
∑ f ( gg0 ) = ∑ f ( g ), (11.2)
g∈ G g∈ G
∑ f ( g −1 ) = ∑ f ( g ). (11.3)
g∈ G g∈ G
This theorem implies that only for compact groups we can safely extend our
proofs of orthogonality for the matrix elements of their representations, for the
equivalence of the group representations to unitary ones, etc. One needs more
care when discussing non-compact groups. For instance, one may still have
left-invariant or right-invariant measures, but their integral on the group may
diverge.
The groups of invertible matrices play a special role among Lie groups for two
reasons:
1n . The only constraint needed to have a group is that the matrices in GL(n, C)
should have non-vanishing determinant, so that we can invert them. This does
not constrain single entries of the matrix and therefore
and
Recall that we discussed real manifolds and therefore we are giving here their
real dimension.
If we restrict ourselves to the group of matrices with unit determinant we
can define
Definition 11.3. The Special Linear group
The proof that this is still a group is straightforward because det ( AB) =
det A · det B, det 1n = 1 and det ( A−1 ) = 1/det A.
Since one of the entries of the matrices in SL(n, F) is fixed in terms of the
others because of the constraint on the determinant we have that
and
Starting from these groups we will now see a series of other groups defined
as subsets satisfying linear or quadratic constraint. The matrices resulting from
these constraints provide the so-called fundamental representation of a group.
As we will see in the following, this is not the only irreducible representation
existing for such groups, but it is fundamental in the sense that the others could
be obtained by decomposing its products (in general a Lie group has an infinite
number of irreducible representations).
Composition of matrices of this form close on matrices of the same form. The
identity is part of UT( p, q) and the inverse is also in the group
A− 1
− A− 1 −1 !
−1 p× p p× p B p×q Cq×q
m = . (11.11)
Oq × p Cq−×1q
Definition 11.6. The subgroup of matrices of Solv(n) having all diagonal ele-
ments equal to 1 is called Nilpotent group, Nil(n).
We can also restrict invertible matrices by requiring that they preserve another
matrix (a metric), that is not necessarily part of the group. We then introduce a
quadratic constraint, where matrices M ∈ GL(n, F) satisfy also a constraint of
the form M T η M = η, or M† η M = η, or M∗ η M = η, for η a matrix that may be
singular or non-singular, antisymmetric or symmetric and, if symmetric, posi-
tive definite or indefinite. All these distinct cases will generate distinct groups,
provided that the constraint allows for group composition to be closed and for
the inverse to satisfy the same constraint (M = 1 always satisfies quadratic
constraints like those mentioned above).
Compact groups
Definition 11.7. The Unitary group
U(n) ≡ { M ∈ GL(n, C) | M† 1n M = 1n }.
118 lie groups
Non-compact groups
If we consider the same type of constraints as those given above, but now for
an indefinite metric of the form
1p
!
η= (11.17)
−1q
Definition 11.11.
U( p, q) ≡ { M ∈ GL( p + q, C) | M† η M = η }.
Definition 11.12.
O( p, q) ≡ { M ∈ GL( p + q, R) | M T η M = η }.
It is very easy to see that the set of parameters labeling the elements of these
groups becomes non-compact. For instance,
!
cosh ψ sinh ψ
= m ∈ O(1, 1), ∀ψ ∈ R. (11.18)
sinh ψ cosh ψ
On 1n
!
where Ω = .
−1n On
1
η= −1q ,
Or
We will not discuss all possibilities here. We close this lecture mentioning
the fact that some of these groups with different definitions are actually iso-
morphic.
For instance, a matrix M ∈ Sp(2,R) has unit determinant. In detail
a b
M= ∈ Sp(2, R) if and only if
c d
a c 0 1 a b 0 1
= (11.19)
b d −1 0 c d −1 0
and this gives ad − bc = 1. But this is the definition of a matrix in the special
linear group of 2-dimensional real matrices and therefore
exercises
1. Take the fundamental representation of SO(2) and show that Zn < SO(2).
The matrices you obtained provide a representation of Zn . Is it reducible?
2. Show that Dn < O(2) by using appropriate matrices in the fundamental
representation of O(2).
3. Consider the matrices in the fundamental representation of SO(3) and
show that S4 < SO(3).
Hint: look for a transformation corresponding to a 3-cycle and one corresponding
to a 4-cycle.
4. Let (qi , pi ), i = 1, . . . , n, be the coordinates and momenta of a classical me-
chanical system with n degrees of freedom. Poisson brackets are defined
as
n
∂ f ∂g ∂g ∂ f
{ f , g} = ∑ i ∂p
− .
i =1 ∂q i ∂qi ∂pi
a) Show that {qi , q j } = { pi , p j } = 0 e {qi , p j } = δji .
b) Show that if we redefine the coordinates with a linear transformation
Q q
=S ,
P p
6. Using the results of the previous exercise show that conjugate elements
of SO(3) are characterized by the angle specifying the rotation.
7. This is just for fun. Use what you just learned to find the conjugacy classes
of S4 .
10. Given M ∈ Sp(2n, R), show that M−1 ∈ Sp(2n, R) and also M T ∈
Sp(2n, R), i.e. prove that if M T ΩM = Ω, then M−T ΩM−1 = Ω as well as
MΩM T = Ω.
1
L = −∂µ Φ† ∂µ Φ + Φ† ηΦ − (Φ† Φ)2 ,
2
φ1 ( x )
φ2 ( x )
where Φ is a complex scalar field with 4 components Φ( x ) =
φ3 ( x )
φ4 ( x )
e η = diag{−1, −1, 1, 1}.
a) What is the group of linear transformations Φ( x ) → SΦ( x ), with
S ∈ Mat(4, C), preserving the kinetic term and the potential, respec-
tively?
122 lie groups
c1
c2
b) For which constant values Φ0 = c3 the Euler–Lagrange equa-
c4
tions are satisfied and what is the residual symmetry group?
(Otherwise, which matrices S satisfy Φ0 → SΦ0 = Φ0 )
12
LIE ALGEBRAS.
As already done for finite groups, also for Lie groups we would like to have a
rapid way to understand their structure and their possible relations. Given their
double nature of groups and manifolds, in order to specify an isomorphism we
need to have their equivalence as topological spaces as well as groups. We
then need to understand when we can deform two groups onto each other in a
continuous way, so that the composition laws are equivalent up to a change of
variables. This last requirement may be very complicated because composition
laws are not necessarily linear. Also, we would like to compare only a finite
number of elements if possible, rather than the infinite number that is defining
a continuous group. For these reasons we now introduce the notion of Lie
algebra, which is going to help us in this tasks.
A Lie algebra is the linearization of the corresponding group around the
identity element. It contains the information that is needed to reconstruct the
group in a neighborhood of the neutral element. Actually, the first useful thing
to note is that linearizing the group G at any of its elements gives the same
result. In fact, since G is a manifold endowed with a group structure, we can
map any element g to any other g0 using the composition of the first with
h = g0 g−1 and in the same way we can map any open neighborhood of g, Ug ,
to an open neighborhood of g0 , Ug0 , by multiplying points of Ug with h from
the left. Since the composition map on a Lie group is a continuous function,
we can do the same for the curves in Ug and eventually for the directional
derivatives defining the tangent space in g. The result is that we can focus on
the tangent space to the group G at the identity without losing in generality,
but simplifying some of the calculations because of the properties of the neutral
element.
123
124 lie algebras
Tg G g 0 ◦ g −1
g g0
Tg0 G Ug0
Ug
G
Figure 21.: Tangent spaces at different points on a group manifold can be related by
group multiplication.
The general procedure to construct the tangent space requires the identification
of curves M (t) ∈ G, passing through the element around which we want to
construct the tangent space, which is the identity. Since we do not want to enter
in a mathematical discussion that would be useless for our purpose, we will
focus on groups of matrices, though obvious generalizations apply to generic
Lie groups. We therefore choose M (t) as a 1-dimensional subgroup of the
group of matrices G, so that
M (0) = 1 (12.1)
and
M ( t1 ) M ( t2 ) = M ( t1 + t2 ), (12.2)
which implies
We then provided explicit maps between a group and its algebra in both direc-
tions:
d
dt
−→
G g
←−
exp
When G = GL(n, C), both maps are easy to compute. In particular, for M ∈
GL(n, C), we can easily obtain the tangent space at the identity by considering
the expansion
M = 1+em+..., e 1, (12.8)
the space of square complex matrices. The inverse operation gives back the
original group close to the identity via the exponential map. By exponentiation
we can indeed obtain n2 different 1-parameter subgroups of GL(n, C):
∞
tk mk
M(t) = exp(tm) = ∑ k!
, (12.10)
k =0
126 lie algebras
where mk is defined by the usual matrix product. This series is always conver-
gent in the space of matrices and we can easily check that M(t) is in fact in
GL(n, C). Its inverse is simply M(t)−1 = M(−t) = exp(−tm). We therefore
conclude that
Definition 12.1. The Lie algebra of the general linear group is
Also for the Special Linear group we can obtain its algebra by expanding the
constraint defining its fundamental representation around the identity:
Definition 12.2.
sl(n, F) = { A ∈ Mat(n, F) | trA = 0}.
In the case of Unitary groups an analogous expansion gives
(1 + e A + . . . ) † (1 + e A + . . . ) = 1 + e ( A + A † ) + . . . , (12.12)
for e 1 and real (any phase can be reabsorbed in the definition of A).
Definition 12.3.
u(n) = { A ∈ Mat(n, C) | A† = − A}.
Obviously dim U(n) = dim u(n) = n2 . This is easy to check for the algebra,
where the n diagonal entries are purely imaginary and the n(n − 1)/2 entries
above the diagonal are complex conjugate of those below the diagonal. We then
n ( n −1)
have only 2 × 2 + n = n2 real degrees of freedom.
Since we obtained the conditions that matrices in the algebra should satisfy
by looking at the linear expansion around the identity, we may ask whether
their exponential, which goes beyond linear level, is really in the group. This
is straightforward to verify because of matrix properties:
where we used the definition of the exponential of a matrix in the first equality
and the fact that A ∈ u(n) in the second equality.
Note also that in Quantum Mechanics the relation between elements of the
Unitary group and the algebra is usually done by considering the exponen-
tial of Hermitian operators, which are related to our group generators by an
imaginary unit.
For the orthogonal group and the special orthogonal group we get that
(1 + e A + . . . ) T (1 + e A + . . . ) = 1 + e ( A + A T ) + . . . , (12.14)
which means that matrices in their algebras are antisymmetric and therefore
with n(n − 1)/2 degrees of freedom. In addition, for the special orthogonal
group we should add the constraint giving unit determinant matrices, but this
is given by the vanishing of the trace, which is guaranteed by their antisymme-
try.
Definition 12.4.
Note that, as expected, the algebras described above have the structure of a
vector space, but differently from the corresponding groups, they do not close
under matrix multiplication. For instance, if A, B ∈ so(n), AB ∈
/ so(n), because
Definition 12.5.
n o
sp(2n, R) = A ∈ Mat(2n, R) | A T Ω = −ΩA ,
1
0
where Ω = .
−1 0
Since Ω T = −Ω, we can also rewrite the constraint as (ΩA) T = ΩA, and
this means that the number of parameters defining the matrices A is equal to
that of 2n × 2n symmetric matrices. Hence
• If we have isomorphic algebras (in a sense that we will specify soon), are
the corresponding groups isomorphic?
We will see that unfortunately the answer to all these questions is NO. However,
the exponential map becomes surjective if the group is connected and compact and
it becomes injective if G is simply connected. Hence we conclude that using the
exponential map we can fully reconstruct a group that is simply connected to
the identity and compact.
12.2.1 Surjectivity
The counterexample has been found by Cartan analyzing the relation between
SL(2, R) and sl(2, R). If A is a matrix in sl(2, R), it has vanishing trace and can
be parameterized by three real parameters a, b, c:
a b+c
A= . (12.17)
b − c −a
The generic matrix in SL(2, R) that can be obtained from an element of the
algebra by exponentiation is going to be exp( A), for an arbitrary choice of
a, b, c. We will now show that such elements, however, do not cover the whole
group. The reason is that they satisfy a constraint on the trace that some of the
elements in the group do not fulfill. In fact, any matrix A of the form (12.17)
can be diagonalized
−1 λ
SAS = , (12.18)
−λ
12.2 from the algebra g to the group G 129
√
by means of the similarity matrix S. Here ±λ = ± a2 + b2 − c2 are the eigen-
values of the matrix A. They are real if a2 + b2 ≥ c2 and purely imaginary
otherwise. Using once more the properties of the exponential map for matri-
ces, we see that
This, on the other hand, is a consequence of the fact that SL(2, R) is connected
and cannot be extended to other groups (for instance in O(n) we have two
disconnected components and we will never get matrices with determinant −1
by exponentiating matrices in the algebra, nor by taking their products).
We already saw that so(n) = o(n), but we can also construct identical algebras
that generate both compact and non-compact groups. This is the case of o(2) '
u(1) ' R. All these algebras have a unique element, which we can choose to
130 lie algebras
0 1
be X = for o(2), X = i for u(1) and X = 1 for R. The group
−1 0
elements are obtained by taking exp(θX ). For U(1) and SO(2) we get that for
any θ = c + 2π Z we get the same element of the group, while for (R+ , ·) the
map is injective.
t2 2
= 1 + t( A + B) + ( A + B2 + 2AB) + O(t3 )
2
t2 t2
= 1 + t( A + B) + ( A + B)2 + [ A, B] + O(t3 )
2 2
t2
= exp t( A + B) + [ A, B] + . . . , (12.24)
2
where ( A + B)2 = A2 + B2 + AB + BA and
[ A, B] ≡ AB − BA. (12.25)
We see that in order to determine the new element of the group from the prod-
uct of exponentials of elements of the algebra we need to know also their com-
mutator, no matter how close we are to the identity.
Note also that
u log u
where ψ(u) = u −1 = 1 + (u − 1) + 61 (u − 1)2 + . . . is regular in u = 1 and
ad X Y ≡ [ X, Y ]. (12.27)
e A Be− A = exp( ad A ) B.
A2 A2
1+A+ +... B 1− A+ +...
2 2
1 2
= B + AB − BA + A B − 2ABA + BA2 + . . . (12.28)
2
= ( B + [ A, B] + [ A, [ A, B]] + . . .) = exp([ A, ·]) B.
[·, ·] : g × g → g
which is
132 lie algebras
1. bilinear: ∀ x, y, z ∈ g and a, b ∈ F
2. antisymmetric
[ x, y] = −[y, x ], ∀ x, y ∈ g;
[ X, [Y, Z ]] + [Y, [ Z, X ]] + [ Z, [ X, Y ]] = 0.
The map we just introduced is simply the commutator for matrices, but in
the general case it is an operation that we can define starting from Lie brackets.
In full generality we can define g from G as its tangent space at the identity T1 G.
This is a vector space on which we can introduce an operation associating a new
vector of the tangent space [ X, Y ] to any other couple of vectors X, Y ∈ T1 G by
taking the generator of the vector field that is obtained from the Lie brackets of
the vector fields generated by the other two. Operatively, we introduce exp(tX )
and exp(tY ), compute the Lie bracket [exp(tX ), exp(tY )] and expand for t → 0.
The element we obtain is [ X, Y ].
exercises
Lie algebras are first of all vector spaces. As such, for any Lie algebra g we
can introduce a basis of n = dim g = dim G vectors {t I } I =1,...,n , so that any
element of g can be written as their linear combination. Lie algebras, however,
are also closed under the product defined by the map in 12.6 (which we will
call commutator in the following). Because of its vector space nature, all the
information contained in the commutator of two generic elements is fixed by
that of the basis vectors:
[t I , t J ] = f I J K tK . (13.1)
Definition 13.1. The coefficients f I J K in (13.1) are called structure constants.
Using the linearity properties of the commutator, we can always express the
result of the commutator of any couple of elements of the algebra X, Y ∈ g in
terms of the structure constants. In detail, we can expand the generic vectors
X and Y as X = x I t I , Y = y J t J , where x I , y J ∈ F, the field on which the algebra
is defined, and do the same for the vector Z, associated to the commutator:
[ X, Y ] = x I y J [t I , t J ] = x I y J f I J K tK = Z ∈ g. (13.2)
From the relation (13.1), defining the structure constants, we can also deduce
some of their properties. First of all, they must be antisymmetric in the first
two indices
fI J K = − fJI K. (13.3)
Then, expanding the Jacobi identity 12.7 in terms of the basis vectors, we can
also deduce that the structure constants must satisfy
1 L
0 = f [ I J L f K] L M = f I J f KL M + f KI L f JL M + f JK L f IL M . (13.4)
3
135
136 lie algebras – properties
Example 13.1. The algebras u(1) ' R are generated by a unique element t1 and
all their structure constants are vanishing.
Example 13.2. The algebra of the Heisenberg group is defined by 3 generators,
with a unique non-trivial commutator:
[ t1 , t2 ] = t3 . (13.5)
[t I , t J ] = e I JK tK . (13.6)
From this introduction, we see that a Lie algebra is specified by the dimen-
sion of the vector space and by the commutator relations. Hence a classification
of inequivalent structure constants allows to classify inequivalent algebras. If
we want to compute two algebras, though, we should also remember that we
can always choose different basis for the same algebra. This means that we can
always change the values of the structure constants by taking linear combina-
tions of the generators:
t̃ I = ∑ c I K tK . (13.7)
K
Since this is just a redefinition of the basis of vectors generating the vector space
g, it does not change the algebra. For this reason, if two algebras g and e g are
connected by a change of basis like (13.7), we say that they are isomorphic (we
will give a more appropriate definition in 14.2). One should be careful on the
choice of the coefficients c I K , though. The algebra g is defined a vector space
on a specific field F, which is usually taken to be C or R. We should then
only take linear combinations where the coefficients c I K are in the same field,
otherwise we are effectively changing the nature and dimension of the algebra.
Later we will see that sometimes establishing an isomorphism between dif-
ferent real Lie algebras using complex change of coordinates can be useful to
study their representations. For this reason we introduce another useful con-
cept: the complexification of a real Lie algebra.
and
so(2, 1) = {span(t̃1 , t̃2 , t̃3 ) , [t̃1 , t̃2 ] = t̃3 , [t̃2 , t̃3 ] = t̃1 , [t̃3 , t̃1 ] = −t̃2 } (13.9)
are not isomorphic as real algebras, but we can map them onto each other if
we allow the use of complex coefficients:
We then say that so(3) and so(2, 1) are different real forms of the same complex
algebra so(3)C ' so(2, 1)C ' sl(2, C).
Definition 13.3. A real form gR of the complex algebra gC is obtained by con-
sidering a complex basis {t I } ∈ gC , I = 1, . . . , dimC gC and restricting the
vector space to the one generated by their real combinations.
Given an algebra g, we can always consider some subspace that is again a vector
space and we can ask whether the result is also an algebra.
Definition 13.4. A subspace h ⊂ g is a subalgebra of g if it is closed with respect
to the bracket operation in g, i.e.
∀ x, y ∈ h, [ x, y] ∈ h.
We can now analyze how various group properties are translated into prop-
erties of their algebras. For instance, in 4.11 and 4.12 we introduced simple
and semi-simple groups. The corresponding definitions for the algebras are
the following:
Definition 13.7. A Lie algebra g is simple is it does not contain any ideal.
Definition 13.8. A Lie algebra g is semi-simple if it does not admit any Abelian
ideal.
Note that while it is obvious that
G simple ⇒ g simple, (13.11)
This is what happens for instance for u(1) = R, so(3) = su(2), so(1, 3) =
sl(2, C), where R = U ](1), SU(2) = SO^ (3), SL(2, C) = SO^(1, 3), but obviously
R 6= U(1), etc.
Summarizing, for any group G there is an algebra g that is obtained as its
tangent space at the identity g = T1 G, while for any algebra g there is a unique
simply connected group G that has g as its tangent space at the identity T1 G.
All the other connected groups K that have g as algebra are of the form
K/H = G, (13.15)
We can understand even more properties of algebras and the relations with
properties of the groups that generated them by introducing a new object: the
Cartan–Killing form.
where X, Y ∈ g.
ad X (t I ) = t J D ( ad X ) J I , (13.16)
where
D ( ad X ) J I ≡ x L f LI J . (13.17)
( X, Y ) = k I J x I y J , (13.18)
140 lie algebras – properties
where
k I J = k J I = f I M L f JL M . (13.19)
Note also that
k I J = ( t I , t J ). (13.20)
Using this form one can prove the following very useful theorem:
Theorem 13.2. Cartan’s criteria:
1. A Lie algebra is semi-simple if and only if det k 6= 0;
2. g semi-simple, real and compact if and only if k is negative definite.
Example 13.4. The algebra so(3) has non-vanishing structure constants f 12 3 =
f 31 2 = f 23 1 = 1 and therefore k11 = f 12 3 f 13 2 + f 13 2 f 12 3 , etc... so that
−2
k= −2 . (13.21)
−2
We see immediately that det k = −8 6= 0 and that it is negative definite. In fact
so(3) is semi-simple and compact.
Example 13.5. In heis(3) the only non-vanishing structure constants are f 12 3 =
− f 21 3 = 1. The Cartan–Killing matrix is then
0
k= 0 , (13.22)
0
which is obviously degenerate. The algebra is not semi-simple. It is straight-
forward to see that a = span(t3 ) is a 1-dimensional vector space that is also an
Abelian subalgebra of g:
[a, g] = 0 ⊂ a ⊂ g. (13.23)
Example 13.6. For so(2, 1) the structure constants are f 12 = f 13 = f 23 1 = 1,
3 2
which gives
2
k= −2 . (13.24)
2
This is non-degenerate, but it is not definite. The algebra is semi-simple, but
not compact.
13.4 casimir operator 141
C2 ≡ k I J t I t J .
Theorem 13.3.
[C2 , X ] = 0, ∀ X ∈ g.
Proof. The theorem follows if we prove that C2 commutes with all the basis
vectors t I .
[C2 , tk ] = k I J [t I t J , tK ] = k I J t I [t J , tK ] + k I J [t I , tK ]t J
= k I J f JK L t I t L + k J I f IK L t L t J (13.25)
= k I J f IK L t J t L + t L t J = k I J k LP f IKP (t J t L + t L t J ),
f I JL ≡ f I J K k KL . (13.26)
exercises
1. Verify that (13.4) follows from the Jacobi Identity for three vectors defin-
ing a basis of the algebra.
2. Compute the Cartan–Killing form for su(1, 1) and su(2) and show that
they are different real forms of sl(2, C).
3. The algebra su(3) in the fundamental representation is determined by
Gell-Mann matrices
0 1 0 0 −i 0 1 0 0
λ1 = 1 0 0 , λ2 = i 0 0 , λ3 = 0 −1 0 ,
0 0 0 0 0 0 0 0 0
0 0 1 0 0 −i 0 0 0
λ4 = 0 0 0 , λ5 = 0 0 0 , λ6 = 0 0 1 ,
1 0 0 i 0 0 0 1 0
0 0 0 1 0 0
1
λ7 = 0 0 − i , λ8 = √ 0 1 0 ,
0 i 0 3 0 0 −2
via Ta = 2i λ a so that Ta† = − Ta and tr( Ta Tb ) = − 12 δab .
a) Compute f abc .
b) Show that su(3) has a proper su(2) ⊕ u(1) subalgebra.
4. Let G = SO(2,2) and its algebra g = so(2, 2).
a) Provide dim g matrices for the basis of elements of g in the funda-
mental representation.
b) Show that H = SO(2) × SO(2) < G and that the two elements a1 , a2 ∈
g corresponding to the generators of the two SO(2) commute.
c) Are there other elements of g that are not in the span of a1 and a2
and that commute with them?
5. Let A = Sp(4, R) and B = SO(4).
a) Give their dimension and define their algebras in terms of matrices.
b) Discuss C = A ∩ B by analyzing a and b.
c) Compute the Cartan–Killing metric of the corresponding algebra c
and discuss if C can be simple, semi-simple and/or compact.
14
R E P R E S E N TAT I O N S O F L I E A L G E B R A S
Since Lie algebras are first of all vector spaces, invertible linear maps respect-
ing the structure of the commutators are changes of basis.
143
144 algebra representations
but
g1 ' g2 ⇒
/ G1 ' G2 . (14.2)
D : g → gl (n, C).
on the spaces of functions on which they act). In the following we will not
be careful about the definition of a proper space of functions on which these
operators act (be it L2 (R) or C ∞ or tempered distributions or else), but always
assume that the operations are legitimate. We are only interested in showing
how some of the most common groups and algebras in physics can be repre-
sented in this way.
We already met the first example, which is the group of translations, isomor-
phic to (R, +). A representation for its 1-dimensional algebra is given by the
derivative of real functions:
d
D ( t1 ) = . (14.7)
dx
∂ ∂
Tij ≡ xi j
− xj i . (14.10)
∂x ∂x
[ Tij , Tkl ] = −δik Tjl + δjk Til − δjl Tik + δil Tjk , (14.11)
In the simplest case, namely so(2), we have only one such generator T =
x ∂y − y ∂ x and this can be simplified using polar coordinates x = r cos θ, y =
r sin θ, leading to
∂ ∂x (r, θ ) ∂y(r, θ )
T= = ∂x + ∂y = −r sin θ ∂ x + r cos θ ∂y . (14.12)
∂θ ∂θ ∂θ
For so(3) we get the expected commutation rules if we define
so that
The same type of operators could be easily introduced for so( p, q), defining
where
1p
!
η= . (14.16)
−1q
The commutator of the generators (14.15) can also be computed rather easily
and gives
The realization of Lie algebras using bosonic operators is very useful in Physics
in the resolution of problems like the harmonic oscillator in Quantum Mechan-
ics and in Quantum models of rotational and vibrational models of nuclei and
molecules.
Assume that we have n bosonic operators bα , α = 1, . . . , n, satisfying the
following commutation relations
We can think of these operators as creation and destruction operators for quan-
tum states, starting from a vacuum satisfying bα |0i = 0.
148 algebra representations
This is clear if one notes that they are in 1-to-1 correspondence with the matrices
representing the same group. For n = 2 for instance
1 0 0 1
T11 = , T12 = ,
0 0 0 0
(14.21)
0 0 0 0
T21 = , T22 =
1 0 0 1
satisfy exactly the same commutation relations and therefore we can establish
an isomorphism between the two algebras. Note also that these generators
can be further split into those of sl(n, C), whose matrix representation has
vanishing trace, plus the identity matrix, that commutes with all the others,
providing a “linear Casimir” C1 = ∑i Tii .
Recall that sl(n, C) is the complexification of su(n) and gl(n, C) is the com-
plexification of u(n), which are their real sections. This implies that we can
relate their representations by taking appropriate real linear combinations.
We now provide some examples that hopefully will clarify the role of these
symmetry generators in Physics.
where, according to the discussion above, T11 is the generator of gl(1, C). This
algebra is obviously Abelian and hence all its irreducible representations are
1-dimensional and labeled by the value of the action of the linear Casimir T11
on the corresponding invariant subspaces of the Hilbert space. A basis of states
generating these invariant spaces is given by
1
| N i = √ ( b † ) N |0i. (14.25)
N!
Since the Hamiltonian is given in terms of the linear Casimir, we can immedi-
ately deduce its eigenvalues and their degeneracies, computing its values on
the invariant spaces of the Hilbert space:
1
H|Ni = N + | N i. (14.26)
2
We now move to the 2-dimensional harmonic oscillator. In this case the Hamil-
tonian is
1 2 1
H= ( p̂ + p̂2y ) + ( x̂2 + ŷ2 ) = b1† b1 + b2† b2 + 1, (14.27)
2 x 2
where
1 1
b1 ≡ √ ( x̂ + i p̂ x ), b2 ≡ √ (ŷ + i p̂y ). (14.28)
2 2
Using bα and bα† operators we can construct the 4 generators of the gl(2, C)
algebra:
T11 = b1† b1 , T12 = b1† b2 , T21 = b2† b1 , and T22 = b2† b2 . (14.29)
1 †
|Ni = b . . . bᆠN |0i, (14.30)
N α1
where N is some normalization factor, and they form an irreducible and com-
pletely symmetric representation of dimension N + 1 (We will come back on
this in section 16.2.3). Once again we see that the Hamiltonian is a function of
150 algebra representations
the linear Casimir N b = b† b1 + b† b2 . This implies that its eigenvalues and their
1 2
degeneracies are also determined by such Casimir.
Note that usually one speaks of the symmetry generators of the harmonic
oscillator as generators of su(n) rather than sl(n, C). The reason is that we
are really interested in unitary transformations that preserve the orthonormal-
ity of the state vectors and this poses obvious restrictions on the actual linear
combinations of the Tij operators that generate admissible symmetries.
and
generate gl(n, C), but now the states forming a basis of the Hilbert space on
which they act are in fully antisymmetric representations
1 †
|Ni = a . . . a†α N |0i. (14.34)
N α1
The check is straightforward also in this case, but pay attention to the commu-
tators:
exercises
2. What about ( TI ) J K = − f I J K ?
Explicitly: consider the matrices TI , with non-zero elements as given above and
check the commutator [ TI , TJ ].
4. Let
D ( Jµν ) = −ηµρ x ρ ∂ν + ηνρ x ρ ∂µ , D ( Pµ ) = ∂µ ,
where µ, ν, ρ = 0, 1, 2, 3. Find the structure constants of the resulting alge-
bra.
commute with all the elements Tij and therefore they are Casimir
operators of degree r.
Try this at least for r = 1, 2, 3.
(1) (2)
Split the so(4) generators in Li = { T23 , T31 , T12 } and Li = { T14 , T24 , T34 }.
( A) ( B)
Compute [ Li , L j ].
(1) (2)
Define Li± ≡ 12 Li ± Li and close the argument.
Show that the generators surviving the projection ∑ j,k φijk Jjk = 0 define a
subalgebra of so(7) (i.e. consider the vector space generated by the linear
combinations of the Jij generators orthogonal to the 7 vectors defined by
∑ j,k φijk Jjk ). Compute its structure constants and the Cartan–Killing form.
Is this one of the algebras we studied so far (so(n), su(n), sp(n))?
15
R E P R E S E N TAT I O N S O F S U ( 2 ) A N D S O ( 3 )
The relevance of the group associated to the invariance of a physical system un-
der spatial rotations in 3 dimensions is obvious. This group can be determined
as the set of transformations preserving the norm of 3-dimensional vectors
~ → RV,
V ~ |V ~ T R T RV
~ |2 → V ~ |2
~ = |V (15.1)
0 0 1 0 −1
D3 ( l1 ) = 0 − 1 , D3 ( l2 ) = 0 , D3 ( l3 ) = 1 0 .
1 0 −1 0 0
(15.3)
153
154 representations of su(2) and so(3)
∑(xi )2 = 1, (15.7)
i
are anti-hermitian D2 (li )† = − D2 (li ) and have vanishing trace. We then con-
clude that D2 (li ) is a 2-dimensional representation of su(2) whose commutators
define the commutators of the algebra. It is also easy to see that the resulting
structure constants are isomorphic to the ones of so(3):
[ D2 (li ), D2 (l j )] = eijk D2 (lk ). (15.11)
We have then established that
su(2) ' so(3). (15.12)
We can now prove that by taking the exponential map of this 2-dimensional
representation of su(2) we can generate back the 2-dimensional representation
of SU(2) that covers the whole group.
A generic element of the algebra can be given in terms of the D2 representa-
tion by A = − 2i θ i σi ∈ su(2). Its exponential gives
∞ n
i i 1
U = exp( A) = ∑ − θ σi
n =0
2 n!
(15.13)
i i 1 1 i j
= 1 − θ σi − θ θ σi σj + . . .
2 2 4
and it can be simplified using that
1 i j 1
θ i θ j σi σj = θ θ σi , σj = θ i θ j δij 12 ,
(15.14)
2 2
which implies
i i 1 1 i 1
U = 12 − θ σi − θ i θ j δij 12 + − θ k σk − θ i θ j δij + . . .
2 8 3! 2 4
! !
1 |~θ |2 i k 1 |~θ |2
= 12 1− + . . . − θ σk 1 − +...
2 4 2 3! 4
|~θ | θi |~θ |
= 12 cos −i σi sin .
2 |~θ | 2
(15.15)
Matching this expression with the one for U given in (15.5), we see that
|~θ | ~θ |~θ |
x4 = cos and ~x = − sin , (15.16)
2 |~θ | 2
156 representations of su(2) and so(3)
Explicitly
i i~
exp − ~θ ·~σ σa exp θ ·~σ = exp ad− i ~θ ·~σ [σa ] =
2 2 2
−θ 3 θ2
0
θ3 θ1 − i θ2
i
exp − 7→ exp θ 3 0 −θ 1 . (15.20)
2 θ + i θ2
1 −θ 3
−θ 2 θ1 0
In fact we can see that the map in (15.20) is not a bijection, because it is not
invertible. Let us consider the simplified case where θ 1 = θ 2 = 0. The map
relates elements of SU(2) and SO(3) as
cos θ 3 − sin θ 3
!
3 /2
e−iθ
3 /2 7→ sin θ 3 cos θ 3 (15.21)
eiθ 1
and this means that different elements in SU(2) may have the same adjoint
representation and therefore be mapped to the same element in SO(3). A simple
instance is obtained by looking at the elements associated to the values θ3 =
0 and θ3 = 2π. While in SU(2) these two elements are distinct and the 2-
dimensional faithful representation shows this explicitly, their 3-dimensional
representation is the same. In fact in SO(3) there is a unique element associated
to such values of the parameters. Explicitly, at θ 3 = 0 we have
while at θ 3 = 2π we have
From (15.24) we see that topologically SO(3) is equivalent to the 3-sphere where
points are identified by a Z2 projection.
158 representations of su(2) and so(3)
x4 < 0 x4 > 0
Z2
Figure 22.: Z2 map identifying points on the 3-sphere. Contractible and non-
contractible cycles.
π1 (SO(3)) = Z2 . (15.25)
This implies that we have 2 classes of equivalence for loops in SO(3): the class
of contractible loops and the class of loops that cannot be contracted to a point.
A pictorial explanation of this fact is the following. The SO(3) manifold is the
set of rotations in 3 dimensions. Any element g ∈ SO(3) can be described by a
rotation of an angle between −π and π around an axis, oriented in a specific
direction in R3 . Obviously rotations by π must be identified with rotations by
−π. This means that we can describe SO(3) as a closed ball in 3 dimensions
of π radius, where opposite points on the surface must be identified. We can
now easily see that a loop that starts from a point on the boundary and runs
in the interior up to the opposite point on the surface of the ball is a closed
loop in SO(3) that can never be contracted to a point. On the other hand if we
have a loop that starts from a point on the boundary, runs in the interior up
15.4 irreducible representations of su(2) and so(3) 159
to another point on the boundary, different from the first, and does the same
thing on the opposite side, we can continuously deform it to a closed loop that
can be contracted to a point by moving different points on the boundary on top
of each other (see fig. 23).
Figure 23.: Contraction of a loop that passes twice through the boundary of the ball
representing SO(3). Red and green dots are points on the boundary, which
are identified. First we move the red dot and its image on top of the green
ones and then we can contract the loop.
This is just a special instance of what happens for any SO(n) group. All
SO(n) spaces are not simply connected and their universal coverings, which
are simply connected, are the Spin groups Spin(n). We conclude that
What we just presented shows that SU(2) and SO(3) representations are related
and that SU(2) representations provide representations for the universal cover-
ing group of SO(3).
In Quantum Mechanics we are interested in unitary representations for the
group elements and hermitian representations for the algebra. The homomor-
phism between SU(2) and SO(3) shows that irreducible representations of SU(2)
are identical to irreducible representations of SO(3) up to a sign. In fact, if we
still want to insist and think of D2 as a representation of SO(3), we can think
of it as a projective representation, whose composition law depends on a phase
(which is invisible in the ray defining a state):
What we can infer from this is that SO(3) admits genuinely projective represen-
tations and that its covering group automatically takes them into account. In
fact, for all practical applications we can actually study the representations of
the latter. For this reason we will now discuss representations of SU(2) and of
its algebra rather than those of SO(3).
In general, a series of theorems by Bargmann1 shows that
Theorem 15.1. A simply connected group without continuous Abelian normal sub-
groups does not admit genuinely projective irreducible representations.
This means that whenever the group we want to represent is not simply
connected, we can use its covering group.
J± ≡ i (l1 ± i l2 ), J3 ≡ i l3 . (15.28)
[ J3 , J± ] = ± J± , [ J+ , J− ] = 2 J3 (15.29)
and the isomorphism of this algebra with sl(2, C) is obvious in the 2-dimensional
representation
1
2 0 0 1 0 0
D2 ( J3 ) = , D2 ( J+ ) = , D2 ( J− ) = , (15.30)
0 − 12 0 0 1 0
We then note that J± |mi are still eigenstates of J3 with eigenvalues increased or
reduced by 1 with respect to |mi:
j + ( j − 1) + ( j − 2) + . . . + j− = 0, (15.34)
which implies that j− = − j and that the vector space has dimension 2j + 1,
with basis
Since the dimension of the vector space V must be an integer number, we also
conclude that
j ∈ N/2. (15.36)
162 representations of su(2) and so(3)
The number j is called the spin of the representation and the vector space on
which it acts is given by
It is interesting to note that we can always find the spin of the representation
of sl(2, C) by using an operator that commutes with the entire set of matrices
providing the representation. By means of Schur’s lemmas we know that on
the common invariant space V this must be proportional to the identity and
therefore have a unique value when acting on any of the vectors of V. An
operator that has this property is the quadratic Casimir. In the sl(2, C) basis
defined by J3 , J± satisfying (15.29), the Cartan–Killing form is
4
k= 4 (15.38)
2
D (C2 )| j, mi = λ( j) | j, mi (15.40)
1 1
D (C2 )| j, − ji = D ( J3 )2 − D ( J3 ) + D ( J+ ) D ( J− ) | j, − ji = j( j + 1)| j, − ji.
2 2
(15.41)
1
λ( j) = j ( j + 1). (15.42)
2
Irreducible representations of sl(2, C) are then distinguished by the eigenvalues
of D (C2 ) and the basis of vectors in each of the invariant spaces is labelled by
the eigenvalues of D ( J3 ).
15.6 matrix representations 163
The procedure we just outlined let us also find a (2j + 1)-dimensional matrix
representation of su(2) explicitly. We start from the basis of normalized vec-
tors providing a basis for the invariant space of the corresponding irreducible
representation of sl(2, C) and then take the linear combinations leading to the
appropriate real form.
Define | j, ji as a norm one state:
We can then proceed in the same way for the other states and eventually define
the form of D2j+1 (la ) as
i
D2j+1 (l1 ) = − D2j+1 ( J+ ) + D2j+1 ( J− ) (15.46)
2
and
1
D2j+1 (l2 ) = − D2j+1 ( J+ ) − D2j+1 ( J− ) (15.47)
2
Note that when j ∈ N we also have a representation for so(3).
Example 15.1. The j = 1/2 representation has highest state | 21 , 12 i. We start by
D2 ( J3 )| 21 , ± 12 i = ± 21 | 12 , ± 21 i. (15.48)
| 12 , − 12 i ≡ D2 ( J− )| 12 , 12 i (15.49)
164 representations of su(2) and so(3)
and
D2 ( J− )| 12 , − 12 i = 0. (15.50)
Analogously
D2 ( J+ )| 12 , 12 i = 0, (15.51)
and
D2 ( J+ )| 21 , − 12 i = D2 ( J+ ) D2 ( J− )| 12 , 21 i
= [ D2 ( J+ ), D2 ( J− )]| 12 , 12 i (15.52)
= 2D2 ( J3 )| 21 , 12 i = | 21 , 21 i.
If the generic vector in Vj= 1 has two components and | 12 , 21 i is the upper com-
2
ponent and | 12 , − 12 i is the lower component, the corresponding representation
2 is given by
1
2 0 0 1 0 0
D2 ( J3 ) = , D (
2 +J ) = , D (
2 −J ) = , (15.53)
0 − 12 0 0 1 0
exercises
There are many groups relevant to physics, but a special role is played by SU(N)
groups. These groups appear naturally in Quantum Mechanics, for instance in
the analysis of the harmonic oscillator and of rotational symmetries, but they
are also at the core of the so-called Standard Model of Particle Physics. This
model is in fact a gauge theory coupled to matter, with gauge group SU (3) ×
SU (2) × U (1) and fundamental particles are sitting in specific representations
of these groups. In addition, there are approximate global symmetries that also
play a role in the analysis of fundamental processes, which are again described
by groups isomorphic to SU(N).
We are not just interested in constructing representations of SU(N) groups,
but also in understanding their products, because it is often useful to under-
stand how a composite system transforms once the transformation properties
of its components are known. Thus, after a brief general discussion, we present
in detail SU( N ) irrepses and their products using Young Tableaux. We stress
that, although some details of the techniques we use here are specific to SU(N)
groups, most of the discussion can be repeated and extended to other groups.
165
166 su(n) irreps and products
D ⊗ D0 = ⊕ j Dj .
If G is finite or if it is a compact Lie group and its irrepses have been classified,
we can use the multiplicity m a of each irreducible fragment to write
D ⊗ D 0 = ⊕ a m a Da . (16.1)
χ D⊗ D0 ( g) = χ D ( g) χ D0 ( g) = ∑ m a χ Da ( g ) (16.2)
a
where c is a label for the invariant subspace for the related irreps, i runs over
the number of times the same representation is repeated and p is the index
labeling the basis vectors. The coefficients Cci ,p| a,m,b,n = h p, ci | a, m; b, ni are
called the Clebsch–Gordan coefficients.
Without giving too many technical details, we now discuss irrepses of the Spe-
cial Unitary group SU(N).
vi 7 → U i j v j , (16.5)
168 su(n) irreps and products
(U ∗ )† U ∗ = (U † U )∗ = 1∗N = 1 N (16.7)
and
w∗ ı̄ 7→ U ∗ ı̄ ̄ w∗ ̄ . (16.9)
Once more, it is customary to refer also to this representation using its dimen-
sion, though in this case we also use a bar to distinguish it from the fundamen-
tal representation and stress that it can be obtained from the former by complex
conjugation:
We now move to the adjoint representation. The product of the two irrepses
N ⊗ N can be described by the set of operators acting on the vector space
defined by rank 2 tensors as
vi ⊗ w∗j ≡ Mi j 7→ U i k U ∗ j l Mk l . (16.13)
16.2 su(n) irrepses and young tableaux 169
w† v 7→ wk∗ U ∗ i k U i l vl = w† U † U v = w† v (16.14)
and thus we can decompose the tensor Mi j into its trace and its traceless parts:
1 i ◦ ◦
Mi j = δj tr M + M i j , with M k k = 0. (16.15)
N
In summary
N ⊗ N = 1 ⊕ ( N 2 − 1 ), (16.16)
where the first one is the trivial representation and the second one is the adjoint.
We recognize that this is really the adjoint representation by the fact that the
dimension is the correct one (dimR SU(N) = N 2 − 1) and by the fact that the
◦
action on M is precisely the one expected by the Adjoint representation, which
takes vectors in the algebra and acts on them by elements of the group from
the left and inverse elements from the right:
◦ ◦ ◦
AdU ( M ) = U M U −1 = U M U † . (16.17)
As we said in the introduction, we would like to see how all irrepses can
be generated by taking products of the fundamental representation. We start
by giving a simple example, where we consider rank 2 tensors built from 2
fundamental representations. These tensors transform as
but what is interesting to note is that these fragments define invariant sub-
spaces under the action of SU(N):
We conclude that
N(N − 1) N(N + 1)
N⊗N = ⊕ . (16.22)
2 2
Using this tensorial method we can extract all irreducible components in a
product, but it is generically too elaborate and complicated. However, a much
more effective method can be put in place by using Young Tableaux.
N ⇔ . (16.23)
Other irrepses follow from its k-fold product, with symmetrization and anti-
symmetrization rules determined by the Young tableau we use to represent it.
For instance the completely symmetric representation is
(N + k − 1)!
··· ⇔ T i1 ...ik = T (i1 ...ik ) ⇔ . (16.24)
| {z } (N − 1)!k!
k
There is an obvious limit on the number of boxes one can pile in vertical,
determined by the fact that in C N we have at most N different indices. This
16.2 su(n) irrepses and young tableaux 171
ei1 ...i N 7→ U i1 k1 . . . U i N k N ek1 ...k N = det U ei1 ...i N = ei1 ...i N . (16.26)
Using this same tensor we can also construct an object transforming in the anti-
fundamental representation, starting from N − 1 vectors in the fundamental
representation. In detail, we can define
which transforms as
P = aT sT , (16.30)
1 2 = 2 1. (16.31)
The other tableau with two boxes, applied to the same tensor gives the
remaining combination
1 =−2 , 1 = 0, 2 = 0. (16.32)
2 1 1 2
172 su(n) irreps and products
N−i+j
D= ∏ h(i, j)
, (16.33)
boxes i,j
where i labels the row, j labels the column and h(i, j) is the corresponding hook
number.
Example 16.1. The SU(4) Young Tableau has dimension D = 140. For
N = 4 the numerator is given by the product of the coefficients in
4 5 6 7
3 4 5 (16.34)
2 3
6 5 3 1
4 3 1 . (16.35)
2 1
T = T ijk ei ⊗ e j ⊗ ek . (16.36)
In order to split T into its irreducible fragments we build the projector operators
associated to the different representations of S3 . There are a total of 6 possible
16.2 su(n) irrepses and young tableaux 173
Their action on the basis of the vector space where rank 3 tensors are defined,
splits V⊗3 N into invariant subspaces whose basis elements are linear combina-
tions of the original ones (we omit the tensor product symbol ⊗ for the sake of
having more compact expressions):
1
Sijk = ei e j e k + ei e k e j + e j e k ei + e j ei e k + e k ei e j + e k e j ei ,
6
1
Aijk = ei e j e k − ei e k e j + e j e k ei − e j ei e k + e k ei e j − e k e j ei ,
6
1
M1,1 ijk = ei e j e k + e j ei e k − e k ei e j − e k e j ei , (16.41)
2
1
M1,2 ijk = √ ei e j e k − 2 ei e k e j − 2 e j e k ei + e j ei e k + e k ei e j + e k e j ei ,
2 3
1
M2,1 ijk = e e e − e j ei e k − e j e k ei + e k e j ei ,
2 i j k
1
M2,2 ijk = √ ei e j e k + 2 ei e k e j − 2 e j e k ei − e j ei e k + e k ei e j + e k e j ei .
2 3
The labels S, A and M stand for symmetric, antisymmetric and mixed sym-
metry, which are the symmetry properties of the corresponding basis. The
projectors in (16.37)–(16.40) select these invariant spaces: P projects into S,
P projects into the 1-dimensional space spanned by A and finally, P 1 2 and
3
A generic tensor can be decomposed by considering all the fragments that re-
main when contracting its T ijk components with the various basis elements.
For instance, the fully symmetric and fully antisymmetric combinations are
1 ijk
T ijk Sijk = T + T ikj + T kij + T jik + T jki + T kji ei ⊗ e j ⊗ ek ,
6
1 ijk
T ijk Aijk = T − T ikj + T kij − T jik + T jki − T kji ei ⊗ e j ⊗ ek , (16.42)
6
while the mixed symmetry combinations give
1 ijk
T ijk M1,1 ijk = T + T jik − T jki − T kji ei ⊗ e j ⊗ ek ,
2
1
T ijk M1,2 ijk = √ T ijk − 2T ikj − 2T kij + T jik + T jki + T kji ei ⊗ e j ⊗ ek ,
2 3
1 ijk
T ijk M2,1 ijk = T − T jik − T kij + T kji ei ⊗ e j ⊗ ek , (16.43)
2
1
T ijk M2,2 ijk = √ T ijk + 2T ikj − 2T kij − T jik + T jki + T kji ei ⊗ e j ⊗ ek .
2 3
One thing, however, that needs to be stressed here is that while M1 and M2
belong to different invariant subspaces of the S3 basis space, they give tensors
associated to the same representation of the permutation group on their indices
and therefore they provide equivalent representations of SU(N). In fact they all
fulfill the same conditions
ijk jki kij
TMa + TMa + TMa = 0, a = 1, 2, (16.44)
Clearly the fact that the antisymmetry is on the first two indices or in the first
and third, or, for that matter, in the second and third, gives obviously the same
SU(N) representation.
Example 16.3. The same kind of decomposition is also useful to analyze invari-
ant subspaces of the Hilbert space of k identical particles transforming under
16.2 su(n) irrepses and young tableaux 175
the action of the fundamental representation of the SU(N) group. The Hilbert
space for each of these particles is H N = C N and the total Hilbert space is
H = (H N )k , which we now decompose using Sk irrepses.
For the sake of simplicity we focus on the decomposition of the 8-dimensional
Hilbert space of k = 3 identical particles, transforming as a doublet of SU(2),
i.e. N = 2. We denote the basis of the 2-dimensional space on which the fun-
damental representation of SU(2) acts by |+i and |−i. The generic element of
the 8-dimensional space H = H2 ⊗ H2 ⊗ H2 is then a linear combination of the
states labelled by | ± ±±i = |±i ⊗ |±i ⊗ |±i, where each of the signs refers
to the corresponding sign in the choice of the basis elements for each of the
2-dimensional Hilbert spaces of the 3 identical particles.
Since k = 3, the SU(2) irrepses acting on H are in correspondence with the
1
(| + +−i + | − ++i − 2| + −+i) , (16.48)
2
1
√ (| − −+i − | + −−i) , (16.49)
2
1
(| − −+i + | + −−i − 2| − +−i) . (16.50)
2
176 su(n) irreps and products
Note that the states with the same spin transform in one of the irreducible
representations ψ Ma of S3 . In conclusion dim H = 8 = d ·D +d ·
D = 1 · 4 + 2 · 2.
The dimension of the irreps is determined by the formula (16.33) and the
number of times it appears is determined by the dimension as a representation
of the corresponding Symmetric group.
This classifies completely irreducible representations. We can now move on
and analyze how to explicitly compute products of representations for the same
group.
The analysis given in the previous section for the construction of the irreducible
representations of SU(N) suggests that Young Tableaux should be used effec-
tively to compute products of representations. In this case we do not discuss
the derivation of the rules that determine how to construct this product and
obtain the Clebsch–Gordan decomposition, but rather explain the result and
provide some examples that can convince us of the effectiveness of the adopted
technique.
1. Label the second tableau in the product with the letters a, b, c, . . . in the 1st, 2nd,
3rd, etc. row:
a a a a
b b b (16.51)
c c c
...
2. Attach all the boxes with an a to the first tableau in all possible different ways, but
such that there are never two a in the same column and such that the resulting
diagram is still a legitimate Young tableau.
16.2 su(n) irrepses and young tableaux 177
⊗ = ⊕ ⇔ 2 × 2 = 1 + 3. (16.53)
⊗ a a = a + + · , (16.54)
a a + = +
a a a
which can be represented by
3 × 3 = 5 + 3 + 1. (16.55)
Example 16.5. In SU(3) the fundamental representation 3 and the anti-fundamental
representation 3 are represented by two different tableaux:
3⇔ and 3⇔ . (16.56)
⊗ = + ·, (16.58)
178 su(n) irreps and products
which corresponds to
3 × 3 = 8 + 1, (16.59)
⊗ = ⊕ ⇔ 3 ⊗ 3 = 6 ⊕ 3. (16.60)
Another interesting lesson we can take home from these examples is that
Young Tableaux corresponding to conjugate irrepses have complementary shapes.
This means that complex conjugation of a given representation is represented
by the complementary tableau, rotated by 180 degrees.
8= = = 8̄ = 8
6= = = 6̄
Figure 24.: Conjugate representations in SU(3). The adjoint representation 8 is real and
its conjugate Young Tableaux has the same shape as the original one. On
the other hand, the representation 6 is complex and its complex conjugate
is represented by a different tableau.
We conclude this lecture, by reminding that you already encountered the Clebsch–
Gordan decomposition for SU(2) in physical problems related to spin and an-
gular momentum. In that case it is also useful to it may be useful to obtain
the exact multiplicative coefficients in the decomposition and this was done as
follows.
16.2 su(n) irrepses and young tableaux 179
~~ ~
D j1 ( g) ⊗ D j2 ( g) = eψ· J(1) ⊗ evecψ· J(2) = 1 + ψ
~ ~J(1) ⊗ 1 + 1 ⊗ ~J(2) + . . . (16.61)
The spaces on which the irrepses are defined, namely V1 and V2 , admit a or-
thonormal basis of eigenstates of J(31) and J(32) :
{| j1 , m1 i} ∈ V1 , {| j2 , m2 i} ∈ V2 . (16.63)
We can then discuss J± . Now the states are mixed by the operators and, starting
from the highest spin state, whose spin is j1 + j2 , one generates 2 states with
spin j1 + j2 − 1:
| j1 , j1 i ⊗ | j2 , j2 − 1i (16.65)
and
| j1 , j1 − 1i ⊗ | j2 , j2 i. (16.66)
| 32 , 32 i = |1, 1i ⊕ | 12 , 12 i. (16.69)
180 su(n) irreps and products
We apply J− to this:
√ √
D ( J− )| 23 , 32 i = 3| 32 , 21 i = 2(|1, 0i ⊗ | 12 , 21 i) + |1, 1i ⊗ | 12 , − 21 i. (16.70)
We deduce that
r
2 1
| 32 , 12 i = (|1, 0i ⊗ | 12 , 21 i) + √ (|1, 1i ⊗ | 12 , − 21 i). (16.71)
3 3
where obviously | 21 , − 21 i = D ( J− )| 12 , 12 i.
exercises
2. We can repeat the analysis of the previous problem by using the isomor-
phism so(4) ' su(2) L ⊕ su(2) R established in an exercise of previous
chapter. (We use here L, R to distinguish the two copies of su(2) in corre-
spondence with Li+ and Li− ). Use the symbol (m, n) = Dm ( g L ) ⊗ Dn ( gR )
for the representation of the elements ( g L , gR ) ∈ SU(2) L × SU(2) R by
means of matrices of dimensions m and n.
Show that the fundamental representation of SO(4) is a couple of dou-
blets of SU(2) and compute their products using the properties of SU(2)
products.
(1)
a) Compute Li on (2, 2) and note that it contains a triplet and a scalar
(the fourth component).
b) Compute (2, 2) ⊗ (2, 2) and decompose the result in its irreducible
parts.
c) Relate this result to the one obtained in the previous exercise.
183
184 poincar é group
( Λ2 , a2 ) ◦ ( Λ1 , a1 ) = ( Λ2 Λ1 , Λ2 a1 + a2 ),
We leave to the reader the check that this composition law satisfies all group
axioms.
The Poincaré group contains several interesting subgroups. The first and
most interesting one is the subset of elements of the form (Λ, 0), which is called
the Lorentz group. Other interesting subgroups are the group of translations,
whose matrices have the special form (14 , a), and the group of rotations, whose
matrices are of the form ( R, 0), with R0 0 = 1, R0 i = Ri 0 = 0 and Ri j Rk j = δik .
Any rotation matrix can actually be parameterized by the unit norm vector n̂
fixing the axis of rotation and by the rotation angle θ:
O(3, 1) = {Λ ∈ Mat(4, R) | Λ T ηΛ = η },
implies
The full Lorentz group has therefore at least 2 disconnected components, cor-
responding to the set of matrices having determinant equal to plus or minus
1.
The subset of matrices with unit determinant contains also the identity ma-
trix and actually forms a subgroup of the original one called
This is also analogous to the SO(N) < O(N) groups. However, we now prove
that SO(3,1) is not connected.
If we focus on the constraint that characterizes the matrices of the group
SO(3,1) we see that the 00 component gives
2 2
−1 = η00 = Λµ 0 Λν 0 ηµν = − Λ0 0 + ∑ Λi 0 , (17.8)
i
which implies
2 2
Λ0 0 = 1 + ∑ Λi 0 ≥ 1. (17.9)
i
Λ0 0 ≥ 1, (17.10)
Λ0 0 ≤ −1. (17.11)
The other components are subsets but not subgroups. One can cover them by
taking the product of elements in SO+ (3, 1) with appropriate elements defining
a discrete subgroup of SO(3,1). In detail, the subset with det Λ = −1 and
Λ0 0 ≥ 1 contains the space parity operator
1
−1
P= , (17.12)
−1
−1
−1
1
T = , (17.13)
1
1
σµ = (12 , σi ). (17.14)
x0 + x3 x1 − i x2
det A( x ) = det = − x µ ηµν x ν . (17.16)
x1 + i x2 x0 − x3
so that Ae† = A,
e which means that A e = xeµ σµ , with xeµ = ΛS µ ν x ν . This in turn
implies that we have a map from SL(2, C) to SO+ (3, 1). We can see that the
map is linear because
xeµ = 12 tr A e( xe)σ̂µ = 1 tr SA( x )S† σ̂µ
2
(17.18)
= 21 tr Sσν S† σ̂µ x ν = ΛS µ ν x ν .
If, in addition, we require that S ∈ SL(2, C) (and hence that det S = 1), then
e = det (SAS† ) = det S(det S)∗ det A = − x µ x ν ηµν . (17.19)
− xeµ xeν ηµν = det A
In addition, SL(2, C) is connected, and the same will be true by the image of
the continuous map given by ΛS . We therefore conclude that
S 7→ ΛS (17.22)
maps elements of SL(2, C) into elements of SO+ (3,1). Finally, since S and −S
are mapped to the same element ΛS we see that the map is 2 to 1.
The fact that such map is a homomorphism follows by the proof that
S 1 S 2 7 → Λ S1 S2 = Λ S1 Λ S2 . (17.23)
This is tedious, but easy to obtain by using the explicit expression for ΛS :
Λ S1 S2 µ ν 1
tr S1 S2 σν S2† S1† σ̂µ = 21 tr S2 σν S2† S1† σ̂µ S1 =
= 2
(17.24)
1 † b S† σ̂ µ S
a
= 2 S2 σν S2 a 1 1 b ,
where we used the cyclic property of the trace, and using that (see the exercises
for a proof of this identity)
Λ S1 S2 µ ν 1
S2 σν S2† c (2 δad δcb ) S1† σ̂µ S1 a
= 4 d b
1 c ( σ̂ ρ ) d ( σ ) b
S2 σν S2† S1† σ̂µ S1 b a
= 4 d c ρ a (17.26)
1 = Λ S1 ρ Λ S2 ν .
1 † ρ 1 tr σ S† σ̂ µ S
µ ρ
= 2 tr S2 σν S2 σ̂ 2 ρ 1
From this theorem we also have a hint of the fact that SL(2,C) provides the
universal covering of SO+ (3,1) and that
Given that SL(2,C) and SO+ (3, 1) are homomorphic, they are also related as
topological spaces.
17.2 the so+ (3, 1) algebra 189
S = u eh , u† u = 1, h† = h. (17.28)
where det u = eiα ∈ U(1) and exp(tr h) ≥ 0. The request of having a unit
determinant for S forces
Thus any SL(2,C) matrix can be written as the product of a matrix u ∈ SU(2)
' S3 and the exponential of a traceless, hermitian, 2 by 2 matrix h, which is
parameterized by three real numbers. We conclude that topologically
Also, since the 2 to 1 map (17.22) identifies matrices S with opposite signs and
only u can change the overall sign, we argue that
As is often the case, also for the SO+ (3, 1) group, it is convenient to determine
representations of the algebra rather than those of the group. For this reason
we now discuss the structure of the algebra generators and their fundamental
representation. The group SO+ (3, 1) can be thought of as a group of matrices.
We can then follow the same derivation of the algebra as in section 12.2, by
expanding matrices Λ ∈ SO(3,1) close to the identity. In detail, we can linearize
the condition Λ T ηΛ = η by choosing Λµ ν = δν + ω µ ν , assuming a small ω:
µ
The so(3, 1) algebra is then determined by the set of matrices ω fulfilling the
condition above, namely
n o
so(3, 1) = ω ∈ Mat(4, R) | ηω = −(ηω ) T . (17.34)
which make the algebra non-compact and whose exponential generates boosts:
cosh ψ sinh ψ
sinh ψ cosh ψ
exp(ψD (k1 )) = , (17.38)
1
1
17.2 the so+ (3, 1) algebra 191
where ψ is the rapidity tanh ψ = − β. Note that since the group SO+ (3,1) is
non-compact, we cannot generate all its elements by exponentiation. However,
being SO+ (3,1) connected, we can always cover the whole underlying manifold
by taking products of boosts and rotations:
Λ = exp ψ ~ · D (~k) exp ~θ · D (~l ) . (17.39)
which show that we have two copies of su(2)C ' sl(2, C).
This simple proof shows that also in this case we can build representations
of so(3, 1) by means of representations of sl(2, C), just like we did for su(2) '
so(3).
192 poincar é group
exercises
h A| Bi = Tr ( A† B).
which leads to
∑(σµ )∗ α β (σµ )γ δ = N δαγ δβδ
µ
and finally to
i
18
I R R E D U C I B L E R E P R E S E N TAT I O N S O F T H E P O I N C A R É
GROUP
Now that we have determined the structure of the Poincaré and Lorentz groups
and of their algebras we can discuss their representations. As we will see in
the following, we will be interested in two different kinds of representations
for these groups, because we will need to employ different strategies in the
description of states and operators of a relativistic quantum mechanical theory.
Let me anticipate that we will need unitary infinite-dimensional representa-
tions to describe physical states, while we will make use of non-unitary, but
finite-dimensional representations for operators.
We restrict our discussion to elements of the Poincaré group that are con-
nected with the identity, i.e.
D ( Pµ ) = ∂µ . (18.3)
The corresponding commutators are those of the Lorentz algebra (17.42) for Jµν
and
193
194 poincar é irrepses
for the others. The last commutator tells us that translation generators trans-
form in the fundamental representation of the Lorentz rotations (and in fact
they carry a vector index µ):
For the sake of simplicity, we are going to consider single particle states, trans-
forming under the action of irreducible representations of the Poincaré group.
The classification of these representations has been worked out first by Wigner,
but one can look for a more detailed discussion in Weinberg’s book ”Quantum
theory of Fields: I”.
As mentioned in the introduction, an irreducible unitary representation of
the Poincaré group must be infinite dimensional. The states providing a ba-
sis for the vector space on which this representation acts form an infinite-
dimensional multiplet and transform onto each other when acted upon by
Poincaré generators. As is always the case, we can decide that a set of com-
muting operators are diagonal on this basis and start labeling the correspond-
ing states by using their eigenvalues. Recall that irreducible representations of
abelian groups are always 1-dimensional.
In the Poincaré algebra, Pµ operators commute among themselves and we
can therefore classify quantum states by using their eigenvalues
P µ | p µ , σ i = i p µ | p µ , σ i, (18.6)
In order to complete the basis of states of the vector space on which a unitary
representation of the Poincaré group acts, we need to provide a representation
for the proper Lorentz group elements. The corresponding operators U (Λ) ≡
U (Λ, 0) should map Pµ eigenstates into states with momentum Λp, because Pµ
transform in the fundamental representation of the Lorentz group and so must
transform its eigenstates. Actually, using Hadamard’s lemma 12.3 and (18.5),
we can see this explicitly by taking the action of
1
U (Λ) = exp ωµν J µν (18.9)
2
196 poincar é irrepses
U † ( Λ ) P ρ U ( Λ ) = U ( Λ −1 ) P ρ U ( Λ ) = P σ ( Λ −1 ) σ ρ = Λ ρ σ P σ , (18.11)
where we used that Λµ ρ ηµν Λν σ = ηρσ and hence (Λ−1 )ρ σ = ηρα (Λ−1 )α β η βσ =
Λσ ρ . This in turn implies that an eigenstate of Pµ with momentum pµ trans-
forms into an eigenstate with momentum Λµ ν pν , when acted upon by U (Λ):
Does this imply that the only result of the action of a Lorentz transformation
on a state | p, σ i is to map it to
Actually, we can see that this cannot be correct, because we did not take into
account the possibility that there may different states whose momenta are in-
variant under the action of Λ and which therefore may mix between then when
acted upon by U (Λ). In order to solve this issue we introduce the so-called little
group and show that irreducible representations of Poincaré group are related
to irreducible representations of this group.
Theorem 18.1. Unitary irreducible representations of the Poincaré group are fixed by
irreducible representations of Iso (k) for fixed k.
Proof. Start from a fixed momentum k, whose little group Iso (k ) is determined
by elements Λ ∈ Iso (k) fulfilling
Λk = k. (18.14)
All eigenstates of the momentum operator with eigenvalue k are then mapped
onto each other by the action of Lorentz transformations
Also D may not be finite dimensional if Iso (k) is not compact. Fortunately this
will not be necessary if one is not interested in non-physical cases.
We can now determine the result of the action of a unitary Lorentz transfor-
mation on a generic state of momentum p. First we find a Lorentz transforma-
tion relating p to k
pµ = Λ p µ ν kν , (18.17)
which we assume does not mix states with the same momentum k:
= U ( Λ Λ p )U ( Λ − 1
Λ p ΛΛ p )| k, σ i,
198 poincar é irrepses
where
Λ− 1 −1
Λ p ΛΛ p k = ΛΛ p Λp = k (18.20)
and therefore
Λ− 1
Λ p ΛΛ p ∈ Iso ( k ). (18.21)
We can use this result to see that the generic unitary representation of a Lorentz
transformation acts on states | p, σ i mixing them according to the little group of
the associated momentum k
where Λ̃ = Λ− 1
Λp ΛΛ p ∈ Iso ( k ). This provides a representation for Λ on each
state | p, σ i.
While the final expression may not be particularly elegant, we reduced the
problem of classifying representations of the Poincaré group to the problem of
classifying representations of the little groups associated to physical momenta,
which can be differentiated by the value of their invariant scalar combinations.
Since the physical interpretation of pµ does not allow for the case p0 < 0 and
since p2 > 0 implies tachyonic states, we can distinguish three physical orbits1 :
• kµ = 0. Iso (k) = SO(3, 1) and its orbit is given by a unique state. It has
the property that all generators of the Poincaré group are represented
by 1-dimensional null matrices and, consequently, all group elements are
represented by identity matrices. This state has momentum and angular
momentum equal to zero and is invariant under Poincaré transformations.
We interpret it as the vacuum.
• k2 < 0. There is a frame where kµ = (k, 0, 0, 0). In this case the little group
consists of space rotations Iso (k) = SO(3), whose covering group is SU(2).
These states describe massive particle and are classified by well known
spin representations D2j+1 and σ = ( j, j3 ). However, since [l3 , Pi ] = e3ij Pj ,
j3 should not be directly interpreted as the eigenvalue of the state with
respect to l3 , but rather with respect W3 , where W µ is the Pauli–Lubanski
operator we will introduce later, when discussing Casimir operators for
the Poincaré group.
1 Also, since we are interested in proper and orthochronous transformations, we should distinguish
p0 > 0 from p0 < 0. However, once more, physics tells us we should only concentrate on p0 > 0.
18.2 finite dimensional representations 199
where δij is the Euclidean metric in 3 dimensions, and we saw that by means of
them we could provide an irreducible 2-dimensional spinorial representation
for the so(3)C algebra by using
1 1 i
DS ( Jij ) = − σij ≡ − [σi , σj ] = eijk − σk = eijk DS (lk ) (18.25)
2 4 2
and then, starting from that, build all the other representations.
For so(3, 1) we introduce 4 by 4 matrices γµ , µ = 0, 1, 2, 3, with analogous
anti-commutation properties:
Definition 18.2. The Clifford algebra Cl (3, 1) is the algebra generated by γµ ,
µ = 0, 1, 2, 3, satisfying
{γµ , γν } = 2 ηµν .
The minimal matrix representation one can give to γµ is 4 by 4 (see the ex-
ercises) and there are obviously various legitimate equivalent choices for their
expression. One of them is the following:
γ0 = i σ1 ⊗ 12 , γi = σ2 ⊗ σ i . (18.26)
and again, just like between SU(2) and SO(3) we have a 2 to 1 map, we also
have a 2 to 1 map between Spin(3, 1) and SO+ (3, 1). We can actually identify
Differently from what happened for SU(2) we can now see that this representa-
tion is not irreducible, though.
Starting from the 4 γ-matrices defining Cl (3, 1) we can always introduce a new
matrix3
γ5 ≡ i γ0 γ1 γ2 γ3 , (18.32)
satisfying
{γ5 , γµ } = 0 (18.33)
and
(γ5 )2 = 1. (18.34)
1
P± ≡ (1 ± γ5 ) , (18.35)
2
which split the vector space on which the spinorial representation acts in two
complementary subspaces. This is actually straightforward to see using the
representation (18.26) for the γ-matrices, which gives
12 O2
P+ = , P− = . (18.36)
O2 12
3 The index 5 is not related to spacetime properties and therefore γ5 = γ5 .
202 poincar é irrepses
We now show that the subspaces im( P± ) are invariant with respect to the action
of the spinorial representation of the Lorentz group DS .
Take an element in any of the two subspaces: η± ∈ Im( P± ). By definition
P± η± = η± and P∓ η± = 0. Hence
γ5 η± = ± η± . (18.37)
We will call spinors in im( P+ ) spinors with positive chirality and those in
im( P− ) spinors with negative chirality.
Actually im( P± ) are vector spaces where irreducible representations of the
sl(2, C) ⊕ sl(2, C) algebra act. In fact we can write
1 1 1
ωµν J µν = ω0i J 0i + ωij J ij = −ω0i ki + eijk ωij lk = li+ ωi+ + li− ωi− , (18.40)
2 2 2
where
1 1
li± ≡ ( l ± i k i ), ωi± ≡ eijk ω jk + 2i ω0i . (18.41)
2 i 2
In the Weyl basis (18.26)
ωi+ − 2i σi
1
− ωµν γµν = . (18.42)
4
ωi− − 2i σi
We then see that Vs = V + ⊕ V − on which we have a split action of li+ and li− .
We also conclude that spinors with positive chirality are objects in the ( 21 , 0)
representation and those with negative chirality in the (0, 12 ). Alternatively,
using the dimensions of the irrepses, we denote them by (2, 1) and (1, 2).
18.3 reducibility of spinor representations 203
Finally, we can show that Spin(3, 1) ' SL(2, C) and that we have a 2 to 1
map between the Spin group and SO+ (3, 1). A generic element of Spin(3, 1) is
represented by4
1 S 0
exp − ωµν γµν = , (18.43)
4 0 ( S −1 ) †
where
i
S = exp − √ ωi+ σi (18.44)
2 2
and S ∈ SL(2, C), with (S−1 )† = exp − 2i ωi− σi , because (ωi+ )∗ = ωi− . As
for SU(2), we can build the adjoint representation on γµ x ν and produce the
2 to 1 map between SL(2, C) and SO+ (3, 1). This map translates elements of
SL(2, C) into ΛS matrices giving the action x µ 7→ ΛS ν x ν and again S and −S
µ
2j1 2j2
During these lectures we stressed several times that an effective way of classi-
fying group representations is given by Casimir operators. Casimir operators
commute with all the generators of a given algebra and therefore allow us to
distinguish irreducible representations by their eigenvalues. This is true also
for the Lorentz and Poincaré group and for this reason we now revisit the dis-
cussion of their irreducible representations in the light of the existence of such
Casimir operators.
1
C1 = Jµν J µν (18.47)
2
and
1 µνρσ
C2 = e Jµν Jρσ . (18.48)
4
The existence of these two quadratic Casimir operators is obviously related
to the fact that its complexification is isomorphic to the direct sum of two
other algebras and therefore we should be able to use their quadratic Casimir
operators to classify irrepses:
Recalling that J0i = k i , Jij = eijk lk and li± = √1 (li ± i k i ) we have that
2
1
2
C1 = − J0i + J J = −|~k|2 + |~l |2 = (~l + )2 + (~l − )2 (18.50)
2 ij ij
and
C2 = eijk J0i Jjk = 2~k · ~l = −i (~l + )2 − (~l − )2 . (18.51)
We then conclude that we can use two labels ( j1 , j2 ) to differentiate the irre-
ducible representations.
18.4 casimir operators 205
In the case of the Poincaré group [C1 , Pµ ] 6= 0 and also [C2 , Pµ ] 6= 0. We then
need to find new Casimir operators. The first one is simply the operator defin-
ing the mass of the state:
P2 | m i = − m2 | m i. (18.52)
This means that W µ are in fact combinations of the Lorentz generators that
leave the momentum invariant and therefore generate the little group.
We then easily prove that W 2 is a good Casimir operator, because W µ is a
vector under Lorentz transformations
and
1 2 2
W2 = J P + Pµ Jµν J νρ Pρ (18.60)
2
is therefore a scalar:
[ Jµν , W 2 ] = 0. (18.61)
which means that irreducible representations are classified by the so(3) Casimir,
as expected.
When acting on massless particle states we have once more the problem with
the infinite quantum numbers not visible in nature. Choose a frame where
pµ = ( p, 0, 0, p) and decompose W µ accordingly
µ µ
W µ = η Pµ + N1 n1 + N2 n2 + Ws sµ , (18.63)
where
W 0 = η |~
P| ~ = η ~P.
and W (18.66)
18.4 casimir operators 207
exercises
Build it explicitly.
labels of sl (2, C) ⊕ sl (2, C) describing them? Which spin states they con-
tain with respect to SO(3) < SO(1,3)?
E3 = O(3) n R3 . (19.1)
M = {~x I , I = 1, . . . , N }. (19.2)
Imagine a molecule where the atoms are in a fixed position. The electronic
levels can be obtained as a function of their configuration and their energy de-
pends on the relative positions of the nuclei. These levels will change even if we
move the nuclei by a little amount. Fix then the average positions h~x I i of the N
209
210 symmetries of molecules and solids
The potential energy of the system depends on the values of q A . Although the
resulting function is strictly valid only for fixed q A , we can still use V (q A ) to
compute the molecule vibrations, because q̇ A is going to be much smaller than
the average speed of the electrons.
The total energy of the system is the sum
H = K + V, (19.4)
1
where K = 2m |~p I |2 and V = V (~x I ).
If we use as coordinates the displacements q A , we can simplify the potential
and expand it around the equilibrium positions:
1 1
K=
2 ∑ q̇2A , V' F q qB ,
2 AB A
(19.5)
A
where we subtracted the constant part V (0) and we stopped at second order in
the expansion of the potential. The equations of motion of this system are
q̈ A + FAB q B = 0. (19.6)
q A = c A cos(ω A t + φ A ). (19.7)
F = S T ∆S, (19.8)
where
λ1
∆=
.. ,
λ A = ω 2A > 0 (19.9)
.
λ3N
19.1 classic vibrations of molecules 211
and the columns of S are given by the 3N eigenvectors c( A) , so that the general
solution can be expressed in terms of the so-called Normal Modes
∑ cA
( B)
qA = α B cos(ω B t + φB ), (19.10)
B
1 1
H= PA PA + ω 2A Q A Q A . (19.11)
2 2
We now understand that the problem of studying molecule vibrations is equiv-
alent to the analysis of its normal modes. This problem becomes purely group
theoretical, once we use the symmetry properties of the molecule.
19.1.1 Symmetries
y3 y3
3 3
x3 x3
y1 y2 y2 y1
1 x1 2 x2 2 x2 1 x1
Since the atom 2 is displaced (by the pink arrow in the picture), the final
configuration is obtained from the original one by permuting the vertices 1 and
2, but also acting on the displacement q2 , which is mapped by means of the
same element of D3 by a reflection around the y axis.
The generic action of G is therefore realized on the q A coordinates by the real
and unitary matrix
D ( g ) = p ( g ) ⊗ R ( g ), (19.12)
D ( g) = ∑ n a D a ( g ), (19.13)
a
q 7→ D ( g)q, (19.14)
D T ( g) FD ( g) = F, (19.15)
or, equivalently
FD ( g) = D ( g) F, ∀ g ∈ G. (19.16)
F mn D n ( g) = D m ( g) F mn , (19.17)
where m and n are the dimensions of the irreducible blocks. Using Schur’s
lemma we can deduce that F mn = 0 whenever m 6= n and is proportional to
the identity on the same irreducible fragments. We conclude that F is block
diagonal on invariant subspaces of the space of displacements and therefore
we can reduce the computation of normal modes to the analysis of irreducible
representations of discrete groups.
19.1 classic vibrations of molecules 213
We call q1 and q2 the displacements, so that the distance between the two
nuclei is
1 1
V= k ( δ − d )2 = k ( q2 − q1 )2 , (19.20)
2 2
so that
1 −1
F=k . (19.21)
−1 1
The symmetry group is the permutation group of two objects S2 ' Z2 , repre-
sented on (q1 , q2 ) by
1 0 0 1
D2 ( 1 ) = , D2 ( σ ) = . (19.22)
0 1 1 0
−1 1
tr( D (σ ) F ) = tr k = −2k = λ1 − λ2 , (19.25)
1 −1
which implies that
λ1 = 0, λ2 = 2k. (19.26)
This is exactly what we should get from the split of the displacements into
the center of mass displacement X = √1 (q1 + q2 ) and the displacement with
2
respect to the center of mass x = √1 ( q2 − q1 ). The Hamiltonian becomes
2
1 2 1
H= Ẋ + ẋ2 + (2k ) x2 . (19.27)
2 2
We see that the null eigenvalue is associated to the center of mass. In fact, its
motion is that of a translation and hence has no vibrational energy. On√the
other hand, x describes precisely a vibrational mode with frequency ω = 2k.
We can actually understand what is described by the normal modes by pro-
jecting the displacement vector into the subspace left invariant by the a-th irre-
ducible representation. This can be done by means of the projector
Na
Pa ≡
#G ∑ χ a ( g ) ∗ D ( g ). (19.28)
g
If one evaluates this projector as a matrix one sees that it is proportional to the
identity only in the block corresponding to the a irreducible representation:
a
( Pa )ij = Na ( Dkk , Dij ) = ∑ δab δij , (19.29)
b
where b runs over the various irrepses contained in D. For the case at hand we
have that
1 1 1 1 1 1
P1 = = √ √ (1, 1) (19.30)
2 1 1 2 1 2
and
1 1 −1 1 1 1
P10 = = √ √ (1, −1) . (19.31)
2 −1 1 2 −1 2
√
Their projection on (q1 , q2 ) gives QCM = q1 + q2 and 2 x = q2 − q1 .
19.1 classic vibrations of molecules 215
y
y1 y2
x1 x2
x
1 2
1 0 0 1
p (1) = p ( ρ ) = , p(σ ) = p(σρ) = . (19.32)
0 1 1 0
This realizes the fact that the equilibrium positions of the nuclei is not affected
by a reflection around the x axis.
Considering the embedding in O(2), the same elements are represented as
1 0 −1 0
R (1) = , R(σ) = ,
0 1 0 1
(19.33)
1 0 −1 0
R(ρ) = , R(σρ) = .
0 −1 0 −1
216 symmetries of molecules and solids
−1
1
1 1
D4 ( 1 ) =
, D4 ( σ ) =
−1
,
1
1 1
−1
1
−1 −1
D4 ( ρ ) = , D4 (ρσ) =
−1
.
1
−1 −1
(19.34)
In this basis
−1
1
0
F = k
−1
. (19.36)
1
0
−1
1
1 0
P1 = , (19.41)
2 −1 1
0
√
which correctly gives the vibration mode in x, with frequency 2k,
1 1
1 0
P10 = , (19.42)
2 1 1
0
which gives the zero mode related to translations in x,
0
1 1 1
P100 = , (19.43)
2 0
1 1
which gives the zero mode related to translations in y, and
0
1 1 −1
P1000 = , (19.44)
2 0
−1 1
which produces anti-clockwise rotations around the origin.
D6 = D1 = D1 0 + D2 + D2 . (19.45)
The corresponding
√ subspaces give the non-trivial dilatation mode, with fre-
quency ω = 3k, the zero energy, rotational mode, the 2-dimensional space of
218 symmetries of molecules and solids
Here are some possible very rough outlines for the solutions of the exercises
provided in the text.
chapter 2
Ai − m2 eij Bj = 0, (A.2)
Bi + m2 eij A j = 0. (A.3)
One has to pay attention to the fact that Dµ φi contains a piece without
δDµ φ a δDν φ a µ
derivatives, so that δφi
= Aµ e ai and δ∂µ φi
= δia δν . We then have
δL δDµ φk µ j
1
δφi
= −δkj δφi
D φ + 6 Cijk + Cjik + Cjki φ j φk
(A.5)
= −δkj Aµ eki D µ φ j + 12 C(ijk) φ j φk .
219
220 solutions
When varying the last term in the Lagrangian one has to take into account
the fact that only the fully symmetrized terms in the C tensor survive,
because they are contracted with three identical fields. Also
δL δDν φ a ν j
= − δaj D φ = −δij D µ φ j . (A.6)
δ∂µ φi δ∂µ φi
3. Euler–Lagrange equations:
chapter 3
1. One by one:
• Q+ with a ◦ b = a/b is not a group because the associative property
fails. Let a, b, c ∈ Q+ , we have that
a
a a
( a ◦ b) ◦ c = ◦c = b = . (A.10)
b c bc
On the other hand
b a ca
a ◦ (b ◦ c) = a ◦ = = . (A.11)
c b b
c
( ab)−1 ba = 1 (A.21)
ba = ab. (A.22)
3. Assume we have 3 elements {1, a, b}. We can fill immediately the lines
and columns where the identity appears
1 a b
1 1 a b
(A.23)
a a
b b
4. Take G = {1, a, b, c}. We start as usual by filling rows and columns asso-
ciated to the identity:
1 a b c
1 1 a b c
a a (A.25)
b b
c c
1 a b c
1 1 a b c
a a 1 (A.26)
b b
c c
and from the rearrangement theorem we immediately fill the rest of the
row:
1 a b c
1 1 a b c
a a 1 c b (A.27)
b b
c c
1 a b c
1 1 a b c
a a 1 c b (A.28)
b b c
c c b
224 solutions
1 a b c 1 a b c
1 1 a b c 1 1 a b c
a a 1 c b a a 1 c b (A.29)
b b c 1 a b b c a 1
c c b a 1 c c b 1 a
These are two distinct groups. We now come back to the other option
above, aa = b. By the rearrangement theorem we complete immediately
the second row and the second column
1 a b c
1 1 a b c
a a b c 1 (A.30)
b b c
c c 1
1 a b c
1 1 a b c
a a b c 1 (A.31)
b b c 1 a
c c 1 a b
1 I J K −1 − I − J − K
1 1 I J K −1 − I − J − K
I I −1 K −J −I 1 −K J
J J −K −1 I −J K 1 −I
K K J −I −1 −K − J I 1 (A.32)
−1 −1 −I −J −K 1 I J K
−I −I 1 −K J I −1 K −J
−J −J K 1 −I J − K −1 I
−K −K −J I 1 K J − I −1
6. Case by case:
• Z4 and S3 have different order, hence they cannot be isomorphic.
#(Z4 ) = 4, #(S3 ) = 3! = 6.
• Z6 and S3 have the same order, but the first is Abelian, while the
second is not. We conclude that they cannot be isomorphic.
• S3 and D3 are isomorphic. In general one can easily prove the
isomorphism by mapping generators between them and showing
that they respect the relations in the presentations. In this case we
can associate a 3-cycle permutation to ρ and a 2-cycle to σ, so that
σρ = ρ−1 σ. An instance that works is σ = (12)(3) ρ = (123).
a = ( i1 , i2 , . . . , i p ), (A.33)
226 solutions
b = ( i p , i p −1 , . . . , i 1 ). (A.34)
as follows from
i p 7 → i p −1 7 → i p
..
. (A.36)
i2 7 → i1 7 → i2
i 7 → i 7 → i
1 k 1
The same argument can be applied to the generic cycle, which is a disjoint
product of cycles.
8. Explicitly:
1 7 → 1 7 → 7 7 → 7
2 7 → 2 7 → 3 7 → 4
3 7 → 4 7 → 6 7 → 6
(34) ◦ (623174) ◦ (34) = 4 7→ 3 7→ 1 7→ 1 = (624173), (A.37)
5 7→ 5 7→ 5 7→ 5
6 7 → 6 7 → 2 7 → 2
7 7→ 7 7→ 4 7→ 3
1 7→ 2 7→ 2 7→ 1
2 7 → 1 7 → 3 7 → 3
(12) ◦ (2)(13)(45) ◦ (12) = 3 7→ 3 7→ 1 7→ 2 = (1)(23)(45), (A.38)
4 7→ 4 7→ 5 7→ 5
5 7→ 5 7→ 4 7→ 4
solutions 227
1 7→ 1 7→ 3 7→ 4
2 7 → 2 7 → 4 7 → 3
(34) ◦ (13)(452) ◦ (34) = 3 7→ 4 7→ 5 7→ 5 = (14)(352). (A.39)
4 7→ 3 7→ 1 7→ 1
5 7→ 5 7→ 2 7→ 2
We see that the result of the application of a 2-cycle (and its inverse) to a generic
permutation p, exchanges in p the position of the two numbers defining such
2-cycle:
chapter 4
1. Cosets have all their elements in common or they are distinct (see The-
orem 4.1). This means that every element of the group must appear in
exactly one distinct coset. Thus, since each coset has the same number
of elements, the number of distinct cosets multiplied by the number of
elements in the cosets is equal to the order of the group.
2. The center of a group G is C = { g ∈ G | gh = hg, ∀h ∈ G }. Let now
Dn = hρ, σ | ρn = σ2 = 1, ρσ = σρn−1 i. From this presentation we
see that any h ∈ Dn can be written as h = σ a ρb , where a = 0, 1 and
b = 1, ..., n. In order to check what is the center we can then see for which
values of a and b the element h commutes with the generators of Dn . The
commutator with σ gives
σ a ρb σ = σ a+1 ρn−b = σσ a ρb ⇔ b = n − b ⇔ b = n/2, (A.44)
which is satisfied only for even n, or when b = 0. The commutator with
ρ gives
a b −1
σ ρ 6= σ a ρb ρ, a = 1, n 6= 2
ρσ a ρb = b + 1 (A.45)
ρ , a=0
We then conclude that
CDn = {1, ρn/2 }, (A.46)
when n is even, and
C Dn = { 1 } , (A.47)
when n is odd. Explicitly:
σ a ρb ρn/2 = ρn−n/2 σ a ρb = ρn/2 σ a ρb . (A.48)
3. We have 9 subgroups of order 2: {1, (12)}, {1, (13)}, {1, (14)}, {1, (23)},
{1, (24)}, {1, (34)}, {1, (12)(34)}, {1, (13)(24)}, {1, (14)(23)}. 4 subgroups
of order 3: {1, (123), (132)}, {1, (124), (142)}, {1, (134), (143)}, {1, (234),
(243)}. 7 subgroups of order 4, 3 of them are isomorphic to Z4 , namely {1,
(1234), (13)(24), (1432)}, {1, (1243), (14)(23), (1342)}, {1, (1324), (12)(34),
(1423)}, and 4 are isomorphic to Z2 × Z2 : {1, (12)(34), (13)(24), (14)(23)},
{1, (12), (34), (12)(34)}, {1, (13), (24), (13)(24)}, {1, (14), (23), (14)(23)}.
Other subgroups of higher order are S3 , D4 and A4 .
solutions 229
we conclude that a, b, c ∈ A8 .
c) H is generated by a e b, while the elements of G come from products
of a, b and c. It is straightforward to see that a and c commute
6. The identity must be in any subgroup. We can then take the other ele-
ments one by one and see if they form a subgroup. An obvious subgroup
is H0 = {1, −1}. For any of the other elements the set {1, Ji } does not
close and, for any given i, needs also −1 = Ji Ji and − Ji = (−1) Ji . The
result is a subgroup Hi = {±1, ± Ji }, where i = 1, 2, 3. Starting from
{1, − Ji } we end up with the same subgroups Hi . Adding a third element
230 solutions
φ( a, b) = ρ, φ( a2 , b2 ) = φ(1, b2 ) = ρ2 ,
φ( a3 , b3 ) = φ( a, 1) = ρ3 , φ( a4 , b4 ) = φ(1, b) = ρ4 , (A.54)
φ( a5 , b5 ) = φ( a, b2 ) = ρ5 , φ( a6 , b6 ) = φ(1, 1) = 1.
chapter 5
A O
A⊕B = . (A.55)
O B
solutions 231
We then have that when A and C have the sane dimensions and so do B
and D
A O C O AC O
( A ⊕ B)(C ⊕ D ) = = = AC ⊕ BD.
O B O D O BD
(A.56)
For the direct product, using double indices as in the text, we have that
We see that
−i −i
−i −i
γ0 = γ1 =
, , (A.59)
−i i
−i i
−1 −i
1 i
γ2 = γ3 =
, . (A.60)
1 i
−1 −i
The product
γ5 = −i γ0 γ1 γ2 γ3 (A.61)
follows by taking the products in each factor of the tensor product defin-
ing the matrices:
1 1
1 1
D R (b) = D R ( ab) =
, . (A.65)
1 1
1 1
Since the group is abelian all its irreducible representations are 1-dimensional.
This implies that we should reduce this 4-dimensional representation into
the sum of 4 1-dimensional representations. In other words, we can put
all its elements in a diagonal form. This is done by computing the eigen-
values and eigenspaces of the generators a and b. From this computation
we get that
1 1 1 1
1 1 1 −1 1 −1 1 1
√ , √ , √ , √ (A.66)
4 1 4 −1 4 1 4 −1
1 1 −1 −1
4. a) The representation given in the text for the generators of the group
is block diagonal, with the first block given by a 3 by 3 matrix and
the second one 1-dimensional. Since all the other elements of A4 are
given by products of these matrices, they all have the same block-
diagonal structure.
b) From the decomposition of the matrices D ( a) and D (b) we already
have two representations, a 3-dimensional one and a 1-dimensional
one. We also know that we always have the trivial representation
D ( a) = D (b) = 1, which has also dimension 1. The last one is obvi-
ously the complex conjugate of the 1-dimensional complex represen-
tation D ( a) = 1, D (b) = e2πi/3 , namely D ( a) = 1, D (b) = e4πi/3 .
c) The elements of H have order 2, hence it is isomorphic to Z2 ×
Z2 . This is a normal subgroup of A4 . Its elements are (12)(34),
(13)(24), (14)(23) and the identity, which we know are closed under
conjugation in S4 and hence also in A4 . (Note that in general conjugacy
classes of Sn and An differ, because elements with the same length of cycles
may be connected by odd cycles, which are not in An . However the only
result may be to split a conjugacy class in multiple conjugacy classes and
hence the argument here goes through) The quotient gives Z3 , generated
by b.
1
P± (ψ ⊗ χ) = (ψ ⊗ χ ± χ ⊗ ψ) (A.68)
2
234 solutions
we get
h i
1
P± P± (ψ ⊗ χ) = P± 2 (ψ ⊗ χ ± χ ⊗ ψ)
1 (A.69)
= 4 [(ψ ⊗ χ ± χ ⊗ ψ) ± (χ ⊗ ψ ± ψ ⊗ χ)]
1
= 2 (ψ ⊗ χ ± χ ⊗ ψ) = P± (ψ ⊗ χ).
2 (1 + D + 1 − D ) = 1 and P+ P− =
1
It is then trivial that P+ + P− =
P− P+ = 0.
6. To be filled in...
chapter 6
1 1 1 1 1 1
( D11 , D11 ) = 1+ + +1+ + = , (A.70)
6 4 4 4 4 2
1√ 1√ 1√ 1√
1
( D11 , D12 ) = 0− 3+ 3+0− 3+ 3 = 0,(A.71)
6 4 4 4 4
1√ 1√ 1√ 1√
1
( D11 , D21 ) = 0− 3+ 3+0+ 3− 3 = 0,(A.72)
6 4 4 4 4
1 1 1 1 1
( D11 , D22 ) = 1− − −1+ + = 0, (A.73)
6 4 4 4 4
1 3 3 3 3 1
( D12 , D12 ) = 0+ + +0+ + = , (A.74)
6 4 4 4 4 2
1 3 3 3 3
( D12 , D21 ) = 0+ + +0− − = 0, (A.75)
6 4 4 4 4
1√ 1√ 1√ 1√
1
( D12 , D22 ) = 0+ 3− 3+0− 3+ 3 = 0,(A.76)
6 4 4 4 4
1 3 3 3 3 1
( D21 , D21 ) = 0+ + +0+ + = , (A.77)
6 4 4 4 4 2
1√ 1√ 1√ 1√
1
( D21 , D22 ) = 0+ 3− 3+0+ 3− 3 = 0,(A.78)
6 4 4 4 4
1 1 1 1 1 1
( D22 , D22 ) = 1+ + +1+ + = . (A.79)
6 4 4 4 4 2
1
( Dij , Dkl ) = δ δ (A.80)
2 ik jl
chapter 7
(χ D1 , χ D2 ) = 1. (A.101)
4.
= trD1 ( g) trD2 ( g) = χ D1 · χ D2 .
(A.103)
5. If D is unitary D † D = 1. Then
1
π i 0 1
π −i 0
D ( J ) = U 0, , 0 = , D (− J ) = U 0, − , 0 = ,
2 0 −i 2 0 i
(A.109)
0
2
π 0 1 2
π −1
D( J ) = U , 0, 0 = , D (− J ) = U − , 0, 0 = ,
2 −1 0 2 1 0
240 solutions
(A.110)
π π 0 i π π 0 −i
D( J 3 ) = U , 0, = , D (− J 3 ) = U , 0, − = .
2 2 i 0 2 2 −i 0
(A.111)
b3 = a3 b ∼ a5 b ∼ ab}. (A.113)
c) All elements of the form an b are off-diagonal and hence χ1 ( an b) =
χ2 ( an b) = χ3 ( an b) = χ4 ( an b) = 0. We then need only to com-
pute the characters of an element for each of the conjugacy classes
C2 , C3 and C4 . We have that χ1 ( a) = χ4 ( a) and χ2 ( a) = χ3 ( a),
χ1 (b2 = a3 ) = χ4 (b2 = a3 ) = −2 and χ2 (b2 = a3 ) = χ3 (b2 = a3 ) = 2,
χ1 ( a2 ) = χ4 ( a2 ) = 2 cos(2π/3) and χ2 ( a2 ) = χ3 ( a2 ) = 2 cos(4π/3).
This implies that the characters of D1 and D4 are identical and there-
fore they are equivalent representations. The same is true for D2 and
D3 . We can also check that (χ1 , χ2 ) = 0:
1
( χ1 , χ2 ) = 12 (1 · 2 · 2 + 2 · 2 cos π/3 · 2 cos(2/3π )
(A.114)
2 · 4 cos(2/3π )2 + (−2)(2) = 0.
+
solutions 241
chapter 8
C1 C2 C3 C4 C5
1 1 1 1 1 1
10 1 −1 1 1 −1
2 2
3 3
30 3
( χ2⊗10 , χ2⊗10 ) = 1
12 ∑ g (χ2⊗10 )∗ ( g)(χ2⊗10 )( g))
(A.115)
1 1
= 12 ∑ g |χ2 ( g)|2 · |χ10 ( g)|2 = 12 ∑ g |χ2 ( g)|2 = (χ2 , χ2 ),
C1 C2 C3 C4 C5
1 1 1 1 1 1
10 1 −1 1 1 −1
2 2 0 2 −1 0
3 3
30 3
C1 C2 C3 C4 C5
1 1 1 1 1 1
10 1 −1 1 1 −1
2 2 0 2 −1 0
3 3 1 −1 0 −1
30 3 −1 −1 0 1