Professional Documents
Culture Documents
A Physicist’s Primer
Q. Ho-Kim
c Q. Ho-Kim (Ho Kim Quang). Group Theory: A Physicist’s Primer.
2014–19 ♠ 191115
Group theory. Symmetries in physics. Finite groups. Lie groups. Lie alge-
bras. Structures of groups. Representations of groups. Representations of
Lie algebras. Groups in particle physics.
Cover illustration: The Palace of the Alhambra in Granada contains a treasure trove of
mosaics, tiles, and other ornaments of wonderful symmetric designs, as exemplified by
this intricate interlacing pattern in the Hall of the Two Sisters shown on the front cover.
(Source: gwydir.demon.co.uk/knots/islamic. Courtesy of Jo Edkins).
Group Theory
A Physicist’s Primer
Q. Ho-Kim
♠ 191115
c Q. Ho-Kim (Ho Kim Quang). Group Theory: A Physicist’s Primer.
Cover illustration from image by Jo Edkins, used with permission.
2014–19 ♠ 191115
Contents
Preface v
iii
iv CONTENTS
C. Manifolds 291–304
References 305–306
Index 307–312
v
Preface
This book is based on the notes for lectures I have given for several years to
advanced-undergraduate and graduate physics students. I aim to explain the
basic elements of group theory from the ground level up to the point where the
student knows enough to feel comfortable with the group applications in physics
literature. I have no pretense of mathematical rigor whatsoever: the necessary
definitions are given and the main theorems stated as clearly and as precisely as
I can, but they are justified (if at all) by plausibility arguments rather than by for-
mal proofs; and abstract concepts are in general illustrated by concrete examples.
The first two chapters contain—beside a summary of the general facts of el-
ementary algebra (set, map, morphism) and linear algebra (vector, matrix, oper-
ator) that one may have learned in requisite mathematics or physics courses—
a detailed description of the structures and representations of finite groups over
real or complex (C) fields. I also state here the result on the classification of fi-
nite simple groups, a recent outstanding achievement in mathematics. Chapter 3
brings a general description of the continuous groups, which of course include
many of the groups familiar to any physics students and bearing the illustrious
names of Galileo, Lorentz, and Poincaré. This prepares the student on a super-
ficial level for the study of a special but very important category of groups, the
compact simple Lie groups and associated algebras, which becomes the focus of
the remaining part of the book. The next two chapters are devoted to sl (2, C ) (the
angular-momentum algebra) and sl (3, C ) (the eightfold-way algebra); they ease
in many key concepts and properties also applicable in the more general context.
In the final two chapters, I present the classical material (due to S. Lie, É. Cartan,
W. Killing, H. Weyl, E. Dynkin, and others) on the structures and representations
of semisimple Lie algebras, and discuss briefly some applications of Lie groups
in the theory of the fundamental particles and fields in physics. The book ends
with three appendices, one on the use of the symmetric group to understand the
irreducible representations of the special unitary Lie groups; the second on the
application of Clifford algebra to construct half-integer spin representations of
the special orthogonal Lie groups; and the third on manifolds. Finally, some of
the sources I have used in preparing my lectures and writing this book are listed
in the References.
Group Structure
In this chapter we give the basics of group theory with emphasis on the finite-
order groups: definitions and a few of the most important properties of group
structure. Information on a group can be gained from an analysis of the rela-
tionships among its elements (multiplication table), or its subsets (subgroups,
cosets, equivalence classes). The various ways in which a group divides (quo-
tient groups), or acts on sets or groups (map, conjugation, multiplication) give us
a deeper understanding about its structure. To illustrate the concepts discussed,
we will use simple examples from point groups and permutation groups. We do
not know about all the groups, but mathematicians now understand very well
a special but important category of groups, called the finite simple groups; we
briefly discuss their characterization towards the end of the chapter.
1
2 CHAPTER 1. GROUP STRUCTURE
Multiplication Table
The essence of a finite group of low order can be captured in a table, called mul-
tiplication table (or group table; first used by Arthur Cayley around 1854), which
enumerates all its elements and gives their pairwise products. The elements are
listed in the first row and the first column (normally in the same order) and
the product of the element labeling row a by the element labeling column b is
recorded at the intersection of row a and column b, like this:
. . . b
.
.
a ab
C3 e a a2
e e a a2
a a a2 e
a2 a2 e a
The groups C1, C2, C3 have a very simple structure: every group element is a
power of a special element called the group generator. This can be generalized
to arbitrary order n: The set consisting of the elements a, a2, . . . , an−1 , and an = e
under multiplication forms an abelian group, called the cyclic group of order n
and denoted Cn . Its elements can be visualized as lying at the vertices of a regular
n-gone on the unit circle, thus forming a cycle. It can be realized for example
with the generator exp (iθn ) representing an counterclockwise rotation on a plane
about the origin by an angle equal to θn = 2π/n; every other group element is a
discrete rotation by an angle equal to a multiple of θn .
| G | = 4: Call e, a, b, c the four elements. Starting with a as a generator, we may
have either b = a2, or b is a new generator. In the first case, with b = a2, we must
have c = a3 and a4 = e to close the group. In the second case, with b a generator,
we must necessarily have c = ab. Either case leads to a possible group.
The first case yields the cyclic group C4, with b = a2, c = a3, a4 = e, equivalent
to the multiplicative group {± e, ± i } defined in Example 1, as seen by identifying
1.1. DEFINITION AND BASIC PROPERTIES 5
e a b c
e e a b c b = a2
C4 a a b c e c = a3 = ab (1.1)
b b c e a
e = a4 = b 2
c c e a b
The only other order-4 group is again an abelian group, the Klein four-group,
denoted by V, with the following multiplication table
e a b c a = bc
e e a b c b = ac
V a a e c b (1.2)
c = ab
b b c e a
c c b a e e = a2 = b 2 = c 2
We see here another way of specifying a group, namely by its generators and
rules : thus V is completely defined by two generators a and b that satisfy the
relations a2 = e, b2 = e, and ba = a−1 b. In this form it is called D2 , so that D2 =
h a, b : a2 = b2 = ( ab)2 = e i. So V is equivalent to D2, an abstract group which can
be realized by all the transformations that leave a rectangle invariant, such that
when some or all of its edges swap places the rectangle appears unchanged (we
call transformations with this property symmetry transformations). One may
take for the symmetry transformations on the rectangle: (i) doing nothing, (ii)
reflection about the median line perpendicular to two edges, (iii) reflection about
the other median line, (iv) counterclockwise rotation of the figure about the center
through an angle of π. This symmetry group is called the group of symmetries of
the rectangle, isomorphic to D2; depending on the specific transformations chosen,
it may carry different names, e.g. C2v, or C2h (the molecules H2O and SF4 have
such a symmetry). Note that the center of the rectangle is left fixed under any
symmetry operation described. The symmetry groups of finite bodies that leave
at least one point of the body fixed are called ‘point groups’. They include rotations
through a definite angle about some axis and mirror reflections in a plane. On the
other hand, the term ‘space group’ refers to a group of an infinite system (e.g. an
infinite crystal lattice) which includes translation. See [Ham] Chapter 1.
C OMMENTS . Let a, b, c be elements of a group G. If ab = ac then b = c; in other
words, if b 6 = c then ab 6 = ac. Similarly with right multiplication of b and c by a.
Each and every element of G appears once and only once in a row or a column
of a multiplication table for group G. Therefore if Jn is a row of the n = | G |
elements of G arranged in some order, then aJn is just a row of the same elements
rearranged in a different order: the binary operation a · Jn acts like a permutation of
the elements. Similar remarks hold for left multiplication Jn a. All the elements of
G are displayed in each row or column always in a different order.
6 CHAPTER 1. GROUP STRUCTURE
+ 0 1 b c 0 1 b c
0 0 1 b c 0 0 0 0 0
1 1 0 c b 1 0 1 b c
b b c 0 1 b 0 b c 1
c c b 1 0 c 0 c 1 b
Addition in F4 Multiplication in F4
From the addition table, we note that ( F4, +) is isomorphic to the Klein group
V∼= Z2 + Z2 , and from the multiplication table that ( F4×, ) is isomorphic to the
cyclic group C3. It turns out to be true in general ([Hu] p. 279) that for q = p m
with any prime p the additive group ( Fq, +) is isomorphic to Z p ⊕ · · · ⊕ Z p (m
summands), and the multiplicative group ( Fq×, ) is cyclic, of order q − 1.
Matrix groups over finite fields. As an example of applications to group theory,
we consider the general linear group GL( n, Fq ) ≡ GL( n, q ) with q = pm , prime p,
and integer m > 1. Since the field over which it is defined is finite, GL( n, q ) must
have finitely many elements. How many elements are there?
If n = 1, GL( n, q ) ∼
= F× ×
q = ( Fq , ) , which has ( q − 1 ) elements.
If n = 2, any matrix M ∈ GL( n, q ) is of the form ac db , with a, b, c, d ∈ Fq ,
such that ad 6 = bc for M to be invertible, i.e. for det M 6 = 0. We know that the
determinant of a matrix is nonzero if and only if the rows of the matrix are linearly
independent (Chap. 2 Sec. 2.1 #10). The first row [ a, b ] can be anything other than
[0, 0], so there are q2 − 1 possibilities for [ a, b ]. The second row [ c, d ] must be
linearly independent from the first, i.e. can be anything other than a multiple of
the first (must not be x[ a, b ] where x ∈ Fq ). So there are q2 − q possibilities for the
second row. Thus, finally, there are ( q2 − 1)( q2 − q ) matrices in GL(2, q ).
These arguments can be generalized to GL( n, q ) for any positive integer n.
Let M ∈ GL( n, q ). The first row of M can be anything other than the zero row,
so there are ( qn − 1) possibilities. The second row must be linearly independent
from the first, i.e. can be anything other than a multiple of the first, so there
are ( qn − q ) possibilities for the second row. Generalizing, the ith row must be
linearly independent from all the first ( i − 1) rows, i.e. it cannot be a linear combi-
nation of the first ( i − 1) rows with combination coefficients in Fq . Since there are
qi−1 linear combinations of the first ( i − 1) rows, we have ( qn − qi−1 ) possibilities
for the ith row. So, in the end, there are ( qn − 1)( qn − q ) · · · ( qn − qn−1 ) matrices
of dimension n × n whose rows are mutually linearly independent. Thus, the
group GL( n, q ) has ∏in=−01 ( qn − qi ) elements.
The special linear group SL( n, F ) is defined to be a subgroup of GL( n, F ), such
that SL( n, F ) = { M ∈ GL( n, F )| det M = 1} . The det function maps GL( n, F ) to
F× , which is a surjective homomorphism. The kernel of this homomorphism
is precisely SL ( n, F ), from which it follows that GL( n, F )/SL( n, F ) ∼ = F× (see
the next two Sections 1.4–5; also First isomorphism theorem in [Hu] p. 44) . So
|SL( n, F )| = |GL( n, F )|/ |F× |, which leads to the order of the finite special linear
group: |SL( n, Fq )| = [ ∏in=−01 ( qn − qi )] / (q − 1).
8 CHAPTER 1. GROUP STRUCTURE
1.2 Subgroups
A group can also be analyzed in terms of subsets of elements; well designed
subsets can sharpen the differences and similarities between the group elements
and yield useful information about the group itself. In physics, this corresponds
to introducing a boundary condition or an interaction in a physical system that
leaves the system with a reduced symmetry. We will consider two such types of
subsets: subgroups, to be discussed in this section, and equivalence classes, to be
studied later in the chapter. The notion of set is of course used in other ways too.
For example, if X is the set of generators of a group G, then one may refer to G by
h Xi to emphasize the special role of X. Thus, h a i = { ak | k ∈ Z } is the signature
of a cyclic group of some order generated by a.
Let G be a group and H a non empty subset of G. If for every a, b ∈ H we
have ab ∈ H, then H is said to be closed under multiplication in G.
Definition 1.2 (Subgroup). A nonempty subset H of a group G is said to be a subgroup
of G if it is closed under the product in G, and is itself a group under the product in G.
This is denoted by H < G.
When the multiplication table for a group G is available, its subgroups can be
identified by simple inspection. Otherwise, a test based on the following result
can be used: A nonempty subset H of a group G is a subgroup if and only if ab−1 ∈ H
for all a, b ∈ H. To prove this proposition, remark that the identity e of G, being
unique, is also the identity of H, and the inverse of a ∈ H is a−1 in G. The direct
proposition is obvious. For the converse, we assume ab−1 ∈ H for all a, b ∈ H.
By this assumption we have e = aa−1 ∈ H for any a ∈ H, and b−1 = eb−1 ∈ H
for any element b ∈ H. Then as we now know that b−1 ∈ H for every b ∈ H, we
have ab = a ( b−1)−1 ∈ H, which means H is closed under the product in G. The
product in H is associative as it is in the group G. Hence H is a subgroup.
Every group G has at least two subgroups: h e i and G itself. They are trivial
subgroups. Any other subgroup that may exist is said to be non-trivial or proper.
E XAMPLE 7: If for an element a of a group G, the set a, a2, . . . , an−1 , an = e (where
n is the smallest positive integer for which an = e) forms a subgroup (the inverse
of ak being an−k , with 0 ≤ k ≤ n), then this subgroup is called a cyclic subgroup
of G, and denoted h a i. The order of element a equals the size of the subgroup h a i.
E XAMPLE 8: Given a group G, the set { c ∈ G | cx = xc for all x ∈ G } is an abelian
subgroup of G, called the center (Zentrum) of G and denoted C ( G ) (or Z( G )). On
the other hand, CG ( x) = { c ∈ G | cx = xc } is called the centralizer of x ∈ G. Two
simple examples are: Z(GL( n, F )) = { a · In | a ∈ F× } for the general linear group;
and Z(SL( n, F )) = { a · In | a ∈ F× , an = 1} for the special linear group.
E XAMPLE 9: For a fixed positive integer n, consider the group G = GL( n, F )
of invertible n × n matrices over the field F. We now verify that the subset H
of invertible matrices with determinant equal to one under matrix multiplica-
tion is a group, and hence a subgroup of G. Let A, B ∈ H, then det ( AB−1) =
det A/ det B = 1, and so AB−1 ∈ H. This suffices for H to be a subgroup of G;
we write this as SL( n, F ) < GL( n, F ). More on these groups in Chapter 6.
1.2. SUBGROUPS 9
Homomorphism
In analyzing group structure, it is important to relate different groups and to
compare them. Here two notions are helpful: maps that preserve group structure,
and distinctiveness of groups.
Definition 1.3 (Homomorphism, Isomorphism). For given groups G and G0 , a func-
tion f : G → G0 is a homomorphism if f ( ab) = f ( a) f (b) for all a, b ∈ G. If f is a
one-to-one correspondence, f is called an isomorphism, and G and G0 are then said to
be isomorphic (denoted G ∼ = G0 ). An isomorphism of a group with itself is called an
automorphism.
The expression ‘ f is a one-to-one correspondence’ (or a bijection) means that
for f : G → G0 there exist both a map g : G0 → G such that g f = e (identity of G)
and a map h : G0 → G such that f h = e0 (identity of G0 ), and that g = h. This map
g is called a two-sided inverse of f . The two-sided inverse of an isomorphism f
is unique and denoted f −1. In terms of group structure, isomorphic groups are
considered the same, whereas non-isomorphic groups, distinct.
Let f : G → G0 be a homomorphism of groups, and e be the identity of G and
e the identity of G0 , then f ( e) = e0 and f ( a−1) = f ( a)−1 for any a ∈ G. We
0
can see this as follows: First, f ( a) = f ( ae) = f ( a) f (e) and so f ( a )−1 f ( a) = f ( e)
and e0 = f ( e) follows. Secondly, e0 = f ( e) = f ( aa−1) = f ( a) f ( a−1), from which
f ( a )−1e0 = f ( a−1), and the final result: f ( a)−1 = f ( a−1).
Definition 1.4 (Kernel, Image). Let f : G → G0 be a homomorphism of groups, and e0
be the identity of G0 . The kernel of f is the set Ker f = { a ∈ G | f ( a) = e0 ∈ G0 } . If S
10 CHAPTER 1. GROUP STRUCTURE
1.3 Cosets
Cosets are sets of equivalent elements in the sense to be defined; we will see
that this concept produces the important result that for finite groups the number-
theoretic properties of their size determine their structure.
Let A be a set of elements a, b, c, . . . ; let R be a subset of the Cartesian product
A × A with the following properties (i) ( a, a) ∈ R for all a ∈ A; (ii) If ( a, b ) ∈ R,
then ( b, a ) ∈ R; and (iii) If ( a, b ) ∈ R and ( b, c ) ∈ R, then ( a, c ) ∈ R. These
three properties are referred to as reflexivity, symmetry, and transitivity. If they
are satisfied by R, then R is said to be an equivalence relation on A. If ( a, b ) ∈ R,
then we say that a is equivalent to b under R, and write a ∼ b or a ≡ b. For any
a ∈ A, the equivalence class of a under R is the set of all those elements of A that
are equivalent to a under R (denoted ā, or [ a ]), that is ā = { x ∈ A | x ∼ a } . Any
representative b ∈ ā defines in turn an equivalence class which coincides with ā,
i.e. b̄ = ā. The particular equivalence relation that interests us now is defined as
follows.
1.3. COSETS 11
Two left cosets (or two right cosets) of a subgroup H either coincide completely, or
else have no elements in common. Let gH and g0 H be two left cosets of H. If there
are no hi , h j ∈ H such that ghi = g0 h j, then gH and g0 H are disjoint, having no
common elements. If on the contrary, there are two elements h i , h j ∈ H such that
ghi = g0 h j, then g−1 g0 = hi h− 1 −1 0
j is in H, which implies g g H = H, or gH = g H.
0
G = H ∪ g2 H ∪ g3 H ∪ . . . ∪ gk H . (1.3)
cosets; only their numbers are equal. In the special case of [ G : H ] = 2, there are
just one left coset and one right coset, and they are the same.
(b) Even when G and H are infinite groups, [ G : H ] may still be finite. This
occurs for example with G = Z, H = h m i, where m ∈ N ∗ , then [ G : H ] = m, so
that we have the disjoint covering Z = 0̄ ∪ 1̄ ∪ · · · ∪ ( m − 1).
(c) The relation | H | ≤ (1/2)| G| can be used for a rapid computer search for
elements of a finite group that do not belong to a subgroup H of G.
(d) A given subgroup of a finite group and its distinct left (or right) cosets
together make up the whole group. Since a group contains in general several
subgroups, it may be partitioned in several different ways, each giving a coset
decomposition of the group.
(e) We can reduce the full symmetry a physical system may have to a sub-
group by imposing a constraint on the system. For example, we can place a rod
through the corners of a square, reducing the possible symmetry transformations,
thus forming a subgroup H. By moving the rod to a new position via a symmetry
operation g, we obtain a new set of operations gH. If we repeat enough times
using different g we can recover all the symmetry transformations the square
would have in the absence of constraints. The sets of operations gH that we need
to recover all the symmetries of the system are just the cosets.
E XAMPLE 15: Groups of order 6. We already have characterized all groups of
orders 1–4, and have learned that all groups of order 5 (or any prime) are cyclic.
Let us now find all distinct groups of order 6. Each of the elements of such a
group, call it G, must have an order equal to one of the divisors of 6 (i.e. 1, or 2,
or 3, or 6).
Assume that G has an element a of order 6, and so G = h a i is a cyclic group.
Any other possible structures can only have elements of order 1, 2, 3. Assume
that there is in G an element of order 3, so that it has the cyclic subgroup h a : a3 =
e i. If G also contains another distinct element b, then it contains the six distinct
elements of the set X = { e, a, a2, b, ba, ba2} , with b of order 2 or 3.
Now, if b3 = e, then b2 must be one of the elements of X; but each such case
would lead to contradictions of the assumptions, and so b cannot have order 3.
Let b2 = e and examine the product ab, which must be one of the six listed
elements. It cannot be e, a, a2, or b, because that would lead to contradictions. Is
ab = ba? If it is so, we have ( ab)6 = a6 b6 = e, that is, the group would contain an
element of order 6, contrary to assumption. So we are left with the last possibility:
ab = ba2. This implies ( ab )2 = abab = ab2 a2 = a3 = e. Element ab is of order
2, which does not contradict any assumptions, and we obtain a second group of
order 6, which is not isomorphic to h a i. It is the lowest-order nonabelian group
that exists. In conclusion, there are just two distinct abstract groups of order 6:
the cyclic group C6 and a dihedral group D3 = h a, b i defined by the algorithm:
a3 = b2 = e, ab = ba−1. (This turns out to be a special case of a general result,
which states that ‘Every group of order 2p, where p is an odd prime, is either the cyclic
group C2p or the dihedral group D p .’ See [Hu] p. 97.)
By taking all possible products of its elements we can establish its multiplica-
14 CHAPTER 1. GROUP STRUCTURE
h a, bi e b ba ba2 a a2
e e b ba ba2 a a2
b b e a a2 ba ba2
ba ba a2 e a ba2 b (1.4)
ba2 ba2 a a2 e b ba
a a ba2 b ba a2 e
a2 a2 ba ba2 b e a
30 20
2 10 3
transitive (see p. 10). The equivalence classes of G under conjugation are called
the conjugacy classes of G; the conjugacy class (or orbit) of an element x ∈ G is
the set x̄ = { gxg−1 | for all g ∈ G } , which should contain only distinct elements.
The number of such elements in x̄ is called its size, | x̄|.
Conjugacy classes have simple properties (which follow from the definition
and the general properties of equivalence classes), namely,
(1) In every group G the identity is a conjugacy class, ē, by itself.
(2) No conjugacy class other than ē is a subgroup.
(3) Every element of G belongs to some conjugacy class.
(4) All elements in a conjugacy class have the same order (( gxg−1)n = 1 iff
xn = 1). The converse is not true in general: different elements having the same
order may be found in different conjugacy classes.
(5) The conjugacy classes of a group G are either disjoint or equal. So they
provide a disjoint covering of the group: G = ∪ ḡ ḡ, so that | G | = ∑ḡ | ḡ| (cf.
| G | = [ G : H ]| H | for cosets of H < G). An important result (which we do not
prove here) is that in a finite group G each conjugacy class has size dividing | G |. The
reason for this is that, while ḡ 6 = ē is not a subgroup of G, its size is the index
[ G : CG ( g)] (where CG ( g) = { x ∈ G | xg = gx} is the centralizer of g ∈ G), which
divides | G |.
E XAMPLE 19: In an abelian group G, each element forms a class by itself. No two
elements are conjugate.
E XAMPLE 20: The dihedral group D3 (h a, b i with a3 = b2 = ( ab)2 = e) has its six
elements distributed among three conjugacy classes: ē = { e }, ā = { a, a2} , and
b̄ = { b, ba, ba2} , of sizes 1, 2, 3, partitioning G according to D3 = ē ∪ ā ∪ b̄, so that
6 = 1 + 2 + 3. All elements of the same order are grouped together in the same
conjugacy class (an exception to the general situation). The center is trivial, h e i.
E XAMPLE 21: In the quaternion group Q8 = {± e, ±i, ± j, ± k } there are five conju-
gacy classes: { 1} , {− 1}, { i, − i }, { j, − j }, { k, − k }, with 3 classes sharing 6 order-4
elements (e.g. i, − i, etc.). The center of Q8 is the set of elements in the conjugacy
classes of size 1, i.e. Z( Q8 ) = { 1, − 1}.
Conjugation can be extended to subgroups. Thus, if H is a subgroup of a
group G, then for any g ∈ G the set gHg−1 = { ghg−1 : h ∈ H } is also a subgroup
of G, called a conjugate subgroup to H. Therefore by Definition 1.7 a subgroup
N < G is normal if and only if it is equal to all its conjugates: N = gNg−1 for all
g ∈ G. For this reason a normal subgroup is also called an invariant subgroup
of G. If x ∈ N, then all gxg−1 are in N (and conversely), which is equivalent to
saying that N < G is normal if and only if it contains elements of G in complete classes.
E XAMPLE 22: Every subgroup of an abelian group is normal, being conjugate
only to itself.
E XAMPLE 23: In the group D3, the order-3 subgroup { e, a, a2} contains the com-
plete classes ē and ā and so is normal; whereas all of its order-2 subgroups { e, b },
{ e, ba } and { e, ba2} only contain some elements of b̄, and so are not normal.
The presence or absence of non-trivial normal subgroups is a key property in
characterizing two important related categories of groups:
1.5. QUOTIENT AND PRODUCT GROUPS 17
E XAMPLE 24: The only abelian simple groups are the cyclic groups of prime order
(having no proper subgroups at all). Cyclic groups with a non-prime order are
neither simple nor semisimple (all their subgroups are normal and abelian). The
nonabelian group D3 has a normal abelian subgroup, and so is neither simple
nor semisimple. Nonabelian simple groups of small order are rare; for example,
there are only two nonabelian simple groups of order less than 200, namely, the
alternating group A5, of order 60, and a subgroup of the symmetric group S7, of
order 168. See p. 28.
C OMMENTS . A symmetry group may contain different types of transformations
(rotations, reflections) and so there has to be some way of reflecting these simi-
larities and differences in its description: this is the reason for having the concept
of conjugacy. In D3 = h a, b i, for example, the elements a and a2 regarded as
planar rotations through different angles about the center of the triangle are ge-
ometrically similar to each other, but quite different from reflections across the
median lines represented by b, ba, and ba2; in no ways a reflection can be made
into a pure rotation, but reflections across different lines are essentially similar, as
are planar rotations about a point through different angles – just different points
of view. Group theory makes this distinction between the two sets precise by
putting them in distinct classes, ā and b̄. In addition, many properties are pre-
served under the group transformations; this will be automatically the case if
those properties derive from elements of a normal subgroup. Thus, the angular
momentum associated with the ā class is conserved under the full group D3.
with eN = N the identity element. It is called the quotient group or factor group,
written G/N and read “quotient of G by N”, or “G modulo N”.
Relationships between homomorphisms and quotient groups give us further
information about the structure of the latter.
Let f : G → G0 be a homomorphism between groups, and e and e0 the re-
spective identities of G and G0 . We now show that the kernel of f , defined by
Ker f = { a ∈ G | f ( g) = e0 } = K is a normal subgroup of G.
(i) For any a, b ∈ K, f ( ab) = f ( a) f (b ) = e0 e0 = e0 , and so ab ∈ K (closure of K).
If for every a ∈ K there is a c such that ac = a (i.e. c = e), then f ( ac) = f ( a ); as
f ( ac) = f ( a) f (c) we have f ( a ) f (c) = f ( a) for every a ∈ K. Hence f ( c) = e0 , so
that c = e ∈ K (identity of K). Finally, for any a ∈ K, we have the inverse a−1 ∈ G,
and e = aa−1 ∈ G. We have f ( e) = f ( aa−1) = f ( a) f (a−1) = e0 f ( a−1). Hence
e0 = e0 f ( a−1), or f ( a−1) = e0 , which says that a−1 ∈ K (existence of inverse).
These are the conditions for K < G.
(ii) For some a ∈ K and any g ∈ G, we have f ( gag−1) = f ( g) f (a ) f ( g−1) =
f ( g)e0 f ( g)−1 = e0 , which means that gag−1 ∈ K for any a ∈ K, i.e. gK g−1 ⊂ K
for all g ∈ G. This is a sufficient condition for the subgroup K of G to be normal,
which we write K / G.
Thus, with the kernel K of f : G → G0 being a normal subgroup of G, the
set of all cosets of K, including K itself, is a group G/K under multiplication
( gK )(hK ) = ghK. Furthermore, the quotient group G/K is isomorphic to G0 .
In the homomorphism between groups f : G → G0 , all the elements ai, i =
1, 2, . . . , k of K are sent to e0 . For some g ∈ G, the k elements ga1 , ga2 , . . . , gak of
G are mapped to f ( gai) = f ( g) f (ai) = f ( g)e0 = f ( g), with 1 ≤ i ≤ k. That is,
if f : ai 7 → e0 , then the same number of elements of G are sent to some element
g0 of G0 . In particular, if b ∈ gK then b = ga with a ∈ K, and f (b ) = f ( ga) =
f ( g)e0 = f ( g), and so f has the same effect on every element of a coset, and it
follows we can define the map f¯ : G/K → G0 by the function f¯( gK ) = f ( g) for
all g ∈ G. Now, since f¯( gK hK ) = f¯( ghK ) = f ( gh) = f ( g) f (h) = f¯( gK ) f¯( hK ),
group multiplication is preserved and f¯ is a homomorphism.
If gK = hK, then f¯( gK ) = f¯( hK ). Conversely, suppose f¯( gK ) = f¯( hK );
then f¯( h−1 gK ) = f¯( h−1K gK ) = f¯( h−1K ) f¯( gK ) = f¯−1 ( hK ) f¯( gK ). Assuming
f¯( gK ) = f¯( hK ), we have f¯( h−1 gK ) = e0 , or h−1 gK = K. Hence gK = hK, or the
mapping is one-to-one, and f¯ is an isomorphism.
In summary: Let f : G → G0 be a homomorphism of groups. Then the kernel
of f , called K = Ker f , is a normal subgroup of G, and the quotient group G/K of G
by K can be formed. The map f¯ : G/K → G0 uniquely defined by f¯( gK ) = f ( g)
for all g ∈ G is an isomorphism, so that the groups G/K and G0 are isomorphic, or
symbolically G/K ∼ = G0 .
C OMMENTS .
(a) If G is a simple group, then it has only h e i and G as normal subgroups, and
G/ h ei ∼
= G and G/G ∼ = h e i are its only quotient groups.
(b) Quotient groups are important for at least two reasons. One, from any
group G, one constructs a new model of group G/H. Two, assuming a finite
1.5. QUOTIENT AND PRODUCT GROUPS 19
group G, one has two groups, H < G and G/H, both related to G and smaller in
size than G, a useful fact in proofs by induction on the size of the group.
E XAMPLE 25: If m > 1 is a fixed integer and a ∈ Z, then the equivalence class
of a under congruence modulo m, which is ā = a + h m i, is a coset of h m i in Z,
and so as sets, Zm = Z/ h m i. Furthermore, as ā + b̄ = a + b (a, b ∈ Z) the group
operations coincide. (Zm is sometimes written Z/mZ, or Z/m.)
E XAMPLE 26: The cyclic group C4 = h a; a4 = e i contains the normal abelian
subgroup H = { e, a2} . Its quotient group C4/H consists of H and M = aH =
{ a, a3} , with the coset products among its elements given by H H = H, H M =
H ( aH ) = ( aH ) H, or H M = M = MH, and MM = H. Define G0 = h c; c2 = e0 i
and the mapping f¯ : H 7 → e0 , aH 7 → c. We see that C4/H is isomorphic to G0 ∼
= C2 .
2 2
E XAMPLE 27: The abelian Klein group V = h a, b; a = b = e, ab = ba i has
two subgroups, H = { e, a } and K = { e, b} , both abelian, cyclic, and normal.
Two quotient groups can be formed, V/H = { H, bH } and V/K = { K, aK }, both
isomorphic to C2. Note that C4 and V are not isomorphic in spite of their similar
decompositions (C4 has one normal subgroup, V has two).
E XAMPLE 28: The cyclic group C6 = h a; a6 = e i has two subgroups, H = { e, a3}
and K = { e, a2, a4 } , both abelian and normal. To them correspond two quotient
groups, C6 /H = { H, aH, a2 H } and C6/K = { K, aK } together with coset mul-
tiplication rules. Define the order-3 cyclic group C3 = { b, b2, b3 = e0 } and the
mapping: H 7 → e0 , aH 7 → b, and a2 H 7 → b2 . This is an isomorphism relating
C6/H with C3. Similarly for C6/K, the mapping f¯ : K 7 → e0 and f¯ : aK 7 → c (where
c2 = e0 ) shows that C6/K ∼ = C 2.
E XAMPLE 29: The nonabelian group D3 = h a, b; a3 = b2 = e, bab−1 = a−1 i has
three order-2 subgroups K i and one order-3 subgroup H, of which only the latter
H = { e, a, a3} is normal (and abelian). Together with the coset bH = { b, ba, ba2} ,
it produces the quotient group D3 /H = { H, bH } , with coset product rules. The
mappings H 7 → e0 and bH 7 → c (with c2 = e0 ) define an isomorphism of groups
D3 /H and C2. Hence we have D3 /H ∼ = C2 .
E XAMPLE 30: Z( Q8 ) = { 1, − 1} is a normal subgroup in Q8 . The factor group
Q8 /Z( Q8 ) is an abelian group isomorphic to V. It consists of the elements given
by the sets {± 1}, {± i }, {± j }, and {± k }.
H̄ = {( h, eK), h ∈ H } and K̄ = {( e H, k ), k ∈ K } .
It can be seen that H̄ and K̄ are normal subgroups of Ḡ, because for any g = ( h, k ),
we can show that g H̄g−1 ⊆ H̄ and gK̄g−1 ⊆ K̄. In addition, Ḡ = H̄ K̄ and
H̄ ∩ K̄ = { e } , where e = ( e H , eK ).
Definition 1.10 (Internal direct product). A group G is said to be the internal direct
product of the groups H and K if the following conditions are satisfied: (i) Both H and
K are normal subgroups of G; (ii) H ∩ K = { e }, where e is the identity of G; and (iii)
G = HK = { hk : h ∈ H, k ∈ K } .
In fact, the internal direct product can be defined in three equivalent ways:
Assuming a group G contains, possibly among others, the subgroups H1, H2 such
that G = H1 H2, then G is said to be an internal direct product of H1 and H2 if one
of the following set of conditions is satisfied:
(1) H1 and H2 are both normal in G, and H1 ∩ H2 = { e }.
1.5. QUOTIENT AND PRODUCT GROUPS 21
gk 1 g−1 = h1 h2 k 1h− 1 −1 −1
2 h 1 = h 1 k 1 h 1 ∈ H1 ,
where in the last step, we have used h2k 1 = k 1 h2. So we have gk 1 g−1 ∈ H1 for
every k 1 ∈ H1. Similarly, gk 2 g−1 ∈ H2 for every k 2 ∈ H2. So both H1 and H2 are
normal subgroups of G.
C OMMENTS . When we say that a group G is an internal direct group of Hi with
i = 1, 2, both H1 and H2 are, by definition, subgroups of G, and G is isomorphic
to the external direct group Ḡ = H1 × H2, but Ḡ does not contain Hi , but only
isomorphic copies of them. Because of this isomorphism, the distinction between
the two cases is often implicitly understood, and the adjectives ‘internal’ and ‘ex-
ternal’ are often omitted. This is one reason for the custom of writing H1 × H2 to
indicate indifferently the internal or external direct product of H1 and H2.
E XAMPLE 34: A ready-made example is provided by the group H × K of any two
groups. It contains two normal subgroups H̄ and K̄, which only have the identity
in common, and such that H × K = H̄ K̄. So, by definition, this group is the
internal direct product of H̄ and K̄.
E XAMPLE 35: The Klein group V has two order-two cyclic subgroups H1 = { e, a }
and H2 = { e, b } with ab = ba, and each of its four elements can be written as
22 CHAPTER 1. GROUP STRUCTURE
the product of one element of H1 and one of H2. Therefore V can be described
as the internal direct product of H1 and H2, which we write as V ∼ = C2 × C2 (or
V∼ = Z2 ⊕ Z2 in the additive notation).
E XAMPLE 36: The cyclic group C6, with the elements a, a2, a3, a4, a5, e = a6, has
two subgroups, namely H1 = { e, a3} and H2 = { e, a2, a4 } . Both are normal in
C6 (because this group is abelian) and have e as the only common element. So
C6 can be regarded as the internal direct group of H1, H2 (we can also check that
each element of C6 is a unique product of the type h 1 h2, that is e = ee, a = a3 a4,
a2 = ea2, a3 = a3 e, a4 = ea4, a5 = a3 a2). Since H1 ∼= C2 and H2 ∼ = C3, we write
∼
C6 = C2 × C3 (or Z2 ⊕ Z3 in the additive notation).
Both H1 and H2 are normal subgroups of C6. The cosets of H1 in C6 are a2 H1 =
{ a , a5 } and a4 H1 = { a, a4} , and we can resolve C6 into the sum of H1 and its
2
y
y 6
A 6
B A
-
O -x O x
C D
B C
Dn has n rotations e, a, a2, . . . , an−1 , and n reflections b, ab, a2b, . . . , an−1 b. Note
that an = b2 = ( ab )2 = e implies: (1) ab = ba−1, ba = a−1 b; (2) ak b = ba−k ,
bak = a−k b. So that (3) ( ak b )2 = e, showing that ak b (k = 0, . . . , n − 1) are indeed
reflections through the different reflection lines; and (4) a j ( ak b ) = ( ak b ) a− j. If n is
odd, there is a reflection for each vertex, the reflection line connecting this vertex
to the midpoint of the opposite side. If n is even, we have a reflection across
the line connecting each pair of opposite vertices, and a reflection across the line
connecting the midpoints of opposite sides (see Fig. 1.2 for n = 3, 4). The number
of reflection elements remains n whether even or odd.
Let h a i = { e, a, a2, . . . , an−1 } ∼ = Cn and h b i = { e, b } ∼
= C2, then Dn = h a ihb i.
The subgroup (of rotations) h a i is normal in Dn , but h b i is not, because for exam-
ple aba−1 = ba−2 is not in h b i. Symbolically, Dn ∼ = Cn o C2, for n ≥ 3.
The structure of Dn can be analyzed further, with the help of the commutation
relations given above, by identifying its conjugacy classes. We have the following:
(1) Odd n ≥ 3: there are ( n + 3) /2 conjugacy classes: { e }; { a±1} , . . . , { a±(n−1)/2 } ;
{ aib |0 ≤ i ≤ n − 1} . (2) Even n = 2m ≥ 4: there are n/2 + 3 conjugacy classes:
{ e }, { am } ; { a±1} , . . . , { a±(m−1) } ; { a2ib |0 ≤ i ≤ m − 1} , { a2i+1b : 0 ≤ i ≤ m − 1} .
Any group is the disjoint union of its conjugacy classes, and the order of the
group is the sum of its class sizes. This gives the partitions: 2n = 1 + 2 + · · · +
2 + n (the summands 2’s occur ( n − 1) /2 times) for odd n; and 2n = 1 + 1 +
2 + · · · + 2 + n/2 + n/2 (the summands 2’s occur m − 1 times) for even n. In the
second case, besides the identity, there is another class of size 1 containing only
an/2 , a 180-degree rotation. In addition, the reflections split into two classes: one
consisting of those (a2i b) with fixed lines connecting opposite vertices and those
(a2i+1 b) with fixed lines connecting midpoints of opposite sides. Finally, there are
differences in the group center and the centralizers depending on the parity of n.
Thus, recalling that the center is made up of the elements of the size-1 conjugacy
classes, we see that the center of Dn with n ≥ 3 is trivial for odd n, and { e, an/2 }
for even n. As for the centralizers, we can see, for example, CG ( b ) = { e, b } for
odd n, and CG ( b ) = { e, an/2, b, an/2 b } for even n. This difference shows up in the
unequal sizes of the reflection conjugacy classes in odd and in even n.
24 CHAPTER 1. GROUP STRUCTURE
(b) In a given cycle, the symbols are read from left to right, starting at any
point in the chain; for example, the cyclic permutations 1 → 2 → 3 → 1 can be
represented by any one of the equivalent cycles: (1 2 3), (2 3 1), or (3 1 2).
(c) A cycle represents an operation on a subset of I n . Two cycles commute
when they have no elements in common (and are said to be disjoint), for example
(1 2 3)(4 7) = (4 7)(1 2 3). Otherwise the order of application of the cycles (as op-
erators) is important, with the operators going from right to left, i.e. the rightmost
cycle acting first, then the next left on the result, and so on down the line, exactly
as in the usual convention in physics for products of operators, ab · x = a · ( b · x).
Example: (1i)(1j )(1i ) = ( ij ), but (1i)(1i )(1j ) = (1j )(i), both acting on { 1, i, j } .
(d) In this notation, the inverse of a k-cycle is (1 2 . . . k )−1 = ( k . . . 2 1), so that
group V; whereas the order-12 normal subgroup is (1) (12)(34) (123), which
S S
is none other than A4. It follows that one can construct two quotient groups from
S4, namely S4 /A4 ∼ = S3 and S4 /V ∼ = C 2.
Symmetric groups are important in group theory because they can be sys-
tematically analyzed and because they contain all the possible structures of finite
groups, as stated in the following theorem.
Theorem 1.2 (Cayley). Every finite group G is isomorphic to a subgroup of the sym-
metric group Sn , with n = | G |.
the action of g has the effect of permuting the order of the elements in the under-
lying set, and so induces a map p : G → S ( G ) defined by x 7 → gx = p g ( x) for all
x ∈ G. The permutation p g thus defined can be seen to satisfy the properties:
(i) p gg0 = p g p g0 , because on the one hand gg0 x = p gg0 ( x), and on the other
hand gg0 x = gp g0 ( x) = p g ( p g0 ( x)) = p g p g0 ( x).
(ii) ex = x = pe ( x), hence pe = (1) is the identity permutation.
(iii) x = p gg−1 ( x) = p g p g−1 ( x) = pe ( x), which defines ( p g )−1 = p g−1 .
This shows that, given a group G of order n, each permutation p g is uniquely
associated with a g ∈ G; and { p g| all g ∈ G } is a subgroup of Sn isomorphic to G.
E XAMPLE 40: C3 and S3. The cyclic group C3, with its three elements a, a2 = b,
a3 = e, is specified by the multiplication table shown on the left-hand side of
Eq. (1.12). The table on the right-hand side shows the permutations in group S3
to which the elements of C3 are mapped: e 7 → pe = (1), a 7 → pa = (123), and
b 7 → pb = (132). By the usual rules of permutation multiplication, we can check:
p2a = pb , p2b = pa , pa pb = pe , in agreement with a2 = b, b2 = a, ab = e. We have
thus shown that C3 is isomorphic to A3, the alternating subgroup of S3.
C3 e a b 1 2 3
e e a b pe 1 2 3
a a b e −→ pa 2 3 1 (1.12)
b b e a pb 3 1 2
E XAMPLE 41: D2 and S4. We show on the left hand side of (1.13) the multipli-
cation table in Eq.(1.2) of the dihedral group D2 , and on the right hand side the
permutations to which the elements of D2 are mapped.
D2 e a b c 1 2 3 4
e e a b c pe 1 2 3 4
a a e c b −→ pa 2 1 4 3 (1.13)
b b c e a pb 3 4 1 2
c c b a e pc 4 3 2 1
C OMMENTS . (a) Cayley’s theorem identifies the structure of any group G of or-
der n with that of an order-n subgroup R of the group of permutations Sn of
degree n. From the examples given above, we note the following properties of
R < Sn which also hold in all generality: (i) Except for the identity p e , the ele-
ments pi ∈ R leave no symbols unchanged. (ii) When resolved into independent
cycles, all the cycles in each element pi must have the same length: pi must be
a product of 1-cycles, or of 2-cycles, etc.. (They reflect the characteristic prop-
erties of the multiplication table.) These very special permutations are said to
be ‘regular’. Subgroups formed only of regular permutations are called regular
permutation subgroups.
(b) Cayley’s theorem implies that the number of distinct (non isomorphic)
groups of a given order is finite, because every one of them is isomorphic to some
subgroup of Sn , and as Sn is finite, it contains only a finite number of subgroups.
(c) We may use the theorem to determine group structures. Finding groups of
order n is equivalent to finding regular subgroups of the same order in Sn , and
we can do this by exploiting the fact that regular permutations, when resolved
into independent cycles, must have all cycles equal in length `, which means that
` must be a divisor of n. And so:
• If n is a prime number, the cycle lengths can only be 1 or n. So the regu-
lar subgroup can only contain the cyclic permutation (12 . . . n ) and its powers.
Therefore, a group of prime order has to be cyclic.
• If n is not a prime, the problem is to find the regular subgroups in Sn con-
taining permutations resolved into cycles whose lengths divide n. As an exam-
ple, take n = 4, which has divisors 4, 2, 1. First consider a degree-4 permutation
(1234), and take its successive powers. The permutations so obtained, (1234),
(13)(24), (1432), and (1234)4 = (1)(2)(3)(4), are the elements of a regular sub-
group of S4 isomorphic to C4. Starting with another 4-cycle leads to no new
structure. Next, take all possible 2-cycles, i.e. (12)(34), (13)(24), and (14)(23), each
of which has a square equal to (1)(2)(3)(4). They form another regular subgroup
of S4, which is isomorphic to D2. This exhausts all possibilities. In conclusion,
there are just two distinct groups of order 4, namely C4 ∼
= Z4 and D2 ∼ = Z2 ⊕ Z2 .
3, and h a3i of order 2. This theorem need not hold for nonabelian groups: For ex-
ample, D3 (size #6, factors 1,2,3,6) has proper subgroups of size 2 (h b i), and 3 (h a i),
but A4 (#12, divisors 1,2,3,4,6,12) has no subgroups of size 6 (but has subgroups
C2 ∼= {(1), (12)(34)} (#2); C3 ∼= {(1), (123)(132)} (#3); as well as subgroups of #4
of the type V ∼ = {(1), (12)(34), (13)(24), (14)(23)}.
Theorem 1.4 (Cauchy). If G is a group whose order is divisible by a prime p, then G
contains an element (and hence a cyclic subgroup) of order p.
Thus, D3 (#6 divisible by 2 and 3) has elements of order 2 or 3. All the groups
of order 8 (e.g. C8, D4, Q8) have at least one element of order 2.
The following theorem extends Cauchy’s theorem and guarantees the exis-
tence of subgroups in groups of prime-power order.
Theorem 1.5 (Sylow). If G is a group whose order is p n m, with n ≥ 1, p prime, and
p, m relatively prime, then G contains a subgroup of order p k for each 1 ≤ k ≤ n.
So, for example, the order of D4 may be written as 23 1, and so D4 must have
nontrivial subgroups of order 2 and 4, which is verified. The symmetric group
S4 has order 23 3, and so must contain subgroups of order 2, 4, 8, and 3, which is
also verified.
At this point, it is useful to recall a few facts seen in previous sections: (1)
Every cyclic group of order n is isomorphic to Zn . (2) Every group with a prime
order p is cyclic, isomorphic to Z p . (3) Every group of order 2p, where p is an
odd prime, is either the cyclic group Z2p or the dihedral group D p . (4) There are
five distinct groups of order 8, namely Z8 , Z4 ⊕ Z2 , Z2 ⊕ Z2 ⊕ Z2 , D4, and Q8
(see Problem 1.7). So we know, up to isomorphism, the structure of all the groups
with order smaller than 12, as shown in the accompanying table.
These results are among the early steps towards a classification of all the finite
groups. This task is far from being completed, but mathematicians have made
a very significant step in identifying all the finite simple groups. This is an out-
standing achievement because finite simple groups are the basic building blocks
30 CHAPTER 1. GROUP STRUCTURE
The sporadic groups are the 26 (and only 26) individual groups that do not
fit into any of the 18 countably infinite families in (1)–(3); they arise usually as
groups of automorphisms of multi-dimensional geometrical configurations. The
smallest sporadic is the Mathieu group M11 of order 7920, and the largest spo-
radic is the Fischer–Griess Monster group M, of order ≈ 8 · 1053, which con-
tains all but 6 of the other sporadic groups. The Monster group (conjectured
by Bernd Fischer and Robert Griess in 1973, and constructed by Griess in 1982)
may be regarded as the automorphism group of a 196 883-dimensional commu-
tative non-associative algebra, called the Griess algebra, also the first nontrivial
tier of the infinite-dimensional vertex operator algebra. It is the most intriguing
of all finite simple groups, with many apparent but still unexplained connections
to modular functions. (For example, the first nontrivial coefficient of a certain
modular function is equal to the sum of the dimensions of the smallest nontrivial
irreducible representations of M: 196 884 = 1 + 196 883. And the pattern contin-
ues on with 21 493 760 = 1 + 196 883 + 21 296 876, relating the second coefficient
of the modular function and the next highest dimension for the Monster group).
The search for an explanation of this completely unexpected relationship between
sporadic groups and modular functions is an object of the moonshine theory, a the-
ory with rich implications in number theory, algebra, and geometry, as well as
wide-ranging ramifications into physics, in particular, quantum field theory and
string theory.
Problems
1.1 Show that the set of all rotations R ẑ ( α ) around the z axis by an angle α acting
on the coordinates ( x, y ) for all values of α forms a (continuous) group.
1.2 Prove that a group is abelian if the order of any of its elements other than the
identity is 2.
1.3 Find the structure of the groups of order 4, 6, 8, and 10 with elements all of
order 2.
1.4 Prove that the intersection of two subgroups of a group of G is a subgroup of
G.
1.5 Use Cayley’s theorem to find all distinct groups of order 6.
1.6 Let G be a group that contains a p-order cyclic subgroup H = h a i, such that
p = | G | /2. Find the coset of H in G, and identify the structure of G for | G | =
4, 6, 8.
1.7 Show that there are five distinct groups of order 8, namely, the cyclic group
C8, the group of symmetries of the square D4 , the quaternion group Q8, and two
product groups, C4 × C2 and D2 × C2.
1.8 Enumerate the conjugacy classes, subgroups and normal subgroups, as well
as the factor groups, of the symmetric groups S2, S3 and S4.
1.9 The dihedral group D4 is the group of the symmetry transformations of a
square, generated by a rotation b through an angle π about a diagonal, and a
32 CHAPTER 1. GROUP STRUCTURE
Group Representations
This chapter begins with a review of the basic concepts and properties of linear
vector spaces essential to a formulation of representation theory and continues
on with a discussion of the different ways in which a group acts on sets or vec-
tor spaces. Groups are abstract: they are symbols and rules. Representations of
groups are concrete: they deal with vectors, operators, and matrices. The aims of
representation theory are to define, construct, and classify all the distinct repre-
sentations of groups. If a system (a differential equation or a physical object, for
example) exhibits some symmetry, then all associated symmetry transformations
form a group, and a deeper understanding of these symmetries gives us better
insight into the system. Mathematicians use representations to study group prop-
erties not readily accessible to groups by themselves, to explore new structures,
or apply group theory in other contexts. Physicists usually encounter groups by
way of their representations and find them in quantum-mechanical applications.
We restrict ourselves in this chapter to finite groups, paving the way for infinite
groups in later chapters.
33
34 CHAPTER 2. GROUP REPRESENTATIONS
by a scalar number (element of a field F). If F is real (R), the vector space is said
to be real; if F is complex (C), we have a complex vector space.
2. Given ψ, φ ∈ V , they are said to be linearly dependent, or parallel to each other,
if αψ + βφ = 0 for some nonzero scalars α and β. If there are no such scalars, the
vectors are linearly independent. A vector space V is n-dimensional if there exist
sets containing n linearly independent vectors but no sets with a greater number
of linearly independent vectors. If n is infinite, V is said to be infinite-dimensional.
3. We assume from here on finite spaces, V = Fn with F = R or C. In such a V any
set of n linearly independent vectors u1, u2, . . . , un span V in the sense that any
vector ψ in V may be written as a linear combination c1 u1 + c2 u2 + · · · + cn un ,
with ci ∈ F. The set { ui} is then said to be complete, or forms a basis of V ; the
numbers ci, called the components of ψ on { ui} , are unique up to normalization.
4. Let V and V 0 be two linear spaces. Suppose there is a map f that assigns to
every vector x in V a vector x0 in V 0, i.e. f : V → V 0, then we say that this map
defines an operator A from V to V 0, such that x0 = Ax. We then write A : V → V 0,
and consider the two statements, on f and A, interchangeable.
An operator A : V → V 0 is said to be linear if A ( αψ + βφ ) = αAψ + βAφ, or
antilinear if A ( αψ + βφ ) = α∗ Aψ + β∗ Aφ for every ψ, φ ∈ V and α, β ∈ C (α∗ is
the complex conjugate of α). If V 0 = F1 , i.e. x0 is an F-scalar, one usually writes
x0 = f ( x), rather than x0 = Ax, and call f a functional, rather than an operator.
5. The sum of the linear operators A : V → V and B : V → V is defined by
def def
( A + B ) ψ = Aψ + Bψ; and the product of A and a scalar α by ( αA )ψ = α ( Aψ)
for every vector ψ ∈ V . So the set of all linear operators in V also forms a linear
vector space. It is of dimension n2 if V is n-dimensional.
6. Let A be a linear operator such that the relation Aψ = 0 for ψ ∈ V holds only
if ψ = 0. Then A is said to be invertible, and its inverse, denoted A−1 , obeys the
relations AA−1φ = φ and A−1 Aψ = ψ for every φ and ψ in V .
7. Let V be a linear inner-product space (cf. #11) of dimension n, and let ui , with
i = 1, . . . , n, be a complete basis. Then we may write any vector ψ in V as a linear
superposition, and any linear operator A : V → V as
n n
ψ= ∑i=1 ui ci , Au j = ∑i=1 ui aij . (2.1)
This system of equations can be solved to give ci and aij . We may regard ci as
the components of a column vector c and aij the entries of an n × n matrix A.
These data, ci and aij , suffice to define ψ and A once the basis vectors ui are
specified. We may refer to c and A as the representations of ψ and A in the basis u.
(Once a basis is fixed, ‘matrix’ and ‘operator’ become practically the same.) All
algebraic operations on abstract vectors and operators are preserved in matrix
representations; to each operation on vectors and operators corresponds a similar
matrix operation. For example, to ψ0 = Aψ corresponds c0 = Ac, and to C = AB
corresponds C = AB. Relations in matrix representations, such as c0 = Ac and
C = AB, are basis-specific.
8. Change of basis . It is sometimes useful
or simpler to work in a different basis.
Let u1, u2, . . . , un and v1, v2, . . . , vn be two bases of V . The change of basis
2.1. LINEAR VECTOR SPACE 35
The relation c(v) = S−1 c(u) is the characteristic transformation property of a col-
umn vector under S, while A(v) = S−1 A(u) S guarantees that A(u) c(u) is also a
column vector, i.e. A(v) c(v) = S−1 A(u) c(u) . Comparing va = ∑m um Sma with
c(v) = S−1 c(u) , we note that the transformations between the basis vectors and
the transformations between the matrices are in inverse relation. The commuta-
tion relation SπV = πU S contained in (2.2) justifies the term intertwining operator
sometimes applied to S. The intertwining of Π V and ΠU by S is illustrated here:
S
V −−−−→ U
ΠV y
Π
y U
V −−−−→ U
S
More generally, two n × n matrices A and B over F are said to be similar if there
exists an invertible matrix M such that B = MAM−1.
9. Let A = [ aij ] be an invertible n × n matrix, and Āij the ( n − 1) × ( n − 1) matrix
obtained by deleting row i and column j from A. Then the determinant of A,
det A ≡ | A |, is given by | A | = ∑nj=1 (− 1)i+ j aij | Āij | (for each i = 1, . . . , n). Let
B = [ bij ] be the n × n matrix with elements bij = (− 1)i+ j | Āij | (called the cofactor
of aij ), then the inverse of A exists, given by A−1 = B/ | A|, if | A | 6 = 0.
10. Let {( xa1, . . . , xan ) | a = 1, . . . , n } be a set of n vectors specified by its compo-
nents in some appropriate basis, and let M be the n × n matrix formed from the
vectors { xa } , a vector on each row. Then the vectors x1 , x2 , . . . , xn are linearly
independent if and only if the determinant of M is non zero.
11. We now begin to study properties that depend on a quadratic quantity, the
bilinear form B ( ψ, φ ). We shall assume V = Cn ; most of the results to be pre-
sented also hold for real vectors upon restriction to R n , with exceptions that will
be noted. The scalar product B ( ψ, φ ) of two vectors ψ and φ in a linear vector
space V is a complex number, denoted by ( φ, ψ ), such that (i) ( ψ, ψ) ≥ 0 with
( ψ, ψ ) = 0 if only if ψ = 0; (ii) ( φ, ψ ) = ( ψ, φ )∗, where ∗ denotes complex conju-
gation; and (iii) ( φ, αψ + βχ ) = α ( φ, ψ) + β hφ, χ), where α, β ∈ C. A product with
property (i) is referred to as a (positive-definite) inner product. Any vector space en-
dowed with an scalar product is called an inner-product space. More technically,
it is a pre-Hilbert or Euclidean space; if also complete, it is called a Hilbert space; see
[Fa].
12. Dirac notation. It is convenient to adopt the Dirac notation in which a scalar
product B ( φ, ψ ) is written as h φ | | ψ i, or h φ |ψi, and call | ψ i a ket vector and h φ |
36 CHAPTER 2. GROUP REPRESENTATIONS
∗
a bra vector. In this notation the conjugation relation becomes h φ | ψi = h ψ | φi ,
which implies that under complex conjugation a ket becomes a bra, and vice
versa, | φ i∗ = h φ |, (h ψ|)∗ = | ψ i, and the relative order of the two vectors in the
product is reversed. The vectors | ψ i and h ψ | are said to be dual to each other, and
belong respectively to space V and its dual, denoted by V ∗ .
13. One calls h ψ | ψi = k ψ k2 the squared length, or squared norm, of ψ (when no
confusion can arise, we write ψ for | ψ i). When k ψ k = 1, the vector ψ is said to
be normalized (to one), or of norm one. Any nonzero vector can be normalized by
dividing it by its own length. Two vectors φ and ψ whose scalar product vanishes
are orthogonal to each other, which we indicate by φ ⊥ ψ.
14. A basis is said to be orthonormal if the basis vectors are mutually orthogonal
and all normalized to 1.
15. Gram–Schmidt Orthogonalization. From any two non parallel vectors we can
produce two orthogonal vectors by subtracting from one its component parallel
to the other. More generally, from n linearly independent vectors φi , the Gram–
Schmidt orthogonalization process produces n mutually orthogonal vectors ψi, each
of which, when divided by its own length, is duly normalized.
16. Conjugations. To any linear operator A corresponds another linear operator,
def
called its adjoint and denoted by A†, such that ψ | A†φ = h Aψ |φi, or equiva-
def
lently ψ | A†φ = hφ | Aψi∗ . The operator A is said to be self-adjoint (or Hermitian)
π ( g) ε x = ε gx = ∑y ε y δy, gx . (2.10)
40 CHAPTER 2. GROUP REPRESENTATIONS
The coefficients of ε y in the y-sum are just the elements of a matrix in the repre-
sentation associated with the homomorphism of G to the permutation group of
X, called the permutational representation. It is related to a finite group itself,
being a subgroup of some symmetric group (e.g. D4 < S4). Matrices in this rep-
resentation can be obtained from the permutation operators p g described in the
previous chapter. For Sn , X may be a set of n letters or symbols.
(4) Let X be the group G itself, so that G × G → G, meaning that the left action
of an element (operator) g ∈ G on another element (considered now a basis vector
in the group space) yields a vector in the same space. Then (2.6) may be re-written
in the form
g gj = ∑ gk Dkj ( g) , g, gi , gk ∈ G , (2.11)
k
where the matrix D( g) is defined as
(
1 if ggj = gk ,
Dkj ( g) = k, j = 1, 2, . . . , | G |. (2.12)
0 if ggj 6 = gk ;
For g = e, Dkj ( e ) = δkj ; and for g 6 = e, all entries of D( g) are 0, except a single
off-diagonal equal to 1 in each row and each column. For any complex-valued
function φ in the group space, we have gi φ ( gj ) = φ ( g− 1
i gj ) . The associated per-
mutation representation defined by the action ( g, x) 7 → gx of G on itself is called
the regular representation of the group G. Its degree is equal to the order | G | of
the group; it is quite unique in that it contains all the basic units (inequivalent ir-
reducible representations) of the group representations, as we shall see. A similar
construction plays a key role in the study of Lie groups.
E XAMPLE 1: To begin, we take the familiar example of rotations in a plane E2
about the origin, forming a continuous group G = { R θ | 0 ≤ θ < 2π } . Let e1 , e2
be the orthogonal unit vectors in E2. Under R θ the basis vectors transform as
π ( Rθ ) preserves orthonormality (in particular, ei e j = e0i e0j = δij ), and sends any
vector V = ei Vi in E2 to V 0 = π ( R θ ) V, with components in the basis e1 , e2 given
by Vj0 = Dji ( θ ) Vi, or explicitly,
The components of the transformed vector in the transformed basis are the same
as the components of the old vector in the old basis: e0i · V 0 = ei · V, which is the
2.2. DEFINITION AND CONSTRUCTIONS 41
E A A2
α1 : 1 1 1
α2 : 1 ω ω2
α3 : 1 ω2 ω
e r s rs
α1 : 1 1 1 1
α2 : 1 1 −1 −1
α3 : 1 −1 1 −1
α4 : 1 −1 −1 1
Figure 2.1: Planar geometric shapes having the symmetries of the dihedral groups
D2, D3 , D4 . Dotted lines indicate the reflection lines.
representation of D4 is specified by
0 −1 0 1
π (r) = , π (s) = .
1 0 1 0
C3 1 2 3
e 1 2 3
a 2 3 1
a2 3 1 2
From this table, we obtain its regular representation (in the basis { e, a, a2} )
1 0 0 0 0 1 0 1 0
2
D ( e ) = 0 1 0 , D ( a ) = 1 0 0 , D ( a ) = 0 0 1 .
0 0 1 0 1 0 1 0 0
Note that (123) and (132) are equal to the matrices D( a ) and D( a2) of C3
in its regular representation. This is because, as discussed in Chapter 1, C3 is
isomorphic to a subgroup of S3 under the mapping a 7 → (123) and a2 7 → (132).
E XAMPLE 8: D2 in function space. Suppose now that π ( D2) acts on a function
space L, which we first assume to be spanned by linear functions of the form
f ( x, y) = ax + by, where a and b are independent real constants. We choose the
basis to be e1 = x and e2 = y. The elements of the group transform e1, e2 in
exactly the same way as in the defining representation and, therefore, the matrix
representation of D2 in this basis is given by the same 2 × 2 matrices.
44 CHAPTER 2. GROUP REPRESENTATIONS
have the same eigenvalues 1, 1, but are not similar. However, if it is true that
A ( g) ∼ B ( g) for every g ∈ G, then there exists an invertible matrix S ( g) such that
B ( g) = S ( g) A( g)S( g)−1 for every g ∈ G, and we can find a matrix S independent
of g such that B ( g) = SA ( g)S−1 for every g ∈ G, which means the representations
2.3. SIMPLE REPRESENTATIONS 45
specify the defining representation of the group D3 over R 2 , and the matrices
√ √
1 −1 − 3 1
R0 = √ , S0 = √1 3
2 3 −1 2 3 −1
define another representation of D3 .√But the latter is equivalent to the first, being
related to it through P = 21 √1 3
by R 0 = PRP−1 and S0 = PSP−1.
− 3 1
Some representations of a group may be considered more ‘basic’ than others.
Some are ‘simple’, others ‘semi-simple’.
D3 irreps dim r s
α1 : 1 1 1
α2 : 1 1 −1
ω 0 0 1
α3 : 2
0 ω −1 1 0
We will use ω = ei2π/n to denote the first primitive nth -root of unity.
• n odd: λ = { 1, ω ±1, ω ±2, . . . , ω ±m } with m = ( n − 1) /2, where we have
used the identity ω m+k = ω −(m−k+1) with k = 1, 2, . . . , m. The eigenvalue λ = 1
gives a representation that is not simple, being decomposable into a direct sum
of two 1D irreps, while of the n − 1 remaining eigenvalues, half (ω − j) give irre-
ducible representations that are equivalent (upon ψ ↔ φ in the basis) to the other
half (ω j). And so, for Dn with odd n, there exist ( n − 1) /2 non-equivalent 2D
irreducible representations.
• n even: λ = {± 1, ω ±1, ω ±2, . . . , ω ±m } , with m = ( n − 2) /2, where we have
used the relation ω m+k+1 = ω −(m−k+1) , with k = 1, 2, . . . , m. The two representa-
tions corresponding to λ = ± 1 are not simple, they are decomposable into direct
sums of 1D irreps. Half of the representations corresponding to the remaining
n − 2 roots are equivalent to the other half. And so for the Dn with even n, there
exist ( n − 2) /2 non-equivalent 2D irreducible representations.
Now, suppose we have a representation ( π, V ) of a group G given in matrix
form D ( g) of degree n = dim V . If V contains no nonzero proper subspaces
stable under G, then π is a simple representation and is indecomposable, not
being expressible as a sum of other representations. If, on the contrary, V contains
a subspace V 1 of dimension n1 different from both 0 and n, then we define its
orthogonal complement W such that V = V 1 ⊕ W (see Sec. 2.1 #21), and in an
appropriate basis for V consisting of ei , i = 1, . . . , n, the vectors ei , i = 1, . . . , n1
span V 1. In order for V 1 to be a G-invariant subspace (that is, π ( g) v ∈ V 1 for
every v ∈ V 1 and every g ∈ G), all the vectors given by
n
π ( g) ei = ∑ e j Dji ( g), ( i = 1, . . . , n1 ) (2.15)
j= 1
⊕a
where each term ai V i in this sum stands for a direct sum V i i = ⊕ j V j of ai repre-
sentatives in the equivalence class. The dimension of the semisimple representa-
tion is the dimension of the underlying space, dim π = ∑κi ai dim V i.
In this section, we have explained the meaning of simplicity and semisim-
plicity of representations. These properties are at the center of the representation
theory and will be the subject of further study in the next sections. We will see
then that the number of simple representations of a finite group G is finite, that
the degree of each has an upper limit (| G |1/2), and that representations of a finite
group G are semisimple, with a unique decomposition as in (2.19), in which the
simple representations that occur are unique, just as are their multiplicities.
E XAMPLE 14: Group S2 has two elements: e = (1)(2) and a = (12). Two ob-
jects can be in either configuration |12i or |21i, which are taken to be a basis in
a two-dimensional space. The operator (12) interchanges |12i, |21i. In this ba-
sis, the state vectors are the standard column vectors (1,0)T and (0,1)T and the
representation of S2 is given by
1 0 0 1
D( e ) = , D( a ) = .
0 1 1 0
Define a new basis, formed by the symmetric combination |+i = |12i + |21i
and the antisymmetric combination |−i = |12i − |21i, which are column vectors
(1, 1)T , (1, − 1)T when written in the standard basis. The group leaves |+i un-
changed and changes |−i by a sign. They form two invariant 1D subspaces, and
together decompose D( S2) into distinct irreps by block diagonalization:
1 0 1 0
D( e ) = , D( a ) = .
0 1 0 −1
E XAMPLE 15: In the preceding example, the vectors |12i and |21i are taken to
form a basis, they can be obtained by applying the group elements e = (1)(2)
and a = (12) on the vector |12i (or, for that matter, |21i ). So there is a unique
correspondence between the basis vectors and the group elements. In the regu-
lar representation, the group elements are used as a basis. Similarly, the vectors
|+i = |12i + |21i and |−i = |12i − |21i can be seen as ( e + a )|12i and ( e − a )|12i,
and therefore, we can regard e + a and e − a as forming another basis; in fact they
constitute, by themselves, two invariant 1D bases.
E XAMPLE 16: The regular representation πR of the group C3 = h a; a3 = e i is
defined by
1 0 0 0 0 1 0 1 0
D ( e ) = 0 1 0 , D ( a ) = 1 0 0 , D ( a 2 ) = 0 0 1 ,
0 0 1 0 1 0 1 0 0
in the basis { e, a, a2} , or in terms of standard column vectors, (1,0,0), (0,1,0) and
(0,0,1). They leave the sums 1 + a + a2, 1 + ωa + ω 2 a2, 1 + ω 2 a + ωa2 invariant
(up to overall constants). We have defined ω a cubic root of 1, which satisfies
50 CHAPTER 2. GROUP REPRESENTATIONS
relates the two bases and simultaneously diagonalizes the three matrices, [ e ], [ a ],
[ a2], resulting in their fully reduced forms D0 = SDS−1:
1 0 0 1 0 0 1 0 0
D0 ( e ) = 0 1 0 , D0 ( a ) = 0 ω 0 , D0 ( a2) = 0 ω 2 0 .
0 0 1 0 0 ω2 0 0 ω
This shows the complete decomposition of πR into the irreps of C3. Now, we write
πR = α1 + α2 + α3 (where αi are the three simple one-dimensional representations
of C3). This illustrates the general property that in any finite group, the regular
representation contains all the irreducible representations of the group.
E XAMPLE 17: We have seen that D3 has 3 irreducible representations in all:
two 1D representations: α1 : r, s 7 → 1, 1; and α2 : r, s 7 → 1, − 1;
one 2D representation: α3.
Any other 2D representations it may have must be semisimple, and may be con-
structed by summing α1 and α2. There are four possibilities: α1 + α1, α2 + α2,
α1 + α2, and α2 + α1. The last two being equivalent, there are thus, in all, 4 two-
dimensional reps, of which only α3 is simple.
{ , } = B satisfies the scalar product axioms (Sec. 2.1 #11) and is G-invariant, that
is B( gx, gy ) = B( x, y ), as we can see here:
{ π ( g) x, π ( g)y} = | G |−1 ∑ h π ( h ) π ( g ) x | π ( h )π ( g ) y i,
h ∈G
(2.21)
= | G | −1 ∑ h π (h0) x| π ( h0) y i = { x, y } ,
0
h ∈G
Thus, redefining the scalar product so that the representation is unitary is equiv-
alent to changing the basis in the underlying vector space by the transformation
S defined by
The first step comes from the definition of ρ, the second and fourth equalities fol-
low from (2.23), and the third from permuting the group elements. Since invari-
ance h ρ (g) x|ρ( g)yi = h x| y i holds for arbitrary x, y ∈ V and g ∈ G, we conclude
that ρ ( G ) ∼ π ( G ) is a unitary representation wrt the standard form h|i.
In sum, h, i is G-invariant in ρ, whereas { , } is G-invariant in π. Whether one
uses π together with { , } , or ρ together with h, i is a matter of choice. These results
have immediate consequences for arbitrary representations of finite groups:
(i) Every representation of a finite group over F = C (or R) can always be chosen to
be unitary (orthogonal, if F = R). Because if given one that is not, we can
find by Theorem 2.2 a unitary representation equivalent to it by a similarity
transformation.
2.5. SCHUR’S LEMMA 53
P ROOF : (a) follows from the fact that the kernel Kerϕ and the image Imϕ
of ϕ are subspaces stable under G. This can be seen as follows. First, Kerϕ =
{ x ∈ V | ϕ ( x) = 0} is an invariant subspace of V with respect to π ( G ) because
given any x ∈ Kerϕ, we have ϕ π ( g) x = ρ ( g) ϕ ( x) = ρ ( g) 0 = 0 for all g ∈ G,
which means π ( g) x ∈ Kerϕ if x ∈ Kerϕ. As π ( G ) is irreducible, Kerϕ must
be either empty (in which case ϕ ( x) 6 = 0 for any x ∈ V ), or the whole of V (in
which case ϕ = 0). Secondly, Imϕ = { y ∈ W | y = ϕ ( x), for some x ∈ V }
is a subspace of W stable under ρ ( G ) because given any y ∈ Imϕ, the vector
ρ ( g) y = ρ ( g) ϕ (x) = ϕ π ( g) x for all g ∈ G, is in Imϕ. As ρ ( G ) is an irreducible
representation, we have either Imϕ = 0 (in which case ϕ = 0), or Imϕ = W (in
which case ϕ is a homomorphism).
Now, from the fact that Kerϕ and Imϕ are G stable subspaces, we see that
either ϕ is the trivial zero map (Imϕ = 0, Kerϕ = V ), or an invertible map (Imϕ =
W , Kerϕ = 0), i.e. a C [ G ] isomorphism, which defines the equivalence of two
simple representations via conjugation, ρ ( G) = ϕ π ( G ) ϕ−1.
(b) As ϕ is a linear mapping on complex space V , it has an eigenvalue λ ∈ C
(cf. Sec. 2.1 #24). Let ψ = ϕ − λIV , then its kernel Kerψ = { x ∈ V | ψx = 0}
= { x ∈ V | ϕ ( x) − λx = 0} must be non empty. So Kerψ is a nonzero sub repre-
sentation of G on V . But by assumption V is irreducible, and Kerψ = V . This
means Ker{ ϕ − λIV } = V , or ϕv − λv = 0 for all v ∈ V , or equivalently ϕ = λIV .
Part (a) of Theorem 2.3 has no converse, but Part (b) has one, which reads
Theorem 2.5 (NSC of Schur-C). Let G be a finite group and V a vector space over C.
Then, a representation ( π, V ) of G is irreducible if and only if every linear transformation
A : V → V that satisfies the commutation relation Aπ ( g) = π ( g) A for all g ∈ G is
scalar.
E XAMPLE 20: Let G be a finite abelian group and ( π, V ) representation of G over
C. Given any fixed element a ∈ G, we have π ( a)π ( g) = π ( g)π (a ) for all g ∈ G
because G is abelian. Assuming now ( π, V ) irreducible, and letting a range over
the entire group, we see that π is scalar and V one-dimensional. So,
• Every irrep of a finite abelian group over C is one-dimensional.
This means in particular that all irreps of Cn and V ∼ = D2 must be one-dimensional.
On the other hand, in the examples we have examined, we have found that
V ∼ = D2 has four 1D representations, and Cn has n 1D representations, and by
Theorem 2.5 they must be irreducible.
C OMMENTS .
(a) Schur’s lemma can be extended to infinite (Lie) groups. It is in particular
useful with Lie algebras in which form it finds many applications in quantum
mechanics.
(b) Theorem 2.3 (a) holds for representation spaces over R too, and so it holds
in general for F = R or C: Let ( π, V ) and ( ρ, W ) be irreducible representations of
a finite group G over a base field F; if ϕ : V → W is an F [ G ] homomorphism, then
either ϕ = 0, or ϕ is an F [ G ] isomorphism. But as it is, Theorem 2.3 (b) does not
hold over R because a linear operator on a real vector space may not have any
real eigenvalues at all (see Section 2.1 #24), and one must insert an extra condi-
tion: If ϕ : V → V is an R [ G ] homomorphism with a real eigenvalue, then ϕ is scalar.
This condition is certainly satisfied if the linear operator of interest is self-adjoint
on an inner-product space over either F = R or C (cf. Section 2.1 #27), and so
Theorem 2.5 can be modified to read
Theorem 2.6 (NSC of Schur-F). Let G be a finite group and V an inner-product space
over F = R or C. Then, a representation ( π, V ) of G is irreducible if and only if every
self-adjoint linear transformation A : V → V that satisfies the relation Aπ ( g) = π ( g) A
for all g ∈ G is scalar.
It is in this form that Schur’s Lemma is most often used in practice.
Dπ ( g) in some chosen basis of order equal to dim V , over base field C (includ-
ing restriction to R when necessary). Although Dπ ( G ) is needed for a complete
description of G on V , or any physical system it could represent, it is an unnec-
essarily unwieldy tool just for identifying simple representations or determining
an arbitrary representation. For this limited purpose, a scalar function encoding
all the information about a representation would suffice, and, in fact, it exists:
Definition 2.5. The character of a finite representation ( π, V ) of a group G over F is
the function χ : G → F defined by the trace of π ( g) on V : χπ ( g) = Tr ( π ( g)) for any
g ∈ G. A simple character is the character of a simple, or irreducible, representation.
Many properties of the character follow from those of the trace. For example,
the cyclicity property Tr ( ABC ) = Tr ( BCA ) implies that χ ( hgh−1) = χ ( g), that is,
χ is constant in each conjugacy class of G; such a function is called a class function.
Another consequence of cyclicity is χ ( g) = Tr (D( g)) = ∑i Dii ( g): χ does not
depend on the basis in which it is calculated. (When there are no ambiguities, the
representation label π in χπ or Dπ is omitted.)
There are two important special cases: (i) χ ( e) = dim V ; and (ii) χ1 ( g) = 1 for
any g ∈ G in the trivial one-dimensional representation (π = 1). We also have
−1 χ ( g) if F = R
χ( g ) = (2.25)
χ ( g )∗ if F = C
Orthogonal Matrices. Let G be a finite group of order | G |, and consider the irre-
ducible representations ( α, V α) and ( β, V β) over C. Let us define for some linear
map B : V α → V β the following operator A on V α :
A = ∑ g∈G β ( g) B α g−1 .
(2.26)
where we have used the fact that hg, with fixed h, ranges over G just as does g.
Applying Schur’s Lemma, we have either (i) α ∼ β, then α ( h) A = Aα (h) for all
h ∈ G which implies A = λI with a scalar λ depending on B; or (ii) α β, then
A = 0. (We shall use the symbol δα,β to mean 1 if α ∼ β, and 0 if α β.)
In an orthogonal basis of V α , α ( g) 7 → Dαij ( g) with 1 ≤ i, j ≤ d α = dim V α , and
the result ∑g β ( g) · B · α ( g−1) = λ I δαβ may be written in matrix form:
β
∑ Dmj ( g) Diα` ( g−1 ) = λij δ`m δαβ , (2.28)
g ∈G
when B is chosen as a matrix with a single nonzero entry: Bji = 1 for some i, j.
The constant λ can be calculated by letting α = β, ` = m, and summing over `,
resulting in λij = δij | G | /dα. We may now write the equation in the form:
dα β
Diα` ( g−1 ) Dmj ( g) = δ`m δij δαβ ,
| G | g∑
1 ≤ i, j, `, m ≤ dα . (2.29)
∈G
dα β
D`αi∗ ( g) Dmj( g) = δ`m δij δαβ 1 ≤ i, j, `, m ≤ dα .
| G | g∑
(2.30)
∈G
p
We may interpret dα / | G | Dijα in this equation as vectors labeled by α, i, j in a
| G |-dimensional vector space; they are mutually orthogonal, and so also linearly
independent. Since the number of linear independent vectors in a linear space
cannot exceed the dimension of that space (Section 2.1 #3), we have
κ
∑α=1 d2α ≤ | G | , (2.31)
Orthogonal Characters. A similar relation for the simple characters can be ob-
tained from (2.30) by letting ` = i and m = j and summing over i, j = 1, . . . , d α
with the result:
1
χα∗ ( g) χ β( g) = δαβ .
| G | ∑ g ∈G
(2.32)
1 nc β
cµ χαµ∗ χµ = δαβ ,
| G | µ∑
(2.33)
=1
58 CHAPTER 2. GROUP REPRESENTATIONS
κ ≤ nc . (2.34)
This interpretation suggests that we may define the inner product of any two
class functions ξ and η in the class function space of a finite group G as follows:
def 1 1 nc
∑ ξ ∗ ( g) η ( g ) = cµ ξ µ∗ ηµ .
| G | µ∑
h ξ, η i = (2.35)
| G | g ∈G =1
Then the statement that the simple characters of a finite group are orthonormal wrt the
inner product (2.35) has a concise expression:
where we have used the orthonormality of the simple characters. The multiplicity
can now be evaluated:
This reduces to (2.37) for g = e, and ∑α χα∗ ( e ) χα( g) = 0 for g 6 = e. With this
equation, one may calculate any one character in terms of the others.
√
Completeness. We have seen before that dα / | G | Dijα may be interpreted as | G |-
component mutually orthogonal vectors. Equation (2.37) states that there are as
many such vectors as there are dimensions and they fill out the entire | G |-dimensional
space. This completeness property of the irreducible representation matrices finds
its expression in the statement
κ dα
dα α
∑ ∑ D ( g) Dijα∗ ( g0 ) = δgg0 . (2.39)
α=1 i,j=1
| G | ij
A similar relation exists for the simple characters, as we now show. Let α be
any representation of G, and [ µ ] some conjugacy class of G; then consider the sum
Aαµ = ∑ α ( h) . (2.40)
h∈[µ]
dα cµ α c ν α∗ cµ cν
∑ |G| dα
χµ δij
dα
χν δij =
|G| ∑ χαµ χαν∗ .
αij α
It follows that
c µ κ α∗ α
| G | α∑
χν χµ = δµν . (2.42)
=1
In the special case [ µ ] = [ e ], this equation reduces to (2.37) and√(2.38). More sig-
nificantly, it shows that the nc κ-dimensional non-zero vectors cµ / | G | χαµ (with
vector label µ and component label α) are orthogonal to one another, so nc ≤ κ.
This inequality, together with (2.34), gives us κ = nc : the number of simple rep-
resentations of G equals the number of its conjugacy classes. To summarize:
When a list of distinct irreducible representations α of a finite group G is complete,
κ 2
√ counting relations (κ = nc and ∑α dα = | G |) are satisfied. The simple characters
the
cµ / | G | χαµ may be regarded either as α-labeled vectors of components µ, or µ-labeled
vectors of components α. Either way they are mutually orthogonal, as shown in (2.33)
and (2.42).
The Character Table. It is customary to display the simple characters χαµ of a finite
group in the form shown in Table 2.1, which may also serve as a tool for finding
unknown entries, even when the representations Dα are not explicitly known.
The columns correspond to the conjugacy classes, labeled [ µ ] and accompanied
by their sizes cµ . The rows correspond to the irreducible representations Dα , so
that the entry at row α, column µ gives the character χαµ . As the number nc of con-
jugacy classes equals the number κ of simple representations (nc = κ), there are
as many columns as rows. Normally, the first row gives the characters (χ1µ = 1 for
all µ) in the trivial one-dimensional representation, and the first column gives the
2.6. MATRICES AND CHARACTERS 61
characters of element e (χα[e] = dim V α = dα for all α). The rows are normalized to
1, and orthogonalized to one another by (2.33). Similarly, the columns satisfy the
orthonormality condition (2.42). In particular, the normalization of the entries in
the first row requires ∑µ cµ = | G |, and similarly, ∑κα d2α = | G | in the first column.
Once completed, the table gives a full account of the characters of the group.
G [ 1 ] 1 . . . [ µ ] c µ . . . [ n c ] c nc
D1 1 ... 1 ... 1
... ... ... ... ... ...
Dα dα ... χαµ ... χαnc
... ... ... ... ... ...
Dκ dκ ... χκµ ... χκnc
C3 : e a a2
C2 : e a
α1 : 1 1 1
S : 1 1
α2 : 1 ω ω2
A : 1 −1
α3 : 1 ω2 ω
with row α1, and similarly the second column is determined by orthonormality
with the first column. Calculate the last two entries of rows α3 and α4 again by
orthogonality with the first two rows.
D2 : e r s rs D2 : e r s rs
α1 : 1 1 1 1 α1 : 1 1 1 1
α2 : 1 1 . . −→ α2 : 1 1 −1 −1
α3 : 1 . . . α3 : 1 −1 1 −1
α4 : 1 . . . α4 : 1 −1 −1 1
E XAMPLE 23: Dihedral groups. First, consider D3 ∼ = S3. It has 6 elements divided
into 3 classes [ e ], [ s ], and [ r ], carrying the (subscript) label [ µ ] (µ = 1, 2, 3):
[1] = { e } ∼ = {(1)} ,
[2] = { s, sr, sr2} ∼ = {(12), (23), (31)}, and
[3] = { r, r2} ∼ = {(123), (321),
of sizes cµ = 1, 3, 2. As κ = nc = 3, there are three simple representations, with
dimensions satisfying the equation d 21 + d22 + d23 = 6, whose unique solution is:
d1 = d2 = 1, d3 = 2. As usual, we know the entries on the first row (1, 1, 1) and
first column (1, 1, 2), and we have to calculate χ22, χ23 , χ32, χ33.
Now, as ( χ22)2 = ( χ2( s ))2 = 1 (because s2 = e), χ22 is real and equal to ± 1.
Orthogonality between rows α1 and α2 requires that 3χ22 + 2χ23 = − 1, which
yields the unique solution χ22 = − 1, χ23 = 1. Entries in the last row are deter-
mined by the orthogonality relations between column [1] and columns [2], [3]:
∑α dα χα2 = 1 + χ22 + 2χ32 = 0 and ∑α dα χα3 = 1 + χ23 + 2χ33 = 0. This gives χ32 = 0
and χ33 = − 1, which completes the character table for D3 . The second row of
the table for D3 can also be found in another way, which proves to be useful
for larger groups. We know from Chapter 1 that D3 has an invariant subgroup,
H = { e, r, r2} with coset M = sH = { s, sr, sr2} . The associated factor group
D3 /H = { H, M } is isomorphic with C2, and so the characters for representation
α2 can be obtained from the characters of C2 by the homomorphism D3 7 → D3 /H,
that is e, r, r2 7 → e ∈ C2, and s, sr, sr2 7 → c ∈ C2. It follows that χ21 = 1, χ23 = 1 and
χ22 = − 1, in agreement with results already found. Entries in the last row, found
from the orthonormality conditions, will complete the table.
We can construct in the same way the character table for D4 (with its conju-
gacy classes known). The results for both D3 and D4 are shown below:
D4 : [ e ]1 [ s ]2 [ rs ]2 [ r ]2 [ r2]1
D3 : [ e ]1 [ s ]3 [ r ]2 α1 : 1 1 1 1 1
α1 : 1 1 1 α2 : 1 −1 −1 1 1
α2 : 1 −1 1 α3 : 1 1 −1 −1 1
α3 : 2 0 −1 α4 : 1 −1 1 −1 1
α5 : 2 0 0 0 −2
2.7. TENSOR-PRODUCT REPRESENTATION 63
From the pattern suggested by these examples, we can write down the char-
acter table for any dihedral group. In fact, we know all about the Dn groups: the
compositions and sizes of their conjugacy classes from Chapter 1 Sec. 1.5; and
their irreducible representations from Example 13, which we now know to be
complete, by ‘counting’ them (κ = nc and ∑κα d2α = | G |). So, we just recall:
• If n is odd, Dn has exactly ( n + 3) /2 irreducible representations, of which 2
are one-dimensional, and ( n − 1) /2 two-dimensional.
• If n is even, Dn has exactly n/2 + 3 irreducible representations, of which 4
are one-dimensional, and ( n − 2) /2 two-dimensional.
E XAMPLE 24: (S3 and permutation rep) From the simple characters, it is easy to
find the decomposition of any representation with known characters. To illus-
trate, take the permutational representation π P of S3 (Example 7 p.43), where
the
traces of the matrices given yield χP ( e ) = 3, χP (12) = 1, and χP (123) = 0,
u1 + cu2 ⊗ v = ( u1 ⊗ v ) + c ( u2 ⊗ v ) ,
(2.43)
u ⊗ v1 + cv2 = ( u ⊗ v1 ) + c ( u ⊗ v2 ) .
(2.44)
( A1 ⊗ B1 )( A2 ⊗ B2 ) = ( A1 A2 ) ⊗ ( B1 B2 ) . (2.46)
α× β β
Di`;jm ( h, k ) = ( Aα ( h ) ⊗ B β ( k ))i`,jm = Aαij ( h ) B`m( k ) . (2.50)
β
χα× β ( h, k ) = Tr ( Dα× β ( h, k )) = Aαii ( h ) B`` ( k ) ,
which is
χα× β ( h, k ) = χα ( h ) · χ β (( k ) h ∈ H, k ∈ K . (2.51)
That is, the characters in the representation α ⊗ β of the direct-product group
H × K are given by the products of the characters in the representations of the
factors. Their inner products can be calculated from (2.35) yielding the result:
where h, i is the bilinear form in (2.35). It is the basis for the following property:
• The representation α ⊗ β of the direct product group H × K over F is irre-
ducible if and only if the representations α and β of H and K over F are both
irreducible. Every irreducible representation of H × K is of this kind.
Recalling that a representation α over F is irreducible if and only if h χα , χα i = 1,
we see that if α and β are both irreducible, then h χα× β, χα× β i = 1, which means
that α ⊗ β is irreducible. Moreover if H has nc conjugacy classes and K has mc
classes, then H × K has nc mc classes, since ( h, k ) ∼ ( h0, k 0 ) implies h ∼ h0 , k ∼ k 0 ,
and vice versa for h, h0 ∈ H and k, k 0 ∈ K. We know that there are as many simple
representations as there are classes, which is nc mc for H × K; and there are pre-
cisely nc mc simple representations α ⊗ β. All possible irreducible representations
of H × K must be of the form α ⊗ β.
Representations of G in U ⊗ V . In physics we often deal with systems of identi-
cal particles existing in representations of some given group G in tensor-product
spaces (for example, space rotations in a two-nucleon system). The problem we
have here is, in mathematical terms, identical to that of the representations of
the product group G × G restricted to its diagonal subgroup {( g, g) | g ∈ G } , and
so, may be simply treated as a special case of that situation, although the carrier
spaces could be physically distinctive with different kinds of degrees of freedom.
66 CHAPTER 2. GROUP REPRESENTATIONS
| ui i ≡ | αii, i = 1, 2, . . . , dα , (2.54)
`
| v i ≡ | β `i, ` = 1, 2, . . . , d β , (2.55)
a
| w i ≡ | αii| β `i = | αi, β `i . (2.56)
They form the orthonormal bases for the spaces U , V , and W = U ⊗ V over C,
and satisfy the orthogonality and completeness relations:
α× β
We shall call the new basis vectors | τγk i with k = 1, . . . , d γ ; τ = 1, . . . , aγ ;
and γ = 1, . . . , κ, subject to the orthonormality and completeness relations:
h τγk|τ 0γ0 k 0 i = δττ0 δγγ0 δkk0 , (2.60)
∑τγk | τγkihτγk| = ID . (2.61)
α× β
Completeness implies that D = dα d β = ∑κγ=1 aγ dγ .
The two orthonormal bases {| αj, βmi} (‘uncoupled’) and {| τγk i} (‘coupled’)
in W are related by a similarity transformation, such that they satisfy the recip-
rocal relations
| αj, βm i = ∑ | τγkihτγk|αj, βmi , (2.62)
τγk
Alternatively, starting from (2.66) we use the completeness of {| αj, βmi} to obtain
γ
Dk0 k ( g) δγγ0 δττ0 = h τγk |(α ⊗ β )( g)|τ 0γ0 k 0 i
= ∑h τγk|αi, β`i hαi, β`|(α ⊗ β)( g)|αj, βmi hαj, βm|τ0γ0 k0 i
β
= ∑ h τγk |αi, β`i · Dijα ( g) D`m ( g) · h αj, βm|τ 0γ0 k 0 i . (2.68)
ij`m
68 CHAPTER 2. GROUP REPRESENTATIONS
If we set j = i and m = `, sum over i and j on both sides of (2.67), and make
use of the orthonormality of the Clebsch-Gordan coefficients, we obtain a relation
between the characters χα× β and χγ ( g):
α× β
κ a κ α× β γ
χα ( g) χ β( g) = ∑γ=1 ∑τγ=1 χγ ( g) = ∑γ=1 aγ χ ( g) . (2.69)
We would have obtained the same result had we started from (2.68). It agrees
α× β
with (2.53), where we set χα× β = ∑γ aγ χγ , as implied by the decomposition
α× β α× β
Dα× β = ⊕ γ aγ Dγ . The multiplicity coefficient aγ , which is the number of
times the irrep Dγ appears in the decomposition of Dα× β , is given by
α× β
aγ = | G | −1 ∑ χα ( g) χ β( g) χγ ( g) = h χγ, χα×β i, (2.70)
g ∈G
with χ = χ∗ if F = C, and χ = χ if F = R.
To summarize: Assuming the irreducible representations α, β of a group G are
known, the representation α ⊗ β of G is given by the tensor product of the reps
of α and β, i.e. Dα× β ( g)=Dα ( g) ⊗ Dβ ( g) with g ∈ G. It is completely reducible
and can be decomposed into a sum of the irreducible representations of G by a
similarity transformation relating the ‘uncoupled’ basis and the ‘coupled’ basis,
determined by an orthogonal matrix the elements of which are called the Clebsch-
Gordan coupling coefficients specific to G. The representations expressed in the
two bases are transformed into each other by the reciprocal relations:
β γ
Dijα ( g) D`m ( g) = ∑ 0 h αi, β`|τγki · Dkk0 ( g) · h τγk0| αj, βmi, (2.71)
τγkk
γ β
Dk0 k ( g) δγγ0 δττ0 = ∑ h τγk|αi, β`i · Dijα ( g) D`m( g) · h αj, βm|τ 0γ0 k0 i, (2.72)
ij`m
α× β
for any element g ∈ G. The decomposition Dα× β = ⊕ γ aγ Dγ is completely
determined by the multiplicities of the inequivalent irreducible representations γ
α× β
through the formula aγ = h χγ, χα× β i.
In fact β ( s ) is equivalent to the diagonal form β : r 7 → diag [1, 1], s 7 → diag [1, − 1].
So β = α1 + α2 is the sum of the two simple one-dimensional representations of
D3, which means we have the complete decomposition α3 ⊗ α3 = α1 ⊕ α2 ⊕ α3.
Another way to obtain this result is to calculate the multiplicities aγ in the
decomposition α3 ⊗ α3 = ∑γ aγ γ, where the sum is over all the simple represen-
tations of D3 . We will use the formula for aγ in the form
α γ
aγ = | G |−1 ∑µ cµ ( χµ3 )2 χµ
together with the necessary data found in the character table for D3. We get 6a1 =
1.22 .1 + 3.02 .1 + 2(− 1)21 = 6; 6a2 = 1.22.1 + 3.02 (− 1) + 2(− 1)21 = 6; and
6a3 = 1.22 .2 + 3.0.0 + 2.(− 1)2(− 1) = 6. Thus, we have aγ = 1 for all three simple
representations, so that α3 ⊗ α3 = α1 ⊕ α2 ⊕ α3, just as we found before.
An even simpler way is to recall that every representation is determined up
to isomorphism by its character. Thus, referring to the character table for D3 ,
we have χα3 ⊗α3 = ( χ3)2 = (4, 0, 1) on the three conjugacy classes; but these
values are the same as given by χ1 + χ2 + χ3. Then, it follows that α3 ⊗ α3 =
α1 ⊕ α2 ⊕ α3. Similarly, α1 ⊗ α3 has the character (2, 0, − 1), and α1 ⊗ α3 ∼ = α3; and
finally, α2 ⊗ α3 = α3 as they have the same character (2, 0, − 1).
E XAMPLE 26: Products are not confined to just two factors: we may have for ex-
ample α× n
3 = α 3 ⊗ . . . ⊗ α 3 , where α 3 is the simple 2-dimensional representation
of S3. What is its decomposition ⊕ γ aγ γ in simple representations? The answer
3
can be found in the character theory. We know that
×n n n
χ = (2, 0, − 1) on the conju-
gacy classes. The character of α3 is 2 , 0, (− 1) , equal to the character of the
decomposition, a1 χ1 + a2 χ2 + 3
a3 χ = ( an1 + a2 +n2a3, a1 − a2, a1 + a2 − a3). We
find: a1 = a2 = 2 n − 1 n
+ (−) /3, a3 = 2 − (−) /3.
Problems
2.1 (Representations in function space) In a two-dimensional Euclidean space,
define a transformation R on the Cartesian coordinates ( x, y ) by x0 = D11 ( R ) x +
D12 ( R )y and y0 = D21 ( R ) x + D22 ( R )y. The set of all homogeneous scalar func-
tions of degree 2 in x, y invariant under R form a linear vector space√V 3 of di-
mension 3 spanned by the basis vectors ψ1 ( x, y ) = x2 , ψ2 ( x, y ) = 2xy and
70 CHAPTER 2. GROUP REPRESENTATIONS
(where we have used group elements to label the representation matrices). But it
can also be defined by the right action
1 : gq = h
gq = ∑h D gh ( q )h , D gh ( q ) =
0 : gq 6 = h
def
{ x, y } = | G |−1 ∑ h D ( h ) x | D ( h ) y i, x, y ∈ V ,
h ∈G
axes, found in C H4 for example). Using the group structure and the orthogonal-
ity and completeness relations, construct the character table for T12. Note that T
is a subgroup of S4, isomorphic to the alternating group A 4 (see Problem 1.10).
2.12 (Character table for S4) Find the characters for the symmetric group S4.
2.13 (Characters for S4) Find the permutational representation on 4 letters for S4;
find its decomposition. What is the product α2 × α4 (where α2 is a nontrivial one-
dimensional representation and α4 a three-dimensional simple representation of
S4)? On the basis of these results, show a way to obtain the characters for S4.
2.14 (Matrix representation of product group) Suppose that a finite group G is
the product group of its two subgroups, G = H × K. Using the matrices and
not the characters, show that the direct product Dα ( H ) × D β ( K ) of two simple
representations of H and K is a simple representation of G.
Lie Groups
and Lie Algebras
73
74 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
geometry combined. Fortunately, a purely algebraic tool, the Lie algebra (a closely
related finite vector space equipped with a bilinear composition rule), can be used
to study properties of the associated Lie groups and extract much of their struc-
ture while avoiding the inherent topological issues; in fact, in cases of physical
interest it encodes almost all of the geometry/topology of the group itself.
This chapter gives an overview of Lie groups and Lie algebras. First we de-
fine the Lie group and give examples, then discuss its algebraic properties (sub-
groups, cosets, invariant subgroups) and its topological properties (compactness,
connectedness, and simple connectedness). We next show when and how a Lie
group linearizes to produce a Lie algebra and describe the reversed path (the ex-
ponential map), and finally introduce general notions of the representation theory
of Lie groups and algebras in preparation for later, more focused, studies.
E XAMPLE 3: Euclidean vector space. The (smooth, real) manifold R n is also a group
(with vector addition), and is a Lie group of dimension n. Its topology structure is
defined by taking all open balls as open sets, and specifying the metric (distance,
norm) d ( x, y) = | x − y | for any elements x, y of the group.
E XAMPLE 4: The general linear groups GL( n, R ) and GL ( n, C ). We already know
2
that the set of all real n × n matrices M( n, R ) can be thought of as R n equipped
with standard topology; it has no group structure. But the subset consisting of
all invertible matrices, G = { A ∈ M( n, R )| det A 6 = 0} , is a group, with matrix
multiplication as the group operation, and an inverse for every element. It is an
2
open subset of smooth manifold R n (‘open’ because given an invertible matrix A
there is a neighborhood U A of A such that every matrix in U A is also invertible).
So G is an n2-dimensional smooth manifold. Moreover, the matrix multiplication
m( A, B ) = AB with A, B ∈ M( n, R ) is given by a smooth (polynomial) function
of their entries and so is a smooth map. The restriction of M × M to G × G is
clearly smooth. As for matrix inversion inv : G → G, the map A 7 → A−1 is also,
by Cramer’s formula and continuous det map, a smooth function of the entries
of A. So G = GL ( n, R ) is a Lie group, of dimension n2.
The action of G = GL( n, R ) on V = R n is described by the transformation
functions f : G × V → V given by
O ( n ) = O ( n, R ) = { A ∈ M( n, R )| AT A = I } .
U ( n )) = U ( n, C ) = { A ∈ M( n, C )| A† A = I } .
is, g = diag [1, . . . , 1; − 1, . . . , − 1], with 1 in the first n diagonal entries and − 1
in the last m, and the bilinear form Bg ( x, y ) = h x, gy in,m on R n+m is given by
h x, gy in,m = x1 y1 + · · · + xn yn − xn+1 yn+1 − · · · − xn+m yn+m . The set of all ( n +
m ) × ( n + m ) real matrices A that preserve this bilinear form Bg ( x, y ) is called
the generalized orthogonal group. This form-invariance condition is written as
AT gA = g (which implies det A = ± 1 among other things), so that the general-
ized orthogonal group is defined by O ( n; m) = { A ∈ GL( n + m, R )| AT gA = g} .
The group SO ( n; m ) is the subset of matrices in O ( n; m ) with det A = 1.
The case O (3; 1) is of particular interest in physics, where it is known as the
Lorentz group. If we write ( x, y, z; t ) for any (space-time) vector v in R 3+1 and
the metric matrix g3,1 = diag [1, 1, 1; − 1], then the Lorentz group is the group of
transformations on R 3+1 that leave the inner product h v, vi3,1 = x2 + y2 + z2 − t2
invariant, or equivalently the group of all 4 × 4 real matrices A that satisfy the
condition AT g3,1 A = g3,1 .
E XAMPLE 10: The Euclidean group E( n ) is the set of bijections of R n that preserve
Euclidean distance. To each element of E( n ) corresponds a map f : R n → R n such
that f ( α; x) = Ax + a where x ∈ R n and α = ( A, a ), with a real invertible n × n
matrix A for a linear map (‘rotation’), and constant a ∈ R n for a non-linear map
(‘space displacement’). (A transformation of the form f ( x) = Ax + a is called
‘affine’.) An α map following a β map (in which x0 = f ( β, x)) is given by
where 1 ascalar,
and 0 is an n-component row vector. It operates on vectors of
the form x1 where x ∈ R n . So the Euclidean group E( n ) can be viewed as a
subgroup of O ( n + 1), and it is then possible to build charts for E( n ) out of any
atlas for O ( n ) showing that E( n ) is a manifold; and as usual matrix multiplication
is smooth. It is globally isomorphic to O ( n ) × R n (R n being here a Lie group),
and so has dimension n ( n − 1) /2 + n = n ( n + 1) /2. So E( n ) is a Lie group of
dimension n ( n + 1) /2; in particular, E(3) has dimension 6.
E XAMPLE 11: The Poincaré group P ( n; 1) is the group of isometries of (Lorentz)
space L = R n+1 equipped with the Lorentz metric g = diag[1, . . . , 1; − 1] and
the bilinear form Bg ( x, y ) = h x, gy i for any x, y ∈ L. An isometry on R n+1 is
represented by the map f : x 7 → x0 specified by x0 = f ( α; x ) = Λx + a with
Λ ∈ O ( n; 1) (‘space-time’ rotation) and a ∈ R n+1 (‘space-time’ displacement).
The Poincaré group is similar on every point to the Euclidean group , except for
80 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
the metric of the space on which it is applied. So just as for E( n ), each element
( Λ, a ) ∈ P ( n; 1) uniquely corresponds to an ( n + 2) × ( n + 2) matrix of the form
Λ a
( Λ ∈ O ( n; 1), a ∈ R n+1 ),
0 1
The matrices X = E12, Y = E23 and Z = E13, where ( Eij )αβ = δiα δjβ , satisfy
Heisenberg’s commutation relations [ X, Y ] = Z, [ X, Z] = 0, [Y, Z] = 0 (hence
the group’s name). H(3) can also be thought of as the subgroup of symmetries of
space of ‘wave functions’ under the transformation f ( x) 7 → e2πi(bx+c) f ( x + a ) for
any square-integrable f .
C OMMENTS . (a) The classical Lie groups are normally understood to be the groups
GL( n, F ), SL( n, F ), O ( n, F ), SO ( n, F ), SO ( p, q; R ), U ( n ), SU( n ), and Sp( n, F ), with
F = R, or C. (b) Some groups similar in structure to Lie groups do not have all of
the required attributes. Some are totally disconnected groups: the Galois group
of an infinite extension of fields, for example, has an underlying space not locally
isomorphic to R n , and so is not a Lie group. Others have infinite-dimensional
manifolds and for this reason are not Lie groups: gauge groups smoothly map-
ping a manifold to a group, groups of diffeomorphisms with Virasoro algebras, or
loop groups with Kac-Moody algebras are examples of great interest to physics.
3.2.1 Compactness
Criteria for compactness of a subspace X take different forms depending on the
space it is in. Thus, a subset X of a topological space is said to be compact if and
3.2. GLOBAL PROPERTIES 81
3.2.2 Connectedness
A space M is said to be disconnected if it can be partitioned into disjoint non-empty
open sets. M is said to be connected if it is not disconnected. There is a closely related
concept: M is said to be pathwise connected if given any two points x and y in M
we can find a continuous path entirely in M joining x to y, i.e. a continuous map
π ( t) in M, 0 ≤ t ≤ 1, with π (0) = x and π (1) = y. If M is locally Euclidean,
as is the case of all the groups we will study, the two concepts, ‘connected’ and
‘pathwise connected’, may be considered the same.
A disconnected Lie group G can be uniquely partitioned into disjoint sets,
called components, such that any two elements of the same component can be
joined by a continuous path, while two elements of different components cannot.
The identity component, called G0, which contains the identity I, plays a crucial
role. We can verify that G0 is closed under multiplication and inversion as fol-
lows. For any two elements g, h of G0, there exist continuous paths entirely in G0
called α ( t) and β ( t ), 0 ≤ t ≤ 1, with α (0) = β (0) = I, α (1) = g, and β (1) = h.
Then α ( t) β (t) is a continuous path lying in G0 joining I to gh, showing that the
product of two elements of G0 is again in G0 . In addition, α ( t)−1 is a continuous
path from I to g−1 and, as G0 is connected, the inverse of an element of G0 is
again in G0. So the connected component containing the identity is a subgroup of
G. In fact, G0 is a normal Lie subgroup. To see this, we need to show that if g ∈ G0
and h ∈ G0, then ghg−1 ∈ G0 . Let α : [0, 1] → G0, with α (0) = I and α (1) = h.
Then the conjugation of α ( t) by g is a continuous map that joins gα (0) g−1 = I to
gα (1) g−1. As G0 is connected to I, the end point ghg−1 must be in G0. In other
words, G0 is invariant under conjugation. Finally, as G0 is a normal subgroup of G,
the set consisting of G0 and its cosets forms a discrete group, which is the quotient
group G/G0. (One often defines the group of components, called π0, of G such that
its order gives the number of connected components. The group G is connected if
and only if π0 ( G ) is trivial, i.e. π0 ( G ) = { 1} , or | π0| = 1.) In summary,
• Let G be a Lie group, and G0 its connected component containing the identity.
Then G0 is a normal subgroup and a Lie group; the quotient group G/G0 is a 0-
dimensional Lie group isomorphic to a discrete group. A study of Lie groups re-
duces to studies of their connected components and their discrete groups.
Now, the connected component can be built up by taking multiple products
of elements near the identity. Let G be a connected Lie group, and U ⊂ G an open
neighborhood of the identity. Assume that U −1 = U (replacing U with U ∩ U −1
if necessary), and define H = ∪n≥1U n = U ∪ U 2 ∪ U 3 · · · (where U = { gi } ,
U 2 = { gi gj } , etc.). H is a subgroup of G (since U n U m ⊆ U n+m ). H is open in
G since it is the union of open sets. Moreover, if for every g ∈ G but g 6 ∈ H,
then gH 6 ⊂ H (since otherwise if h1, h2 ∈ H are such that gh1 = h2, then g =
h2 h− 1
1 ∈ H, in contradiction). Thus H is the complement of the union of all cosets
not containing H, so it is closed. Since H is a non-empty, open and closed subset
of G, and since G is connected, we have G = H. Therefore,
• If G is a connected Lie group and U an open neighborhood of the identity, then U
generates the whole G. This result shows the important role of U in the local
properties of a connected group.
3.2. GLOBAL PROPERTIES 83
3.2.3 Simply-Connectedness
A Lie group G is said to be simply-connected if every closed curve in G can be continu-
ously varied so that it contracts to a point in G. To show that G is simply-connected,
one verifies that for every closed loop lying in G defined by α ( t), 0 ≤ t ≤ 1,
α (0) = α (1), there exists a family of continuous curves α ( s, t), 0 ≤ s, t ≤ 1, lying
84 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
in G, such that (i) The curve is a closed curve: α ( s, 0) = α ( s, 1) for each value
of s; (ii) The curve α (0, t) coincides with the given loop α ( t ) for all t; (iii) When
s = 1, the curve contracts to a point: α (1, t) = α (1, 0) for all t. (Whenever G is
connected, simply-connectedness informs us on its ‘shape’ and on the presence
or absence of ‘holes’ in space G. This being the case, one finds it useful to intro-
duce the fundamental group π1 of G, such that G is simply-connected if and only if
π1 is trivial, π1 ( G ) = { 1} .)
E XAMPLE 21: In Euclidean space R n , let r ( t) with t ∈ [0, 1] and r (0) = r (1) = 0,
define a loop passing through the origin. Then we can define the continuous map
p ( s, t) = (1 − s)r(t), with 0 ≤ s ≤ 1, that represents a family of closed curves that,
as s varies, deform the loop continuously from r ( t) for s = 0 to a point r = 0 for
s = 1. So, R n is simply-connected.
E XAMPLE 22: In R 2, the points given by eiφ , with real φ, define the circle S1. With
0 ≤ t ≤ 1 consider the following curves in S1: (1) φ = t (1 − t) defines a curve that
can be deformed to a point; (2) φ = 2πt goes around the circle once and cannot
be deformed continuously to a point; (3) φ = 2πmt, with integer m > 0, define
a family of closed curves of m loops that cannot be deformed into one another.
So S1 is not simply-connected, it is infinitely connected (i.e. π1 = Z). However,
Sn−1 = { x ∈ R n || x| = 1} for any n > 2 is simply-connected (π1 = { 1} ). It
follows that SU (2), a connected Lie group topologically equivalent to S3, is also
simply-connected.
E XAMPLE 23: Let R + be the additive group of real numbers. It is non compact,
simply-connected, and contains a discrete invariant subgroup { n } = a (0, ± 1,
± 2, . . . ) (any real a). The circle S1 is then isomorphic to R + / { n} and multiply
connected. If S1 is associated with the rotation group SO (2), all rotations through
θ + 2πn are identified. On the segment [0, 2π ) of R + , with 2π identified with 0,
SO (2) is single-valued.
C OMMENTS . The importance of the topic discussed here arises from the following
facts. If the group manifold is simply-connected, every continuous function on
the group is single-valued, and there exists a 1-to-1 correspondence between the
representations of G and those of its Lie algebra (introduced later). On the other
hand, if the group manifold is multiply-connected, then m-valued functions can
exist, and we expect to have m-valued representations. The difficulty in study-
ing multiple-valued representations is resolved by the fact that for any multiply
connected group G, there exists a simply-connected group G e called the universal
e
covering group for G, such that G can be mapped to G by homomorphism, i.e. G
is isomorphic to the quotient group G/N e where N is an invariant subgroup of G. e
Every irreducible representation (Definition 3.8) of G, whether single or multiple-
valued, is a single-valued representation of G, e and one needs to study just G. e
In Table 3.1 we give a summary of the global properties (compactness, con-
nectedness, simply-connectedness) of some of the Lie groups discussed in this
chapter. The real Lie groups are listed first, followed by the complex Lie groups.
In each row we indicate, in order, the name of a group, its compactness (CMP
(Yes or No)), connectedness (CNN), simply-connectedness (SCNN), and finally the
fundamental group π1 (G) for connected groups (Y). The order of the component
group |π0 | (the number of components) is also given; simply-connectedness and
3.3. LOCAL PROPERTIES 85
π1 (G) are given whenever the group is connected. In the last row, one finds the
pages where more information relevant to each column can be found. In the trivial
cases, there are possibly exceptions to the displayed general rules: GL(1, R) = R× ,
GL(1, C) = C× , SL(1, R) = SL(1, C) = {1}, and O(1) = SO(1) = SU(1) = {1}.
Note that U(n), SU(n), and Sp(n) are real Lie groups.
Table 3.1: Global properties of some Lie groups discussed in this chapter.
where here and in the following summation over repeated indices is implicit.
Substituting this expression in (3.3) we obtain by expansion
0 0 0 ∂ f j ( β; x0 ) ∂
F ( x ) ≈ F ( x ) − δαµ F ( x0) . (3.5)
∂β µ ∂x0j
β=0
where, on the RHS, θ is infinitesimal and M is the generator of the matrix group.
E XAMPLE 27: Let F ( x) = x + c, with a constant c.
(a) If x0 = f ((0, a); x) = x + a: we have X1 = 0 and X2 = − ∂/∂x = − ∂ x.
From X1 , X2 , we get F 0 ( x) = (1 − a∂ x )( x + c ) = x + c − a.
(b) If x0 = f ((α, 0); x) = eα x, then X1 = − x∂ x and X2 = 0, and we get
F 0 ( x) = ( x + c) + (− αx∂ x + 12 α2 x∂ x x∂ x + · · · ) x = c + x(1 − α + 12 α2 + · · · ), which
is F 0 ( x) = e−α x + c. This tells us that we can get the full transformation if we
know the local transformation specified by Xµ .
Commutation relations. As the generators Xµ are derived from the elements of a
connected Lie group G in a neighborhood U of the identity, one may expect that
the vector space spanned by Xµ has a structure arising from the group laws. Let
α, β be arbitrary elements in U, where α may be approximated to
1
α ≈ I + δαµ Xµ + δαµ Xµ δαν Xν + · · · ,
2
and similarly for β. We have, up to second order, the products
[ Xµ , Xν ] = cλµν Xλ , (3.12)
where the expansion coefficients cλµν , called structure constants (in the given ba-
sis), are constant numbers. If Xµ† = ± Xµ , then both LHS and RHS of (3.12)
are skew-Hermitian, and so if Xµ are Hermitian the c’s are imaginary, but if Xµ
are skew-Hermitian the c’s are real. The properties implicit in the definition of
the commutator (namely, skew symmetry [ A, B ] = −[ B, A], and the Jacobi identity
3.3. LOCAL PROPERTIES 89
Given Xµ in terms of the composition function φ, as in (3.9), one can readily prove
the results just stated and obtain explicit expressions of the structure constants in
terms of φ ( β, α ) as follows. As Θ µλ is non-singular, we can define its inverse by
Ψµν ( α ) Θνλ ( α ) = δµλ for arbitrary α. Using Xµ ( α ) = − Θ µλ ( α ) ∂λ in the commu-
tation relation, we have
[ Xµ , Xν ] = Θ µσ ( ∂σ Θ νλ ) − Θ νσ ( ∂σ Θ µλ ) ∂λ .
Introducing Ψρκ Θ κλ = δρλ on the right-hand side we immediately obtain
[ Xµ , Xν ] = cκµν Θ κλ ∂λ = cκµν Xκ , (3.15)
where
∂Θµρ ( α ) ∂Θνρ ( α )
cκµν = Θ νσ ( α ) − Θ µσ ( α ) Ψρκ ( α ) (3.16)
∂ασ ∂ασ
are in fact independent of α, as can be shown by differentiation. So cκστ can be
calculated at any point α and, in particular, at the identity (α = 0), where one
has Θ (0)µν = Ψ (0)µν = δµν . Thus, setting α = 0 in (3.16), one gets the explicit
expression of the structure constants of the group of transformation
2 κ
κ ∂ φ ( β, α ) ∂2 φκ ( β, α )
cµν = − . (3.17)
∂β µ ∂αν ∂β ν ∂αµ α= β=0
x0
@
3
α @ δα
@
-
R
@
x α + dα x0 + dx0
two real numbers, also a real, is commutative and associative. The vector space
spanned by the generators of a Lie group of transformation defined in the pre-
ceding section, which comes equipped with a product, is also an algebra, called
Lie algebra, although the product is neither commutative nor associative.
E XAMPLE 33: o(V ), o( n, R ) and o( n, C ). Let V be a linear vector space with a bi-
linear symmetric inner product h v, wi = v T w defined on V . Then, with the linear
and bracket operations defined as in gl(V ), the orthogonal Lie algebra o(V ) consists
of all linear operators X on V under which the bilinear form h, i is invariant, so
that h Xv, w i + h v, Xwi = 0 for all v, w in V . If X and Y satisfy this invariance
condition, so does [ X, Y ]. In particular, for V = Fn , we may take h v, wi = ∑i vi wi
for any v, w ∈ Fn , so that the invariance relation reads ∑ Xij v j wi + vi Xij w j = 0.
If it is satisfied by X, it is also by cX for any c ∈ F. The orthogonal Lie al-
gebra o( n, F ) consists of the skew-symmetric n × n matrices X = − XT over F:
o( n, F ) = { X ∈ gl( n, F )|XT + X = 0} . Note that the expected subalgebra of
o( n, F ) defined by the additional restriction Tr X = 0 is unnecessary because
XT + X = 0 implies that all diagonal entries are 0; so o( n, F ) and so( n, F ) are
the same space of matrices, meaning o( n, F ) ≡ so ( n, F ).
C OMMENTS . (a) In general, Lie groups having the same identity component have
the same Lie algebra (see p. 97 for an explanation of this correspondence). For
example, the Lie groups O ( n) and SO ( n), which coincide in neighborhoods of
the identity I, have the same Lie algebra o( n, R ) = so ( n, R ). (b) As we have
seen before, the orthogonal group is defined by the form invariance condition
h Av, Awi = h v, w i, or the matrix relation AT A = I. If A is in the identity compo-
nent, we may write A = I + M + · · · , where | Mij | 1, the two relations become
h Mv, wi + h v, Mwi = 0 and MT + M = 0.
E XAMPLE 34: u( n ) and su( n ). Let V be a complex vector space endowed with a
positive-definite sesquilinear symmetric inner product h v, w i = v † w on V . The
unitary Lie algebra u(V ) consists of all the linear operators X on V satisfying the
(infinitesimal) invariance (unitarity) property h Xv, w i + h v, Xw i = 0 for all v, w
in V . The Lie algebra u(V ) is defined over R, but not over C (in the sense that if
X 6 = 0 has the invariance property, so does aX if a is real, but not if a is complex).
So u(V ) is a real Lie algebra.
In V = Cn with inner product h v, wi = ∑i v∗i wi (∗ denoting complex conju-
gation), the unitary Lie algebra u( n ) consists of the n × n complex matrices that
satisfy X† + X = 0 (X is then said to be skew-Hermitian or skew-adjoint). The
special unitary Lie algebra su ( n ) consists of elements of u( n ) with Tr X = 0. Note
that su ( n ), just like u( n ), is a real Lie algebra, even though it involves complex
matrices; and also that it is not the same as so ( n, C ).
E XAMPLE 35: sp(V ), sp( n, F ), and sp( n ). Let V be a vector space over F = R
or C equipped with a definition of a non-singular skew-symmetric bilinear form
B ( v, w). Then, the symplectic Lie algebra sp (V ) consists of all linear operators X act-
ing on V that have the infinitesimal invariance property B ( Xv, w) + B(v, Xw) = 0
for all v, w in V . The symplectic Lie algebras sp( n, R ) and sp ( n, C ) of R 2n and C2n ,
in which B ( v, w) = v T Jw, are set of 2n × 2n matrices that satisfy XT J + JX = 0
(cf. Example 8). For n = 1, invariance means Tr X = 0, and so sp(1, F ) is identical
with sl (2, F ). The matrices that are simultaneously in sp( n, C ) and u(2n ) form the
compact symplectic Lie algebra, denoted usp( n ) or sp( n ).
E XAMPLE 36: Consider V = R n+m with metric gn,m , so that h v, w iL = vT gn,m w
for any v, w ∈ V . The generalized orthogonal Lie algebra o( n, m; R ) consists of those
3.4. LIE ALGEBRAS 93
3.4.2 Properties
We discuss here a number of concepts and properties of Lie algebras, with exam-
ples based on structures of lower dimensions.
Structure constants. In a vector space g of dimension η, a set of η linearly in-
dependent vectors { xµ | µ ∈ N η } may be chosen as a basis, and any vector x in
this linear space is of the form x = ∑µ aµ xµ , where aµ are real for a real space,
and complex for a complex space. The Lie algebra, also denoted g, based on this
vector space can then be defined by a multiplication table for xµ which enumer-
ates all [ xµ , xν ] = ∑λ cλµν xλ , where the coefficients cλµν , called the structure con-
stants of the Lie algebra g (in the chosen basis), obey the relations cλµν = − cλνµ and
cαµν cκαλ + cανλ cκαµ + cαλµ cκαν = 0 derived from the axioms defining the Lie bracket
(see also (3.13) and (3.14)). This multiplication table, or an equivalent list of the
η 3 structure constants (1 ≤ λ, µ, ν ≤ η and taking into account symmetries), com-
pletely defines the Lie algebra and implies many of its properties. But in practice
this approach is rarely followed.
E XAMPLE 37: If a Lie algebra is abelian, its structure constants vanish in any basis:
cλµν = 0 for all indices.
E XAMPLE 38: Given a Lie algebra g (of dimension η) and a Lie subalgebra h (of
dimension ζ) by definition closed under both linear and bracket operations, we
can choose a basis for g, such that xµ , µ = 1, . . . , ζ span h. As h is closed under
[, ], we must have cλµν = 0; µ, ν = 1, . . . , ζ; λ = ζ + 1, . . . , η in that basis.
E XAMPLE 39: Dimension one. Rx and Cx, defined in any basis x of a one-dimen-
sional vector space, are real and complex abelian Lie algebras with the trivial
bracket [ x, x ] = 0. Examples of one-dimensional
matrix Lie algebras are so (2)
(basis X = 10 −10 ) and so (1, 1) (basis X = 01 10 ).
E XAMPLE 40: Dimension two. Let x1 and x2 be a basis of a two-dimensional Lie
algebra g. Then, in general [ x1 , x2 ] = ax1 + bx2 , where a, b are real or complex
scalars. If both a and b vanish, we have an abelian algebra. If not, assume b 6 = 0,
then y1 = x2 /b and y2 = ax1 + bx2 satisfy [ y1, y2 ] = y2. So, there are just two
distinct two-dimensional real or complex Lie algebras:
( i) [ x1, x2 ] = 0 ;
(3.22)
( ii) [ x1 , x2 ] = x2 .
The first is the direct sum of two one-dimensional algebras, Fx1 ⊕ Fx2 , and the
second is the unique two-dimensional nonabelian algebra over either R, or C. An
example of ( ii) is the affine algebra of the line aff(1) (cf. Example 25).
94 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
Algebraic concepts. Let g and h be Lie algebras. The symbol [g, h ] denotes the
linear span of all [ x, y ] with x ∈ g and y ∈ h. Using this notation, we define a
Lie subalgebra of a Lie algebra g as a subspace, say h, of g that is closed under the
bracket operation, [h, h ] ≤ h, with the same linear and bracket operations as for
g. Several kinds of subalgebras are important enough to be given names:
• A Lie ideal is a Lie subalgebra q of a Lie algebra g such that [g, q] ≤ q, that
is to say, [ x, y ] ∈ q for every x ∈ g and y ∈ q.
• The center of g is defined as Z(g) = { z ∈ g|[ z, x] = 0 for all x ∈ g} . It is
an ideal of g. If g is abelian, Z(g) = g; and conversely Z(g) = g implies
an abelian g. A semisimple Lie algebra can have no center, Z(g) = 0; see
Sec. 3.5.2.
• The derived algebra D g of the Lie algebra g is the subspace spanned by all
Lie products: D g = {[ x, y ]|x, y ∈ g} , or in short hand, D g = [g, g], itself an
algebra and an ideal of g. The Lie algebra g is abelian if and only if D g = 0.
We say g is perfect if D g = g. Every semisimple Lie algebra is perfect (but
the converse is not true).
Just as for Lie groups, it is useful to have a way of distinguishing different Lie
algebras, which is given by the following
If Lie algebras g and g0 are related by a Lie isomorphism φ, we say that g and
g0 are isomorphic, or they are ‘the same up to isomorphism’; and write g ∼ = g0 .
The inverse map φ−1 is then also a Lie algebra isomorphism.
For example, when we said that the nonabelian two-dimensional Lie algebra
is unique, we implied ‘up to isomorphism’. Thus, the two-dimensional Lie al-
gebra defined by [ y1, y2 ] = y1 is isomorphic to the Lie algebra defined in (3.22)
if we let x1 7 → − y2 and x2 7 → − y1. This involves a simple change of the basis in
the underlying vector space. Another kind isomorphism calls for a less obvious
change of the base field, as we now explain.
If g is a finite-dimensional real Lie algebra, its complex extension is obtained by
tensoring g with C over R, which we write as gC = g ⊗ R C = g ⊕ ig, where i is
the complex unit. This means one forms linear combinations x + iy with x, y ∈ g,
and then (a, b ∈ R) define ( a + ib )( x + iy ) and [ x + iy, x0 + iy0 ] in the usual way.
The resulting structure gC is a complex Lie algebra, called the complexification
(or complex extension) of the real Lie algebra g. Conversely, we say that g is a
real form of gC (one among several possible real forms associated with a given
complex Lie algebra gC ). A basis for g over R is also one for gC over C.
E XAMPLE 41: Dimension three. It is convenient to classify the algebras of a given
dimension according to the dimensions of their derived algebras. In the present
case, dim g = 3, there are 4 possibilities: dim D g = dim [g, g] = 0, 1, 2, 3.
Let x1 , x2 , x3 be a basis of the real (or complex) space g. If dim D g = 0, g is
3.4. LIE ALGEBRAS 95
[ R 1, R 2 ] = R 3, [ R 2, R 3 ] = R 1, [ R 3, R 1 ] = R 2 ;
and define algebras of the type (3.27). It is evident that the complex extension
o(3)C = o(3) + io(3) of the real Lie algebra o(3) is isomorphic to o(3, C ).
E XAMPLE 45: su (2). This real Lie algebra consists of the 2 × 2 trace-free skew-
Hermitian matrices, expressible in the (canonical) basis given by
1 0 −i 1 0 −1 1 −i 0
S1 = 2 , S2 = 2 , S3 = 2 .
−i 0 1 0 0 i
They obey the Lie products [ S1, S2 ] = S3, [ S2, S3 ] = S1, [ S3, S1 ] = S2, which
show that su (2) ∼
= so (3) (this isomorphism arises from the fact that the Lie groups
SU (2) and SO (3) look the same near I, see p.97 et seq.). Note for reference that
these matrices may be written S j = − iσj /2, with j = 1, 2, 3, in terms of the stan-
dard Pauli matrices σj normalized such that σi σj = δij + ieijk σk . Explicitly,
0 1 0 −i 1 0
σ1 = , σ2 = , σ3 = . (3.30)
1 0 i 0 0 −1
E XAMPLE 46: o(2, 1; R ). This generalized orthogonal real Lie algebra consists of
all 3 × 3 real matrices L such that LT g2,1 + Lg2,1 = 0, with the metric g2,1 =
diag [1, 1, − 1]. We take as a basis
0 0 0 0 0 1 0 −1 0
L1 = 0 0 1 , L2 = 0 0 0 , L3 = 1 0 0 .
0 1 0 1 0 0 0 0 0
[ L1 , L2 ] = L3 , [ L 2, L 3 ] = − L 1, [ L3, L1 ] = − L2.
After a change of basis such that L1 ↔ L3, these relations are of the type (3.28),
and, as mentioned above, are equivalent to (3.27) when complexified. Therefore,
o(2, 1; R )C ∼
= o(3, C ).
E XAMPLE 47: sl (2, R ). This algebra is the space of all 2 × 2 trace-free real matrices;
a convenient basis is
1 0 0 1 0 0
H= , E= , F= , (3.31)
0 −1 0 0 1 0
From this definition, one sees that its properties are mostly the same as for ex-
ponentials of numbers, with the differences arising from the non-commutativity
of matrices. Let X, Y ∈ M( n, F ), and eX , eY be defined as in (3.32), then the exp
map has the following properties:
1. e0 = I.
2. If XY = YX, then eX eY = eX+Y . Hence eX e−X = I, or (eX )−1 = e−X .
3. If S is an invertible matrix, then exp( SXS−1) = S exp( X) S−1.
4. (exp X)T = exp ( XT ) (where ()T denotes transposition).
5. If γX ( s ) = esX , then γX ( s )γX ( t ) = γX ( s + t ) for any s, t ∈ R, or C.
2
6. In M( n, F ) viewed as a space Fn , the s-function γX ( s ), with a fixed X,
describes a smooth curve with tangent vector at s given by the derivative
0 ( s ) = Xγ ( s ) = γ ( s ) X; in particular, at the identity, it is γ0 ( 0 ) = X.
γX X X X
7. det (exp X) = exp(Tr X).
−1
The continuous curve γX ( t ) = exp tX, satisfying γX (0) = I, γX ( t ) = γX (− t)
and γ ( s )γ (t) = γ ( s + t ) for all s, t ∈ R and fixed X, is a group, called the one-
parameter subgroup in G determined by X. It maps R → G : t 7 → etX for each
X ∈ M( n, R ). Given these properties, for every X ∈ M( n, R ) we take a real
variable t to form the series ∑∞ m m
m=0 t X /m!, which is an invertible matrix exp tX
for all t, and so lies in GL( n, R ). Thus exp defines a continuous homomorphic
map exp : R 3 t → etX ∈ GL( n, R ), and every continuous homomorphism R →
GL( n, R ) is of this form.
A closed subgroup of GL( n, R ) is a Lie group called linear group (see p. 76).
Let G ⊂ GL( n, R ) be a linear group, and define the set of real matrices
(making use of the series (3.32)). Since X = limt→0 (etX − I ) /t = detX /dt |0, it is
the tangent space to G at the identity. The dimension of the linear group G is then
defined to be the dimension of the vector space Lie( G )
We can also calculate the group commutator of the elements eX , eY of G:
2 [X,Y ]+···
etX etY e−tX e−tY = I + t2[ X, Y ] + · · · = et . (3.33)
Here the dots stand for terms of degree three and higher. Given these relations,
it is evident that since eX and eY are in G, so is eX eY e−X e−Y , and therefore [ X, Y ]
is in Lie( G ) (cf. arguments leading to (3.12)). In other words, if we have X, Y ∈
Lie( G ), it follows that [ X, Y ] ∈ Lie( G ). The vector space Lie( G ) is closed under
the Lie bracket, and so by definition is a Lie algebra, Lie ( G ) = g. In particular, if
Lie( G ) = 0, then G is a discrete group, and vice versa.
3.5. BACK TO LIE GROUPS 99
E XAMPLE 55: Orthogonal Lie group, algebra. Let γ ( t) with γ (0) = I be a curve
consisting of orthogonal matrices in O ( n, F ), meaning that γT ( t )γ (t) = I for all
t. It follows, upon differentiation wrt t and setting t = 0, that γ0 (0)T + γ0 (0) = 0,
and so X = γ0 (0) is in o( n, F ). On the other hand, take any X ∈ o( n, F ) (so
that XT + X = 0), then the product (exp( tX ))T exp( tX ) = exp ( tX T ) exp( tX ) has
the derivative exp( tXT )( XT + X) exp( tX) identically equal to 0 for all t, which
means that (exp( tX ))T exp ( tX ) is constant for all t, which is I if we let t = 0.
Hence (exp( tX ))T exp( tX ) = I, or exp( tX) lies in O ( n, F ) for all t. In conclusion,
Lie O ( n, F ) = o( n, F ) = { X | X ∈ gl ( n, F ); XT + X = 0} .
because the Lie group SO( n, F ) is the connected component of O ( n, F ), so that the
two groups have the same properties near the identity, and therefore the same Lie
algebra. This is an example of an important general result in Lie group theory:
• Given the connected components G0 and G00 of linear groups G and G0 , if G0 =
G00 then Lie( G ) = Lie ( G0 ); and vice versa.
E XAMPLE 56: Unitary Lie group and algebra. What is the Lie algebra of the Lie
group U( n )? A matrix A is unitary if and only if A† = A−1, and etX is unitary
† †
if and only if (etX )† = e−tX , or etX = e−tX (that (etX )† = etX can be seen from
†
the power series of exp). For this to hold, it suffices X = − X. Conversely, if
†
etX = e−tX , it is necessary that X† = − X, as seen by differentiating at t = 0. In
other words, etX is unitary if only if X is skew-Hermitian, or
Lie U( n ) = u( n ) = { X | X ∈ gl( n, C ); X† + X = 0} .
The special unitary group SU( n ) is a subgroup defined by the additional condi-
tion det etX = 1 for all real t,which holds iff Tr X = 0 in the tangent vector space.
We write this as: Lie SU( n ) = su ( n ) = { X | X ∈ u( n ); Tr X = 0} .
E XAMPLE 57: Symplectic Lie group and algebra. A symplectic Lie group consists
of elements A satisfying the condition AT J A = J. The corresponding relation
T
etX JetX = J holds for all real t iff XT J + J X = 0. And we have
Lie Sp( n, F ) = sp( n, F ) = { X | X ∈ gl (2n, F ); XT J + J X = 0} .
because it shows that the group law of a connected Lie group G can be recovered from
the Lie bracket in Lie algebra g (but not necessarily the overall geometry of G) by
mapping µ ( X, Y ) 7 → exp( X) exp(Y ).
From Lie algebra back to Lie group. The reversed path (from Lie algebra to Lie
group) is more complex, and for this reason, we just enumerate here a few of
the most important results concerning connected Lie groups obtainable from Lie
algebras (see [FH], [Sa], [Ja] for details and proofs):
• The product rule in a connected Lie group G admitting Lie algebra g = Lie( G )
can be recovered from the Lie bracket in g via (3.34).
• Any Lie algebra is isomorphic to a subalgebra of gl ( n ) (Ado’s theorem). Hence,
every Lie algebra is equivalent to a matrix Lie algebra, in contrast to the
fact that not all Lie groups are matrix Lie groups.
• Any finite-dimensional Lie algebra is isomorphic to the Lie algebra Lie ( G ) of some
Lie group G (Lie’s third theorem).
• For any finite-dimensional Lie algebra g, there is, up to isomorphism, a unique
connected, simply-connected Lie group G such that Lie( G ) = g. Any other
connected Lie group G0 with Lie ( G0 ) = g must be a quotient group of the
form G/Z for some discrete invariant subgroup Z in G.
E XAMPLE 64: Consider the Lie algebra u(2), with basis S0 and Si (i=1,2,3), just as
in Examples 45 and 62. We calculate the elements of the Lie group G = U (2),
admitting Lie(G) = u(2) as the Lie algebra by exponential mapping. Let X =
θ0 S0 + ∑3i=1 θi Si, with real numbers θ0, θi , be an arbitrary element in u(2), then
the exp-map X 7 → exp X produces an element in U (2):
X → exp θ0 S0 + ∑i θi Si = exp ( θ0S0) exp ∑i θi Si
where we have used [ S0, Si ] = 0 in the last equality. To transform this into an
explicit matrix form, we will need ( ∑3i=1 θi σi )2 = θ 2, as shown here:
3 2 3
∑ θi σi = ∑ θi θj σi σj = ∑ θi θj ( δij + ieijk σk ) = ∑ θi2 .
i= 1 i,j i,j i= 1
(The minus sign in − i n · σ comes from the sign in S j = − iσj /2.) Thus, any ele-
ment of U (2) is specified by four real parameters: one specifying a phase factor
exp(iθ0), and three determining a 2-by-2 unimodular unitary matrix, a result im-
plying that U (2) is expressible as a direct product of U (1) and SU (2).
In the U (2) matrix given above, call A the factor belonging to SU (2), and
re-express its elements in terms of r0 = cos( θ/2) and ri = ni sin( θ/2), with i =
1, 2, 3. If we regard r0, ri as the Cartesian coordinates in a 4-dimensional Euclidean
space, then, as det A = r20 + r21 + r22 + r23 = 1, the group space is the unit sphere
S3, and so SU (2) is both compact and simply-connected.
Again, let A ( n, θ ) = exp ∑i θi Si ∈ SU (2), then A ( n, 0) = diag[1, 1] = I2 is
the identity, and A ( n, ± 2π ) = − I2 for arbitrary n. Further, we also have A ( n, θ +
4π ) = A − n, −( θ + 2π ) = A ( n, θ ). It follows that in the parameter space the
group elements occupy a sphere of radius θ = 2π, centered at the identity I2,
and with all points on the spherical shell θ = 2π identified with − I2. I2 and − I2
form a discrete invariant subgroup Z = { I2, − I2} of order 2 in SU (2). Hence,
the Lie algebra su (2) gives rise to two distinct Lie groups: SU (2)/ { I2} ≡ SU (2) and
SU (2)/Z ∼ = SO (3). Any pair of elements ± A ∈ SU (2) corresponds to a single
element R ∈ SO (3), and the map SU (2) → SO (3) is two-to-one.
The relationship between some matrix Lie groups and matrix Lie algebras are
summarized in the following Table 3.2.
A, X, and I are real square matrices, whereas B and Z are complex square matrices, all
of the size n × n, except for the symplectic groups/algebras when they are of 2n × 2n, as
is the skew-symmetric metric matrix J. The dimensions in the last columns are over R
(or C) for real (or complex) groups/algebras. S and SS refer to the simplicity and semi-
simplicity of the Lie algebra (Y for yes). Exceptional cases: so(2) and so(2, C) are one-
dimensional abelian; so(4) and so(4, C) are semisimple, but not simple. Decompositions:
∗ gl(n, R) = RI ⊕ sl(n, R); † u(n) = iRI ⊕ su(u); ‡ gl(n, C) = CI ⊕ sl(n, C).
3.6. REPRESENTATIONS 105
3.6 Representations
We now discuss in general terms the representations of Lie groups and Lie alge-
bras. As these two structures are closely related, the latter being relatively the
simpler of the two, it is important to establish how, what, and when representa-
tions of Lie algebras can inform us about the representations of Lie groups. In
general, we will assume representations to be complex, but will consider both real
and complex Lie groups and Lie algebras. The scope of our present study of the
representations of Lie groups and Lie algebras is determined by their dimensions:
we will be dealing only with Lie groups, Lie algebras, and their representations
of finite dimensions.
ad Xi · Xj = Xk R kj ( Xi ) = Xk ckij ; (3.38)
" # " #
2 0 0 0 0 0
adH = 0 0 0 , adX− = −1 0 0 .
0 0 −2 0 2 0
In the canonical basis { S1, S2, S3 } , in which [ S1, S2 ] = S3 (cf. Example 45), we
have the non-zero matrix elements: (adS1)23 = − 1, (adS1 )32 = + 1; (adS2)13 =
+ 1, (adS2 )31 = − 1; (adS3 )12 = − 1, (adS3)21 = + 1. All other elements are 0.
Note for reference (adS1 )2 = diagonal[0, − 1, −1], (adS2)2 = diagonal[− 1, 0, −1],
(adS3)2 = diagonal[− 1, −1, 0]. And so one can find Tr (adSi · adS j )2 = − 2δij.
From the relation π ( x) = dΠ (etx) /dt |0 for all x ∈ g between the representa-
tions of a Lie group G and its Lie algebra Lie ( G ) = g, one can expect that every
G-morphism is a g-morphism. The converse holds if G is connected and simply-
connected, so that φG (V , W ) = φg (V , W).
An isomorphism of representations is a morphism that is invertible. Two repre-
sentations related by an isomorphism are said to be isomorphic or equivalent to one
another, and can be regarded as the same.
Let V be a representation of a Lie group G (or Lie algebra g). Then a subrepre-
sentation is a vector subspace W ⊂ V invariant under the action: Π ( g)W ⊂ W
for all g ∈ G (respectively, π ( x)W ⊂ W for all x ∈ g). A subrepresentation W of
V is called proper if it is neither { 0} nor V itself.
Definition 3.8. A non-zero representation V is said to be irreducible or simple, iff it
has no proper subrepresentations. Otherwise, it is called reducible.
110 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
1
| G | g∑
B( v, w ) = B Π ( g)v, Π ( g)w , (3.39)
∈G
This inner product is positive definite: B̄( v, v ) > 0 (integral of a positive function:
the Haar measure is positive,and, on a compact group, positive definite); and
G-invariant: B̄ Π ( h)v, Π (h)w = B̄ ( v, w). The representation Π is unitary with
respect to the inner product B̄, and therefore completely reducible:
• Every finite-dimensional representation of a compact Lie group is either irre-
ducible, or completely reducible.
The existence of the Haar measure for compact Lie groups implies that many
properties of the representations of finite groups have parallels in the representa-
tions of compact Lie groups. In particular, characters can be defined, with familiar
properties. Let ui be a basis of a vector space V , and Π ( g) : V → V the linear oper-
ators corresponding to the elements g of a compact real Lie group G. A character
of representation V is defined to be a function on the group:
From this definition, it is clear that χV does not depend on the choice of basis in
V , and that χV = 1 for the trivial representation
R V = C.
Now, define the inner product h f 1, f 2 i = G f 1∗ ( g) f 2( g)dµ ( g) with the Haar
measure µ ( g) on G. With respect to this inner product, the characters are orthog-
onal (vectors) — if V, W are non-isomorphic complex irreducible representations
of G, then h χV, χW i = 0. Suppose V is a complex representation of a compact
real Lie group G, then V is irreducible if and only if h χV , χV i = 1. As V is com-
pletely reducible, it can be uniquely decomposed as V = ∑i ai Vi, where Vi are
pairwise non-isomorphic simple representations, with the multiplicities given by
112 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
ai = h χV , χVi i. These properties are just the same as for finite groups. So, if we
restrict ourselves Rto compact Lie groups, we can carry over, with modifications
(e.g. | G |−1 ∑g 7 → G dµ ( g)), the results derived in Chapter 2 for finite groups.
Among the Lie groups already mentioned, O ( n ), SO( n ), U( n ), SU( n ), and
Sp( n ) are compact groups (see Example 14 in Sec. 3.2.1).
Tensor Products of Representations. In Chapter 2 we define products of repre-
sentations of finite groups. That definition remains valid as well for Lie groups.
We recall it now, and extend it to Lie algebras.
Definition 3.11 (One Lie group). If G is a Lie group, and ( Π 1, U ), ( Π 2, V ) are rep-
resentations of G, then the tensor product of Π 1 and Π 2 is a representation
of G acting
on U ⊗ V defined by Π ( g) ∑( u ⊗ v ) = ∑ Π 1 ( g) u ⊗ Π 2( g) v for all g ∈ G. The usual
notation is Π ( g) = Π 1 ⊗ Π 2( g).
Definition 3.12 (Two Lie groups). Let G and H be Lie groups, ( Π 1, U ) a representa-
tion of G, and ( Π 2, V ) a representation of H. Then the tensor product of Π 1 and Π 2 is
a representation of G × H acting on U ⊗ V defined by
Π ( g, h) ∑( u ⊗ v ) = ∑ Π 1( g) u ⊗ Π 2( h )v
Definition 3.13 (One Lie algebra). Let g be a Lie algebra, and let ( π1, U ) and ( π2, V )
be representations of g. Then the tensor product π1 ⊗ π2 is a representation of g acting
on U ⊗ V defined by π1 ⊗ π2 ( x) = π1 ( x) ⊗ I + I ⊗ π2 ( x) for all x ∈ g.
To see this is so, let α ( t) = etx be some one-parameter curve in a Lie group G,
with x = α0 (0) and α (0) = 1. Then, as π ( x) = d Πα (t)/dt|0, we have
d
π1 ⊗ π2 ( x)( u ⊗ v ) = Π 1 (etx ) u ⊗ Π 2(etx ) v
dt 0
Definition 3.14 (Two Lie algebras). Let g and h be Lie algebras. If ( π1, U ) is a repre-
sentation of g, and ( π2, V ) a representation of h, then the tensor product of π1 and π2,
denoted π1 ⊗ π2, is a representation of g ⊕ h acting on U ⊗ V , defined by
π1 ⊗ π2 ( x, y ) = π1 ( x) ⊗ I + I ⊗ π2 ( y )
The trace form is clearly symmetric; that it is also an invariant form can be shown
by recalling that π ( xy ) = π ( x) π (y) when taking the trace of the representations
of both sides of the identity [z, xy ] = [z, x] y + x [z, y ]. The trace form on a real Lie
algebra has all the properties of an inner product except possibly positive defi-
niteness. But the trace form on a complex Lie algebra cannot be an inner product,
since it lacks the sesquilinearity of inner products on complex vector space.
The importance of the trace form lies in the fact that if the trace form of a Lie
algebra g in a representation (π, V ) is non-degenerate, then g is decomposable into a
direct sum of the form g = Z(g) ⊕ gss , where gss is semisimple and Z(g) the center
of g (i.e. [ Z(g), g] = 0).
E XAMPLE 73: Here we calculate the trace form of some classical Lie algebras g in
the standard representation (π, V ), making use of properties listed in Table 3.2. We
use notations x ∈ g, π ( x) = X, i = 1, 2, . . . , n. For gl ( n, F ), the standard rep-
resentation is V = Fn , so one has B ( x, y ) = ∑ij Xij Yji , which is non-degenerate.
Since sl ( n, F ) < gl ( n, F ), it follows that B on sl ( n, F ) is also non-degenerate. For
114 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
In calculating the trace of matrix representatives R, the matrices (of order η) can
evidently be expressed in any basis. Given a basis { Xi, i = 1, 2, . . . , η } of g, the
Killing form of any pair Xi and Xj is
K ( Xi, Xj ) = Tr adXi · adXj = ∑ cki` c`jk . (3.45)
k`
One can calculate K ( Xi, Xj ) using either the Lie brackets among the Xi or the
structure constants cki` . Since cki` are tensors of type (2, 1), the cross-contraction
of the two upper indices in the pair of c yields a second-rank covariant tensor,
called the Cartan–Killing metric, and denoted κij = K ( Xi, Xj ). As c can be made
real for all real or complex Lie algebras, [ κij ] is real and symmetric, and therefore
can be brought into a diagonal form. If one or several of the diagonal elements
turn out to be zero, the metric tensor is singular, making the corresponding bi-
linear form K degenerate. So, if g has an abelian ideal, there will be directions in
which ad is zero and K degenerate.
If the Killing form is non-degenerate, the matrix [ κij ] is non-singular, and an
inverse [ κ ij ] exists, such that κij κ jk = δi k = κ kj κ ji . These second-rank tensors,
κij and κ ij , are symmetric in their respective indices; they are nonsingular and
dual, and may be used to raise and lower indices of tensors. In particular, we can
define structure constants that are completely antisymmetric in its three indices, as
in cijk = c`ij κ`k .
Just as any other trace form, the Killing form may be non-degenerate, degen-
erate, or even zero. Following are some examples.
E XAMPLE 74: aff(1). We obtain κ11 = 1, κ12 = 0, κ22 = 0, and det κ = 0, so [ κ ] is
singular. For any X = a1 X1 + a2 X2 and Y = b1 X1 + b2 X2 , we get K ( X, Y ) = a1 b1 .
So for an arbitrary Y, one may have K ( X, Y ) = 0 with a non-zero X = a2 X2 . The
Killing form on aff(1) is degenerate.
E XAMPLE 75: The Heisenberg algebra H (3) = { x, y, z } is defined by the commu-
tation relations [ x, y ] = z, [ x, z ] = 0, [ y, z ] = 0. In the adjoint representation we
3.6. REPRESENTATIONS 115
have the only non-zero elements (adx)zy = 1 and (ady )zx = − 1, together with
[adz ] = 0. Clearly, the Killing form K ( u, v) = 0 for any u, v ∈ H (3).
E XAMPLE 76: so (3). Recalling that adR i = R i, with R i the standard representa-
tion as in Example 68, we calculate the Cartan-Killing matrix κij = Tr ( R i R j ) =
− 2δij, and det κ = − 8, so [ κ ] is non-singular. Let X = aR x + bR y + cR z and X0 =
a0 R x + b0 R y + c0 R z be any elements of o(3). Then K ( X, X0) = − 2( aa0 + bb0 + cc0 ),
which nullifies for arbitrary X0 only if X = 0; so, this form is non-degenerate. In
particular K ( X, X ) = − 2( a2 + b2 + c2 ), to be compared with the orthogonality
relation ∑ cTi ci , where ci are the column vectors in X. This result holds in general:
the metric tensor of the orthogonal (real) Lie algebra is negative-definite.
E XAMPLE 77: sl (2, C ). With the adXα in Example 71, the matrix καβ can be calcu-
lated to give κ HH = 8, κ+− = κ−+ = 4; all other components are zero; it follows
det [ καβ ] = − 128. For any X = aH + bX+ + cX− in sl (2, C ), with complex num-
bers a, b, c, we have the Killing form K ( X, X0) = 8aa0 + 4( bc0 + cb0). So K ( X, X0) =
0 for arbitrary X0 only if X = 0. It is non-degenerate. We note further that the
trace form in the standard representation is given by BHH = 2, B+− = B−+ = 1;
all other components are zero. So that Bad ( x, y ) = 4Bstandard ( x, y ). This is so be-
cause on a simple Lie algebra (over C) any two symmetric invariant bilinear forms are
proportional to each other. See also the example of sl ( n, C ) discussed below.
E XAMPLE 78: su (2). To go to su(2) (recalling sl = su ⊕ i su), we require X† = − X,
which implies a∗ = − a and c = − b∗. Accordingly, letting a = iα, b = β + iγ and
c = − β + iγ, with real α, β, γ, we have the Killing form K ( X, X ) = − 8( α2 + β2 +
γ2 ), which is negative definite for arbitrary real α, β, γ. In the basis Si = − iσi/2
we have κij = − 2δij, and det [ κij ] = − 8 (cf. Example 71). Since Tr ( SiS j ) = − 21 δij ,
we again have Tr (adSi · adS j ) = 4Tr ( SiS j ). Also cf. case of so (3).
The Killing form is the basis of an effective test (Cartan criterion) for semisim-
plicity of Lie algebras, which has several important implications.
• A Lie algebra is semisimple if and only if its Killing form is non-degenerate.
Arguments leading to this theorem can be found for example in [FH] p. 479, or
[Ja] p. 69. (1) Suppose the Killing form K is nondegenerate. As with any other
form, this implies that g is reductive, i.e. g = Z(g) + gss, where [ Z(g), gss ] = 0. But
with nondegenerate K, the center must be 0; because otherwise, there would be an
x0 ∈ Z(g) so that ad x0 = 0 and K ( x0, y ) = 0 contrary to assumption. Therefore,
Z(g) = 0 and g is semisimple. (2) Suppose Lie algebra g is not semisimple. Then
it has a non-zero abelian ideal a (so that [a, a] = 0 and [a, g] = a). Choose a basis
{ Xi } for g. Take a non-zeroA in a and any X in g, and calculate the Killing form
K ( A, X ) from the formula X, [ A, Xk ] = X` (adA · adX)` k . If Xk ∈ a, then the
LHS
vanishes,
[ X, [ A, Xk ]] = 0, because a is an abelian ideal. If Xk 6 ∈ a, then
X, [ A, Xk ] ∈ a, so that the sum on the RHS contains only terms with X` ∈ a,
and (adA · adX)` k = 0 for ` = k. Hence, K ( A, X) = Tr (adA · adX) = 0, and the
Killing form K is degenerate.
E XAMPLE 79: gl( n, F ). Let x = cIn with c ∈ C and In the n × n identity. Clearly,
x ∈ gl( n, C ); or more precisely x ∈ Z[gl]. Since adx = 0 the Killing form K ( x, x)
is zero. On the other hand, we see that the trace form B ( x, x) = Tr ( x2 ) = nc2 6 = 0.
Since Tr x = nc, we verify that 2nTr( x2 ) − 2(Tr x)2 = 0.
116 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
can be associated with observable quantities. For example, the generators of rota-
tion algebra so (3) may be identified with the angular momentum; while those of
the hidden symmetry so(4) in the hydrogen atom, with the angular momentum
and the Runge–Lenz vector. As Emmy Noether has shown (1918), symmetry im-
plies the existence of conservation laws: so, in any physical system invariant un-
der a continuous symmetry that is global (acting the same way everywhere and at
all times), there exists an associated conserved ‘charge’, i.e. a time-independent
quantity. Thus, invariance of the laws of physics under spatial displacements
implies conserved momentum; time-translational invariance implies conserved
energy; and invariance under phase changes of the wave functions of charged
particles implies conserved electric charge. As the states of a physical system can
be uniquely identified by the values of its complete set of conserved quantities,
much of the effort in the study of a quantum system is to identify the symmetry
transformations that leave the system invariant (or nearly invariant), and to find
the irreducible representations of these symmetry groups. Symmetry, invariance,
and conservation are closely related concepts.
In quantum physics, a state is represented by a vector in a Hilbert space H .
Given a symmetry group G of the physical system and T a linear operator rep-
resenting any element of G, for every allowed state ψ of the system the vector
Tψ is also an allowed state, and so also is ψ + Tψ, by the superposition princi-
ple. It follows that a basis of an irreducible representation of G can be formed by
appropriate linear combinations ∑T Tψ ; conversely, allowed physical states are
expressible in terms of the basis vectors of irreducible representations. Given this
relationship, the tools of group theory can be applied to uncover properties of
a symmetric system, as for example in rotational-invariant atoms, where energy
spectra, branching rules, and selection rules can be inferred with confidence.
Furthermore, as probabilities of events, rather than vectors themselves, are
measurable, the requirement of conservation of the inner product itself would be
unnecessarily stringent; instead it suffices to have the equality of probabilities for
equivalent events in two reference frames: |h φ |ψi|2 = |h φ0 | ψ0 i|2 for all vectors
ψ, φ in H . The conservation of the probability implies that, up to an arbitrary
constant phase factor, the mapping ρ : ψ → ψ0 must be either linear and unitary, or
antilinear and antiunitary.
An operator U : ψ → ψ0 = Uψ is said to be linear and unitary if it satisfies the
conditions U | αψ + βφ i = αU | ψi + βU |φ i and hUφ |Uψi = h φ | ψi; and an operator
T : ψ → ψ0 = Tψ is said to be antilinear and antiunitary if it satisfies the conditions
T | αψ + βφ i = α∗ T | ψ i + β∗ T | φ i and h Tφ | Tψi = h φ | ψi∗ = h ψ | φi.
E XAMPLE 81: An example of an antilinear antiunitary operator is the time rever-
sal operator T : t → − t; it effectively reverses the motion (the linear and angu-
lar momentum) leaving the position unchanged. An operator O transforms as
O 0 = T O T −1, so that a matrix element such as h ψ (t)|O|φ(t)i becomes under
time reversal h Tψ (t)|O 0| Tφ (t)i = h ψ (− t)|O|φ(− t)i∗.
The Lie algebra g ∼= u (H) of a unitary group G ∼
= U (H) consists of all the lin-
ear operators on Hilbert space H that are skew-Hermitian (A† = − A). As usual,
to each A ∈ g is associated a one-parameter (unitary) group U A ( t ) = exp ( t A)
such that A = limt→0 [(U A( t ) − 1) /t ]. Physicists generally prefer, by a simple
118 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
for any two elements g, h of a group and some real angle ϑ ( g, h). Accordingly, the
general bracket relations are modified by adding an extra multiple of the identity:
Xi , Xj = ckij Xk + bij I. (3.47)
If the constants bij cannot be eliminated by redefining the Xi ’s or the phases ϑ’s,
then they must have physical significance and be kept as part of the algebra. The
presence of such irreducible constants (in the algebra) or phases (in the group)
implies the existence of a superselection rule that all the states of a physical system
must obey (which forbids, for example, the superposition of physical states of
different masses).
Non-relativistic quantum physics is invariant under the isometries in R 3 × R 1
— space and time displacements, space rotations, and transformations between
uniformly moving frames (called boost) — which constitute the Galilean group.
Their effects on the position x and time t are: x → x0 = Rx + a + vt, and t → t0 =
t + τ, where R is a rotation, a a space displacement, τ a time displacement, and
v the constant velocity of a moving coordinate frame. Their group representation
in H is given by the operators on H :
The ten parameters θi , ai, vi , and τ are the (real) parameters of the symmetry
transformations generated by the corresponding skew-Hermitian operators R i,
Ti, Bi , and D. In the physicist’s convention, one would rather take the Hermitian
operators Ji = iR i, Pi = iTi, K i = − iBi, and H = − iD. These operators obey the
Lie product rule, as listed in the following multiplication table (where each entry
at row i and column j shows the value of the bracket [ Xi , Xj ]):
Jj Kj Pj H
Ji ieijk Jk ieijk K k ieijk Pk 0
Ki ieijk K k 0 iδij λI iPi
Pi ieijk Pk − iδijλI 0 0
H 0 − iPj 0 0
3.6. REPRESENTATIONS 119
From this table, a few points to note: (1) The operators K i do not commute
with the time displacement operator H, and so are not conserved; but both Ji
and Pi do commute with H, and so have conserved eigenvalues. (2) There exist
several subalgebras, and an irreducible constant (λ in [ Pi, K j ]) both indicative that
Galilean symmetry might not be at the fundamental level of nature symmetry. To
see how the constant λ arises, let us make in succession a translation x → x + a,
then a boost x → x + vt in the x-direction, resulting in x → x + vt + a. This
corresponds in operators on Hilbert space:
exp (ivK x) exp(− iaPx) = exp(iλav/2) exp i( vK x − aPx) ,
with the presence of a phase factor.
In relativity, space and time are unified as Minkowski spacetime R 3+1 and the
symmetry group of relativistic quantum mechanics is elevated to the Poincaré
group P(3; 1). It is a ten-dimensional non-abelian non-compact Lie group of all
isometries in R 3+1 generated by the spacetime translation four-vector Pµ , and
the rotation-boost antisymmetric tensor Mµ,ν (where µ, ν = 1, 2, 3, 0), obeying the
following bracket rules:
i[ Mµν , Mλκ ] = gνλ Mµκ − gµλ Mνκ − gκµ Mλν + gκν Mλµ , (3.48)
i[ Pµ, Mλκ ] = gµλ Pκ − gµκ Pλ , (3.49)
[ Pµ, Pν ] = 0. (3.50)
P(3; 1) is the semiproduct of the abelian subgroup of translation h Pµ i and the
noncompact nonabelian Lorentz subgroup SO (3; 1) also known as the homoge-
neous Lorentz group. The identity component of the latter, called SO∗ (3; 1), is
the connected and simply connected group of the ordinary spatial rotations Mij
and Lorentz boost Mi0 , and can be generated by exponentiation of h Mij, Mi0 i.
To see how the Poincaré group compares with the Galilean group, let us define
Pi = Pi , H = P0, Ji = M23, M31, M12 , and K i = M10, M20, M30 . Then the 10
generators of the Poincaré algebra have the Lie products listed below:
Jj Kj Pj H
Ji ieijk Jk ieijk K k ieijk Pk 0
Ki ieijk K k − ieijk Jk iδij H iPi
Pi ieijk Pk − iδij H 0 0
H 0 − iPj 0 0
Now, call M the scalar operator giving the total mass of the system, so that the
total energy operator (or Hamiltonian) H is H = M + H 0, where H 0 is the kinetic
plus potential energy (setting the speed of light c = 1). Let v be the typical veloc-
ity and m the typical mass, then, in the limit v 1, we get H = O( m ), Ji = O(1),
Pi = O( v ), and K i = O( v−1), and we recover the Galilean algebra provided we
identify λI with M. This irreducible constant is just a byproduct of the reduction
of Poincaré symmetry to a lower level.
The three non-compact groups just mentioned, Galileo, Poincaré, and Lorentz,
act on infinite-dimensional Hilbert spaces, and admit infinite-dimensional repre-
sentations; and so will not be considered further in the rest of this book.
120 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
Problems
3.1 (Rotations, O(2), SO(2)) A rotation R ( θ ) in the plane R 2 about the origin
through an angle θ, is defined by a real matrix R ij , such that
i j cos θ − sin θ
R ( θ )e = e R ji ( θ ) , R( θ ) = ,
sin θ cos θ
" # 0 a b c
0 a b
0 a 0 0 d e
A= , B= 0 0 c , C=
0 0 0 0 0 f
0 0 0
0 0 0 0
3.13 (Is given Y equal eX with real X?) Are the following (real) matrices of the
form eX ? If so, compute X.
1 0 0 1 a 0 0 −1 1 1 1 1
, , , , , ,
0 0 1 0 0 b 1 0 0 1 −1 1
3.14 (Dimension-three algebras) Find the generators of rotation acting on (a) three
dimensional Euclidean space of metric g = diag[1, 1, 1]; and (b) three dimensional
Lorentzian space of metric g = diag[1, 1, − 1].
3.15 (Exponential of rotation) (a) Referring to Problem 3.1, write the rotation ma-
trix for a very small angle δθ as R( δθ ) = I + δθR: sum the expansion series that
defines exp ( θR ). (b) When the rotation is applied on the Cartesian coordinates
x1 , x2 of a two-dimensional Euclidean space, we can define the corresponding
Lie group of rotations, generated by a differential operator X. Find X and calcu-
late ϕX ( θ ) = exp (− θX ) (the minus sign in exp (− θX ) arises because the operator
is applied on the coordinates, not the basis vectors)
3.16 (Lorentz boost) (a) Referring to the relevant discussion in the chapter, show
that for any one-dimensional Lie group, with the composition function φ ( β, α ),
one can find a new parameter α0 such that the composition function is additive,
i.e. φ ( β0, α0 ) = β0 + α0 (i.e. one-parameter Lie groups are abelian).
(b) In a two-dimensional space with metric g = diag (1, − 1), define the Lorentz
boost L ( v) along the x axis by the matrix
p
γ γv
L ( v) = where 0 ≤ v < 1, γ = 1/ 1 − v2.
γv γ
(i) Show that L ( v 2) L ( v1) = L ( φ (v2, v1 )), with φ ( v2, v1 ) to be found. (ii) Find the
new parameter ω ( v), such that L̄( ω2) L̄( ω1) = L̄ ( ω2 + ω1). (iii) Find the group
generator X such that the exponential map exp ( ωX ) reproduces L̄ ( ω ).
3.17 (Heisenberg group) Define the sets of matrices:
(" # ) (" # )
1 a b 0 α β
H= 0 1 c ; a, b, c ∈ R ; h= 0 0 γ ; α, β, γ ∈ R .
0 0 1 0 0 0
(a) Show that H is a Lie group, and find its center. (b) Show that for any X ∈ h,
the exponential eX ∈ H, and conversely if X is any matrix such that etX ∈ H, then
122 CHAPTER 3. LIE GROUPS & LIE ALGEBRAS
X = detX /dt |t=0 is in h. Define a basis for h, together with its algebra. (c) Find the
center C (h) of h. Using the BCH formula, calculate exp X for an arbitrary X ∈ h.
3.18 (Euclidean group E(2)) In classical physics, space is assumed to be homo-
geneous and isotropic, so that physical phenomena should not depend on the
specific location or orientation of the physical system. In mathematical terms,
this means a physical Euclidean space, with its symmetries described by the
Euclidean group, which consists of uniform translations and uniform rotations.
Take, as an example, the Euclidean group E (2) defined by the following transfor-
mations of the Cartesian coordinates x1 , x2 in R 2:
x01 = x1 cos θ − x2 sin θ + α1; x02 = x1 sin θ + x2 cos θ + α2,
in terms of a displacement distance (− ∞ < αi < ∞; i = 1, 2) and a rotation
angle (0 ≤ θ ≤ 2π). We call A ( θ, α) the operator effecting this transformation:
A ( θ, α) xi = x0i = R ij ( θ ) x j + αi .
(a) Verify that A ( θ, α) form a Lie group, and that A ( θ, 0) and A (0, α) are its
subgroups. Calculate the transformation functions as well as the composition
functions of E (2). (b) Calculate the generators of E (2) and their brackets.
3.19 (Linearization of E(2)) Let L be the linear space of (column) vectors of the
form ( x1, x2 , 1)T , where x1 , x2 are the Cartesian coordinates in two-dimensional
Euclidean space. The object of this problem is to show that E(2) is isomorphic
to a matrix Lie group, by showing that E(2) group and algebra elements (Prob-
lem 3.18) are expressible as 3 × 3 matrices in space L, denoted M(3, R ), with
appropriate properties.
(a) Find the generators of E(2) and their brackets as elements of M(3, R ).
(b) Prove that A ( θ, α) = A (0, α) A ( θ, 0). (c) Show that the exponential series
of exp( θX0) and exp( α1 X1 + α2 X2 ) sum to A ( θ, 0) and A (0, α) respectively. (d)
What is A ( θ, 0) A(0, α)? Compare it to A (0, α) A (θ, 0). (e) In question (b), we have
proved that exp ( α1X1 + α2 X2 ) exp( θX0) is identical to the group element A ( θ, α).
But the theory tells us that, in general, it is the map exp( θX0 + α1 X1 + α2 X2 ) that
should reproduce the group elements. How do you explain the difference?
3.20 (Invariant subgroup and factor group of E(2)) With the notations and results
in Problem 3.19, answer the following questions:
(a) Prove the relation eθX0 T( α) e−θX0 = T[ R (θ )α]. From this result, discuss the
meaning of the commutation relations between X0 and X1 , X2 . (b) Prove that T =
{ A (0, α)} is the invariant subgroup of E(2), and that the factor group E(2) /T is
isomorphic to SO(2). (c) Is E(2) simple, semi-simple, compact?
sl(2, C )
and Associated Groups
This chapter is devoted to a study of the Lie algebra sl (2, C ) and the Lie groups
SO (3) and SU (2). As sl (2, C ) is isomorphic to so (3)C and su (2)C, studying one
is studying at the same time the others. Further, as sl (2, C ) turns out to be a
Lie subalgebra of all simple and semi-simple Lie algebras, this simplest of all Lie
algebras is a model for and a part of the study of the more complex cases. It also
has importance in its own right, in particular in physics, where it underlies the
concepts of angular momentum, spin, and isospin.
We will discuss first sl (2, C ) and classify all of its representations. We next
consider the groups SO(3) and SU (2), and their representations, and the direct-
product representations, which occur frequently in studies of physical systems.
[ h, e ] = 2e, [ h, f ] = − 2 f , [ e, f ] = h. (4.1)
123
124 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
2. sl (2, C ) is a simple Lie algebra, because the only nonzero ideal it has is itself
(having its derived algebra, D g = [g, g], equal to itself, D g = g; cf. Sec. 3.5.2).
Being a simple Lie algebra, sl (2, C ) may be considered a basic building block of
the higher-dimensional Lie algebras.
3. sl (2, C ) is evidently the complex extension of sl (2, R ), and isomorphic to the
complexification su (2)C of the real Lie algebra su (2) (and so also to so(3)C, the
complexification of so(3)), under the correspondence h 7 → 2iS3, e 7 → iS1 − S2,
f 7 → iS1 + S2, where S j = − iσj /2, as defined in Chapter 3, Examples 45 and 48.
The close relationship we see here among these algebras arises from the fact that
Lie algebra is a local theory.
4. Which Lie groups G admit sl (2, C ) as Lie algebra? An answer is the connected
and simply connected group G = SL (2, C ) in a one-to-one correspondence with
sl (2, C ); but it is not the only one, because as SL (2, C ) has as center the subgroup
{ I, − I } of scalar matrices, one may also have G = SL (2, C )/{ I, − I }.
For a similar reason, the Lie groups having su (2) as Lie algebra include, not
only the connected, simply connected group SU (2) which is in a one-to-one corre-
spondence with su (2), but also the quotient SU (2)/ { I, − I } (since SU (2) contains
the invariant discrete subgroup { I, − I }).
The simple Lie algebras su (2)C and so(3)C are mutually isomorphic, but it is
not expected that their associated Lie groups to be so, because whereas SU (2) is
simply connected, SO (3) is not. In fact SO (3) ∼ = SU (2)/{ I, − I } (see Example 64
in Chapter 3; we will elaborate on this point later on).
5. Given the close relationships among all the simple Lie algebras of dimension
three, on the one hand, and between each of them and its associated Lie groups,
on the other hand, when we study any one of them, for example sl (2, C ) as we
choose to do here, we are in effect studying also all the others.
4.2 Representations
As one may recall, a representation ( π, V ) of a Lie algebra g is a vector space V together
with a morphism π : g → gl (V ), or equivalently a linear action of g on V by mapping
g × V → V , which assigns to each element x of g an operator π ( x) : V → V ,
preserving both linearity and the product laws. So, if H, E, F are linear operators
on a vector space V obeying the bracket relations
Hv = µv, v ∈ V, (4.3)
HEv = ( µ + 2) Ev ,
(4.4)
HFv = ( µ − 2) Fv .
Evk has the same eigenvalue as vk−1 , and so must be proportional to it, which
we write as Evk = ck vk−1 . In particular c0 = 0 and c1 = λ (this is so because
c1 v0 = Ev1 = EFv0 = [ E, F ]v0 = λv0).
With the commutation relations (4.2) we do the following calculations:
R ik R j` δk` = δij ,
(4.15)
R i` R jm R kn e`mn = eijk .
and the angle ω through which the (counterclockwise) rotation is made. Since
n can be defined by two angles (the polar and azimuthal angles θ and ϕ), every
rotation, denoted R ( ω, n), is specified by three parameters ω, θ, ϕ with values in
the ranges
0 ≤ ω ≤ π, 0 ≤ θ ≤ π, 0 ≤ ϕ < 2π. (4.16)
The limit ω = 0 corresponds to the group identity element R (0, n) ≡ I. The
values of the parameters in their ranges of definition give a unique assignment of
the group elements, except for ω = π, when we have the redundancy
R ( π, n) = R ( π, −n). (4.17)
From geometric considerations, we also expect that, when ω exceeds its normal
limits, R must satisfy the following global conditions:
The group parameter space consists of all points inside and on a sphere of
radius π about the origin. To each vector ω = ωn drawn from the origin there
corresponds a (counterclockwise) rotation around the vector ω through an angle
equal to the distance ω from the origin. This assignment is unique, except on the
surface ω = π, where two diametrically opposite points give the same rotation
(cf. (4.17)). Thus, SO (3) as space is a compact sphere.
Rotations about the coordinate axes ei : When n lies along a coordinate axis ei , the
rotation through an angle ψ will be given by one of the matrices:
1 0 0
R (ψ, e1 ) = 0 cos ψ − sin ψ , (4.19a)
0 sin ψ cos ψ
cos ψ 0 sin ψ
R (ψ, e2 ) = 0 1 0 , (4.19b)
− sin ψ cos ψ
cos ψ − sin ψ 0
R (ψ, e3 ) = sin ψ cos ψ 0 . (4.19c)
0 0 1
We have fixed the sense of rotation, and hence the signs of the entries, according
to the right-hand rule.
General rotation ( ω, n): Given a rotation R ( ω, ei ), we can obtain the rotation
R ( ω, n) for any n by rotating ei to n with an operation S, such that n = Sei.
E XAMPLE 3: Using R ( ω, ei ) given in Eq. (4.19), we calculate R ( ϕ, e3 ) R (θ, e2 ), now
called S, and find its inverse S−1:
cφ cθ − sφ cφ sθ cφ cθ sφ cθ − sθ
S = sφ cθ cφ sφ sθ , S −1 = − s φ cφ 0 , (4.20)
− sθ 0 cθ cφ sθ sφ sθ cθ
132 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
All matrices for rotations through the same angle are related by similarity trans-
formations that rotate the axis of one rotation into the axes of the others. So all
rotations through a given angle ω belong to a conjugacy class of SO (3); they cor-
respond to points on a spherical shell of radius ω in the parameter space.
The Euler angles. A rotation R in three-dimensional space can also be specified
by the relative orientation of two Cartesian systems of coordinates, say, (e1 , e2 , e3)
and (e01 , e02 , e03 ). Three successive (counterclockwise) rotations bring (e1, e2 , e3 ) to
(X, Y, Z), then to (L, M, N), and finally to (e01 , e02 , e03 ), in detail as follows:
(a) Rotate about e3 through α, for any value of α between 0 and 2π. This brings
e 7 → OX, e2 7 → OY, leaving e3 ≡ OZ unchanged.
1
In this form, we see that R ( α, β, γ) results from three successive operations: (a)
First, rotate about e3 through γ; (b) Then rotate about e2 through β; (c) Finally
rotate about e3 through α. The angles α, β, γ, called the Euler angles, have the
ranges already specified: 0 ≤ α, γ < 2π and 0 ≤ β ≤ π. This expression is very
convenient because it gives an arbitrary rotation as a product of rotations about
fixed axes, which are given by simple expressions, as in (4.19).
C OMMENTS .
1. The assignment of the Euler angles to rotations is not always unique. Thus,
R ( α, 0, γ) is function only of the sum α + γ, and R ( α, π, γ) only of the difference
α − γ, as can be seen from (4.24).
4.4. PARAMETERS OF ROTATION 133
2. As the expressions (4.21) and (4.24) for rotations in the two parameterizations
are well defined, we can find the relations between the variables α, β, γ and ω, θ, ϕ
for any given rotation:
ϕ = ( π + α − γ ) /2 ,
tan( β/2)
tan θ = , (4.25)
sin[( α + γ ) /2]
β α+γ
cos ω = 2 cos2 cos2 −1 .
2 2
SO (3) near the identity. Take any element R ∈ SO (3), and let it approach the
identity, so that it may take a linearized form, R ≈ I + M, in accord with the
condition det R = 1 with all entries in M infinitesimal. Requiring R T R = I means
that MT = − M, and so M has the general form:
0 − θc θb
M = θc 0 − θa , (4.26)
− θb θa 0
in terms of three real variables θa , θb , θc . The signs of the matrix elements have
been chosen so that in the appropriate limits they are in accord with the right-
hand rule we have been using. For example, when θb = θc = 0, (4.26) agrees with
R ( ω1, e1 ) − I in (4.19a), for small ω1 provided θa and ω1 are identified. For the
same reason, θb and θc are identified with the angles of rotation about the axes e2
and e3 , i.e. θb ≡ ω2 and θc ≡ ω3.
Therefore, R ∈ SO (3) near the identity can be written as
R = I + ( ω1R1 + ω2 R2 + ω3 R3) , (4.27)
where R 1 = R x, R 2 = R y, and R 3 = R z are the 3 × 3 matrices defined in the
previous chapter, Sec. 3.4. We know that they are a basis for the real Lie algebra
so (3), which upon complexification is isomorphic to Lie algebra sl (2, C ) by the
maps 2iR 1 7 → F + E, 2R 2 7 → F − E, and 2iR 3 7 → H.
The self-adjoint objects Li = iR i are more commonly used in physics, where
they represent the orbital angular momentum:
0 0 0 0 0 i 0 −i 0
L1 = 0 0 − i , L2 = 0 0 0 , L3 = i 0 0 . (4.28)
0 i 0 −i 0 0 0 0 0
These matrices can be succinctly written as ( Lk )`m = − iek`m , where ek`m is the
totally anti-symmetric unit tensor of rank 3. They have several properties of in-
terest. First let us rewrite (4.15) as R i` R jm e`ms = eijk R ks , and substitute in it eijk
for i( Li ) jk to obtain
R i` (iLn )`m R jm = i( Lk )ij R kn , (4.29)
which is equivalent to (recalling that R T = R −1)
RL j R −1 = Lk R kj ( j = 1, 2, 3). (4.30)
134 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
Comparing this relation with Rej = ek R kj , we see that each L j behaves as a three-
dimensional vector under the adjoint action of R. Now, when R approaches the
identity, (4.30) becomes
( I − iωi Li ) L j ( I + iω` L` ) = Lk ( I − ωi Li )kj . (4.31)
As ( Lk )`m = − iek`m , we have the commutation relations:
[ Li, L j ] = ieijk Lk (with L†i = Li ) . (4.32)
These relations derived in a particular (the standard) representation of so (3)C will
retain their form in any other representation, as well as in abstract algebra.
Conversely, for every element ∑ ωi Li of so (3)C, the mapping exp · ∑ ωi Li pro-
duces an element of the rotation group. In particular, the series for exp(− iωLi)
can be summed, and the matrices (4.19) recovered as exp(− iωL1) = R ( ω, e1 ),
exp(− iωL2) = R ( ω, e2 ), and exp(− iωL3) = R ( ω, e3 ).
E XAMPLE 4: Calculation of exp(− iωL3). Let ψ = − iω and E0 = diag[0, 0, 1].
Note that L23 = I − E0 = diagonal[1, 1, 0] and L33 = L3. The series for exp(− iωL3)
can be separated into odd and even powers, and summed separately:
ψ3 ψ2 ψ4
exp (− iωL3) = E0 + L3 ψ + + · · · + L23 1 + + +···
3! 2! 4!
= E0 + L3 sinh ψ + L23 cosh ψ
= E0 − iL3 sin ω + L23 cos ω = R ( ω, e3 ) .
So, one recovers (4.19c).
Given these results, the simplest way to show that all elements of SO (3) are
of the form exp (− i ∑ ωi Li ) is to start from (4.30), that is from SL3 S−1 = Lk Sk3 ,
where S ( θ, ϕ ) is (cf. Example 3) the orthogonal transformation that rotates e3 to
the direction n specified by its angles θ and ϕ,
n = S ( θ, ϕ )e3 = ej S j3 ( θ, ϕ ) . (4.33)
There are three free parameters, making SU (2) a three-parameter group. The
two complex numbers a and b are also referred to as the Cayley–Klein parameters.
If each element of SU (2) is denoted by ( a, b ), then the identity element of SU (2)
is (1, 0) ≡ I, the inverse of ( a, b ) is ( a∗, − b ), and the product rule, obtained by the
usual matrix multiplication, is ( a, b ) · ( c, d ) = ( ac − bd ∗ , ad + bc∗ ).
When only unitarity is required, the group is enlarged to the unitary group
U (2) consisting of elements of the form
0 a b
A =e iλ
, | a |2 + | b |2 = 1, λ∗ = λ . (4.37)
− b ∗ a∗
If a and b are fixed, U (2) is reduced to the group of unimodular complex numbers
z, with | z | = 1, having the structure of U (1). On the other hand, if λ is fixed, at 0
for example, the group is reduced to SU (2). The map f : U (1) × SU (2) → U (2)
is a homomorphism that induces the isomorphism (U (1) × SU (2))/Z ∼ = U (2),
where (U (1), SU (2)) 3 Z = {(1, I ), (−1, − I )} ∼= 2
Z is Ker f .
The condition | a |2 + | b |2 = 1 makes it clear that U (2) and SU (2) are com-
pact groups, because, as spaces, they are equivalent to S3 within C2 ∼ = R4. U (2)
preserves the inner product of any two vectors in C2 on which the group acts; it
follows that the relative angle and relative phase of any two such vectors, and
their norm are all preserved. This invariance does not depend on the value of the
determinant of the group element; but if det A = 1 is also satisfied, the overall
phase is the same for all vectors, that is ψ0 = Aψ for all ψ.
SU ( 2) near the identity. Consider an element A ∈ SU (2) near the identity: A ≈
I − iM, where all entries of the matrix M are small. (Note the explicit imaginary
unit i, and A ≈ I, implementing the condition det A = 1.) Then, the unitarity
condition A† A = I means M† = M, or Mij∗ = Mji in any basis. The general form
of M subject to this requirement and consistent with (4.36) is
1 ω3 ω1 − iω2
M= , (4.38)
2 ω1 + iω2 − ω3
where ωi are real and infinitesimal, and the overall numerical factor is conven-
tional. It may be written as M = 21 ∑3j=1 ω j σj , where σj are the self-adjoint Pauli
matrices defined by the relations
(with I = diagonal[1, 1]). The standard form of the σi ’s (with a diagonal σ3) is:
0 1 0 −i 1 0
σ1 = , σ2 = , σ3 = , (4.41)
1 0 i 0 0 −1
On the abstract level, the Ji ’s defined by (4.42) are the basis of the Lie algebra
su (2)C, which is manifestly isomorphic with so (3)C defined in (4.32) through the
correspondence Ji 7 → Li .
Exponential mapping. For every element X = ∑ ωi Ji of su (2), the map exp · X
produces an element of SU (2). In the standard representation, the Ji ’s are evi-
dently cyclic (in particular, ( n · σ )2 = I with a unit vector n), and so exp X has a
closed form, as already seen in Chapter 3. This result can be derived in a differ-
ent way, similar to that of (4.34) making use of the fact that SU (2) transformations
may be regarded as rotations in C2 (see also below). By this, we mean that there
exists an operator T in SU (2) whose adjoint action on exp( ω J3) gives exp( ωnk Jk ),
or equivalently rotates σ3 to a certain direction n, such that
Tσ3 T −1 = σk nk = σ1 n1 + σ2 n2 + σ3 n3 , (4.43)
where the nk ’s are the components of the unit vector n (defined as in Example 3
in terms of its polar and azimuthal angles θ and φ). We find
cos( θ/2) − e−iφ sin( θ/2)
T = iφ , T −1 = T † . (4.44)
e sin( θ/2) cos ( θ/2)
As σ3 is diagonal given by diag [1, − 1], the matrix exp(− iωσ3/2) is also diagonal,
given by diag [e−iω/2 , eiω/2 ]. Matrix multiplication gives us
ω −1 ω ω
T exp − i σ3 T = cos − i n · σ sin , (4.45)
2 2 2
which is just the 2 × 2 matrix U ( ω, n) found in the last chapter. It gives the stan-
dard representation of any element of SU (2) in the angle and axis parameters.
For reference, note also the following properties:
U ( ω, n) U ( ω0, n ) = U ( ω + ω 0, n ) ;
U (4π − ω, − n) = − U (2π − ω, − n) = U ( ω, n) ; (4.46)
U (2π, n) = − I ; U (4π, n) = I .
U ( ω, n) = exp (− iω n · J ) , (4.47)
4.4. PARAMETERS OF ROTATION 137
with Ji obeying (4.42), and the group parameters having the ranges of values
0 ≤ ω < 2π, 0 ≤ θ ≤ π, 0 ≤ φ < 2π. The equivalent formula in terms of the
Euler angles is analogous to (4.35):
SO (3) versus SU ( 2). We have mentioned before that SU (2) acts on C2 in a way
similar to SO (3) rotating in R 3. To examine this relationship in more detail, let
x1 , x2 , x3 be the Cartesian components of a vector x ∈ R 3 . Under SO (3), they
transform as
xi 7 → x0i = R ij xj , R ij R ik = δjk . (4.50)
Tσj T −1 = σi R ij , (4.51)
a result which generalizes (4.43). This equation says that to each SO (3) rotation
R, there correspond two equivalent operations, ± T, in SU (2). Hence, SO (3) is a
two-to-one homomorphic image of SU (2). This implies in particular that not all the
representations of SU (2) are representations of SO (3), but all the representations
of SO (3) must necessarily be present among those of SU (2).
To elaborate further, let us work with the angle-axis (ω; n) parameterization.
As we recall, the group SO (3) is compact, and its three-parameter space consists
of all points ωn for which 0 ≤ ω ≤ π, i.e. it is a spherical ball of radius π centered
at the point associated with the identity, and each pair of antipodal points on the
spherical surface (of radius π) are identified with the same group element. On
the other hand, the three-dimensional manifold underlying SU (2) consists of all
points ωn for which 0 ≤ ω ≤ 2π, i.e. it is a spherical ball of radius 2π. The group
operations are in one-to-one correspondence with elements of the group within a
138 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
radius 2π from the origin. As U (2π, n) = − U (0, n) = − I, all points on the surface
of the spheroid must be identified with the group element − I.
It is clear that any two points in the parameter sphere of SU (2) (or equiva-
lently two group elements) can be connected by a line lying completely within
the parameter space. So SU (2) is connected. In addition, any closed loop in the
spherical ball beginning and ending (for example) at the center of the sphere can
always be continuously deformed to the origin, whether the path touches the
sphere surface or not. Hence, SU (2) is simply connected. This is to be contrasted
with the situation in SO(3): Here, if a closed path beginning at the origin cuts
the sphere surface (with the intersection identified with the antipodal point) and
returns to the origin, it cannot be continuously deformed to a point at the ori-
gin without breaking the loop, and so SO (3) is not simply connected, it is doubly
connected, but possesses no other multiple-connectedness.
In summary, SO (3) and SU (2) are simple, compact, connected three-parameter Lie
groups. The group SO (3) is doubly connected, but the group SU (2) is simply connected,
and so provides the universal cover for all groups with the Lie algebra defined by the
bracket relations (4.42).
J3 | jm i = | jm i m , m = − j, − j + 1, . . . , j − 1, j ;
(4.52)
J± | jm i = | j m ± 1i [( j ∓ m )( j ± m + 1)]1/2 .
j− m
For reference, note that | jm i = a j−m J− | jj i, where ak = [(2j − k ) !/ (2j )!k! ]1/2.
The expression (4.52) involves a specific choice of phase, referred to in physics
literature as the Condon–Shortley convention. One can check that J12 + J22 + J32 is
the Casimir operator, with invariant value j ( j + 1) on V (2j) .
4.5. ROTATION MATRICES 139
j 1
ψm =
Nm ∑ ξ| ⊗ ·{z· · ⊗ ξ} ⊗ η| ⊗ ·{z· · ⊗ η} , (4.54)
P
j+ m j− m
where one sums over all permutations of the 2j factors in the summand (produc-
2 terms). This expression for the ψ j ’s, as a sum of tensor products, shows
ing Nm m
j j0
that they are orthonormalized vectors in (C2 )×n space, so that ψm † ψm0 = δjj0 δmm0 .
E XAMPLE 5: We label the rows and the columns in the order m = j, j − 1, . . . , − j,
and use a simplified notation in which Ji stands for π j [ Ji ].
j = 1/2: The standard representation. Order of rows and columns: ( 12 , − 12 ).
1/2 0 0 1 0 0
J3 = , J+ = , J− = ,
0 − 1/2 0 0 1 0
1 0 1/2 −i 0 − i/2
J1 = ( J+ + J− ) = , J2 = ( J+ − J− ) = .
2 1/2 0 2 i/2 0
Note that both J1 and J3 are real, but J2 = ( J+ − J− ) /2i is pure imaginary, as gen-
erally expected. These matrices are equivalent, up to a similarity transformation,
to the defining matrices (4.28) of SO (3).
C OMMENTS . The ideas about SU (2) we are presenting here can be nicely illus-
trated by the rotational symmetries of charge space (isospin T) and coordinate
space (spin, angular momentum J). In the first case, nuclear isospin invariance
requires that each nuclear energy level in a light nucleus be part of a multiplet
of energy levels in 2T + 1 nuclei with the same mass number but with (charge)
T3 varying from − T to + T. In the second case, rotational invariance manifests
itself in energy sequences J = 0, 1, 2, 3, . . . observed in any deformed nucleus or
diatomic molecule.
140 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
Then, the representation matrix D j for the unitary operator U [ R (α, β, γ )] in the
canonical orthonormal basis {| jmi} of space V (2j) is defined by
j
U [ R (α, β, γ)]| jm i = | jm0i Dm0m [ R (α, β, γ)], (4.56)
j
Dm0 m [ R (α, β, γ )] = h jm0|U [ R (α, β, γ)]| jm i
0 j
= e−iαm dm0 m ( β )e−iγm , (4.57)
j
dm0 m ( β ) = h jm0|e−iβJ2 | jm i .
j j
For simplicity, we will write Dm0 m ( α, β, γ ) instead of Dm0 m [ R (α, β, γ )], and, as
usual, summed over repeated indices. With a purely imaginary J2 = −i/2( E − F )
and the Condon–Shortley convention for the | jm i’s (cf. (4.52)), all matrix elements
j
dm0 m ( β ) are real-valued. It is our main task to evaluate this object.
Some general facts can be deduced from (4.57): As the Ji ’s are self-adjoint,
the operator U [ R ] defined in (4.55) is unitary. Given a unitary U [ R ] and an or-
thonormal basis for V (2j) , the representation matrix D j [ R ] is unitary, D j† D j = I,
so that
j† j j
Dm0 m ( α, β, γ ) = Dm0 m ( α, β, γ )−1 = Dm0 m (− γ, − β, −α) . (4.58)
j j j
Together with the reality of d m0 m ( β ) it also follows that d m0 m ( β ) = dmm0 (− β).
j
Since R (0, n) is the group identity element, Dm0 m [ R (0, n)] = δm0 m . In addition,
from (4.57) we have D j (2π, β = 0, γ = 0) = D j [ R (2π, e3 )], and so
j
Dm0 m [ R (2π, e3 )] = e−i2πm δm0 m = (− 1)2m δm0m = (− 1)2jδm0 m . (4.59)
j
sin[(2j + 1) ω/2]
χ j ( ω ) = Tr D j [ R (ω, e3 )] = ∑ e−imω = . (4.60)
m=− j
sin( ω/2)
The group character χ j ( ω ) carries two labels, j for the representation and ω for
the class. As the trace is basis independent, this result (4.60) holds true in any
basis. The character of the (single-valued) j-representation of SO (3) corresponds
to integral values of j.
C OMMENTS . By definition, any element R in SU (2) or in SO (3) must be such
that det U [ R ] = 1. We can directly check that this property is satisfied in the
irreducible representation (4.52), and so also in any other representation. Recall-
ing that det Π j (ex ) = exp(Tr π j ( x)) for all x ∈ g, we see that det U [ R ] = 1 iff
Tr π ( x) = 0, a condition clearly satisfied, because π j ( x) = ω1 J1 + ω2 J2 + ω3 J3
and because every Ji is traceless in any π j .
E XAMPLE 7: With the full matrix D j given by (4.57) in the canonical basis, we just
need to evaluate its non-trivial part, d j , by summing the series involving powers
of π j ( J2), except for the case j = 0, which is trivial: d 0( β ) = 1.
j = 1/2 : The power series for the exponential e−iβJ2 , with J2 = σ2 /2 given in
Example 5, sums up to be
In particular, one has d 1/2 (2π ) = − I for β = 2π, and d 1/2 (4π ) = I for β = 4π.
The matrix for the rotation through β about e2 is D1/2 [ R ( β, e2 )] = d1/2 ( β ), from
which one can obtain the matrix for the rotation through β about any axis n by
an appropriate transformation: D1/2 [ R ( β, n)] = D [ S ] D [ R (β, e2 )] D [S]−1. Now,
for a full revolution β = 2π, one would expect R (2π, n) = R (0, n), but, instead,
one finds D1/2 [ R (2π, n)] = D [ S ] D [ R(2π, e2 )] D [S]−1 = − I, although one has of
course D1/2 [ R (0, n)] = I: Rotations through one complete revolution (or an odd
number of revolutions) are mapped to − I (rather than I); and two (or an even
number of) revolutions are mapped to I. This is an example of the situations
discussed before (page 140).
The group character of π 1/2 can be evaluated from D1/2 ( α, β, γ ):
χ1/2 = cos ( β/2) e−i(α+γ)/2 + e+i(α+γ)/2
= 2 cos ( β/2) cos[( α + γ ) /2] = 2 cos( ω/2) .
142 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
To obtain the result in the last line, we used the equivalence relations between the
two sets of parameters. This agrees with the general formula (4.60) for j = 1/2.
E XAMPLE 8: j = 1 : Using the representation of J2 given in Example 6, with the
cyclic property J23 = J2, we sum the exponential series of e−iβJ2 to get
√
√c2 − 2cs s2
√ β β
d1( β ) = 2cs c√
2 − s2 − 2cs ( c = cos , s = sin )
2 2 2
s 2cs c2
(rows and columns in the order + 1, 0, − 1). In particular, one has d 1 (2π ) = I and
D1 [ R (2π, n)] = I, leading to the conclusion that j = 1 corresponds to a faithful
(single-valued) representation of the group SO (3). The group character is
χ1 = c2 e−i(α+γ) + e+i(α+γ) + c2 − s2
α+γ
= 4c2 cos2 − ( c2 + s2 ) = 2 cos ω + 1 ,
2
in accord with the general formula (4.60). We have used the trigonometric iden-
tity cos 2ψ = 2 cos2 ( ψ ) − 1.
0 0p
0 ξ j+m η j−m (2j )! 0 0
ψm =∑p ∑ a j+ m − µ b m − m + µ c µ d j− m − µ
0
( j + m )!(j − m )! µ 0
m0
p (4.61)
( j + m0 ) ! ( j − m0 ) ! ( j + m ) ! ( j − m ) !
× .
µ! ( j − m0 − µ ) ! (j + m − µ ) ! (m0 − m + µ ) !
0j j j
This equation may be reshaped in the form (cf. (4.11)) ψm = ∑m0 ψm0 Dm0 m ( a, b )
so that D j ( a, b ) can be identified with the representation matrix D j [ R (a, b)] of the
4.5. ROTATION MATRICES 143
The sum is taken over all possible values of µ as long as none of the arguments
of the factorials where µ appears are negative.
Special Cases.
• a 6 = 0, b = 0, then µ = 0 and m0 = m:
j j
Dm0 m ( a, 0)δm0m a j+m ( a∗ ) j−m ; in particular, Dm0 m (± 1, 0) = δm0 m (± 1)2j.
• m0 = j, then µ = 0:
1/2
j (2j )!
Djm ( a, b) = a j+ m b j− m .
( j + m) !( j − m) !
• m = − j, then µ = 0:
1/2
j (2j )! 0 0
Dm0 ,− j ( a, b ) = ( a ∗ ) j− m b j+ m .
( j + m ) ! ( j − m0 ) !
0
C OMMENTS . Given these results, we can show that the unitary representations
D j are irreducible. For this to be the case, any matrix that commutes with all D j ’s
is necessarily a multiple of the unit matrix (Schur’s lemma). First, if a matrix M
commutes with the matrix D j ( a, 0), it must be diagonal: Mmm0 = cm δmm0 . If M
commutes with the general unitary matrix D j ( a, b ), then the ( jm)-component of
j j
MD and DM yields c j Djm = Djm cm for all allowed m. As we have seen above,
j
Djm 6 = 0, so cm = c j for all m. Thus, a matrix that commutes with all matrices of
the representation must be a multiple of the unit matrix, and so the representations
j
Dmm0 ( a, b ) are irreducible. Also, as D j have different dimensions for different j,
they are not equivalent to one another.
Expression of D j in terms of Euler angles. With a, b as given in (4.49) the repre-
sentation matrix can now be expressed in the Euler angles:
j 0 j
Dm0 m ( α, β, γ ) = e−im α dmm0 ( β ) e−imγ
j 0 [( j + m0 ) ! ( j − m0 ) ! ( j + m ) ! ( j − m ) ! ]1/2
d m0 m ( β ) = ∑ (−)µ+m −m µ! (m0 − m + µ ) !( j − m0 − µ ) !( j + m − µ ) !
µ=0
0 0
β 2j+m−m −2µ β m −m+2µ
× cos sin . (4.63)
2 2
144 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
Its transpose is
j [( j + m0 ) ! ( j − m0 ) ! (j + m ) ! ( j − m ) ! ]1/2
dmm0 ( β ) = ∑ (−)ν ν! (m0 − m + ν ) !( j − m0 − ν ) !( j + m − ν ) !
ν= 0
0 0
β 2j+m−m −2ν β m −m+2ν
× cos sin . (4.64)
2 2
These formulas are due to Eugene Wigner, and are called the Wigner formula(s).
The matrix d j obeys the following general identities:
j j
dm0 m ( β ) = d−m,−m0 ( β ) (4.65)
m0−m j
= (−) dmm0 ( β ) (4.66)
m0−m j j
= (−) dm0 m (− β ) = dmm0 (− β ) (4.67)
j 2j
dm0 m (2π ) = (−) δm0 m . (4.68)
j
d m0 m ( π ) = (−) j−m δm0 ,−m (4.69)
j j j+ m 0 j
d m0 m ( π − β) = (−) j−m d−m0 ,m ( β ) = (−) d m 0 ,− m ( β ) (4.70)
They directly follow from the Wigner formula in its various equivalent forms.
(4.67) is an expression of the equivalence of the change β 7 → − β and matrix
transposition. (4.69) follows from (4.64) because [cos( π/2))]k 6 = 0 iff k = 0, or
ν = j + ( m − m0 ) /2; and ( j − m0 − ν ) ! 6 = 0 iff j − m0 − ν = −( m0 + m ) /2 = 0. Fi-
j j j
nally, (4.70) follows from dm0 m ( π − β ) = h m0 |e−i(π− β) J2 | m i = dm0 m00 ( π )dm00m (− β ),
and the preceding general identities.
Special Cases. Let c ≡ cos ( β/2) and s ≡ sin( β/2).
(i) m = m0 = 0; then j = ` an integer:
" #2
` ` ! c`−µ sµ
d`00 ( β ) = ∑ (−)µ . (4.71)
µ=0
µ! (` − µ ) !
E XAMPLE 9: In the list of matrices given below the entries of the first and last rows
of a matrix d j, as well as the entries of its first and last columns, are evaluated with
(4.72). The matrix elements are given, as usual, in the decreasing order of m. If j
j
is an integer, then d00 is given by (4.71); for other j, d j are calculated with (4.63).
Notations: c ≡ cos ( β/2) and s ≡ sin( β/2). Besides d 0 ( β ) = 1, we have
4.6. DIRECT-PRODUCT REPRESENTATIONS 145
c −s
(i) j = 1/2 : d1/2 ( β ) = .
s c
2 √
√c − 2cs s2
√
(ii) j = 1 : d1 ( β ) = 2cs c√ 2 − s2 − 2cs.
s2 2cs c2
3
(iii) j = /2 :
3 √ 2 √ 2 3
√ c − 3c s 3cs √− s 2
3c2 s c3 − 2cs2 − 2c2s + s3
√3cs2 .
3/2
d ( β) = √
3cs2 2c2 s − s3 c3 − 2cs2 − 3c s
√ 2 √ 2
s3 3cs 3c s c3
C OMMENTS .
1. There is a simple but important physical consequence which follows directly
from the present discussion. The physics of a system can be described by an
action integral that depends on the state function ψ in the form ψ† O ψ. If a trans-
formation group leaves the action integral invariant, then the generators of the
group produce conserved quantities. Whether the system is invariant under a
direct product group, like U (1) × SU (2), or a group, like U (2) itself, has implica-
tions on the quantum numbers. In either case, the group representation may be
labeled by ( n, j ), with a U (1) (integral) quantum number n, and an SU (2) (inte-
gral or half-integral) quantum number j. If the symmetry group is U (1) × SU (2),
then n and j are independent conserved quantum numbers; but if it is U (2), then
invariance requires 2j + n = 2Q be an even number, and it is Q that is the con-
served number.
2. The Lie algebra su (2) < sl (2, C ) shows that its irreducible representations can
be characterized by integral or half-integral values of j. In quantum physics, su (2)
is usually associated with symmetries in coordinate or charge space, then j is
interpreted as the spin (angular momentum), or isospin quantum number. We
know that there exist in nature two kinds of particles: the bosons, which obey
Bose–Einstein statistics and have integral spins, and the fermions, which obey
Fermi–Dirac statistics and have half-integral spin. The isospins of bosons and
fermions may be either integral (π mesons, Σ baryons) or half-integral (K mesons,
the nucleons).
| jm i ⊗ | j 0 m0 i ≡ | m, m0 i , (4.73)
where m, m0 take all allowed values, and the labels j, j 0 are implicit in the short-
ened form on the RHS. In this product space, a rotation R has a representation
defined by (again with implicit summation over repeated indices)
j j0
Π ( R)|m, m0i = | n, n0 i Dnm ( R ) Dn0m0 ( R ) . (4.74)
j× j0 j j0 0
Symbolically, Dnn0 ,mm0 = Dnm Dn0 m0 . The product representation D j× j is re-
ducible (unless either j or j 0 is zero), and one problem is to find its decomposition
into irreducible representations (cf. Example 2).
Consider a rotation R (δω, n) around an arbitrary axis n:
0 0
D j× j [ R (δω, n)] = D j [ R (δω, n)] D j [ R (δω, n)]. (4.75)
Assuming δω infinitesimal, we can expand both sides to the first few terms:
0
h i h 0
i
Ij× j0 − iδωD j× j ( Jn ) = Ij − iδωD j ( jn ) ⊗ Ij0 − iδωD j ( jn0 ) (4.76)
where Ir is the identity matrix and Dr ( Jn ) the matrix representing the operator
π r ( Jn ) in space V (2r) , with Jn = J · n. Equating the first-order terms in δω on
0 0
both sides of the equation produces D j× j ( Jn ) = D j ( jn ) ⊗ Ij0 + Ij ⊗ D j ( jn0 ), where
on the right-hand side we are reminded that we are dealing with direct (and not
ordinary) products of matrices. This being understood, we write the operators in
0
the product space V (2j) ⊗ V (2j ) in a simplified form: Jk = jk + jk0 , with (k = 1, 2, 3),
rather than in the more precise form: Jk = jk ⊗ I + I ⊗ jk0 . It is important when
using the abbreviation to recall the specific space on which each operator applies.
4.6. DIRECT-PRODUCT REPRESENTATIONS 147
Each angular momentum, jk or jk0 , satisfies (4.42) in its own space, and commutes
with the other, whereas an angular momentum Jk = jk + jk0 in the product space
obeys its own commutation relations.
A point worth noting, illustrated for SU (2) and su (2) but generalizable, is
that the representations Π of a Lie group G and π of the Lie algebra Lie( G ) = g
in direct-product space are quite different from each other. Thus we have
j× j0 j j0
Π ( g)|m, m0i = | nn0 i Dnn0,mm0 ( g) = | nn0i Dnm ( g) Dn0m0 ( g), (4.77)
J 2 → J ( J + 1) , J3 → M ,
2
j → j ( j + 1) , j02 → j 0 ( j 0 + 1) .
With J and M taking all (still to be specified) allowed values, the set of vectors
|( jj 0) J M i gives a second basis for the product space, and so each vector is ex-
pressible as a linear combination of the product vectors (4.73). Since the eigen-
values j and j 0 are constant in both bases, we may confine ourselves to just the
0
subspace V (2j) ⊗ V (2j ) with fixed j and j 0 . For this reason, the labels j, j 0 may be
suppressed, as in | J M i ≡ |( jj0) J M i.
For a given J, the 2J + 1 vectors | J M i with M in the range − J ≤ M ≤ J span
an irreducible subspace. Therefore if the vector | J M i for a particular value (for
example, the highest value) of M is known, then all the other vectors in V (2J ) can
be constructed by repeated application of the raising/lowering operators J± on
the known vector.
On the other hand, recalling that M = m + m0 , all the vectors with the same
value of M but different values of J can be constructed from the set of vectors
| mm0i, with different m and m0 added up to M, as we will see in the following
example. Let g( M ) be the degree of degeneracy of M in this space, i.e. the numbers
of vectors | J M i having the same value of M; it is equal to the number of pairs
( m, m0 ) such that M = m + m0 . Knowing g( M ) we can find the allowed values of
J for fixed M, and hence the range of values of J for any given M. This description
should become clearer through the following example.
E XAMPLE 10: Let j = 3 and j 0 = 2, which implies − 3 ≤ m ≤ 3 and − 2 ≤ m0 ≤ 2,
and so − 5 ≤ M ≤ 5. We list in the following table g( M ), M, all pairs ( m, m0 ) with
M = m + m0 , and allowed values of J. It tells us which | mm0i’s are found in | J M i,
148 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
g( M ) M ( m, m0) J
1 5 (3, 2) 5
2 4 (3, 1) (2, 2) 5, 4
3 3 (3, 0) (2, 1) (1, 2) 5, 4, 3
4 2 (3, − 1) (2, 0) (1, 1) (0, 2) 5, 4, 3, 2
5 1 (3, − 2) (2, −1) (1, 0) (0, 1) (−1, 2) 5, 4, 3, 2, 1
5 0 (2, − 2) (1, −1) (0, 0) (− 1, 1) (− 2, 2) 5, 4, 3, 2, 1
5 −1 (1, − 2) (0, −1) (−1, 0) (−2, 1) (−3, 2) 5, 4, 3, 2, 1
4 −2 (0, − 2) (−1, −1) (−2, 0) (−3, 1) 5, 4, 3, 2
3 −3 (− 1, −2) (−2, − 1) (− 3, 0) 5, 4, 3
2 −4 (− 2, −2), (−3, −1) 5, 4
1 −5 (− 3, −2) 5
g( M ) = 0 if | M | > j + j 0
= j + j + 1 − | M | if | j − j 0 | ≤ | M | ≤ j + j 0
0 (4.79)
= 2 min( j, j 0) + 1 if 0 ≤ | M | ≤ | j − j 0 | .
j+ j0
N= ∑ (2J + 1)
J = j− j0
1
= { 2(( j − j 0 ) + 1} + { 2( j + j 0 ) + 1} (2j 0 + 1)
2
= (2j + 1)(2j 0 + 1) .
This state counting provides a consistency check of the results obtained above.
Thus, for fixed j and j 0 , there are two independent sets of orthonormal vectors:
|( jj0) J, M i : | j − j 0 | ≤ J ≤ j + j 0 , −J ≤ M ≤ J ; (4.80)
0 0 0 0 0
| jm, j m i : − j ≤ m ≤ j, −j ≤ m ≤ j ; (4.81)
0
each spans the product space V (2j) ⊗ V (2j ) (i.e. each is complete). The two bases
4.6. DIRECT-PRODUCT REPRESENTATIONS 149
j+ j0 J
0 0
| jm, j m i = ∑ ∑ |( jj 0) J, M ih J M| jm, j0m0 i , (4.82)
J =|j− j0 | M =− J
j j0
0
|( jj ) J, M i = ∑ ∑ | jm, j 0m0 ih jm, j0m0 | J M i , (4.83)
m=− j m0 =− j0
So in this convention, the unitary transformation and its inverse are specified by
the same set of real coefficients: h J M | jm, j 0m0 i = h jm, j 0m0 | J M i.
Properties of the Clebsch-Gordan coefficients for SU ( 2).
(i) Angular momentum selection rule:
j j0
∑ ∑ h J 0 M0 | jm, j 0m0 ih jm, j0m0 | J M i = δ J 0 J δM0 M
m=− j m0 =− j0
j+ j0 J
∑ ∑ h jn, j 0n0 |( jj 0) J, M ih J M| jm, j0m0 i = δnm δn0 m0 .
J =|j− j0 | M =− J
where each summand D J occurring only once. It is the same as the relation given
in Example 2. When the decomposition is written out fully, we have what is
known as the Clebsch-Gordan series:
j+ j0
j j0 J
Dmn ( R )Dm0n0 ( R ) = ∑ h jm, j0m0 | J M i DMN ( R )h J N | jn, j0n0 i . (4.86)
J =|j− j0 |
J j j0
δ J J 0 DMM0 ( R ) = ∑0 h J M | jm, j 0m0 i Dmn ( R )Dm0n0 ( R )h jn, j0n0 | J 0 M0 i , (4.87)
mm nn0
++ −+ +− −−
11 1 0 0 0
q q
1 1
10 0 2 2 0
1, − 1 0 0 0 1
q q
1 1
00 0 2 − 2 0
4.6. DIRECT-PRODUCT REPRESENTATIONS 151
The weight diagrams. We introduce here a graphical tool, the weight diagram
of a representation, which gives a graphical distribution of the weights of the
representation in weight space. For SU (2), it is simply a linear plot, but it is more
complicated in higher-dimensional groups.
For example, the irreducible representation V (4) has the weight diagram:
s s s s s
−2 −1 0 1 2 m
s s
e s
−1 0 1
As another example, let us take V (2) ⊗ V (3). The eigenvalues for J3 in V (2) are
0 and ± 1; and in V (3), ± 3/2 and ± 1/2, which means the twelve eigenvalues in
V (2) ⊗ V (3) are ± 5/2, ± 3/2 (twice), and ± 1/2 (three times), as shown:
s se s
h
e se
h se s
− 52 − 32 − 12 1
2
3
2
5
2
An analysis along the same lines yields the decomposition of a typical tensor
product Syma V (1) ⊗ Symb V (1), or equivalently V (a) ⊗ V (b), according to
We can check this equation by calculating the dimensions on both sides, and see
that they agree. It is remarkable that the multiplicity of every component is one,
and that the weights (with respect to J3) of all composing representations occupy
every point of the set 12 Z.
Problems
4.1 (Casimir invariant) Let h, e, f be the canonical basis of sl (2, C ). Show that
C = e f + f e + 12 h2 is an invariant of the algebra. Express C in terms of J1, J2, J3 ,
with h = 2J3, e = J1 + iJ2, and f = J1 − iJ2.
4.2 (Brackets) Let h, e, f be the canonical basis of sl (2, C ). Calculate the brackets
[ h, em ], [ h, f m ], [ f , em], and [ e, f m].
4.3 (Spin-one representation) Let ξ, η ∈ C2 transform into ξ 0 = aξ + cη, η 0 = bξ +
dη under SU (2) (so that d = a∗ , c = − b∗, and aa∗ + bb∗ = 1). Let ψ+ = ξ 2 , ψ0 =
√
2ξη, and ψ− = η 2. Show that the ψµ ’s transform as a spin-one representation;
find the representation matrix D1 in terms of the Euler angles.
4.4 (Quaternion groups) A quaternion q is a number that can be written as q =
∑3µ=0 xµ qµ where xµ are real numbers, and qµ = { q0, qi } (with µ = 0, . . . , 3 and
i = 1, 2, 3) are generalized non commuting complex numbers defined such that
q0 q0 = q0, q0 qi = qi q0 = qi , qi q j = − δijq0 + eijk qk ; and q∗0 = q0, q∗i = − qi under
complex conjugation.
(a) Let H be a one-dimensional quaternion-valued vector space, with typical
vector η. The set of all non-zero mappings η 0 = Qη form a Lie group, called
GL(1,Q). Find its generators and its Lie algebra.
(b) Show that { Q ∈ GL (1, H )|Q∗ Q = 1} is a Lie group. Find its associated
Lie algebra.
(c) Evaluate the exponential mapping of the above Lie algebras.
4.5 (det R = 1) Show that the invariance of the tensor eijk under rotations is
equivalent to det R = 1 for any orthogonal real 3 × 3 matrix R.
4.6 (Invariant axis) Show that a proper rotation in an odd-dimensional space pos-
sesses an invariant axis.
4.7 (Similarity transformation) Consider the group of pure rotation in R 3: { R ∈
M(3, R 3)| RT R = I; det R = 1} .
(a) Find the direction of the axis and the angle of rotation in terms of the matrix
elements R ij .
(b) Find the transformation U that diagonalizes R, so that U −1 RU = Λ is a
diagonal matrix.
(c) Find the transformation that gives Λ in a real orthogonal basis.
4.8 (Angle-axis rotation matrix) Write out the 3 × 3 matrix of rotation R ( ω, n)
in terms of the rotation angle ω and the spherical coordinates θ, ϕ of the axis n,
or alternatively in terms of ω and the Cartesian components n1, n2 , n3 of n in an
orthonormal basis { ei } , with n2 = 1.
4.6. DIRECT-PRODUCT REPRESENTATIONS 153
sin α 1 − cos α
x0 = x + ( α × x) + [ α × ( α × x)].
α α2
4.13 (Commutation of angular momentum) (a) Show that the sequence of in-
finitesimal rotations by the directed angles α, β, − α, and − β on the position
vector x is equivalent to the rotation by the directed angle β × α on x.
(b) In Hilbert space, a rotation by α is represented by the unitary operator
U ( α) = e A, where A = α · X with Xi denoting the generators of rotation. Follow-
ing the argument used in (a), derive the commutation relations for Xi (see also
Sec. 3.3, Chapter 3).
4.14 (Exponential mapping) Re-express R ( ω, n) = exp(− iωnk Lk ) in a non-trans-
cendental form by summing the power series of the exponential. Lk are 3 × 3
matrices of the angular momentum operators in a Cartesian basis.
4.15 (From cartesian basis to spherical basis) Show that the matrices Jk ≡ D1 ( Jk )
in the spin j representation (in which D1 ( J3) is diagonal) are equivalent by a sim-
ilarity transformation to the defining matrices Lk of the algebra so (3) given in the
preceding problem.
4.16 (Non standard parameterizations) In the text, we emphasized the standard
parameterization ( ω1, ω2, ω3 ) of SU (2). But there are times when it is more ad-
vantageous to use a non-standard parameterization (e.g. to make the exponential
mapping simpler). In SU (2) the use of a parameterization associated with the
operators J± = J1 ± iJ2 has a further advantage when we need to study transi-
tions between states in the same irreducible representation. In this problem, we
examine just such mappings involving the elements J± , J3.
(a) Evaluate Uo ( ω+, ω− , ω3 ) = exp[− i(ω+ J+ + ω− J− + ω3 J3 )] and relate the
parameters ω+ , ω− , ω3 to the standard parameters ω1, ω2, ω3.
(b) Evaluate Ua ( α+, α− , α3 ) = exp[− i(α+ J+ + α− J− )] exp[− iα3 J3 ], and dis-
cuss the characteristics of the parameters.
(c) Repeat with Ub ( β +, β − , β 3 ) = exp [− iβ + J+ ] exp[− iβ 3 J3 ] exp[− iβ − J− ].
j
4.17 (Matrices in non standard parameters) Find Dm0 m [U ] for the different U
found in Problem 4.16.
4.18 (j = 1/2 rotation matrices) (a) Write down the matrices for the same general
rotation in the Euler angles, R ( α, β, γ), and the angle-axis parameters, R (ω, n),
both in the two-dimensional j = 1/2 angular-momentum basis.
154 CHAPTER 4. sl (2, C ) AND ASSOCIATED LIE GROUPS
(b) By comparing the two expressions, D1/2 [ R (α, β, γ)] and D1/2 [ R (ω, n)], de-
rive the relationship between the two sets of parameters.
4.19 (Schwinger’s model) J. Schwinger gave a model of the algebra of angular
momentum based on the algebra of two independent harmonic oscillators repre-
sented by ( a+ , a†+ ) and ( a−, a†− ), which satisfy the commutation relations typical
of uncoupled harmonic oscillators:
h i h i
a+ , a†+ = 1, a− , a†− = 1,
h i h i
a+ , a†− = 0, a− , a†+ = 0.
(a) Define the number operators N+ = a†+ a+ and N− = a†− a− . Show that
they commute with one another, and thus have common eigenkets | n+ , n− i of
respective eigenvalues n+ , n− where the vacuum ket |0, 0i ≡ |0i is defined by
a+ |0i = 0, and a− |0i = 0. Construct the normalized ket | n+ , n− i from the nor-
malized vacuum state |0, 0i.
(b) Define the operators J+ = a†+ a− , J− = a†− a+ , J3 = 21 ( a†+ a+ − a†− a− ), and
J = 12 ( J+ J− + J− J+ ) + J32 . Derive their bracket rules.
2
5.1 Structure
5.2 Representations
We will apply the basic ideas and techniques developed in the last chapter for
sl (2, C ) to study sl (3, C ). As we are dealing now with a larger space — an eight-
dimensional sl (3, C ) rather than the three-dimensional sl (2, C ) — we need to
generalize old ideas and introduce new concepts. The basis for our study is an
elementary result from linear algebra, namely, that commuting diagonalizable
matrices are simultaneously diagonalizable. By the end of this chapter, we will
have almost all the tools we need for studying other simple or semi-simple Lie
algebras. The Lie group SU (3), which admits sl (3, C ) as the Lie algebra, plays
an important role in physics, first as the (approximate) symmetry of the quark
flavor, where it was referred to as the eightfold way, then later as the (exact) sym-
metry of the quark and gluon color, which gave rise to a gauge theory of strong
interactions called quantum chromodynamics (QCD).
5.1 Structure
After a brief review of some key properties of sl (2, C ), we will take up sl (3, C ),
define a standard basis for it, analyze its structure, and consider its roots and
weights, phrasing the results so as to be generalizable to other Lie algebras.
Basic results on sl (2, C ). To begin, we recall the most important properties of the
complex semi-simple algebra sl (2, C ) which, as we shall see, plays an essential
role in the analysis of sl (3, C ) and other simple Lie algebras. Let h, e, and f be the
canonical basis for sl (2, C ) satisfying the bracket relations
[ h, e ] = 2e, [ h, f ] = − 2 f , [ e, f ] = h . (5.1)
155
156 CHAPTER 5. LIE ALGEBRA sl (3, C )
adh z = α ( h) z, z ∈ g, (5.3)
π ( h )u = µ ( h )u, u ∈ V. (5.4)
Elementary properties. From Table 5.1 several properties of g = sl (3, C ) are ap-
parent. First, t1 and t2 commute with each other, and, as we have noted above
concerning the matrices T1 and T2, it is the largest number of mutually commut-
ing elements this algebra can have. They define a Cartan sub algebra (denoted h)
of g, as is called the maximal abelian sub Lie algebra of a semisimple Lie algebra,
with adh simultaneously diagonalizable for all h ∈ h. The dimension of a Cartan
sub algebra, called the rank ` of the Lie algebra (` = 2 in this case), is independent
of the specific choice of its span. (Other possible choices for h can be identified
from the entries of Table 5.1.)
Secondly, the elements t1, e1 , f 1 satisfy the bracket relations
They form a sub algebra of sl (3, C ) isomorphic to sl (2, C ) via the correspondence
of t1, e1, f 1 with the elements h, e, f in (5.1). It is not the only sub algebra isomor-
phic to sl (2, C ) in sl (3, C ), as t2, e2, f 2 also obey similar relations
These two sl (2, C ) sub algebras are referred to as s1 and s2 . Together they generate
the whole algebra g since the remaining two elements, e3 and f 3, are fixed by
of the Lie algebra itself. It is useful to note that the matrix adx for any element x
of the algebra can be obtained from Table 5.1.
The adjoint representation: roots. Non-zero weights of the adjoint representa-
tion are called roots. Generalizing Eq. (5.3), we have
Definition 5.2 (Roots of sl (3, C )). Let h be an element of the Cartan sub algebra h of
g = sl (3, C ). A non-zero linear function α ( h ) is called a root of g with respect to h if
there exists a non-zero element z ∈ g such that adh z = α ( h )z. The element z is called
a root vector corresponding to the root α.
Just as for the weights of any other representation, α ( h) is a linear function
of h ∈ h (which means α ∈ h∗ ), and can be specified in terms of an ordered pair
( α1, α2 ), with α1 = α ( t1) and α2 = α ( t2) relative to the basis { t1, t2 } of h.
Simple roots. From Table 5.1, one can establish the complete root system of
sl (3, C ). The product rules given there tell us that ei, f i , with i = 1, 2, 3, are the 6
joint eigenvectors of adt1 and adt2, with the corresponding 6 roots given below:
These roots α(1) and α(2) are called the simple roots of the algebra relative to
the given h, forming the set Σ = { α(1), α(2) } ; they have the property that all the
roots of the algebra can be expressed as linear combinations of them, with integral
coefficients which are either all greater than or equal to zero (positive roots), or else
all less than or equal to zero (negative roots). This can be verified:
(2, − 1) = α(1)
(− 1, 2) = α(2)
(1, 1) = α(1) + α(2) ≡ α(3)
(5.10)
(− 2, 1) = − α(1)
(1, − 2) = − α(2)
(− 1, −1) = − α(1) − α(2) = − α(3) .
The set of all the roots (± α(1), ± α(2), ± α(3)) of g is a finite subset of h∗ denoted
by ∆; and its subset of positive roots (which are α(1) , α(2) , and α(3) ) is called ∆+ .
(0 ∈ h∗ is not considered a root, i.e. 0 6 ∈ ∆, although g0 is often used to denote h.)
160 CHAPTER 5. LIE ALGEBRA sl (3, C )
From their explicit expressions, we observe that the roots of sl (3, C ) have the
following properties (cf. [Ja] Chap. IV), which hold in general for other semisim-
ple Lie algebras (in particular, sl (2, C )) as well:
(i) If α is a root, the only integral multiples kα of α that are nonzero roots are
α and − α (i.e. ± 2α, ± 3α, . . . are not roots).
(ii) To each root corresponds a unique linearly independent element of g with
that root, so that we have the direct-sum decomposition:
g = h ⊕ (⊕ α∈∆ gα ) . (5.11)
h∗0
(iii) All the roots are real; they lie in and span a real subspace of the general
weight space h∗ , i.e. more precisely h∗0 = { α ∈ h∗ | α = a1α(1) + a2 α(2) ; ai ∈ R } ,
where α(1) , α(2) are the simple roots of g. Moreover, the rank ` of g is defined to be
dim h = dim h∗ ; and the dimension of g is dim g = dim h + ∑α dim gα = | Σ | + |∆|.
The Killing form on g and h. The Killing form K on g is a bilinear trace form in the
adjoint representation: K ( x, y) = Tr (adx · ady ) for any x, y ∈ g. We will use the
notation ( x : y ) ≡ K ( x, y). As we have already explained in Chapter 3 Sect. 3.6.4,
to calculate the Killing form of any x, y ∈ g, (i) Pick a basis { zi } for g; (ii) Calculate
[ x, zi ] = zk aki and [ y, zi] = zk bki ; (iii) Calculate [ x, [ y, z j ]] = zi aik bkj . The coefficient
of the zi with i = j on the right-hand side, if nonzero, contributes to the trace; (iv)
Repeat for every possible value of j, and obtain ( x : y ) = Tr (adx · ady ) = a jk bkj .
If [ x, [ y, z j ]] does not re-generate z j for any value of j, then ( x : y ) = 0. The Killing
form ( x : y ) for arbitrary x, y ∈ g can be expressed as a linear combination of
κij = ( zi : z j ) in any basis { zi} of g, as is shown, using a general property of
representations:
π ∑i ai zi π ∑j bj z j = ∑ij ai bj π ( zi ) π (z j) .
So, to calculate the Killing form K ( x, y) for any x, y ∈ g it suffices to know the
Killing forms of all the pairs of elements in some basis. The Killing forms for the
basis elements of sl (3, C ) given in Table 5.1 are listed in Table 5.2.
The Killing form on sl (3, C ) has the following properties:
(i) ( h : zα ) = 0 for any h ∈ h and any zα ∈ (g − h ).
(ii) ( zα : z β ) = 0, unless α + β = 0, with zα , z β ∈ (g − h).
(iii) The determinant of the matrix [ κij ] is non-zero: the Killing form on sl (3, C )
is non-degenerate, as is expected from Cartan’s criterion applied to a simple Lie
algebra. This also holds true for the restriction of K ( x, y) to CSA h.
(iv) K ( x, y) = Tr (adx · ady ) = 6 Tr ( xy ) for all x, y ∈ sl (3, C ).
Correspondence between h∗ and h. As the Killing form on h is non degenerate,
we can establish a one-to-one correspondence between h and h∗ : Let $ ( h) be an
element of h∗ , a linear functional of h ∈ h. Then there exists a unique element
h$ ∈ h defined such that the Killing form of h $ and h is precisely $ ( h ) ∈ h∗ :
def
( h$ : h ) = $ ( h) . (5.12)
This formula determines $ from any given h $ . Inversely, given a root $, how
do we find the element h$ in h? First, consider the simple roots. Let h = at1 + bt2
5.1. STRUCTURE 161
Table 5.2: The Killing forms on sl (3, C ). Each entry gives K ( a, b) ≡ ( a : b ), with
row label a and column label b.
be an arbitrary element of h; then from (5.9) the simple roots as linear functions
of a and b are given by
α(1) ( at1 + bt2) = 2a − b, α(2) ( at1 + bt2) = − a + 2b . (5.13)
As the corresponding elements hα(i) live on the CSA h, they must have the form
hα(i) = ci t1 + di t2 and, by definition, must satisfy (5.12):
2a − b = ( hα(1) : at1 + bt2)
= c1 a ( t1 : t1 ) + d1 b ( t2 : t2 ) + ( c1b + d1 a )(t1 : t2)
= (12c1 − 6d1) a + (12d1 − 6c1) b ;
− a + 2b = ( hα(2) : at1 + bt2)
= (12c2 − 6d2) a + (12d2 − 6c2) b ,
where we have used the values of the Killing forms ( t1 : t1 ) = ( t2 : t2 ) = 12 and
( t1 : t2 ) = − 6. The unique solution to these equations for ci and di, with arbitrary
a, b, are c1 = 1/6, d1 = 0, and c2 = 0, d2 = 1/6. (Note its independence from a, b,
and h.) The results gives the correspondence between two bases of h:
1 1
hα(1) =
t , hα(2) = t2 . (5.14)
6 1 6
The vectors h$ corresponding to any other root $ ∈ ∆ can be obtained by
using linearity, namely, h−α(i) = − hα(i) and hα(1) +α(2) = hα(1) + hα(2) . What we
have obtained is a new basis for CSA h; for example,
adhα(1) · e1 = [ hα(1) , e1 ] = α(1) ( hα(1) ) e1 ,
adhα(2) · e2 = [ hα(2) , e2 ] = α(2) ( hα(1) ) e2 . (5.15)
It follows that
α(1) ( hα(1) ) = α(2) ( hα(2) ) = α(3) ( hα(3) ) = 1/3, ,
α(1) ( hα(3) ) = α(2) ( hα(3) ) = − α(1) ( hα(2) ) = 1/6 .
162 CHAPTER 5. LIE ALGEBRA sl (3, C )
which, of course, can also be obtained from (5.14) and Table 5.2.
Inner product on the real root space h∗0 . The Killing form is defined on h. We now
want to have a similar bilinear form on h∗ . Let h∗0 be the subspace of h∗ spanned
by the simple roots. Take $, σ any two vectors of h∗0 and define an inner product of
$, σ via the bilinear form of the corresponding elements h $, hσ :
def
h $, σ i = ( h$ : hσ ) . (5.16)
α(2 ) α(3 )
KA
A
A
(1 ) A (1 )
−α A -α
A
A
A
A
−α(3)
AU −α(2)
Figure 5.1: The root system ∆ plotted on sl(3, C)’s root space.
−1 t t1 t2 t1
A e3 A
e2 K
A A
t 0 A t
f e
t1 f
t -1 t t Af
1 A A
−2 0 2 −1
A A
A f2 AU
−1 t t1 −1 t t −2
(a) (b)
[ h, [ eα , e β ]] = [ eα , [ h, e β]] + [[ h, eα], e β ]
= [ eα , β ( h )eβ] + [ α (h)eα, e β ]
= ( α ( h) + β ( h )) [eα, e β ] . (5.21)
The condition that emerges, namely, adh [ eα, e β ] = ( α ( h) + β ( h )) [eα, e β ] for any
element h of h, can be satisfied in one of two ways:
(i) [ eα , e β ] is zero, but α + β 6 = 0;
(ii) [ eα , e β ] is a nonzero root-vector with
(a) α + β 6 = 0, or (b) α + β = 0.
Case (i) describes the situation where α + β is not a root, and so [ eα , e β ] = 0.
In case (iia) α + β is a root in ∆, and [ eα, e β ] must be proportional to eα+ β , with
a multiplying constant, Nαβ , yet to be determined. In other words, [ eα , e β ] is again
an eigenvector for h, with eigenvalue α + β.
164 CHAPTER 5. LIE ALGEBRA sl (3, C )
[ e α , e −α ] = ( e α : e −α ) h α . (5.22)
t -t -0
AA
A
A
t -Ath -t -0
A
A
A
t -AAt -0
Figure 5.3: The action of adgα on gβ produces gα+ β , h, or 0. This action is indicated by the
arrows; the central dot represents h; any other dot represents a g β for some β ∈ ∆.
To check that these relations agree with the multiplication Table 5.1, it suffices
to recall that ( eα : e−α ) = 6 for all α ∈ ∆; choose h 1 = t1 = 6hα(1) and h2 = t2 =
5.2. REPRESENTATIONS 165
6hα(2) ; and recall the roots α(1) = (2, − 1), α(2) = (− 1, 2), and α(3) = (1, 1) written
in the basis { α (h 1), α (h2)} of h∗ . Taking ei and e−i (with the subscripts i standing
for α(i) ) to be ei and f i respectively, we find the nonzero structure constants:
N1,2 = − N−1,−2 = 1,
N−1,3 = − N1,−3 = 1,
N2,−3 = − N−2,3 = 1 .
which in fact turn out to be true in other Lie algebras as well (cf. Problem 5.4).
Standard normalization. Of the admissible bases, those associated with a fixed
(positive) root α play a special role as they span a three-dimensional sub algebra.
Let us choose them with the following normalization:
eα = eα ,
f α = [2/α(hα)( eα : e−α )] e−α , (5.29)
h̃α = [2/α(hα)] hα .
Then (5.24) and (5.25) (with h set to be h̃α ) assume the canonical form (5.1):
[ h̃α , eα ] = 2eα ,
[ h̃α , f α ] = − 2 f α, (5.30)
[ eα , f α ] = h̃α .
Hence h h̃α, eα , f α i is isomorphic to sl (2, C ). Noting that h̃α (also called a co-root)
differs from hα by a constant, one has from (5.12) α ( h̃α ) = 2.
5.2 Representations
Quick review. Many of the general remarks we made in the last chapter about the
representations of SU (2) and of its associated algebras apply to our present case
as well. In particular, just as the group SU (2) is connected and simply connected,
so too is the group SU (3), and so there is a one-to-one correspondence between
the finite-dimensional representations of the Lie group SU (3) and those of the Lie
algebra su (3). On the other hand, as the complex representations of the real Lie
algebra su (3) are in one-to-one correspondence with the complex representations
of the complexified algebra su (3)C, and as su (3)C ∼ = sl (3, C ), there must also be a
one-to-one correspondence between the finite-dimensional complex representations Π of
SU (3) and the finite-dimensional complex representations π of sl (3, C ), which is ex-
pressed by Π (ex) = eπ(x) for all x ∈ sl (3, C ). The representation Π is irreducible
166 CHAPTER 5. LIE ALGEBRA sl (3, C )
HZα v = ( Zα H + [ H, Zα ]) v
= (µ ( h) + α ( h )) Zα v .
Hi E1 v = µ ( hi) + α(1) ( hi ) E1 v ,
Hi E2 v = µ ( hi) + α(2) ( hi ) E2 v .
We already know that µ ( h 1) = m1 and µ ( h2) = m2 are integers (or zeros) and
the simple roots relative to h1 and h2 are given by ( α(1) ( h1), α(1) ( h2)) = (2, − 1)
and α(2) ( h1), α(2) ( h2) = (− 1, 2). So, with the weight of v given by ( m1, m2 ),
these equations tell us that E1 v has the weight ( m1 + 2, m2 − 1), and E2 v has the
weight ( m1 − 1, m2 + 2). The ladder operators Ei ’s flip the parity of the weight
vector, from (− 1)m1+m2 for v to (− 1)m1+m2 +1 for Ei v.
Ordering the weights. Just as for sl (2, C ), the irreducible finite-dimensional rep-
resentations of sl (3, C ) are identified by their highest weights. The additional
complication here is that CSA h is two-dimensional, and so one must introduce
some convention for ordering the weights:
Definition 5.3 (Ordering of the weights; highest weight). Let λ and µ be any two
weights of a representation π of sl (3, C ). Then λ is said to be higher than µ if λ − µ can
be written as a positive root of sl (3, C ): λ − µ ∈ ∆+ . This relationship will be indicated
by λ µ. A weight λ of π is said to be a highest weight if λ µ for all other weights
µ of π. Similarly, λ is a lowest weight if µ λ for all µ of π.
The positive roots of sl (3, C ) relative to the CSA h h 1, h2 i are α(1) = (2, − 1),
α (2)= (− 1, 2), and α(3) = (1, 1). The condition that λ − µ = aα(1) + bα(2) be
positive means a ≥ 0 and b ≥ 0. In particular, α(3) , considered a weight, is higher
than both α(1) and α(2) .
If µ? is a highest weight of π, then µ? + α is not a weight for any α ∈ ∆+ .
On the other hand, a weight µ of π must be of the form µ? − kα, with k ≥ 0; all
weights lie in a sequence, or string, µ? , µ? − α, . . . , µ? − nα (n ≥ 0), spaced by
some α ∈ ∆+ . Since π is assumed finite-dimensional, the string is finite and n
must be a finite non-negative integer. Several steps are needed that lead us to the
main theorem concerning the simple representations of sl (3, C ):
1. Let us recall our normalization factors for si < g:
For each α(i) , consider the sequence of vectors v0, v1, . . . , vni generated from v0
according to vk = Fi vk−1 = ( Fi )k v0 with k = 1, 2, . . . , ni ; finiteness of the rep-
resentation requires Fi vni = vni +1 = 0. Each vk is an eigenvector of Hi with
eigenvalue µk = µ? ( hi) − kα(i) ( hi ) where 0 ≤ k ≤ ni.
168 CHAPTER 5. LIE ALGEBRA sl (3, C )
As Ei vni +1 = tni vni vanishes because vni+1 = 0, it follows from vni 6 = 0 that
tni = 0, which implies ni = 2 µ? ( hi)/α(i) ( hi). As α(i) ( hi ) = 2 for both i = 1, 2, we
have ni = µ? ( hi), and so µ? = ( n1, n2 ) is a pair of non-negative integers. In each
i-direction on h∗ , the vectors v0, Fi v0, . . . , ( Fi )ni v0, generated from v0 and attached
to the eigenvalues ni , ni − 2, . . . , − ni, define an (ni + 1)-dimensional irreducible rep-
resentation of si , a sub representation of g.
2. Now let µ be any weight of π, and α any positive root of the algebra, and
p, q be non-negative integers. Let Sα ( µ ): µ + pα, . . . , µ, . . . , µ − qα be a string of
weights such that µ + ( p + 1) α and µ − ( q + 1) α are not weights. (p and q are
then the numbers of times the root α can be added to and subtracted from the
weight µ while still remaining a weight of the representation.) The corresponding
weight vectors will be built by applying ( Eα )k and ( Fα )k on the weight vector v
of weight µ. Taking α to be a simple root α(i) , we identify the highest weight with
µ? ( hi ) = µ ( hi) + pi α(i) ( hi), and the corresponding lowest weight with − µ? ( hi ) =
µ ( hi) − qi α(i) ( hi). It then follows that
2 µ? ( h i ) 2 µ ( hi)
qi + p i = = µ? ( h i ) , qi − p i = = µ ( h i) , (5.36)
α (i) ( h i ) α (i) ( h i )
which fixes the numbers pi = [ µ? ( hi) − µ ( hi)] /2 and qi = [ µ? ( hi) + µ ( hi)] /2.
These are the constraints on the string of weights based on µ ( h i) and of length
qi + pi + 1 in the representation admitting µ? ( hi) as its highest weight:
sum of its weight spaces, V = ⊕ Vµ, where each Vµ is a joint eigenspace for π ( h1) and
π ( h2), with h1 and h2 being the chosen basis for CSA h.
2. Every irreducible representation of sl(3, C ) has a unique highest weight, say λ,
of the form λ = ( n1, n2 ) with non-negative integers n1, n2 relative to the chosen CSA
{ h1, h2 } . The representation is then denoted by πn1 ,n2 or V (n1 ,n2 ) .
3. Conversely, every pair of non-negative integers, n1 and n2, determines an irre-
ducible representation πλ admitting λ = ( n1, n2 ) as its highest weight. Two irreducible
representations of sl (3, C ) with the same highest weight are equivalent, that is to say, πλ
is uniquely determined by ( n1, n2 ).
These vectors are the weight vectors of weights ε 1 = (1, 0), ε 2 = (− 1, 1), and
ε3 = (0, − 1) relative to the CSA h = h h1, h2 i (read off from the diagonal matrices
T1, T2 on p. 157). Since Ei ξ 1 = 0 for i = 1, 2, 3, ξ 1 is the highest-weight vector,
with weight ε 1 = (1, 0). We can also check that ε 1 − ε 2 = α(1); ε 2 − ε 3 = α(2) ;
ε 1 − ε 3 = α(1) + α(2) (with α(1) = (2, − 1) and α(2) = (− 1, 2)) which shows that
ε 1 ε 2 and ε 1 ε 3, or ε 1 is the highest weight. The standard representation is
indeed the irreducible representation π1,0. Further, since ε 2 ε 3 and ε 1 ε 3, the
lowest weight in π1,0 is ε 3.
Suppose, conversely, that we are asked to construct the representation with
the pair of integers (1, 0) ≡ µ? as its highest weight. The procedure we will follow
is based on our discussion in §1 and §2 (pp. 167–168). We have, by assumption,
( H1, H2 ) u0 = µ? u0, and E1 u0 = 0, E2 u0 = 0. The string of weights starting from
µ? spaced by α(i) must end with the weight µ? − ni α(i) . Since n1 = µ? ( h1) = 1,
the weight string spaced by α(1) is µ? , µ? − α(1) , corresponding to the sequence
of vectors u0, F1 u0 (which forms a doublet of s2 ); and so ( F1)2 u0 = 0 because
170 CHAPTER 5. LIE ALGEBRA sl (3, C )
α(2)K
A
α(2K
A)
ε2 A ε1 A −ε3
sA s s
A A - A A
A (1 )
α A A
A A A -(1)
UA s s UA s α
ε3 −ε1 −ε2
Figure 5.4: Finding the weights of the fundamental representations π1,0 and π0,1 of
sl(3, C) through applications of root vectors Fi on the weight space. The weight vectors
ξ 1 , ξ 2 , ξ 3 of π1,0 have the weights ε1 , ε2 , ε3 ; and the weight vectors ξ 3 , ξ 2 , ξ 1 of π0,1 have the
weights −ε3 , −ε2 , −ε1 .
µ + pα(2), . . . , µ, . . . , µ − qα(2) ;
( E2) p v, . . . , v, . . . , ( F2)q v .
µ ( h1) = 1, which implies q = 1. It follows that we can build from µ = (1, − 1) the
string consisting only of µ, µ − α(1) , with the corresponding weight vectors F2 v0
and F1 F2v0 (a doublet of s1 ). As in the previous case, we can show that either F1
or F2 applied on F1 F2 v0 leads to no new vectors.
In conclusion, the irreducible representation π0,1 is spanned by v0, F2 v0, and
F1 F2 v0, with weights − ε 3, − ε 2, − ε 1; as shown in the right half of Fig.5.4.
We see that the diagrams for π1,0 and π0,1 are their mutually reflected images.
This is so because the two representations are dual (or conjugate) to each other,
with the lowest weight of one being the negative of the highest weight of the
other. (See Problem 5.9. In the case of sl (2, C ), the weights for any representation
π are symmetric about the origin, and so each π is its own dual.) The two repre-
sentations V (1,0) and V (0,1) are mapped into one another by an automorphism of
sl (3, C ), which sends each element Z to − ZT (the negative of transposed Z). This
(1,0)
relationship is indicated by the notations π1,0 = π0,1 and V = V (0,1). For each
T
Z = π ( z ), there corresponds π ( z ) = − Z in V . In particular, the elements of the
Cartan sub algebra in V are
−1 0 0 0 0 0
π ( h1) = 0 1 0 , π ( h 2 ) = 0 − 1 0 . (5.37)
0 0 0 0 0 1
The weight vectors are ξ 1, ξ 2 , and ξ 3 (with covariant, or lower, labels) carrying
the weights (− 1, 0), (1, − 1), and (0, 1). The highest weight is (0, 1).
This method of finding the fundamental representations, by repeated applica-
tions of the lowering operators on the highest-weight vector, can be used to find
any other simple representations. Some more examples are given in Figs.5.5–5.6
and Problems 5.7–5.8. Each such representation is reducible under the action of
any sub Lie algebras sα , the decomposition being of the form V = ⊕ Wi, where
the Wi ’s are sα -irreducible representations equivalent to complete α-strings of
weights. Another approach to finding higher-dimensional representations is to
form tensor products of lower-dimensional irreducible representations and de-
compose them (since they are fully reducible) into irreducible subspaces by the
symmetry arguments, as illustrated in the following examples. This type of con-
struction is explained in more detail in Chap. 7 and Appendix A of this book.
E XAMPLE 3: π1,0 ⊗ π1,0, π1,0 ⊗ π0,1, and π0,1 ⊗ π0,1. Degree-two tensor products
of V ≡ V (1,0) and V ≡ V (0,1) provide us with the simplest examples. Recall
that V (1,0) has the weight vectors ξ i with weights ε 1 = (1, 0), ε 2 = (− 1, 1), and
ε 3 = (0, − 1); and V (0,1) has the vectors ξ i with weights − ε 1, − ε 2, − ε 3.
(i) V ⊗ V contains the weights ε i + ε j , which are (2, 0), (0, 1), (1, − 1), (− 2, 2),
(− 1, 0), and (0, − 2), as well as (0, 1), (1, − 1), and (− 1, 0). It is clearly not ir-
reducible. Let ξ i ξ j be any element in V ⊗ V , then it may be written as ξ i ξ j =
1 i j j i 1 i j j i
2 ( ξ ξ + ξ ξ ) + 2 ( ξ ξ − ξ ξ ) , and so it splits into a second rank contravariant
tensor Sij = ξ i ξ j + ξ j ξ i and a covariant vector ηk = ekmn ξ m ξ n . Since Sij contains
the highest-weight vector ξ 1 ξ 1 with weight (2, 0), from which applications of F1
and F2 can produce the remaining five components, or weight vectors, it is just
172 CHAPTER 5. LIE ALGEBRA sl (3, C )
α(2 ) 0, 2
−2, 2 0, 1 2, 0 α(2 )
r KAA r r r
A A A KA A
A A
A A - A
A
AU r AU r α(1) −1, 1 r AU r 1, 0
A
−1, 0 A 1, −1 A A - α(1 )
A A
AU r r r AU r
0, −2 −2, 0 0, −1 2, −2
Figure 5.5: The sextuplet representation π2,0 and its dual π0,2 . The multiplets of the sub
algebras s1 and s2 can be identified from the displayed (in frames) weights.
α(2 )
−1, 2 K
A t t 1, 1
A A
A A
t AU f
t AU t -α(1)
−2, 1 A A 0, 0 2, −1
A A
A AU t
−1, −1U t 1, −2
Figure 5.6: The adjoint (octet) representation π1,1 . Representation space V decomposes
into ⊕4s=1 Ws , where Ws are s1 (or s2 ) irreducible representations.
3, 0 0, 3
r r r r r
A A A A
AU r AU r AU r r AU r
A A A A
AU r AU r r AU r AU r
A A A A
AU r 0, −3 −3, 0r AU r AU r AU r
Figure 5.7: The representation π3,0 = 10 and its dual π0,3 = 10.
they all appear in triangular (rather than hexagonal) patterns, and are evidently
irreducible. This observation has a generalization: The weights of the symmetric
powers Symn V (1,0) and Symn V (0,1) occur with multiplicity 1, and are all irreducible:
Symn V (1,0) = V (n,0), and Symn V (0,1) = V (0,n). In turn V (n,0) and V (0,m) are the
foundation for constructing arbitrary irreducible representations V (n,m) . In fact,
V (n,0) ⊗ V (0,m) is completely reducible, decomposable into a direct sum. To sum-
marize:
(a) (b)
Y Y.....
K0 K+
...
... ...s̄ 1 .........................................
. ...
. .
.
.
... .... ...
1
.. . .
u
.. . .
d ....................................................................
. .. .
. ..
3 ...
... ....
. ..
... ....
. .
η
.. . ..
... ... 8.. ...
−............................................................................. π+
.. .... .
.
....
..
....
. 0 π
. ..0
.. .. ... ..
..
.. .....
..
.. .... .. π .
ū
. . ..
........................................................................
d̄
. .
... .
. .. . .. .
.. .... .. . .. .
− 23 s
... .. ... . .. .
−1
..
K−
.......................................
K̄0
I3 ........
I3
− 12 0 1
2 −1 − 12 0 1
1
2
(c) (d)
Y..... Y......
n p ∆.− 0 + ++
....................................∆
................................∆ ....................................∆
.
1 .........................................
. ... . 1 .. . ... . ..........
.
. . . .
.. .. . . . .
.. .
. .. . . . .. . .
.. . ... .. . .
. .. . . .. .
.. . .. . ..
. . ∗0
. . ..
Λ .. .
−.............................................................................
.. . .
.Σ
. .
Σ+ Σ∗−......................................................................................Σ∗+
.
. .. .
0 Σ 0
. . . .
.. . ..0 Σ . ..
.. . .. . .. . . .
.. .
... . .. . . ..
. . .. . .. . .. .
..
. .
∗− ............................................. ∗0
.. . .. . .
−1 Ξ−
.........................................
Ξ0 −1 Ξ Ξ
.. .
.. .
I3
........ .. .
−1 − 12 0 1 .. .
2 1 .. .
−
−2 ..
Ω
I3
..........
− 32 −1 − 12 0 1
2 1 3
2
Figure 5.8: sl(3, C) in particle physics: (a) Quarks, antiquarks in the fundamental repre-
sentations. (b) Pseudo-scalar mesons as quark-antiquark states in an octet representation.
(c) Light baryons as tri-quark states of an octet. (d) Heavy baryons in a decuplet. Here, the
√
Cartan basis consists of the isospin I3 = 1/2λ3 and the hypercharge Y = 1/ 3λ8 . Adapted
from Elementary Particles and Their Interactions by Ho-Kim Quang and Pham Xuan-Yem.
Springer (1998).
3 ⊗ 3 = 3 ⊕ 6, 3 ⊗ 3 = 3 ⊕ 6, 3 ⊗ 3 = 1 ⊕ 8.
Note that 6 = Sym2 3 is the symmetric part of 3 ⊗ 3. Take as another example the
order-3 tensor product of 3, which can be reduced in two successive steps to
3 ⊗ 3 ⊗ 3 = 3 ⊕ 6 ⊗ 3 = 1 ⊕ 8 ⊕ 8 ⊕ 10
The singlets in 3 ⊗ 3 and 3 ⊗ 3 ⊗ 3, though congruent, are not the same. In the
first case, it is a symmetric scalar product of vectors, ξ i ξ i ; in the second, it is an
antisymmetric scalar product eijk ξ i ξ j ξ k . Similarly, the octets 8 in the two tensor
products are completely unrelated.
(b) Flavor and color symmetries. Most strongly-interacting particles can be re-
garded as being made up of several quarks of three kinds (or flavors) called u, d
and s. (Actually there are more flavors, but these are the lightest ones.) u, d, s
PROBLEMS 175
Problems
5.1 (Matrices Eij ) The matrices Eij have a single non-zero entry, equal to 1, at row
i, column j, i.e. ( Eij )ab = δia δjb . Rewrite the basic elements of sl (3, C ) in terms of
3 × 3 matrices Eij , and check their commutation relations.
5.2 (Spin and hypercharge) In physics, the CSA of su (3)C is often defined with a
different basis: tz = 1/2t1 and y = 1/3t1 + 2/3t2. In the standard representation,
√
Tz = π ( tz) = 1/2 λ3 and Y = π ( y) = 1/ 3 λ8, where the λi are the well-known
3 × 3 Gell-Mann self-adjoint matrices. The remaining elements λi are related to
the elements used in the chapter in the following way: λ1 ± iλ2 = 2t± , λ6 ± iλ7 =
2u± , λ4 ± iλ5 = 2v±, where t+ = e1, t− = f 1, u+ = e2, u− = f 2, v+ = e3, and
v− = f 3.
176 CHAPTER 5. LIE ALGEBRA sl (3, C )
6.1 Basics
6.2 Roots and Root Spaces
6.3 String of Roots
6.4 System of Roots
6.5 Cartan Matrix
6.6 Dynkin Diagram
6.7 Classification of Simple Algebras
This chapter generalizes our previous study of the simple Lie algebras sl (2; C )
and sl (3; C ) to all other complex simple or semisimple Lie algebras. After a brief
review of some of the concepts already seen, we shall discuss the basic proper-
ties of the semisimple (and simple, the limiting case) Lie algebras (in particular
their roots and root spaces). The essence of the structure of a semisimple Lie
algebra is encapsulated in a fundamental system of roots, or equivalently a Car-
tan matrix, or a planar diagram, called the Dynkin diagram. This representation
is in fact the tool used to show that there are only a finite number of classes of
simple Lie algebras over the complex numbers, which can thus be identified and
systematically studied. The main point of this chapter is to present this elegant
and important result in mathematics. It was first obtained by Wilhelm Killing
in 1888–1890, and made more rigorous by Élie Joseph Cartan in 1894 (who also
classified semisimple real Lie algebras). Important contributions were later made
by Hermann Weyl, Eugene Dynkin, Claude Chevalley, and Jean-Pierre Serre.
6.1 Basics
To begin, at the risk of being repetitive, we will recall (Chapters 3–5) the defini-
tion of simple and semisimple Lie algebras over complex numbers, and of their
Cartan sub algebras. Next, we will discuss the roots and root spaces of complex
semisimple Lie algebras, with the aim of applying their properties to uncover
177
178 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
their existence and structure. For a more detailed and more rigorous treatment,
the reader is referred to [Ja] and [Sa].
Let g be a complex Lie algebra. A sub Lie algebra h of g is called an ideal in g
if it is invariant with respect to g, that is, if [ h, g] ∈ h for all h ∈ h and all g ∈ g. A
Lie algebra g is simple if dim g ≥ 2 and if it contains no proper ideal. In other words,
it is its own ideal: [g, g] = g. A Lie algebra g is semisimple if it has no nonzero abelian
ideal. A Lie algebra is semisimple if only if it can be written as a direct sum of
simple Lie algebras. In particular if this sum reduces to a single term, then it is
simple. So, semisimple Lie algebras include every simple Lie algebra, but also
many others.
In a semisimple algebra of dim g ≥ 2, every element xi (called generator) of
a given basis { xi } of g has a nonzero commutator with some other element xj .
The structure constants then carry a lot of information, which is encapsulated in
a criterion for semisimplicity (referred to as the Cartan criterion): A Lie algebra g
is semisimple if and only if its Killing form is non-degenerate. Put it in another way,
to show that g is semisimple it suffices to check that there are no bases { xi } for g
in which the Killing form K ( xi, xj ) = κij is singular, i.e. in which the determinant
of [ κij ] is zero. That is why throughout this section we need just require semi-
simplicity, rather than the more restrictive condition of simplicity.
We have seen in previous chapters that sl (2; C ) has a special matrix H = 2J3,
and sl (3; C ) has two commuting matrices T1 and T2 that play a key role in their
study. Now, for a general complex semisimple Lie algebra g, we shall follow the same
strategy and seek among its elements those that can play a similar role. That is,
we want to find an abelian subalgebra h that acts diagonally on one faithful (and
so on any) representation of g and that contains the greatest possible information
about g. Such a set of elements is defined as follows:
The dimension of the Cartan subalgebra (dim h) is called the rank (`) of the
algebra, and so ` = dim h. A semisimple Lie algebra always possesses a Cartan
subalgebra; it is essentially unique in the sense that any two CSA’s are equivalent
by automorphism and so must have the same dimension. For example, in su (2)C,
the CSA could be defined by J3, or J2, or J1; and in sl (3; C ), it could be chosen to
be the pair of matrices T1 and T2, or equivalently Tz (charge) and Y (hypercharge),
or any other pair of commuting independent matrices in sl (3; C ).
Just as all h in h commute, so do all adh, since ad[ h, h 0] = [adh, adh0 ]. Since all
adh’s commute, and each adh is diagonalizable (by construction), the adh’s are
simultaneously diagonalizable (from linear algebra). This leads us naturally to a
study of the simultaneous eigenvalue problem for all the adh’s, with h ∈ h, as a
way to classifying the remaining elements of g.
6.2. ROOTS AND ROOT SPACES 179
Definition 6.2 (Roots). Let g be a complex semisimple Lie algebra with Cartan subal-
gebra h; let h act on g by the adjoint representation such that
adh z = α ( h) z, h ∈ h, z ∈ g. (6.1)
α ( h) is a complex linear functional of h (so α is in h∗ , the space dual to h). The nonzero
eigenvalues in (6.1) are called the roots of g with respect to the CSA h. If α 6 = 0, the
eigenvector z is called a root vector corresponding to the root α, and the space of all such
vectors for a fixed α is the root space to α, called gα .
The set of all the roots, α, β, . . . , will be denoted by ∆, and thus ∆ is a finite
subset of h∗ that does not include { 0} , although h itself is the eigenspace for the
action of adh corresponding to the eigenvalue 0. To include { 0} in our discussion,
it is useful to adopt the notations ∆0 = ∆ { 0} and g0 = h. So the action of adh
S
The action of adh preserves each subspace in the direct sum, acting on each term
by scalar multiplication of 0 or α (by (6.1)). In addition, for any zα ∈ gα , z β ∈ gβ
with α, β ∈ ∆0, we have adh zα , z β = α ( h ) + β ( h ) zα , z β , which means that
Thus, adgα sends any gβ to a root space or h, or else to 0, preserving the direct-sum
decomposition: We then say that the Cartan decomposition is adg-invariant.
One can give any space of linear operators a bilinear trace form as inner prod-
uct; for Lie algebras g, a convenient definition is the Killing form: K ( x, y) =
Trg (ad( x) ad( y )) for any x, y ∈ g. As before, we will also use the symbol ( x : y )
for K ( x, y). When two elements x and y of g satisfy ( x : y ) = 0, we say that they
are orthogonal to each other (relative to the Killing form) and write x ⊥ y.
Many properties of the roots of complex Lie algebras are geometric in nature
and arise from the concept of bilinear trace form K. They will be presently dis-
cussed and numbered P1, P2,. . . for reference. Cf. [Ja] Chap. IV.
P1. gα ⊥ gβ relative to the Killing form for any α, β ∈ ∆0 such that α + β 6 = 0.
This can be seen by examining the process of calculating Tr (ad( xα )ad( y β ))
through the computation of xα , [ y β, z ] for given xα ∈ gα , y β ∈ gβ , and any
z ∈ gν , and identifying the nonzero coefficients
of z in the resulting sum as the
contributions to Tr (ad( xα )ad( y β )). As gα , [gβ , gν ] ⊂ gα+ β+ν by (6.3), it is clear
that no terms proportional to z can exist unless α + β = 0. So ( xα : y β ) = 0 if
180 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
h ⊕ gα
K= h ∗ 0
⊕ gα 0 ∗
h gα g− α gβ
h 0
K= gα 0 0 ∗ 0
g− α ∗
gβ 0
P6. The number of linearly independent roots of g is equal to its rank `, or dim h.
h∗ is spanned by the (non-zero) roots, hence its dimension is given by the
largest number of linearly independent roots. The assertion that dim h∗ = dim h
(= `) follows from the isomorphism of dual vector spaces, which says to each
nonzero h ∈ h there exists a nonzero α ∈ h∗ . Let us assume, in the contrary, that
there exists a non-zero h ∈ h, such that α ( h) = 0 for every root α (so that the
subspace of h∗ spanned by the roots has dimensionality less than `). Then (6.4)
implies ( h : k ) = 0 for every k ∈ h and h 6 = 0, which contradicts P4.
We recall (from Chap. 5, Sec. 5.1) a definition of the inner product on h∗ . For
every element $ of h∗ , we can define a unique element h$ of h such that
def
( h $ : k ) = $ ( k) , for all k ∈ h. (6.5)
The mapping h∗ 3 $ ↔ h$ ∈ h is an isomorphism from h∗
to h, allowed by the
non-degeneracy of K |h. If $ is a root, h$ is called a root generator of $.
If $, σ are any two elements of h∗ , and h$ , hσ their corresponding elements in
h, then the bilinear form on h∗ is defined by
def
h $, σ i = ( h$ : hσ ) . (6.6)
It follows that h $, σ i = $ ( h σ) = σ ( h$), and so h, i is a non-degenerate, symmetric
bilinear form on h∗ . As there exists a unique non-zero root generator h α for each
α ∈ ∆, we have ( hα : hα ) 6 = 0. Hence ( hα : hα ) = α ( hα) 6 = 0, and so h α, α i 6 = 0
for every (non-zero) α ∈ ∆. This property is referred to as ‘the non-isotropy of the
roots’ relative to h, i.
P7. For any h ∈ h and α ∈ ∆, let xα ∈ gα and x−α ∈ g−α be the root vectors obeying
the equations [ h, xα ] = α ( h )xα and [ h, x−α ] = − α (h) x−α. Then, x±α must also satisfy
[ xα , x−α ] = ( xα : x−α ) hα , (6.7)
where hα ∈ h is the root generator corresponding to the root α.
Just as for sl (3, C ), the proof
runs as follows:
By the invariance of K, we have
for any h ∈ h [ xα , x−α ] : h = xα : [ x−α , h ] = α ( h)( xα : x−α ) (which is 6 = 0 by
P5). Since α ( h) = ( h α : h ) by definition, the equation also reads [ xα , x−α ] : h =
( hα : h )( xα : x−α ). Then (6.7) follows from the non-degeneracy of K on h (which
when applied here tells us that ( h α : h ) 6 = 0 for all h ∈ h).
P8. The root space gα has dimension 1 for every root α ∈ ∆, whereas gkα is zero for each
k = 2, 3, . . . .
Given a root α, let x±α and hα be as defined in P7. Consider a subspace w of g
spanned by hα , x−α and gkα (with k = 1, 2, 3, . . . ), that is, w = Chα ⊕ Cx−α ⊕ k gkα ,
where the summands have the dimensions 1, 1, d α , d2α , . . . , respectively.
By P7, (6.1), and (6.3), w is invariant under x±α and hα , that is to say, ad x±α
and ad hα send every z ∈ w either to 0 or w. So it is meaningful to calculate the
traces of ad x±α and ad hα over w. As the adjoint action of h α ∈ h acts diagonally
on w, we have the trace of ad h α over w given by
Trw (adhα ) = α ( hα)[− 1 + dα + 2d2α + · · · ]. (6.8)
182 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
where dkα = dim gkα ≥ 0. But by (6.7), ( xα : x−α )adhα = [adxα , adx−α ], and so
Tr (adhα ) = 0 since ( xα : x−α ) 6 = 0 (dual vectors). It follows that
α ( hα)[− 1 + dα + 2d2α + 3d3α + · · · ] = 0.
Since α ( hα ) = h α, α i 6 = 0 (non-isotropy of the roots), this can be satisfied only if
dα = 1 (so gα = Cxα ), and d2α = d3α = · · · = 0. By the reflection symmetry of ∆,
we also have d−α = 1 and d−kα = 0 for k = 2, 3, . . . .
In conclusion, for each α in ∆, g±α = Cx±α , and g±kα = 0 if k ≥ 2. If α is a root,
the multiple kα is a root only for k = 1 or − 1.
For each root α ∈ ∆, the elements xα ∈ gα , x−α ∈ g−α and hα ∈ [gα , g−α ] of g
defined in P7 obey the bracket relations
[ hα , xα ] = α ( hα) xα ; [ hα, x−α ] = − α (hα ) x−α ; [ xα , x−α ] = ( xα : x−α ) hα. (6.9)
We see that the adjoint action of h α carries each xα and each x−α into itself, and so
hα , xα and x−α generate a three-dimensional sub Lie algebra of g. Let us change
their normalization, and define:
2 2
h̃α = hα , eα = xα , f α = ξ x−α ; where ξ = . (6.10)
α ( hα) α ( hα )( xα : x−α )
The re-normalized elements obey the bracket relations:
[ h̃α , eα ] = 2eα , [ h̃α, f α ] = − 2 f α, [ eα, f α ] = h̃α . (6.11)
h̃α , eα , and f α form a canonical basis for a three-dimensional subalgebra in g iso-
morphic to sl (2, C ), which we have studied in Chapter 4. It is significant that
whereas hα ∈ h (so also h̃α ) is uniquely determined from each given α ∈ ∆, with
a fixed normalization, α ( h̃α) = 2, there is some arbitrariness in the root vectors:
eα can be replaced by aeα , and f α by b f α , with a, b ∈ C provided ab = 1, without
causing any changes in their bracket relations. For any α ∈ h∗ , let α̃ = 2α/ hα, αi,
then h̃α is related to α̃ in precisely the same way as h α is related to α. (We call α̃ the
coroot and h̃α the coroot generator of the root α.) Thus, we have α̃( h ) = ( h̃α : h )
and h α̃, β̃ i = ( h̃α : h̃ β ). The symmetry between roots and co-roots is captured in
the identity α̃( h β ) = β ( h̃α ). Information about the independent roots suffices in
practice, so we state the result as follows:
P9. For every independent (nonzero) root α of g, there is a three-dimensional simple sub
Lie algebra sα isomorphic to sl (2, C ), defined by the canonical basis h h̃α , eα , f α i together
with the canonical commutation relations (6.11).
E XAMPLE 1: For the simple Lie algebra g = sl(3, C ), one can choose for the CSA
h = { t1, t2 } . There are six roots—two linearly independent α1, α2 plus four others
α3 = α1 + α2, − α1, − α2, − α3. sl (3, C ) admits the decomposition to the direct sum
g = h ⊕ h e1i ⊕ h e2i ⊕ h e3i ⊕ h f 1i ⊕ h f 2i ⊕ h f 3i. The subalgebras s1 = h h1, e1, f 1 i
and s2 = h h2, e2, f 2 i associated with the independent roots α1 and α2 are isomor-
phic to sl(2, C ). It turns out that h i = ti are coroot generators, related to the root
generators hαi by hi = 6hαi . The non-zero Killing forms are ( h i : hi ) = 12 with
i = 1, 2, ( h1 : h2 ) = − 6, ( e j : f j ) = 6 with j = 1, 2, 3. We have the inner products
on h∗ : h αi, αi i = 1/3 for i = 1, 2, and h α1, α2 i = − 1/6, h α1, α3 i = h α2, α3 i = 1/6.
6.3. STRING OF ROOTS 183
With γ ( h̃α) ∈ Z, this tells us that ( h̃α : h̃ β ) ∈ Z , and so, ( h̃α : h̃α ) = ∑γ γ2 ( h̃α ) ∈
N (i.e. non-negative integers). From the relationship between h̃α and hα (in par-
ticular, ( hα : hα )( h̃α : h̃α ) = 4 and 2( hα : h β ) = β ( h̃α )( hα : hα )), it follows that
( hα : hα ), and hence also ( hα : h β ), are both rational numbers for all roots α, β ∈ ∆.
Let h0 be the real subspace of h spanned by the h α ’s, where α ∈ ∆, i.e. the set
of all elements of the form ∑α∈∆ Rhα . If x, y are any two elements in h0 ⊂ h,
then ( x : y ) = ∑αβ (Rhα : Rh β ) is real. Further, ( x : x) = ∑γ ( γ ( x))2 ≥ 0 (since
γ ( x) ∈ R) is equal to zero if and only if γ ( x) = 0 for all γ ∈ ∆, that is if and
only if x = 0 by P4. This shows that: The restriction of the Killing form to h0 is real,
positive-definite, and so h0 is a Euclidean space.
The isomorphism h ↔ h∗ defined by the relation ( h α : h β ) = h α, β i gives
the correspondence between the real vector space h0 and the R-span of ∆, called
∆R ∼ = h∗0 . As ( hα : h β ) defined on h0 is a rational number, so is h α, β i defined on
∗
h0 . Further, as h α, α i = ∑γ h α, γ i2, and h α, γ i is a rational number for α, γ ∈ h∗0 ,
we have h α, α i ≥ 0, and it is zero only if h α, γ i = 0 for all γ, that is if and only if
α = 0, since the roots span h∗ . In conclusion:
The R-span of ∆, denoted ∆R ∼ = h∗0 , is a Euclidean subspace of h∗ with respect to the
symmetric, positive-definite inner product h, i. For any α, β ∈ h∗0 , the product h α, β i is a
rational number (Q), and | α |2 = h α, α i is positive-definite.
In terms of roots, the Cartan integer may be rewritten as
2β ( hα ) 2h α, β i
aαβ = β ( h̃α ) = = , (6.13)
α ( h α) h α, α i
where α, β ∈ h∗0 and hα ∈ h0 . In particular aαα = 2 for all α, but in general
a βα 6 = aαβ and takes one of the values 0, ± 1, ± 2, ±3, by P12. As the roots are
vectors in a Euclidean space, ordinary geometrical notions apply: angles between
vectors can be defined and Schwarz’s inequality can be invoked. Let θ be the
angle (0 ≤ θ ≤ π) between the two roots α and β on Euclidean space h∗0 , then
4h α, β i2
aαβ a βα = = 4 cos2 θ ≤ 4 . (6.14)
h α, αih β, βi
The possible values of aαβ a βα are 0, 1, 2, 3, 4. The value 4 means cos θ = ± 1 corre-
sponding to the trivial cases (θ = 0 or β = α, and θ = π or β = − α). The value 0
corresponds to orthogonality of α and β, or equivalently a βα = aαβ = 0, with no
186 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
restrictions on the ratio | β | /|α|. For the remaining cases, let us assume first that
a βα ≤ 0, or equivalently θ ≥ π/2. Then of the a βα and aαβ , one is − 1 and the
other is − 1, or − 2, or − 3, corresponding to the angle θ = 2π/3, 3π/4, or 5π/6,
and the ratio aαβ /a βα = | β |2/ | α |2 = 1 or 2 or 3 or their reciprocals. For the case
a βα ≥ 0 (θ ≤ π/2), we remove the signs of the a’s and replace θ with π − θ, which
leads to the angles π/6, π/4 and π/3. The accompanying table and Fig. 6.1 show
the four non trivial cases, with roots α and β 6 = ± α.
"" ◦ bb
" ◦ 45◦ 90◦ T 120◦ @ 135◦ ◦
" 30 60 T @ b150
P14. aαβ a βα for any roots α, β 6 = ± α has one of the values: 0, 1, 2, 3. The angle between
two such roots on h∗0 has one of the values: 30◦ , 45◦ , 60◦ , 90◦ , 120◦ , 135◦ , and 150◦ .
Root system. The roots α, β, . . . of a semisimple Lie algebra g wrt CSA h have the
following essential properties:
(i) 2h β, α i/ h α, αi is an integer (called Cartan integer, denoted by aαβ );
(ii) β − aαβ α is also a root; and
(iii) α and r α are both roots only for r = ± 1.
The notions of roots and root spaces may be carried to an abstract level by
defining, without referring to any group or algebra, a root system in a Euclidean
vector space V with respect to a positive-definite inner-product h, i as a finite non-
empty subset of V whose elements have properties (i)–(iii). Thus, ∆, described
above, is an example of a root system in h∗0 . The rank of a root system is the dimen-
sion of V (equal to the rank of g as defined before, ` = dim h). As examples, we
now determine all root systems of rank one and two.
6.4. SYSTEM OF ROOTS 187
E XAMPLE 3: Rank one. There is only one rank-one root system, denoted by A 1.
It consists of two vectors α and − α. The Cartan integers are aαα = a−α,−α = 2,
a−α,α = aα,−α = − 2. It will be identified with the root system of the Lie algebra
a1 = sl (2, C ). This is so because this algebra has a CSA h H i of rank 1, with Lie
brackets [ H, E±] = ± 2E±, implying a pair of root ± α such that ± α ( H ) = ± 2.
Here, the root system lies on h∗0 = R H.
A1 q -
E XAMPLE 4: Rank two. There are four distinct root-systems of rank two shown in
Fig. 6.2. Their labels are the names of the root systems of the corresponding Lie
algebras in the Cartan–Killing classification. A 1 ⊕ A 1 corresponds to a decom-
posable algebra, as we will see later. The other cases (A 2, B2, G 2) give simple Lie
algebras; they satisfy conditions (i), (ii), (iii) of root systems. There are no other
systems of rank two.
β
6
A1 ⊕ A1 q -α | β | : | α | arbitrary ratio;
α⊥β
?
β
KA | β | : | α | = 1;
A2 Aq - α
A angle between adjacent vectors
UA = π/3
β √
I 6
@
B2 @q - α | β | : | α | = 2;
@ angle between adjacent vectors
?@R = π/4
6
β H
YH KA
* √
G2 HAq - | β | : | α | = 3;
H α
AUA HH
angle between adjacent vectors
j
= π/6
?
Figure 6.2: The four root-systems of rank two; α and β are the two simple roots.
In any vector space, it is always practically useful to have a basis to work with.
It is also the case with h∗0 we are dealing with here.
Let ( α1, α2, . . . , α` ) be a basis for h∗ consisting of independent roots (we al-
ready know ` = dim h∗ by P6). We want to check that every root β can be written
as β = ∑i qi αi, with qi ∈ Q. If this is true, then h β, α j i = ∑i qi h αi, α j i for any
188 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
`
2h β, α ji 2 h α i, α j i
h α j, α j i
= ∑ j j qi , j = 1, 2, . . . , `. (6.15)
i= 1 h α , α i
2 h α i, α j i 2`
det = det [h αi, α j i] 6 = 0 ,
h α j, α j i ∏ j h α j, α j i
Definition 6.3 (Simple roots). Relative to a given ordering in h∗0 , a root α is said to be
simple (or fundamental) if it is positive and cannot be written as the sum (β + γ) of
two positive roots (β and γ).
P16. Let Σ be the set of all simple roots of g relative to a fixed ordering in h∗0 . It has the
following properties:
(a) For any distinct simple roots α, β ∈ Σ , α − β is not a root.
(b) For any distinct simple roots α, β ∈ Σ , we have h α, β i ≤ 0.
(c) The set Σ is a basis for h∗0 .
(d) For every positive root β not in Σ, there exists an element α ∈ Σ such that β − α
is a positive root.
6.4. SYSTEM OF ROOTS 189
As we assume all the roots up to level n are known, we know the number q.
Then we can calculate the number p = q − ∑i a ji k i , and hence roots of levels
n + 1, . . . , n + p. There are new (positive) roots if and only if p > 0; they are
β + α j , . . . , β + pα j. The following examples illustrate the procedure.
E XAMPLE 9: Lie (A 2) = sl (3, C ). We have aii = 2, a12 = a21 = − 1. The roots of
level one are the simple roots α1 and α2. Consider the strings Sα2 ( α1) and Sα1 ( α2).
As α1 − α2 is not a root, q = 0 in both cases. For β = α1, the only non-zero
coefficient is k 1 = 1, and so in Sα2 ( α1), we have p = q − a21 k 1 = −(−1) · 1 = 1,
which means that Sα2 ( α1) consists of α1 and α1 + α2. We obtain similarly p = 1
for Sα1 ( α2), and so this string consists of the roots α2 and α2 + α1. As 2α1 and 2α2
are not roots, the sum α1 + α2 is the only root at level two.
To examine level three, consider the string Sα1 ( γ ), where γ = α1 + α2. Since
γ − α1 = α2 is a positive root, but not γ − 2α1, we know that q = 1. We calculate
p = q − ( a21k 1 + a22 k 2 ) = 1 − (− 1 + 2) = 0. We can do a similar calculation for
Sα2 ( γ ). The result shows that there are no roots at level three or higher.
. Conclusion: The roots of sl (3, C ) are ± α1, ± α2, ±( α1 + α2 ).
√
E XAMPLE 10: Lie (G 2) = g2 . Calling α2 the long root (| α2| = 3| α1|), the Cartan
integers for g2 are a12 = − 3, a21 = − 1. With α1 and α2 known as level-one
positive roots, we consider the strings of roots based on them to obtain the roots
at the next level. Since α1 − α2 is not a root, q = 0 for both strings Sα j ( β ). For the
string Sα2 ( α1), we have β = α1 (so k 1 = 1), we have p = 0 − a21k 1 = 1, and so
p = 1, and
S α 2 ( α 1 ) : α 1, α 1 + α 2 .
192 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
0, 1
Y
H
H
6 HH
-
−1, 1
1, 0
- H-H
α2 −3, 2
YH
H KA * 3, −1
H
HHA
HA - 2, −1 α1
H
s
A HH
A H
AU
HH
j
Figure 6.3: Path in h∗0 tracking the positive roots of g2 . The coordinates of the simple roots
are given by the Cartan matrix column entries : α1 = (2, −1) and α2 = (−3, 2).
hi, ei , f i obey, just as in (6.11), the Lie brackets [ h i, ei ] = 2ei , [ hi, f i ] = − 2 f i, and
[ ei, f i ] = hi . But we also have for i 6 = j: [ ei, f j ] = 0 since αi − α j is not a root; further
[ hi, e j ] = [ h̃αi , eα j ] = α j ( h̃αi ) eα j = aij e j ; and finally [ hi, f j ] = − α j( h̃αi ) f j = − aij f j .
(Note: no summation over repeated indices here and in the following equations.)
Thus, we have for i, j = 1, 2, . . . , `:
hi , h j = 0
hi , e j = aij e j
(6.17)
hi , f j = − aij f j
ei , f j = δij hi.
The set { h1, h2, . . . , h` } span the CSA h. The 3` elements ei , f i , hi do not span
the whole Lie algebra g, but generate g in the following sense: We can write
each positive root β as a sum of simple roots in the form β = αi1 + αi2 + · · · +
αik such that each partial sum αi1 + αi2 + ·h· · + αim , 1 ≤ m ≤ k, isi a root. In
such a representation, let us define e β = eik , [ eik−1 , . . . , [ ei2 , ei1 ] . . . ] and f β =
h i
f ik , [ f ik−1 , . . . , [ f i2 , f i1 ] . . . ] for each positive root β. Then the set
{ hi (1 ≤ i ≤ `); e β , f β ; with β ∈ ∆+ }
forms a basis for g (dim g = ` + 2| ∆+ | = ` + | ∆ |). The multiplication table for the
elements of g in this basis is rational, and is completely determined by the Cartan matrix.
The following example shows how this is done.
E XAMPLE 11: Rank-two Lie algebra sl (3, C ) has the set of positive roots ∆+ con-
sisting of α1, α2 and α3 = α1 + α2. Take the FSR Σ = { α1, α2 } , and the CSA
h = { h1 = h̃α1 , h2 = h̃α2 } . To ± α1 correspond the root vectors e1 and f 1; and
to ± α2, the root vectors e2 and f 2. Finally, we associate with ± α3 the vectors
e3 = [e1, e2 ] and f 3 = − [ f 1, f 2 ] (conventional sign, an unimportant detail). So a
complete basis for the algebra consists of h 1, h2, e1, f 1, e2, f 2, e3, f 3. The six ele-
ments hi , ei , f i with i = 1, 2 obey (6.17) with the Cartan integers a11 = a22 = 2 and
a12 = a21 = − 1. We just need to consider relations involving e3 = [e1, e2 ] and
f 3 = − [ f 1, f 2 ]. For hi = h1, or h2, we have
[e1, f 3 ] = a12 f 2 = − f 2,
[e2, f 3 ] = − a21 f 1 = f 1,
[ f 1, e3 ] = − a12e2 = e2,
[ f 2, e3 ] = a21 e1 = − e1.
194 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
Finally, using these relations, we get [e3, f 3 ] = −( a21h1 + a12h2 ) = h1 + h2. The
results of these calculations agree with Table 5.1 of the last chapter. Together with
Example 9, this example illustrates that a set of simple roots (FSR) determines all
the roots of a Lie algebra, and so defines the Lie algebra itself.
Figure 6.4: Basic Dynkin links. Solid circles (•) denote the shorter roots.
For each FSR = { α1, α2 } (Example 4), or associated Cartan matrix (Sect. 6.5),
draw a two-point diagram according to the following information:
A 1 ⊕ A 1: a12 = 0 7 → unconnected vertices of equal weights.
A2: a12 a21 = 1 7 → vertices of equal weights connected by a line.
B2 : a12 a21 = 2 7 → vertices connected by two lines; | α1|2 = 2| α2|2.
G2: a12 a21 = 3 7 → vertices connected by three lines; | α2|2 = 3| α1|2.
Conversely, a (connected) Dynkin diagram determines the associated FSR, or
the Cartan matrix. From a given diagram we obtain for two vertices i, j the num-
ber of connecting lines n, which is equal to aij a ji . If n = 0, it means aij = a ji = 0. If
6.6. DYNKIN DIAGRAM 195
n = 1, we must have aij = a ji = − 1. If n > 1, the two roots have unequal lengths.
Supposing α j is the long root (
), the relation aij /a ji = | α j |2/ | αi|2 gives us the
Cartan integers a ji = − 1 and aij = − n. Applying this argument to every part of
the diagram we obtain the corresponding Cartan matrix. If the weight | αi|2 for
√
every vertex is also known, then one can calculate h αi, α j i = − 21 n | αi | | α j| for
i 6 = j, n = 1, 2, 3. In this way, the Dynkin diagram determines the associated FSR
(up to similarity transformations in h∗0 ). So, to find all the FSRs one just has to
draw all the ‘allowed’ Dynkin diagrams.
E XAMPLE 12: b3 . Suppose we are given the following Cartan matrix (for a Lie
algebra called b3 ), what is the corresponding Dynkin diagram?
2 −2 0 1 2 2
A [b3 ] = − 1 2 − 1 ⇐⇒ u e e
0 −1 2 α1 α2 α3
The product a12 a21 = 2 and the ratio a12 /a21 = 2 tell us that the roots α1 and
α2 are connected by two lines and that | α2| > | α1|. The product a23 a32 = 1 and
the ratio a23 /a32 = 1 imply that α2 and α3 are of equal lengths and joined by one
line. Finally, a13 a31 = 0 means that α1 and α3 are not directly joined.
One can also trace the reversed path from the diagram back to the matrix.
From the diagram, we know that of the three roots, α1 is the shortest, whereas
| α2| = | α3|. Furthermore the numbers of lines yield 2 = a12 a21, 1 = a23 a32 and
0 = a13 a31, from which one deduces the Cartan integers a12 = − 2, a21 = − 1,
a23 = a32 = − 1, and a13 = a31 = 0, and hence the Cartan matrix A [b3 ].
In the classification problem, we are concerned with the basic building blocks,
which for systems of roots means the indecomposable fundamental systems of
roots. An FSR Σ = { αi, . . . , α` } is said to be indecomposable if it is impossible
to partition it into two non-overlapping subsets Σ 0 and Σ 00 such that h αi, α j i = 0,
or aij = 0, for every αi ∈ Σ 0, α j ∈ Σ 00 , which means the associated Cartan matrix
cannot be written in a block-diagonal form. On the Dynkin diagram, the equiv-
alent condition is connectedness, under which there exists a sequence αi1 , . . . , αi`
such that every two consecutive points are joined by at least one edge; i.e. there
are no separate pieces, or subdiagrams. Finally, the indecomposability of an FSR
implies and is implied by the simplicity of the corresponding Lie algebra. See
[Ja] pp. 127–128. And so, the problem of finding all the simple Lie algebras is the
same as that of finding all the indecomposable FSR’s on a Euclidean space, in turn
equivalent to finding all the connected Dynkin diagrams. Formulated in this way,
the problem is essentially geometric, if one temporarily ignores the weights of the
vertices.
The defining conditions for FSRs in Definition 6.4 — (i) h α, αi > 0, (ii) aij ≤ 0
for i 6 = j, and (iii) aij a ji = 0, 1, 2, 3 for i 6 = j — severely limit the number of such
systems and the associated Dynkin diagrams (which are then said to be allowed).
Those conditions, translated into geometric terms, imply the following general
rules for the Dynkin diagrams (see [Ja], [Sa]):
196 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
• Rule 1. An allowed diagram has more vertices than links, and cannot contain
closed polygons.
• Rule 2. In an allowed diagram, no more than three lines can leave a vertex.
• Rule 3. If a diagram Γ containing a simple A 2 chain is allowed, the diagram Γ0
obtained from it by shrinking the A 2 chain to a point is also allowed. Conversely,
if Γ is not allowed, neither is Γ0 .
C OMMENTS . A link is what connects two vertices; a link may have one, two, or
three lines (called, respectively, A 2, B2 and G 2 links; see Fig. 6.4). Rule 1 is needed
to prove Rules 2 and 3. A simple A 2 chain referred to in Rule 3 is a diagram (or
any part of it) in which each vertex is connected to the next by an A 2 link (i.e. a
single line). These rules will be examined in Problem 6.3.
We will now apply these rules to find the allowed connected diagrams.
(1) A simple A 2 chain of any length (as in Fig. 6.6(a)) does not break any rules,
and so yields an allowed connected diagram.
(2) The basic G 2 link shown in Fig. 6.4 cannot be part of any larger graph,
because any additional lines leaving either vertex would violate Rule 2 . The only
allowed configuration containing a G 2 link is:
e e
(3) By Rules 2 and 3 the graphs on the left-hand side of Fig. 6.5, all with four
lines leaving a vertex, are not allowed, and so neither are the corresponding ex-
panded graphs on the right-hand side, with a simple A 2 chain inserted, either as
stand-alone diagrams or as parts of larger diagrams. This means that a connected
allowed diagram cannot contain more than one B2 link (as in (a)), or more than
one branch (as in (b)), or simultaneously a B2 link and a branch (as in (c)).
(a) e e e e e e e ··· e e e
e e e e
(b)
Q
Q e QQe e e ··· e
e
eQ Qe e Q
Qe
e e
(c) e
e e e e e ··· e
e
Q
Qe Q
Qe
Figure 6.5: Non allowed configurations. Each graph in the left column is obtained by
shrinking to a point the intermediate one-line linked chain of the corresponding graph in
the right column.
(4) At this point, we have reduced the possible candidates to four basic con-
figurations shown in Fig. 6.6. An allowed Dynkin diagram must have one of the
following configurations:
6.6. DYNKIN DIAGRAM 197
√
(a) f f f··· f f
√
(b) f f
(c)
α1 α2 α p−1 α p βq βq−1 β2 β1
f f f. . . f f f f. . . f f
fγ1
(d) fγ2
fγr−1
f f f f f f f f f
α1 α2 α p−1 δ βq−1 β2 β1
Figure 6.6: Allowed general configurations. Diagrams (a) and (b) pass all possible tests;
but diagrams (c) and (d) are still subjected to further restrictions, as described in the text,
where we will make use of the vertex labels displayed here.
Configuration Fig. 6.6 (a) is subject to no other restrictions, and any number of
linked vertices is allowed, which can be checked by displaying the corresponding
non singular Cartan matrix. Configuration Fig. 6.6 (b) is the only one allowed
with a G 2 link, as already discussed.
Configurations Fig. 6.6 (c) and Fig. 6.6 (d) cannot contain any other double-line
or branching because of the restrictions discussed in (3) p. 196. Whether there are
further restrictions or not must be checked (e.g., by verifying the non-singularity
of the Cartan determinant, or the positivity of the inner product; see Problem 6.4).
It turns out, in fact (cf. Comments p. 198), that the following rules hold:
(i) From Fig. 6.6 (c): configuration in Fig. 6.7 (c.1) is not allowed; but are al-
lowed configurations in Fig. 6.7 (c.2) (as a chain with an unlimited number of A 2
links) and in Fig. 6.7 (c.3) (as a finite chain).
(ii) From Fig. 6.6(d): configuration in Fig. 6.8(d.1)–(d.2) are not allowed as
subdiagrams; but are allowed configurations in Fig. 6.8 (d.3) (as an unlimited
chain) and in Fig. 6.8 (d.4)–(d.6) (as finite chains).
In summary, an allowed indecomposable FSR must correspond to a connected
Dynkin diagram that is either a member of one of the three unlimited chains in
Figs. 6.6 (a), 6.7 (c.2), and 6.8 (d.3); or one of the five configurations in Figs. 6.6 (b),
198 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
(c.1) f f f f f ×
√
(c.2) f f··· f f f
√
(c.3) f f f f
Figure 6.7: Chains with one B2 link: (c.1) is not allowed; (c.2)–(c.3) are.
6.7 (c.3), and 6.8 (d.4)–(d.6). The systems that contain just simple A 2 links (also
said to be simply laced) are shown Figs. 6.6 (a) and 6.8 (d.3)–(d.6). The fact that for
every connected allowed Dynkin diagram there exists a corresponding FSR, and
hence a simple complex Lie algebra, will be proved by explicit construction, as
we will see in the next section.
E XAMPLE 13: Consider Fig. 6.7 (c.1) as an example, in which one assigns relative
weights 1, 1, 2, 2, 2 to the vertices in order from left to right. The matrix with
entries given by the Cartan integers calculated from the diagram is
2 −1 0 0 0
−1 2 −2 0 0
0 −1 2 −1 0
.
0 0 −1 2 −1
0 0 0 −1 2
If all entries in the first row (r1) are replaced by the numbers obtained by the
combination of rows r1 + 2 ( r2) + 3 ( r3) + 2 ( r4) + r5, then we will have a matrix
with one row with all entries equal to zero. Hence the matrix is singular, and the
associated Dynkin diagram not allowed.
C OMMENTS . We now sketch the arguments (found in [Ja] pp. 132–133), based
on the general properties of the simple roots, that lead to the results given in
Figs. 6.7–6.8. In Fig. 6.6 the nodes are labeled by the simple roots ui , all normed
to one, h ui, ui i = 1, and spanning, in each case, a Euclidean space h∗0 . With i 6 = j,
√ √
one has 2h ui, u j i = − 1, − 2, − 3 for resp. A 2, B2 , G 2 links.
p q
Consider first the case of Fig. 6.6(c) , and define φ = ∑1 iαi and ψ = ∑1 jβ j ,
with p, q ≥ 1, which have norms | φ |2 = p ( p + 1) /2 and | ψ |2 = q ( q + 1) /2 and in-
ner product h φ, ψ i2 = p2 q2 /2 . By Schwarz’s identity (|h φ, ψi|2 < | φ |2| ψ |2 where
φ ∦ ψ), we have 2pq < ( p + 1)( q + 1), or equivalently ( p − 1)( q − 1) < 2 . Hence
the only possible positive integers p, q are: (i) p = 1 with q arbitrary (or q = 1, p
arbitrary), and (ii) p = 2, q = 2 . They are represented resp. by the diagrams in
Fig. 6.7 (c.2) and (c.3), which are allowed. But Fig. 6.7 (c.1) is not allowed, because
it has p = 2, q = 3.
6.6. DYNKIN DIAGRAM 199
f
(d.1) f f f f f ×
f
(d.2) f f f f f f f f ×
f √
(d.3) f f··· f f
H
Hf
f
√
(d.4) f f f f f
f
√
(d.5) f f f f f f
f
√
(d.6) f f f f f f f
Figure 6.8: Configurations with one branching (d.1)–(d.2) are not allowed; but (d.3)–(d.6),
in which one of the branches has length one, are allowed.
p− 1 q−1
For the case of Fig. 6.6(d), define φ = ∑1 iαi , ψ = ∑1 jβ j and χ = ∑r1−1 kγk,
where p, q, r ≥ 2. The vectors φ, ψ, χ lie in mutually orthogonal subspaces of h∗0 ,
which do not contain the (normed) simple root δ. This root, on the other hand, is
linked to α p−1 , βq−1 and γr−1 . If θ1, θ2 , θ3 are the angles between δ and the vectors
φ, ψ, and χ, then we have the inequality relation cos2 θ1 + cos2 θ2 + cos2 θ3 < 1.
Now, noting 2h ui, ui+1 i = − 1 for A 2 links, we get simultaneous conditions on
p, q, r:
cos2 θ1 = h φ, δ i2/ h φ, φ i = ( p − 1) /2p,
cos2 θ2 = h ψ, δi2/ h ψ, ψi = ( q − 1) /2q,
cos2 θ3 = h χ, δi2/ h χ, χi = ( r − 1) /2r.
This leads to the restriction p −1 + q−1 + r−1 > 1. As p, q, r are interchangeable,
we may assume p ≥ q ≥ r (≥ 2). It follows that r = 2 is the only possible value
of r, and, finally, there remain four configurations: (i) r = q = 2, with p arbitrary,
and (ii) r = 2, q = 3, with p = 3, 4, 5. They are represented resp. by the diagrams
in Figs. 6.8 (d.3), (d.4), (d.5), and (d.6). The diagrams in Figs. 6.8 (d.1) and (d.2),
which have r = q = p = 3 and r = 2, q = 3, p = 6, are disallowed.
200 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
B` e e··· e e u ` = 2, 3, 4, . . .
C` u u··· u u e ` = 3, 4, 5, . . .
e
D` e e··· e
e ` = 4, 5, 6, . . .
H
He
G2 u e `=2
F4 u u e e `=4
e
E6 e e e e e `=6
e
E7 e e e e e e `=7
e
E8 e e e e e e e `=8
The restrictions on the values of the parameter ` in the unlimited series are
intended to avoid double counting, or to remove non-allowed diagrams. A 1, B1,
and C1 correspond to the same single-vertex diagram. C2 and B2 are represented
by the same double-line two-point diagram. As for the D` series, each diagram
must end with a branching, and so D1 would be empty, while D2 is a non-allowed
disjoint two-point diagram, representing A 1 ⊕ A 1. Finally, D3 is identical to A 3
(a single-edge three-point chain). So, we end up with a complete list of distinct
indecomposable allowed fundamental system of roots and associated connected
Dynkin diagrams.
The symmetries observed in some diagrams are manifestations of automor-
phisms (self-equivalences) of the associated Lie algebras, with implications on
their representations. Thus, A ` with ` ≥ 2 is symmetric under the reflection
which maps nodes to nodes α1+ j ↔ α`− j . E6 has a similar symmetry. The system
D` is invariant if the two short branches are interchanged, and the diagram D4
has an even higher symmetry as it remains invariant under the full symmetric
6.7. CLASSIFICATION OF LIE ALGEBRAS 201
group S3 acting on the three end points. In contrast, the diagrams B` , C` , and the
exceptional systems (apart from E6) do not have any such symmetry.
By assigning a weight to every vertex in the Dynkin diagrams, we can write
down the corresponding Cartan matrices, as discussed before. Since this corre-
spondence is not affected by multiplying all the roots of a given FSR by a constant,
it suffices to normalize one (short) root to 1 and assign appropriate weights to all
other vertices relative to that vertex. Thus, the vertices in A ` , D` , E6, E7, and E8
all carry the same weight. In B` , C` , and F4 the short roots have weights equal to
1 and the long roots have weights equal to 2 . G 2 is the only system with a long
root of weight 3 relative to its weight-one short root.
Having now the list of all allowed root systems, we proceed to show there
exist corresponding simple Lie algebras by actually constructing them. We shall
adopt the following notations: ei are canonical basis vectors of space V; ε i are
basis in the dual space V ∗ , such that ε i (ej ) = δij. We also use ai, bi , . . . for real or
complex numbers. Every classical Lie algebra is a sub Lie algebra of the general
linear Lie algebra gl ( n, C ). The space of all (n × n)-matrices admits as a basis
the set of n2 matrices [ Eij ] (with i, j = 1, 2, . . . n), which are ( n × n )-matrices with
1 at the ij-position and 0’s everywhere else; so elements of the matrix [ Eij ] are
( Eij )rs = δir δjs . The [ Eij ]’s obey the product rule Eij Ekm = Eim δjk , or equivalently,
So, the `(` + 1) matrices Eij with i 6 = j are the root vectors with roots relative to
H given by the linear functions αij ( H ) = ai − a j (or αij = ε i − ε j). Then, a basis
for Lie algebra sl ( n, C ) = h ⊕ α∈∆ gα may consist of h1, . . . , h` and Eij with i 6 = j.
L
where we have used ∑ni ai = 0 and ∑ni a0i = 0. Note (i) ( H : H 0 ) = 2n Tr( H H 0);
and (ii) the Pythagorean form of the norm ( H : H ) = 2n ∑ a2i .
6.7. CLASSIFICATION OF LIE ALGEBRAS 203
1
h αµ, αν i = αν ( Hαµ ) = (2δµν − δµ,ν+1 − δµ,ν−1 ) for µ, ν = 1, . . . , `. (6.22)
2n
Normalizing Hαµ such that hµ ≡ h̃αµ = Hαµ 2/ h αµ, αµ i, we get hµ = eµ − eµ+1 ,
and hence the Cartan integers aµν = αν ( hµ ), which are aµµ = 2, aµν = − 1 if
µ = ν ± 1, and aµν = 0 otherwise. They are the entries to the Cartan matrix:
−1
2 0
−1 2 −1
0 −1 2
···
A[sl(` + 1, C)] = . (6.23)
···
2 −1 0
−1 2 −1
0 −1 2
It is represented by a Dynkin diagram (with the same weights for all points),
which turns out to be identical to diagram A ` given in Table 6.1 :
A` f f f··· f f
α1 α2 α3 α`−1 α`
The structure of the algebra can be completely determined from the ` sub
algebras sµ = h hµ, eµ , f µ i, where hµ = eµ − eµ+1 , eµ = Eµ,µ+1 , and f µ = Eµ+1,µ
for µ = 1, 2, . . . , `, associated with the simple roots αµ = ε µ − ε µ+1 composing the
fundamental system Σ = { αµ; µ = 1, 2, . . . , `} . In conclusion, the complex matrix
Lie algebra sl (` + 1, C ) is equivalent to the complex simple Lie algebra, now called a` ,
represented by the Dynkin diagram A ` .
2. Symplectic Lie algebra sp(`, C ).
The matrix Lie algebra sp(`, C ) consists of complex matrices M of order n = 2`
that satisfy the condition MT J + J M = 0, where J and M are similarly partitioned
into (` × `)-blocks, as follows
M1 M2 0 I`
M= and J= ,
M3 M4 − I` 0
204 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
Accordingly, one may choose for sp(`, C ) a basis consisting of the following
`(2` + 1) real matrices (with i, j = 1, 2, . . . , `):
For the CSA h, one chooses as basis the set of ` diagonal matrices
This says that Eij1 (i 6 = j), Eij2 (i ≤ j), and Eij3 (i ≤ j) are the root vectors of sp(`, C )
with respect to h corresponding to the root functions
ij ij ij
α 1 ( H ) = ai − a j , α 2 ( H ) = ai + a j , α3 ( H ) = −( ai + a j ), (6.29)
with the indicated restrictions on the ranges of the indices. They define the set ∆.
Restriction ai ∈ R defines the set ∆R ⊂ h∗0 . Once some ordering in h∗0 is chosen,
we may specify the fundamental root system Σ and the set ∆+ of positive roots:
Σ: α i = ε i − ε i+ 1 ( i < `); α` = 2ε ` ;
∆+ : εi − ε j , εi + ε j ( i < j ≤ `); 2ε i (1 ≤ i ≤ `).
( H : H0 ) = ∑ α ( H ) · α ( H0 )
α ∈∆
= ∑i6= j ( ai − a j )( a0i − a0j ) + 2 ∑i≤ j ( ai + a j )( a0i + a0j )
`
= 4(` + 1) ∑i=1 ai a0i = ( n + 2) Tr( H H0 ) . (6.30)
form ( Hαi : H ) = 4(` + 1) ∑k ak ck for the LHS, and solving the resulting equations
for ci to obtain
ei − ei+1
Hαi = , ( i < `) (6.31)
4(` + 1)
e`
Hα` = . (6.32)
2(` + 1)
The inner products on h∗0 for the simple roots can now be calculated, with
h αi, α j i = αi ( Hα j ), leading to the results
which just shows that sp(`, C ) is the Lie algebra c` associated with the Dynkin diagram
1 1 1 1 2
C` v v··· v v f
α1 α2 α`−2 α`−1 α`
along the lines we have followed problematic. To avoid this difficulty, we go over
to a different basis for V via a unitary transformation, giving us a new set of ma-
trices obeying a different condition, but the algebra they represent is equivalent
in every way to the original one.
Given M ∈ so ( n, C ), let U be a unitary matrix on Cn , and define N = U † MU.
Then the condition on M, namely MT + M = 0, becomes for N:
N T U T U + U T U N = 0. (6.37)
where I` is the identity matrix of order `. With N partitioned in the same way,
N1 N2
N= , (6.39)
N3 N4
the condition (6.37) tells us that the diagonal blocks are the negative transposes
of each other (N4T = − N1), and that the off-diagonal blocks are skew symmetric
(N2T = − N2, N3T = − N3). There appear similarities and differences with the
symplectic case already seen, which we can exploit to define an appropriate basis
consisting of the following `(2` − 1) matrices (with i, j = 1, 2, . . . , `):
(Note that while N and E are matrices on C2` ; E1, E2, E3, and e are considered
objects defined on C` .) The diagonal matrices ei (with i = 1, 2, . . . , `) are chosen
as a basis for the Cartan sub algebra of the algebra,
h = { H | H = ∑i ci ei } , and of
its restriction on the real field, h0 = { ∑i aiei ai ∈ R } . Relative to this CSA, we
have the root vectors and corresponding roots:
ij
Eij1 : α1 = εi − ε j , i 6 = j,
ij
Eij2 : α2 = εi + ε j , i < j,
ij
Eij3 : α3 = −( εi + ε j ), i < j.
ij ij ij
(Note that ± 2ε i are not roots.) Altogether α1 , α2 , α3 form the set of roots ∆, with
its restriction to reals denoted by ∆R ⊂ h∗0 . The fundamental system Σ and the
6.7. CLASSIFICATION OF LIE ALGEBRAS 207
Note that all simple roots have the same length | αi|2 = 1/2(` − 1), and the
fundamental-coroot generators are h̃αi = ei − ei+1 ( i < `), and h̃α` = e`−1 + e` .
So, the structure of so (2`, C ) can be encoded in the Cartan matrix
−1
2 0
−1 2 −1
0
−1 2
···
A[so(2`, C)] =
··· .
(6.45)
2 −1 0 0
−1 2 −1 −1
0 −1 2 0
0 −1 0 2
208 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
In conclusion, the even-order orthogonal Lie algebra so (2`, C ) is equivalent to the com-
plex simple Lie algebra, now called d` , associated with the Dynkin diagram D` :
f α`−1
f f··· f f
D` HH f α`
α1 α2 α`−3 α`−2
E XAMPLE 14: For ` = 1, i.e. so (2, C ), this formalism gives us N = a0 −0a and
√
U = 1/ 2 −i i − 1
†
0 −1
−1 , where a ∈ C, which leads to M = UNU = ia 1 0 . (As
already noted, so (2, C ) ∼
= C is not a semisimple Lie algebra.)
Odd-order orthogonal Lie algebra so (2` + 1, C ). Now, the matrices N of order
n = 2` + 1 are partitioned in submatrices in the following way:
B C1 C2
N = D1 N1 N2 , (6.46)
D2 N3 N4
where B is a complex number, Ci are 1 × ` matrices, Di are ` × 1 matrices, whereas
the Ni are ` × ` matrices, all of complex numbers. Taking into account the extra
dimension in this case, we choose for unitary transformation the matrix U:
√
1 0 0
1 2 0 0
U= √ 0 iI` − iI` , U T U = 0 0 I` , (6.47)
2 0 −I −I 0 I 0
` ` `
Here, as in all other classical algebras, we have a Pythagorean norm: ∑i a2i . Also,
note that ( H : H 0 ) = ( n − 2) Tr( H H0 ), exactly the same formula as for even order.
The root generators corresponding to the simple roots are
ei − ei+1
Hαi = ( i < `), (6.49)
2(2` − 1)
e`
Hα` = . (6.50)
2(2` − 1)
The inner products on h∗0 can then be calculated as before, leading to the results:
So α` is the short root: | α` |2 = 1/2 | αi |2 for every i < `; and the fundamental
co-root generators are h̃αi = ei − ei+1 (with i < `) and h̃α` = 2e` .
The Cartan matrix and associated Dynkin diagram for so (2` + 1, C ) are:
−1
2 0
−1 2 −1
0 −1 2
···
A[so(2` + 1, C)] = . (6.54)
2 −1 0 0
−1 2 −1 0
0 −1 2 −1
0 0 −2 2
2 2 2 2 1
B` f f··· f f v
α1 α2 α`−2 α`−1 α`
In conclusion, the odd-order orthogonal Lie algebra so(2` + 1, C ) is the same as the com-
plex simple Lie algebra b` associated with the Dynkin diagram B` .
C OMMENTS . Once the sets ∆ and Σ of a Lie algebra g are known, one can fol-
low the steps described in Example 9 to find the Cartan integers, bypassing the
computation of inner products. To illustrate, we consider two examples. In the
case of A ` , we see that αi + αi+1 (1 ≤ i ≤ ` − 1) are roots (being equal to αi,i+2 ),
210 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
but αi + 2αi+1 are not, and neither are αi + α j for j ≥ i + 2. This means that for
two adjacent roots in Σ we have q = 0 and p = 1 (as in P10-P11) and the Cartan
integer ai,i+1 is − 1, and that the non-adjacent fundamental roots are orthogonal
to each other and the corresponding Cartan integers are 0. As a second exam-
ple, consider C` : For i = 1, 2, . . . , ` − 1, αi + αi+1 are roots, but αi + 2αi+1 are not;
so one has q = 0, p = 1, which implies ai+1,i = − 1. On the other hand, both
α` + α`−1 = ε ` + ε `−1 and α` + 2α`−1 = 2ε ` are roots, but not α` + 3α`−1, hence
a`−1,` = − 2.
4. Exceptional Lie algebra g2 .
Let ei be the standard basis in C3 or R 3 , and ε i the corresponding dual basis.
Then { H = ∑3i aiei | a1 + a2 + a3 = 0} defines a two-dimensional CSA h (if ai
are complex) or h0 (if ai are real). The dual basis h∗ is similarly subject to the
restriction ε 1 + ε 2 + ε 3 = 0. The root system ∆ consists of twelve elements, namely,
± ε i and ±( εi − ε j), with i, j = 1, 2, 3; and so we have the dimension of the algebra:
dimg2 = dimh + | ∆ | = 14. The subset of positive roots ∆+ consists of ε 1, ε 2, − ε 3,
ε 1 − ε 2, ε 1 − ε 3, and ε 2 − ε 3.
We choose the simple roots to be α1 = ε 2 and α2 = ε 1 − ε 2, so that all roots
may be written in the form n1 α1 + n2 α2, where n1, n2 are both non-negative or
non-positive integers. In particular, the maximal root is ε 1 − ε 3 = 3α1 + 2α2.
The Killing form is given by ( H : H 0 ) = 8 ∑3i ai a0i , where H, H 0 ∈ h0 . From
this, one obtains the root generators Hα defined by ( Hα : H ) = α ( H ), and the
inner product h, i on h∗0 , as follows.
Fundamental root generators:
Hα1 = − 1/24 e1 + 1/12 e2 − 1/24 e3 and
Hα2 = 1/8 e1 − 1/8 e2 .
Inner products in h∗0 : h α1, α1 i = α1 ( Hα1 ) = 1/12, h α2, α2 i = α2 ( Hα2 ) = 3/12,
h α , α2 i = α1 ( Hα2 ) = − 1/8. Thus, α2 is the long root: | α2|2 = 3| α1|2.
1
1 3
2 −3 u e
A [g2 ] =
−1 2 α1 α2
Σ: α1 = 1/2( ε 1 − ε 2 − ε 3 − ε 4), α2 = ε 4, α3 = ε 3 − ε 4, α4 = ε 2 − ε 3.
∆+ : ε i, ε i + ε j, ε i − ε j (for i < j ), 1/2( ε 1 ± ε 2 ± ε 3 ± ε 4) .
There are 24 positive roots, of which ε 1 + ε 2 = 2α1 + 4α2 + 3α3 + 2α4 is the maxi-
mal root. Killing form on h0: ( H : H 0 ) = 18 ∑i ai a0i .
6.7. CLASSIFICATION OF LIE ALGEBRAS 211
roots.
Fundamental co-roots:
h̃α1 = e1 − e2 − e3 − e4 , h̃α2 = 2e4 , h̃α3 = e3 − e4 , h̃α4 = e2 − e3 .
The Cartan integers, obtained from the formula aij = α j ( h̃αi ), are displayed in
the matrix form and represented by a Dynkin diagram:
−1
2 0 0
1 1 2 2
−1 2 −2 0 u u e e
A [ f4 ] =
0 −1 2 −1 α1 α2 α3 α4
0 0 −1 2
2 0 −1 0 0 0
0 2 0 −1 0 0
−1 0 2 −1 0 0
A [ e6 ] =
0 −1 −1 2 −1 0
0 0 0 −1 2 −1
0 0 0 0 −1 2
1
f
α2
1 1 1 1 1
E 6: f f f f f
α1 α3 α4 α5 α6
1
f
α2
1 1 1 1 1 1
E7: f f f f f f
α1 α3 α4 α5 α6 α7
1
f
α2
1 1 1 1 1 1 1
E8: f f f f f f f
α1 α3 α4 α5 α6 α7 α8
Figure 6.9: Dynkin diagrams for the systems E6 , E7 , and E8 . Note that, with this choice of
basis (where the branching occurs at the third node), omitting the last node of E8 produces
E7 , and omitting the last node of E7 produces E6 .
−1
2 0 0 0 0 0 0
0
2 0 −1 0 0 0 0
−1 0 2 −1 0 0 0 0
0 −1 −1 2 −1 0 0 0
A [ e8 ] = . (6.55)
0
0 0 −1 2 −1 0 0
0
0 0 0 −1 2 −1 0
0 0 0 0 0 −1 2 −1
0 0 0 0 0 0 −1 2
To each FSR there corresponds exactly one simple complex Lie algebra (and
so a unique complex compact Lie group) of some complex dimension, but there
are in general several real forms of the same real dimension. For example to E8
are associated a complex Lie algebra of complex dimension 248 and three real
forms of real dimension 248. The Lie algebra e8 contains as subalgebras all the
exceptional Lie algebras as well as several classical Lie algebras.
A classical Lie group G acting on a vector space V may generally be viewed as
a group of automorphisms in V preserving a bilinear form B ( x, y ) where x, y ∈ V.
For Lie algebra Lie ( G ) the equivalent statement would refer to infinitesimal in-
variance of systems of a form similar to ( x, dy ) + ( dx, y) = 0 on R n for O ( n ).
Killing (1889) showed that G 2 could be realized as a group of local transforma-
tions on R 5 , and Cartan (1894) proved that similar groups of transformations on
R 15, R 16, R 27 , and R 29 exist for F4, E6, E7 and E8, respectively. Alternatively, the
exceptional groups can be regarded as isometry groups over a noncommutative,
nonassociative, normed division algebra over the real numbers, called the octo-
nion algebra (see [Ba], [CS]). The octonions (discovered independently by John
Graves (1843) and Arthur Cayley (1845)) are an 8-dimensional algebra with basis
1, e1, . . . , e7 , in which each unit, except 1, squares to − 1, that is both noncommu-
tative and nonassociative.
214 CHAPTER 6. SIMPLE LIE ALGEBRAS: STRUCTURE
Problems
6.1 (Find ∆ for b2 ) Find all the roots of the b2 Lie algebra.
6.2 (Find ∆+ for d4 ) We have seen in Sect. 6.5 that the set of roots of a simple Lie
algebra g can be found from its given fundamental system Σ and Cartan matrix
A. Use this method to find all positive roots of the d4 Lie algebra.
6.3 (Rules for Dynkin diagrams) Identification of the allowed Dynkin diagrams
(those that are consistent with the definition of a fundamental system of roots
(FSR)) is based on a set of general rules derived from the definition of FSR of
semisimple Lie algebras. Prove:
(a) Rule 1: A Dynkin diagram contains more vertices than joined pairs. It
cannot have closed polygons. (A closed polygon, or a loop, is a set of points
sequentially connected and the last point is also connected to the first.)
(b) Rule 2: The maximum number of lines issuing from a vertex is 3.
(c) Rule 3: If an allowed Dynkin diagram Γ contains an A 2 simple chain as a
sub diagram, then the diagram obtained from Γ by replacing the simple chain by
a vertex point is also an allowed Dynkin diagram
6.4 (Non allowed diagrams) Show that configurations Fig. 6.7(c.1), Fig. 6.8(d.1)-
(d.2) are not allowed by showing that the bilinear form h, i defined in the respec-
tive spaces h∗0 is not positive definite.
6.5 (Generators of g2 ) Given a 2 × 2 Cartan matrix with entries a11 = a22 = 2,
a12 = − 3, a21 = − 1, identify the generators of the corresponding Lie algebra.
6.6 (Multiplication table for g2 ) With the same definitions as in Problem 5, estab-
lish the multiplication table for the generators of the Lie algebra g2 .
6.7 (Positivity of norms in the classical series) Check that vectors in the root
spaces associated with diagrams A ` , B` , C` , and D` have positive definite norm
squares, as expected of Euclidean spaces.
6.8 (Positivity of norms in the exceptional algebras) Same question as in the pre-
ceding problem for the five exceptional algebras.
6.9 (Maximal roots) Find the maximal roots of e6, e7 , and e8.
6.10 (Data on g2 ) Rank-two
√ 1 Lie algebra g2 is defined by the simple roots α1 and
2 2 ◦
α related by | α | = 3| α | and separated by an angle of 150 . Introduce the
basis ei (i = 1, 2, 3) for h0 and the dual basis ε i (i = 1, 2, 3) for h∗0 obeying ε i ( e j ) =
δij and ε 1 + ε 2 + ε 3 = 0. Express the fundamental coroot generators h i and the
fundamental roots and weights αi and ω i, where i = 1, 2 and ω i ( h j ) = δij , in the
appropriate basis.
215
216 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
def
root hi = h̃αi ) is uniquely associated to a simple root αi ∈ Σ . The roots lie in a
real subspace h∗0 of h∗ .
As in Chapters 3–5, we call π : g → gl(V ) a finite-dimensional representation
of g on the complex vector space V , such that each element z in g is mapped to a
linear operator π ( z ) ≡ Z in V .
As a Cartan sub Lie algebra is abelian, one may define joint eigenvectors v
of all the operators H ≡ π ( h ) for h in the given CSA h. Such a vector v ∈ V ,
if nonzero, is called a weight vector, and the corresponding eigenvalue µ ( h ) for
π ( h) is a linear function of h on h, and so is an element of h∗ , called the weight of
v: π ( h)v = µ ( h ) v, or Hv = µ ( h)v . For a given µ ∈ h∗ , the weight space Vµ (which
may be zero) is a subspace of V consisting of 0 and all weight vectors of weight µ.
When Vµ 6 = 0, µ is said to be a weight in the representation π. Further, the dimension
of Vµ is called the multiplicity of µ. Thus V admits a direct-sum decomposition
V = ⊕ µ Vµ , where h acts diagonally on each Vµ by scalar multiplication by µ. The
weight spectrum of the adjoint representation of g consists of 0 (of multiplicity `)
and non-zero weights (each of multiplicity 1), called the roots of g.
Let us now select a basis for h. We know that generally in semisimple Lie al-
gebras, to each α ∈ ∆+ correspond one-dimensional dual root spaces gα and g−α ,
such that the direct sum sα = gα ⊕ g−α ⊕ [gα , g−α ] determines a three-dimensional
subalgebra of g isomorphic to sl (2, C ). And so, in particular, for every (simple)
root αi in the fundamental system Σ that spans h∗ , we have one such subalgebra
of g: si ≡ sαi with i = 1, 2, . . . , `. Now in each subalgebra si , we can pick a basis
consisting of the elements ei ≡ eαi ∈ gαi , f i ≡ f αi ∈ g−αi , and hi ≡ [ei , f i ], normal-
ized such that αi ( hi ) = 2. With this standard normalization, h i, ei , f i satisfy the
canonical commutation relations for sl (2, C ).
We can now choose as a basis for the CSA those ` elements h 1, h2, . . . , h` in h
that obey the two conditions h i = [ei , f i ] and αi ( hi ) = 2. (In fact, they are pre-
cisely the fundamental co-root generators h̃αi .) As far as g itself is concerned, the
remaining (dim g − `) elements needed to complete a basis for the whole algebra
may be chosen to be either (i) ei, f i (i = 1, 2, . . . , `) together with their distinct
multiple commutators; or (ii) eα ∈ gα , f α ∈ g−α for all α ∈ ∆+ . We will have
occasions to use one or the other type of basis in our discussion.
This choice of basis { hi} for h has important implications. From our analysis
of the sl (2, C ) representations, we know that each π ( h i) is diagonalizable on the
restriction of π to si , with integral eigenvalues. As all π ( h i)’s commute, they are
simultaneously diagonalizable. The joint eigenvectors v are finite in number, and span
V . Each of the corresponding eigenvalues, µ, is a linear function of hi , and so lives
in h∗ , or, more precisely in h∗0 because µ ( hi) ∈ Z. It will soon become apparent
that π (g) has the structure of a direct sum, V = ⊕ a Wa , where Wa are the sub
representations in irreducible π under a given si = { hi, ei , f i } .
Just as for sl (3, C ), there is a Lemma concerning weight spaces of any semisim-
ple Lie algebra g, which says that If α is any root and zα the corresponding root vector,
then π ( zα ) ≡ Zα maps weight space Vµ into weight space Vµ+α . In other words, if
v is a weight vector of representation π with weight µ, then Zα v, if not zero,
is again a weight vector of π with weight µ + α, as is clear from the equation
H Zα v = ( µ + α )(h) Zαv.
7.1. GENERAL PROPERTIES 217
2h µ, αi
wα : µ 7 → µ0 , where µ0 = µ − α = µ − µ ( h̃α ) α , (7.3)
h α, α i
W w0 w1 w2 w3 w4 w5
w0 w0 w1 w2 w3 w4 w5
w1 w1 w0 w4 w5 w2 w3
w2 w2 w5 w0 w4 w3 w1
w3 w3 w4 w5 w0 w1 w2
w4 w4 w3 w1 w2 w5 w0
w5 w5 w2 w3 w1 w0 w4
hµ, α1 i = 0
w1
α2
w0
ω2
hµ, α2 i = 0
ω1
w4 α1
w2
w3 w5
Figure 7.1: The Weyl chambers of h∗0 for a2 are labeled by the operators wi that
take the fundamental Weyl chamber w0 to the Weyl chamber wi . Also indicated
are the simple roots α1, α2, and the fundamental weights ω 1, ω 2.
As an alternative to the simple roots αi, one may pick the fundamental weights
ω i to form a basis for h∗0 , so that Λ = { ∑i mi ω i | mi ∈ Z } . It is called the Dynkin
basis (each ω i is uniquely associated to a node αi in the Dynkin diagram). The
condition ω i ( h j) = 0 for i 6 = j means that the ω i’s are the first weights on the
edges of a cone called the fundamental Weyl chamber on h∗0 . Thus, one may describe
Λd as the intersection of Λ and the closed fundamental Weyl chamber. The sum
δ = ∑`1 ω i gives the lowest strongly dominant weight.
Any weight (or root) µ may be written either as µ = ∑i k i αi in the basis { αi} ,
or as µ = ∑i mi ω i in the basis { ω i} . But only in the latter are the coefficients always
integral, which is a clear advantage. So, in the Dynkin basis we may write any
weight (or root) µ = m1 ω 1 + m2 ω 2 + · · · + m` ω ` as µ = ( m1, m2, . . . , m` ), with
mi = µ ( hi) ∈ Z. Correspondence between the two bases is determined from the
Cartan matrix by the relations αi = ∑j ω j a ji and k i = ∑j bij m j , where a ji = αi ( h j)
is the element A ji of the Cartan matrix, and bij is the element A− 1
ij of its inverse.
Note that the components of the simple roots αi wrt CSA { h1, . . . } are given by
the entries in the i-th column of the Cartan matrix.
220 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
E XAMPLE 2: From the data on the algebraic structure (Chapter 6) we can deter-
mine its fundamental weights for any simple Lie algebra. Take a2 as an example.
From the Cartan matrix, we obtain the two simple roots: α1 = ω 1 a11 + ω 2 a21 =
2ω 1 − ω 2 and α2 = ω 1 a12 + ω 2 a22 = − ω 1 + 2ω 2, and by inversion, the fun-
damental weights: ω 1 = 2/3α1 + 1/3α2 and ω 2 = 1/3α1 + 2/3α2. On the other
hand, writing ei for the ith coordinate in R n or Cn , and ε i the ith coordinate
function, we know that α1 = ε 1 − ε 2 and α2 = ε 1 + 2ε 2, and so ω 1 = ε 1 and
ω 2 = ε 1 + ε 2. A more direct way to find ω i goes by their definition: We already
know that h1 = e1 − e2 , h2 = e2 − e3 on h0. Since ω i live on h∗0 , they must have
the form ω i = ai ε 1 + bi ε 2 + ci ε 3, with ε 1 + ε 2 + ε 3 = 0. By duality ε i (ej ) = δij and
ω i ( h j ) = δij , with i (or j ) = 1, 2 , we find again ω 1 = ε 1 and ω 2 = ε 1 + ε 2. These
ε1 = 2/3α1 + 1/3α2 = ω 1 = ω1
ε2 = − /3α + /3α = − ω + ω = ω 1 − α1
1 1 1 2 1 2 (7.5)
ε3 = − 1/3α1 − 2/3α2 = − ω 2 = ω 1 − α1 − α2
The various bases for h∗0 we have considered are shown in Fig.7.2.
α2 = ε2 -ε3 ε1 -ε3
e
ω2 e
ε2 ω1 = ε1
ε2 -ε1 u α1 = ε1 -ε2
ε3
ε3 -ε1 ε3 -ε2
Figure 7.2: Vectors in a small sector of h∗0 space of a2 , with various relative angles
given by θ ( ω 1, ω 2 ) = 60◦ , θ ( α1, α2 ) = 120◦ , and θ ( ε 1, ε 2) = 120◦ . The darkened
cone is the fundamental Weyl chamber.
same line, then the number of weights on each row first increases then decreases
symmetrically as the height increases, giving a vaguely spindle-shaped diagram.
The weight system of any irreducible representation π is invariant under the
Weyl group W, that is, if µ is a weight of π, then so is w · µ for any w ∈ W. A
weight µ and its conjugates w · µ have the same multiplicity and are considered
equivalent. The different equivalent weights form a Weyl orbit (W-orbit), their
number being called the orbital size.
Just as the weights of an irrep of a2 lie on a triangle or a hexagon in the weight
lattice Λ, so the weight systems of irreps of any simple Lie algebra g lie on con-
vex (polygonal) hulks in Λ. As we have seen, the Weyl group W of g sends each
weight µ into a weight of the same representation. In particular, if µ occupies a
corner of the polygon, its mirror image w · µ occupies an opposite corner since
µ and w · µ are the extreme weights of some string of weights. Starting with the
highest weight λ, we can generate other corner weights with every w in W; re-
peating the process, we can identify all other weights and their associated strings,
and so obtain the weight system πλ . These remarks show how the Weyl group
can be useful in finding all the weights of a given irrep.
The simplest of the irreps of the Lie algebra a2 are the fundamental repre-
sentations (1, 0) and (0, 1), which we have already seen in Examples 1–2 of this
chapter and Fig. 5.4 of Chapter 5. We now consider a few others.
E XAMPLE 7: π1,1 of a2 = sl (3, C ). Subtracting α1 and α2 from the dominant
weight λ = (1, 1) associated with the weight vector v produces two doublets:
(1, 1), (−1, 2): v, F1 v (s1 doublet),
(1, 1), (2, −1): v, F2 v (s2 doublet).
From (− 1, 2) subtracting α2 twice gives a triplet:
(− 1, 2), (0, 0)00, (1,-2): F1 v, F2 F1 v, F22 F1v (s2 triplet).
From (1, − 2) subtracting α1 once gives
(1, − 2), (−1, −1): F22 F1 v, F1 F22 F1v (s1 doublet)
On the other hand, subtracting α1 twice from (2, − 1) gives
(2, − 1), (0, 0)0, (-2,1): F2 v, F1 F2 v, F12 F2 v (s1 triplet).
Finally, from (− 2, 1) subtracting α2 once gives
(− 2, 1), (−1, −1): F12 F2 v, F2 F12 F2v (s2 doublet).
The end line is the all negative weight (-1,-1). To verify that the weight system just
obtained is complete, we can show that it transforms into itself under the known
Weyl group (described in Example 1). It is illustrated in Fig. 7.3.
Each of the weights (0,0) and (− 1, −1) is reached from two different paths.
What is their multiplicity? The weight (− 1, − 1) is the mirror image of (1,1) under
the Weyl transformation w1 w2 w1, and so it has the multiplicity of one.
Now, let’s look at (0,0). We have |(0, 0)0i = F1 F2 v as a member of an s1 triplet,
and |(0, 0)00 i = F2 F1 v a member of an s2 triplet. Let us define the orthonormal
vectors |(0, 0)0i of the s1 -singlet and |(0, 0)1i of the s1 -triplet. Then by defini-
tion |(0, 0)0i = |(0, 0)1i; and |(0, 0)00i must be a linear combination of the form
|(0, 0)00 i = F2 F1 v = a |(0, 0)1i + b |(0, 0)0i, such that a2 + b2 = 2.
√ Applying E1 on √ this vector, we get on the RHS aE1|(0, 0)1i + bE1|(0, 0)0i =
a 2|(2, −1)i = a 2F2 v, where we have used (7.6), whereas √ on the LHS we have
√
E1 F2 F1v = F2 E1 F1 v = F2 v. Equating RHS and LHS, a 2 = 1, we get a = 1/ 2
7.2. IRREDUCIBLE REPRESENTATIONS 225
α(2 )
−1, 2 K
AAu u 1, 1
1,1 A A
A A
-1,2 2,-1 A A
0,0 0,0 u UA u
h UA u -α(1)
−2, 1 A A 0, 0 2, −1
1,-2 -2,1 A A
A A
−1, −1 U A u UA u 1, −2
-1,-1
Figure 7.3: Two views of the weight system of the adjoint representation of a2 .
Note the two-fold degeneracy of (0, 0).
(with conventional sign and phase). So |(0, 0)0i and |(0, 0)00 i are independent, and
the weight (0,0) of π11 has multiplicity two, which implies that the dimension of
π11 is equal to eight.
E XAMPLE 8: π3,0 of a2. Following the usual steps, we obtain the following:
(30), (11), (−1, 2), (−3, 3): v, F1 v, F12v, F13v (s1 quadruplet);
(11), (2, −1): F1 v, F2 F1 v (s2 doublet);
(− 1, 2), (00), (1, −2): F12 v, F2 F12v, F22 F12 v (s2 triplet);
(− 3, 3), (−2, 1), (−1, −1), (0, −3): F13 v, F2 F13v, F22 F13 v (s2 quadruplet);
(2, − 1), (00), (−2, 1): F2 F1 v, F1 F2 F1 v, F12 F2 F1 v (s1 triplet);
(1, − 2), (−1, −1): F22 F12v, F1 F22 F12 v (s1 doublet).
Again this is the complete Weyl invariant weight system having (3,0) as the
highest weight, as illustrated in the following diagram (cf. Fig. 5.8(d) Chapter 5).
30
11
-1,2 2,-1
-3,3 00
1,-2 -2,1
-1,-1
0,-3
notes the induced representation on the r-th exterior power (the r-fold antisym-
metric product) r V . If µ1, µ2 , . . . are the weights of π for the vectors v 1, v2, . . . ,
V
then the weights of r π are the r-wise sums of distinct weights µi given by the
V
sums µi1 + µi2 + · · · + µir subject to the restrictions i1 < i2 < · · · < ir , and have
the products vi1 ∧ vi2 · · · ∧ vir as weight vectors. Note that all indices must be
different to satisfy antisymmetry, and i1 < i2 < · · · to avoid double counting.
Similar considerations apply mutatis mutandis to the completely symmetric pow-
ers, named Symr π and Symr V , the weights of which are given by the r-wise sums
µi1 + µi2 + · · · + µir subject to the restrictions i1 ≤ i2 ≤ · · · ≤ ir .
• Weight space:
Weight lattice: Λ = { ∑n1 mi ε i | mi ∈ Z } / ( ∑n1 ε i = 0).
Dominant weights: λ = ∑n1 mi ε i, with integers m1 ≥ m2 ≥ · · · ≥ mn ≥ 0, so
that λ ( hi) are all non-negative integers.
Fundamental weights: ω 1 = ε 1, ω 2 = ε 1 + ε 2,. . . , ω ` = ε 1 + ε 2 + · · · + ε ` , in
accordance with ω i ( h j) = δij . Note also that the lowest strongly dominant weight
is given by δ = ∑`1 ω i = ∑`1 kε `+1−k.
The root space admits several bases tied together by relations to be given here.
The simple roots may be written in the Dynkin basis { ω 1, . . . , ω ` } as αi = ∑j ω j a ji
where a ji = αi ( h j ) (as already indicated before), or explicitly,
α1 = ε 1 − ε 2 = (2, − 1, 0, . . . ),
α2 = ε 2 − ε 3 = (− 1, 2, −1, 0, . . . ),
α3 = ε 3 − ε 4 = (0, − 1, 2, −1, 0, . . . ), ..., ...,
α`−1 = ε `−1 − ε ` = ( . . . , 0, − 1, 2, − 1), and
α` = ε ` − ε `+1 = ( . . . , 0, − 1, 2).
Upon inverting, we obtain the relations
ε1 = ω 1,
ε2 = ω 2 − ω 1 = ω 1 − α1,
ε3 = ω 3 − ω 2 = ω 1 − α1 − α2, ..., ...,
ε ` = ω ` − ω `−1 = ω 1 − α1 − · · · − α`−1 ,
ε `+1 = − ω ` = ω 1 − α1 − · · · − α` .
• The classification of the irreducible representations of the special linear Lie algebra
sl ( n, C ) has a neat solution: all of its irreps can be built up from its standard repre-
sentation, the latter being defined by traceless matrices of order n.
To begin, it is useful to recall the results concerning the case of sl (3, C ). In
its standard representation, each element is represented by a 3 × 3 traceless com-
plex matrix, and a basis of the CSA may be taken to be 3 × 3 diagonal matri-
ces, e.g. π ( h1) = diag [1, − 1, 0] and π ( h2) = diag [0, 1, − 1]. The basis of C3 is
the usual column vectors ξ i , which are joint eigenvectors of π ( h 1), π ( h2) with
weights ε 1 = (1, 0), ε 2 = (− 1, 1) and ε 3 = (0, − 1), where the components are
given by ε i ( h j ). Alternatively we may write ε 1 = ω 1, ε 2 = ω 2 − ω 1 and ε 3 = − ω 2.
The highest weight is ε 1 corresponding to the highest-weight vector ξ 1 = u. Then
we can identify ξ 2 and ξ 3 with F1 u and F2 F1 u, respectively, where F1 and F2 are
the lowering operators associated with the two simple roots α1 and α2 of sl (3, C ).
As ε 1 = ω 1, this is just the irrep called π1,0.
Now, in the general case of sl ( n, C ), each element h in h is represented on Cn
by a diagonal n × n matrix of the form H = diag [ a1, . . . , an ], where ai ∈ C and
∑n1 ai = 0. Its eigenvectors are the standard basis vectors in Cn , with eigenvalues
a1, . . . , an , so the weights, as elements of h∗ , are ε 1, ε 2, . . . , ε n (because ε i ( H ) = ai ).
As ε 1 = ω 1, ε 2 = ω 1 − α1, etc., we see that ε 1 is the highest weight (whereas
ε n is the lowest), and so the standard representation is π1,0,... . The representation
space Cn is spanned by v, F1v, F2 F1 v,. . . , F` · · · F2 F1 v, where v is the highest-weight
vector, and Fi = Fαi are the lowering operators associated with the simple root αi .
The dual π ∗ to π1,0,... is a representation such that π ∗ ( H ) = − H T , so its
weights are − ε `+1, − ε `, . . . , − ε 1, or ω ` , ω ` − α` , . . . , ω ` − ( α` + α`−1 + · · · + α1 ).
228 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
Its highest weight being − ε `+1 = ω ` indicates it must be the irrep π...,0,1 . Both
π1,0,... and π...,0,1 of course have the same dimension n.
Other fundamental representations (those having a fundamental weight as high-
est weight) all are exterior products of π1,0,... , called ϕ1 for short. For example,
take 2 ϕ1. It has `(` + 1) /2 weights, of the form ε i + ε j with 1 ≤ i < j ≤ ` + 1,
V
A` f f f··· f f
ϕ1 ϕ2 ϕ3 ϕ`−1 ϕ`
Vi
Figure 7.4: The fundamental representations ϕi = ϕ1 = π0,...,1,0,... , with i =
2, . . . , `, of sl ( n, C ) associated to the nodes in the Dynkin diagram.
with i > 1) are not irreducible although they contain the irrep π0,...,0,m,0,... (with m
in the rth position) when it is decomposed into a direct sum.
The general result on the representations of the special linear Lie algebra can
be stated as follows: Every irreducible representation of sl (` + 1, C ) has its high-
est weight of the form ∑i mi ω i = ( m1 + m2 + · · · + m` ) ε 1 + · · · + m` ε ` , where
m1, . . . , m` are natural numbers N; conversely to every `-tuple of non-negative in-
tegers m1, . . . , m` corresponds a unique irreducible representation of sl (` + 1, C )
having highest weight ( m1, m2, . . . , m` ). It occurs in the tensor product
Symm1 ϕ1 ⊗ Symm2 ϕ2 ⊗ · · · ⊗ Symm` ϕ` = πm1 ,m2 ,...,m` ⊕ · · · , (7.7)
where ϕi are the fundamental representations of the algebra, all obtainable from
the standard representation: ϕi = i ϕ1.
V
ε 2 + ε 4, and ε 3 + ε 4. Noting that they are symmetric about the origin, we see
that ϕ2 = π0,1,0 is self-dual, being isomorphic with ( 2 ϕ1 )∗ = 2 ϕ∗1 . Finally,
V V
V3
ϕ3 = ϕ1 has weights ε 1 + ε 2 + ε 3, ε 1 + ε 2 + ε 4, ε 1 + ε 3 + ε 4, and ε 2 + ε 3 + ε 4, and
so ϕ3 = π0,0,1. Since ε 1 + ε 2 + ε 3 + ε 4 = 0, we see that 3 ϕ1 = ϕ∗1 . In conclusion,
V
@ s @ @ @
@ @ @ @ @ @ @ @
6
@ @ @ @ @ @ @ @
@s @ε 2
@ s @ @s @ε 2
@ s @
@ @ @ @ @
@ @ @ @ @ @
@
@ @ @ @ @ @ @ @
@ s @ @ @
@ @ @ 0 @ε 1 s
@ @ @sf0 - @ε 1
@ @
@@ @ @@ @ @ @ @ @@
@ @ @
@ @ @ @ @ @
@s Rs @
@ @s
@ @ @ @ @ @ @ @s @
@ @ @ @ @ @ @ @
@ @ s @@ @ @ @ @ @
@ @ @
@ @ @ @
@ @ @ @
Figure 7.5: Lie algebra sp(2, C ). (a) LHS: Root diagram (| ∆ | = 8), arrows indicat-
ing the simple roots. (b) RHS: Weight diagram of 2 ϕ1, with arrows indicating
V
@ @ ε2 @ @ @ ε2 @
@s @ @s ε1 + ε2 @s @s @s ε1 + ε2
@ @ @ @ @
6 @
@ @ @ @ @ @
@ @ @ @ @ @
@ @ 0 @ ε1 @s @ 0 @s ε1
@ @ @ @ @ @
@ @ @ @ @ @
@ @ @ @ @ @
@s @ R
@@s ε1 − ε2 @s @s R
@@s ε1 − ε2
@ @ @ @ @ @
@ @ @ @ @ @
Figure 7.6: (a) LHS: Root diagram (| ∆ | = 4) for so (4, C ); note the isomorphism
so (4, C ) ∼
= sl (2, C ) × sl (2, C ). (b) RHS: Root diagram (| ∆ | = 8) for so (5, C ). Ar-
rows indicate the simple roots.
• The irreps: (1) With ϕ1 being the standard rep, ϕi = i ϕ1 are the irreps of
V
Symm1 ϕ1 ⊗ · · ·⊗ Symm`−2 ϕ`−2 ⊗ Symm`−1 ϕ`−1 ⊗ Sym(m` −m`−1 )/2 π2ω` if m`−1 +
m` is even and m` ≥ m`−1 ; (2c) Symm1 ϕ1 ⊗ · · · ⊗ Symm`−2 ϕ`−2 ⊗ Symm`−1 πω`−1 ⊗
Symm` πω` if m`−1 + m` is odd.
The irreducible representations of types (2a) and (2b) have weights in the inte-
gral sublattice Z { ε 1, . . . , ε ` } of Λ. On the other hand, those of type (2c) are built on
the irreps of highest weights ω `−1 and ω ` , which lie in the half-integral sublattice
and correspond to the two nodes that branch off one end of the Dynkin diagram.
These reps πω`−1 and πω` are the two half-spin representations of so(2`, C ), both
of dimension 2`−1 . Clifford algebras are a more appropriate approach to study
them, as we will explain in Appendix B.
E XAMPLE 11: We now illustrate the above results for some lower rank algebras.
For ` = 1, d1 = so(2, C ) ∼ = C is not semisimple (Chapter 6); b1 = so (3, C ) has
the same system of roots as sl (2, C ), namely {± ε1} . But while the fundamental
weight is ε 1 for a1 , it is 1/2ε 1 for b1 . So, the standard representation of b1 is a spin
representation of an orthogonal Lie algebra.
E XAMPLE 12: Rank 2 includes d2 = so (4, C ) and b2 = so(5, C ). Lie algebra
so (4, C ) has two simple roots, α1 = ε 1 − ε 2 and α2 = ε 1 + ε 2 in a disconnected
7.2. IRREDUCIBLE REPRESENTATIONS 233
Dynkin diagram. Its roots, ±( ε1 − ε 2 ) and ±( ε1 + ε 2), form the root system ∆
shown in Fig.7.6(a). They lie on two orthogonal lines, which means that so (4, C )
is reducible, being decomposable into the direct sum of its two si subalgebras, one
of roots ±( ε1 − ε 2 ) and the other of roots ±( ε1 + ε 2 ). The fundamental coroots of
so (4, C ) are h1 = e1 − e2 and h2 = e1 + e2 , from which follow the fundamental
weights ω 1 = 1/2( ε 1 − ε 2) and ω 2 = 1/2( ε 1 + ε 2). The representation ϕ1 is so (4, C )
itself spanned by the basis vectors ξ i of C4 with weights ± ε 1 and ± ε 2. The second
exterior product 2 ϕ1 has weights ±( ε1 + ε 2 ), ±( ε1 − ε 2), and 0 (multiplicity 2),
V
shown in the weight diagram in Fig.7.7. Both are reducible and could be under-
stood in terms of the representations of sl (2, C ) × sl (2, C ).
@ @ ε2 @ @ @ ε2 @
@ @s @ ε1 + ε2 @s @ @s ε1 + ε2
@ @ @ @ @ @
@ @ @ @ @ @
@ @ @ @ @ @
@s @ 0 @s ε1 @ @sg0 @ ε1
@ @ @ @ @ @
@ R
@ @ @ @ @
@ @ @ @ @ @
@ @s @ ε1 − ε2 @s @ @s ε1 − ε2
@ @ @ @ @ @
@ @ @ @ @ @
Figure 7.7: Representations of so(4, C ). (a) LHS: Weight diagram of the standard
representation ϕ1 = 4. (b) RHS: Weight diagram of the exterior product repre-
sentation 2 ϕ1 = 6. Arrows indicate the fundamental weights.
V
The other rank-2 orthogonal Lie algebra, b2 = so (5, C ), has two simple roots,
of unequal sizes, given by α1 = ε 1 − ε 2 and α2 = ε 2. Its root system ∆ =
{± ε1, ± ε 2, ±( ε1 + ε 2 ), ±(ε1 − ε 2 )} is illustrated in Fig.7.6(b). Its fundamental co-
roots are h1 = e1 − e2 and h2 = 2e2 , from which are computed the fundamen-
tal weights ω 1 = ε 1 and ω 2 = 1/2( ε 1 + ε 2). The standard representation ϕ1 of
so (5, C ), of highest weight ε 1, has weights ± ε 1, ± ε 2 and 0. Its second exterior
product 2 ϕ1 has weights ± ε 1, ± ε 2, ±( ε1 + ε 2), ±( ε1 − ε 2 ), and 0 (multiplicity
V
@ @ ε2 @ @ @ ε2 @
@ @s @ ε1 + ε2 @s @s @s ε1 + ε2
@ @ @ @ @ @
@ @ @ @ @ @
@ @ @ @ @ @
@s @s 0 -
@s ε1 @s @sg0 @s ε1
@ @ @ @ @ @
@ @ @ @ @ @
@ @ @ @ @ @
@ @s @ ε1 − ε2 @s @s @s ε1 − ε2
@ @ @ @ @ @
@ @ @ @ @ @
r ω2
1
α2 r r r r r rω
ε1
r e -rα1 , ε 2 r r r
3
r rε r r r r
Figure 7.9: g2 . (a) LHS: Root diagram with all 12 roots. α1 = ε 2 is the short root
and α2 = ε 1 − ε 2 the long one. (b) RHS: Weight diagram of the standard (seven-
dimensional) representation ϕ1 = π1,0; it is self-dual, with ω 1 = (1, 0) = − ε 3 as
the highest weight, and − ω 1 = ε 3 the lowest.
E XAMPLE 15: Construction and decomposition of tensor product π1,0 ⊗ π1,0. The
tensor-product representation π1,0 ⊗ π1,0 splits into a symmetric part and an an-
tisymmetric part, as in ab = 1/2( ab + ba ) + 1/2( ab − ba ) or, symbolically, using
Young diagrams (described in Appendix A):
236 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
1
= Sym2 π1,0 ⊕ 2
^
1 ⊗ 2 = 1 2 ⊕ π1,0.
2
V2
The antisymmetric part, denoted π1,0, corresponds to a representation con-
sisting of the 21 antisymmetric products of two vectors, one from each of the two
copies of π1,0, with corresponding weights given by pairwise sums of distinct
weights (0 and ± ε i) of π1,0. It is reducible and decomposes into a direct sum of
two irreducible representations, the standard representation π1,0 and the adjoint
representation π0,1. So we write 2 π1,0 ∼ = π1,0 ⊕ π0,1.
V
2
The symmetric part, denoted Sym π1,0, consists of 28 symmetric products of
vectors of the two spaces π1,0, with weights given by pairwise sums of weights
of the component vectors. Among these, 2ω 1 = 2( ε 1 + ε 2) is the highest weight
and − 2(ε1 + ε 2 ) the lowest; and the weight 0 has the multiplicity of 4. Sym2 π1,0
decomposes into a direct sum of two irreducible representations according to
Sym2 π1,0 ∼ = C ⊕ π2,0, where C ∼ = π0,0 and π2,0 a 27-dimensional irreducible rep-
resentation of g2 with highest weight 2( ε 1 + ε 2 ).
In conclusion, π1,0 ⊗ π1,0 is reducible, with the decomposition into irreps,
In fact, in general any irrep πab of highest weight aω 1 + bω2 can be obtained from
decomposition of some tensor power of the standard representation.
r r r r
r dr dr r r dr dr r
dr f
dr dr r dr f
dh
r dr r
r dr dr r r dr dr r
r r r r
π1,0 ⊕ π0,1 = 7 + 14. (b) RHS: Weight diagram of Sym2 ϕ1 decomposable into
π0,0 ⊕ π2,0 = 1 + 27.
7.2. IRREDUCIBLE REPRESENTATIONS 237
CSA : h = C4 ; h0 = R 4 . Basis : { ei ; i = 1 − 4} .
Σ (4) : α1 = 1/2( ε 1 − ε 2 − ε 3 − ε 4), α2 = ε 4, α3 = ε 3 − ε 4, α4 = ε 2 − ε 3.
∆+ (24) : ε i , ε i + ε j, ε i − ε j ( i < j ), 1/2( ε 1 ± ε 2 ± ε 3 ± ε 4 ) .
FCR (4) : h1 = e1 − e2 − e3 − e4 , h2 = 2e4 , h3 = e3 − e4 , h4 = e2 − e3 .
FW (4) : ω i ( h j) = δij ; ω 1 = (1, 0, 0, 0), ω 2 = 1/2(3, 1, 1, 1), ω3 = (2, 1, 1, 0),
ω 4 = (1, 1, 0, 0);
δ = ∑ ω i = 1/2(11, 5, 3, 1) all written in basis ( ε 1, ε 2, ε 3, ε 4 ).
CSA : h = C 6 ; h0 = R 6 .
√
Σ ( 6) : α1 = 1/2 ( ε 1 − ε 2 − ε 3 − ε 4 − ε 5 + 3ε 6 ), α2 = ε 1 + ε 2,
α 3 = ε 2 − ε 1 , α 4 = ε 3 − ε 2, α 5 = ε 4 − ε 3, α 6 = ε 5 − ε 4.
√
∆+ (36) : ε i + ε j |i< j≤5 , ε i − ε j | j<i≤5 , 1/2(± ε1 ± ε 2 ± ε 3 ± ε 4 ± ε 5 + 3ε 6)
(with even number of minus signs) .
√
FCR (6) : h1 = 1/2 (e1 − e2 − e3 − e4 − e5 + 3e6 ), h2 = e1 + e2 ,
hi = ei−1 − ei−2 ( i = 3, 4, 5, 6).
√ √
FW(6) : ω 1 = (0, 0, 0, 0, 0, 2/ 3), ω 2 = 1/2(1, 1, 1, 1, 1, 3),
√ √
ω 3 = 1/2(− 1, 1, 1, 1, 1, 5/ 3), ω 4 = (0, 0, 1, 1, 1, 3),
√ √
ω 5 = (0, 0, 0, 1, 1, 2/ 3), ω 6 = (0, 0, 0, 0, 1, 1/ 3).
√
∑ ωi : δ = (0, 1, 2, 3, 4, 4 3) in basis ( ε 1, ε 2, ε 3, ε 4, ε 5, ε 6 ).
The smallest irreps have the dimensions 1, 27 (twice), 78, 351 (four times), 650,
1728 (twice), 2430, 2925, 3003 (twice). The existence of multiple non-isomorphic
representations for some dimensions can be understood in part from the symme-
try of the Dynkin diagram for E6. In the configuration of the Dynkin diagram
consistent with the definitions of αi and ω i given above (where the last node α2 is
linked to the middle node α4, and α1, α3 are symmetric with α5, α6 resp.) the fun-
damental irreps of highest weights ω 1, ω 2, · · · , ω 6 have dimensions 27, 78, 351,
2925, 351, 27. Of all the exceptional algebras, e6 is the only one that has any non
self-dual irreps (making it a possible candidate for a flavor-chiral particle theory).
238 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
CSA : h = C 7 , h0 = R 7 .
√
Σ (7) : α1 = 1/2 ( ε 1 − ε 2 − · · · − ε 6 + 2 ε 7 ), α2 = ε 1 + ε 2,
αi = ε i−1 − ε i−2 ( i = 3, 4, . . . , 7).
√
∆+ (63) : εi + ε j ( i < j ≤ 6); εi − ε j ( j < i ≤ 6); 2 ε7.
√
1/2 (± ε1 ± ε 2 ± · · · ± ε 6 + 2 ε 7 ) (odd number of minus signs).
√
FCR (7) : h1 = 1/2 (e1 − e2 − · · · − e6 + 2 e7 ), h2 = e1 + e2 ,
hi = ei−1 − ei−2 ( i = 3, 4, 5, 6, 7).
√ √
FW (7) : ω 1 = (0, . . . , 0, 2), ω 2 = 1/2(1, 1, 1, 1, 1, 1, 2 2),
√ √
ω 3 = 1/2(− 1, 1, 1, 1, 1, 1, 3 2), ω 4 = (0, 0, 1, 1, 1, 1, 2 2),
√ √
ω 5 = (0, 0, 0, 1, 1, 1, 3/ 2), ω 6 = (0, . . ., 0, 1, 1, 2),
√
ω 7 = (0, . . . , 0, 1, 1/ 2).
√
∑ ωi : δ = (0, 1, 2, 3, 4, 5, 17/ 2).
CSA : h = C 8 , h0 = R 8 .
Σ (8) : α1 = 1/2 ( ε 1 − ε 2 − · · · − ε 7 + ε 8), α2 = ε 1 + ε 2,
αi = ε i−1 − ε i−2 ( i = 3, 4, . . . , 8).
∆+ (120) : 1/2(± ε1 ± ε 2 ± · · · ± ε 7 + ε 8) (even number of minus signs),
εi + ε j ( i < j ≤ 8); εi − ε j ( j < i ≤ 8).
FCR (8) : h1 = 1/2 (e1 − e2 − · · · − e7 + e8 ), h2 = e1 + e2 ,
hi = ei−1 − ei−2 ( i = 3, 4, . . . , 8).
FW(8) : ω 1 = (0, . . . , 0, 2), ω 2 = 1/2(1, . . . , 1, 5),
ω 3 = 1/2(− 1, 1, . . . , 1, 7), ω 4 = (0, 0, 1, . . . , 1, 5),
ω 5 = (0, 0, 0, 1, . . . , 1, 4), ω 6 = (0, . . . , 0, 1, 1, 1, 3),
ω 7 = (0, . . . , 0, 1, 1, 2), ω 8 = (0, . . . , 0, 1, 1).
∑ ωi : δ = (0, 1, 2, 3, 4, 5, 6, 23).
7.3.1 Preliminaries
Let h be a Cartan sub algebra of a simple Lie algebra g, and π a representation of
g on vector space V . Under the action of h, V decomposes into a direct sum of the
weight spaces Vµ , each attached to a weight µ of π relative to h, with a dimension
equal to nµ , the multiplicity of the weight µ. We write accordingly:
V = ⊕ µ Vµ , dim Vµ = nµ . (7.8)
For each positive root α ∈ ∆+ , one may define a sub algebra sα of g of type
a1 ∼
= sl (2, C ) spanned by the operators Hα and E±α (with E−α ≡ Fα ), which obey
the standard commutation relations [ Eα , E−α ] = Hα and [ Hα, E±α ] = ± α ( Hα) E±α ,
where Hα is normalized such that α ( Hα ) = 2.
Before considering the FMF and WCF, we will state a few relevant results,
labeled R1, R2, . . . as follows.
R1. In any irreducible representation πn of Lie algebra a1 = { H, E, F } on vector space
V , let vk be the eigenvectors with eigenvalues µk of H and k = 0, 1, . . . , n. Then vk are
also eigenvectors of EF with eigenvalues tk . The following relations hold:
(a) TrV H = µ0 + µ1 + · · · + µn = 0.
(b) tk − tk−1 = µk (with t−1 = 0), so that tk = µ0 + µ1 + · · · + µk .
(a) Since µk are symmetric about 0 and all of multiplicity of one, their sum
vanishes. (b) Since EF and FE commute with H, they also admit v k as eigenvec-
tors with eigenvalues tk and t0k . From Fvk = vk+1 and EFvk = tk vk , we deduce
EFvk−1 = Evk = tk−1 vk−1 . Then from these three relations, we get FEvk = tk−1 vk ,
and so t0k = tk−1 . The commutation relation [ E, F ] = H acting on v k produces
tk − tk−1 = µk . This is an iterative relation for tk , starting with t0 = µ0; its solu-
tion is tk = µ0 + µ1 + · · · + µk , as stated. Further, as we know that µk = n − 2k,
the sum yields tk = ( k + 1)( n − k ). Noting from this result that tn = 0, we also
obtain TrV H = 0, already found in (a).
R2. In any representation π of a semisimple Lie algebra g, let Vρ be the weight space
for the weight ρ, and let Hα , Eα , Fα define the sub Lie algebra a1 ∼
= sα < g for any root
240 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
α 0 of g. With Trρ denoting a trace on Vρ , let tα,ρ = Trρ Eα Fα and t0α,ρ = Trρ Fα Eα .
Then, if ρ + α is a weight, we have t0α,ρ = tα,ρ+α ; and if not, t0α,ρ = 0. Furthermore,
tα,ρ − t0α,ρ = nρ ρ ( Hα ), where nρ = dimVρ.
R2 evidently generalizes R1 (b) to the situation where sα is a subalgebra. If
ρ and ρ + α are weights of π, then from the cyclicity property of the trace, we
have Trρ Fα Eα = Trρ+α Eα Fα (by inserting intermediate states). Taking the trace of
both sides of Eα Fα − Fα Eα = Hα and noting that Tr ρ Hα = nρ ρ ( Hα ), we obtain the
stated result.
R3. For each root α 0, the action of sα < g on the representation space V decomposes
it into irreducible subspaces Wa (a = 1, 2, . . . ), such that each Wa is itself a direct sum of
weight spaces for Hα . We express this fact in the form
V = ⊕ a Wa , Wa = ⊕ jWρa+ jα , (7.9)
k
tα,µ = ∑ nµ+iα ( µ + iα )( Hα). (7.11)
i= 0
Letting k → ∞, we obtain the relation given in R5. This result says that the trace
of Eα Fα on any weight space Vµ is the sum of its weight and of all higher weights
factoring in their multiplicities.
In the basis { Xi} , one may write [ Xi , Z] = ∑j cij Xj , or inverting this with Xi0 , we
get the constant cij = ([ Xi, Z] : X0j ) = ( Xi : [ Z, X0j ]), in which the second equality
follows from invariance of the Killing form, ([ a, b] : c ) = ( a : [ b, c ]). This gives
cij = −([ X0j, Z] : Xi ), or equivalently, [ X0j , Z] = − ∑i cij Xi0 . All the terms in the
expansion of [ C, Z ] cancel in pairs, leading to [ C, Z ] = 0 for every Z in π (g).
(3) Now if the representation π is irreducible, then by Schur’s lemma, C acts as
a scalar multiplication in V by a constant which may serve to characterize π.
In any representation π of g, a natural choice of basis is the set of elements
{ Hi, i = 1, 2, . . . , `; Eα , α ∈ ∆ } , where Eα are the root vectors corresponding to
the (positive or negative) roots α, and Hi = Hαi , are the fundamental co-roots
242 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
normalized in the usual way, αi ( Hi ) = 2. Then the dual basis, obtained from its
definition as X0j = Xk [ κ −1]kj , where [ κ −1] is the inverse of the matrix [ κ ], with ele-
ments given by the Killing products κij = ( Xi : Xj ), can be easily calculated since
we know that [ κ ] is block-diagonal by the orthogonality of the CSA and various
gα ⊕ g−α , and so restrictions of [ κ ] to the disjoint subspaces can be calculated sep-
arately. The elements Hi0 dual to Hi are defined by the condition ( Hi : Hj0 ) = δij
on the restriction of the Killing form to h. One finds Hi0 = Hk bki | αk |2 /2, where
| α |2 = h α, α i, and bki are the elements of the inverse Cartan matrix.
For any α ∈ ∆, one determines the element Hα in h normalized such that
α ( Hα ) = 2. Then E±α are found as root vectors with the roots ± 2 relative to Hα .
We already know that ( Eα : Eβ ) = 0 unless α + β = 0, and so E±α are dual with
respect to the Killing form. All we need now is to consistently normalize them. By
the invariance property of the Killing form, we have ( Hα : H ) = α ( H )(Eα : Fα ).
Recalling now that ( Hα : H ) = α ( H )(2/ |α|2), we get the form ( Eα : Fα ) = 2/ | α|2.
Hence the dual to Eα is given by Eα0 = (| α|2/2) E−α, with Fα ≡ E−α .
In the bases { Hi, i = 1, . . . , `; Eα , α ∈ ∆ } and { Hi0, i = 1, . . . , `; Eα0 , α ∈ ∆ } just
defined, the (quadratic) Casimir operator assumes the form
`
C= ∑ Hi Hi0 + ∑ Eα Eα0 . (7.12)
i= 1 α ∈∆
This is a convenient basis for C because every term Xi Xi0 acts as multiplication by
a scalar on any weight space, and this scalar can be calculated in terms of weight
multiplicities. In addition, if the representation is irreducible, C = cI where c is
a (Casimir) constant characteristic of the representation. Taking the trace on both
sides of (7.12) then produces a relation among the weight multiplicities.
E XAMPLE 16: The Lie algebra a1 has one-dimensional CSA h spanned by the unit
vector ε: there exist one fundamental root α1 = 2ε and two roots α = ± α1 (so
that | α |2 = 4). Its Cartan matrix A = 2 has the inverse B = 1/2. So we have
H10 = H1 B | α |2/2 = H1, and Eα0 = 2E−α . It follows that the Casimir operator
is C = H12 + 2( Eα1 E−α1 + E−α1 Eα1 ). The canonical generators correspond to the
angular momentum operators, H1 = 2J3 and E± α = J± = J1 ± iJ2 ; hence the
Casimir operator is C = 4 J32 + 1/2( J+ J− + J− J+ ) = 4J 2 .
It will be seen below that for an irreducible representation of any Lie algebra
of highest weight λ, with half the sum of positive roots given by δ, the Casimir
invariant equals c ( λ) = h λ, λ + 2δ i. In the case of a1, we have δ = 1/2α1 and
λ = mω 1 = ( m/2)α1, where m is an N number, and so we get the inner product
h λ, λ + 2δ i = m/2( m/2 + 1)h α1, α1 i, which, with | α1|2 = 4, takes the form 4j ( j + 1)
found in the angular momentum algebra. This implies j = m/2 = 0, 1/2, 1, . . . .
E XAMPLE 17: For a2, the quadratic Casimir invariant of irrep λ = ( m1, m2 ) is
given by c ((m1, m2 )) = 1/9(3m1 + 3m2 + m21 + m22 + m1 m2 ).
C OMMENTS .
(a) The value of c depends on the normalization of Xi , or of αi . A related quan-
tity, given by I ( λ) = [dim( πλ ) /dim(g)] c(λ), called the index of representation,
appears in the renormalization group equations of quantum field theory.
7.3. DIMENSION OF REPRESENTATION 243
(b) C defined here is called the quadratic Casimir operator. For a` one may
also define a third-order operator C3 of the same kind by d ijk Xi Xj Xk where dijk is
a completely symmetric invariant tensor. It is involved in the studies of anomalies
in particle theory. Casimir operators of higher orders do not have such simple
explicit expressions, but may be written as homogeneous polynomials of higher
orders of the basis elements. They do not exist in all orders for all Lie algebras
(C exists for all simple Lie algebras, but C3 only for a` ). Thus, for the simple Lie
algebras in the classical series, they exist only in orders of 2, 3, . . . , ` + 1 for a` ; of
2, 4, . . . , 2` for b` and c` ; and, finally, of 2, 4, . . . , 2 ` − 2 and ` for d` .
Note that an explicit expression for Hi0 is not needed in this calculation.
Recalling that ρ ( Hα) equals (2/ | α|2)h ρ, αi for any ρ ∈ h∗ , and the expression
Eα0 = (| α|2/2) Fα, we can use (7.10) to write
∞
| α |2
Trµ ( ∑ Eα Eα0 ) = ∑ tα,µ = ∑ ∑ nµ+iα h µ + iα, α i. (7.14)
α ∈∆ α ∈∆
2 α ∈∆ i = 0
The two terms with ± α and i = 0 cancel out, leaving us with
∞ ∞
∑ ∑ nµ+iα h µ + iα, α i + ∑ ∑ nµ+iα h µ + iα, α i
α∈∆+ i=1 α∈∆− i=1
∞ ∞
= ∑ ∑ nµ+iα h µ + iα, α i − ∑ ∑ nµ−iα h µ − iα, α i.
α∈∆+ i=1 α∈∆+ i=1
Using the identity given in R4, we can see that the second term reduces to the
double sum ∑α0 ∑∞ i=0 nµ+iα h µ + iα, α i . It follows that
∞
Trµ ( ∑ Eα Eα0 ) = nµ ∑ h µ, α i + 2 ∑ ∑ nµ+iα h µ + iα, α i. (7.15)
α ∈∆ α∈∆+ α∈∆+ i=1
The trace of the Casimir operator C on Vµ in any representation π then follows
from the above results, together with the definition δ = 1/2 ∑α∈∆+ α :
∞
Trµ C = nµ h µ, µ + 2δ i + 2 ∑+ ∑ nµ+iα h µ + iα, α i. (7.16)
α ∈∆ i = 1
244 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
c = h λ, λ + 2δ i = | λ + δ |2 − | δ |2. (7.17)
χπ ( h ) = ∑ nµ exp µ ( h ), (7.19)
µ∈Λ
where nµ are the multiplicities of the weights µ (i.e. the dimensions of the weight
spaces Vµ ). As h is normally restricted to h0 (the real part of h), we may replace
µ with 2πiµ and exp µ ( h) with exp 2πµ (ih), making apparent the Cartan subal-
gebra ih0 of the compact form of g. Nevertheless, to keep it short, we keep using
the simpler notation without making the factor 2πi explicit.
Hermann Weyl has given a simple formula (WCF) for the character of a finite-
dimensional irreducible representation πλ of a semisimple Lie algebra with high-
est weight λ as the quotient of two sums of exponentials:
for any element h of h0. The summation runs over all elements w of the Weyl
group W of the Lie algebra g; and δ is half the sum of the positive roots (equal to
the sum of the fundamental weights). The power of this formula lies in the fact
that it does not depend on the details of the representation, only on its highest
weight λ, and on two characteristics of the Lie algebra, namely δ and W.
For a Lie algebra g of rank ` and Cartan sub algebra h, the weights ρ of any
representation of g are integral linear functions of h0 , i.e. such that all values
ρ ( h̃α ), with α ∈ ∆, are integers. All such integral forms constitute a lattice Λ for
g, a free abelian subgroup of h∗0 generated by the fundamental weights ω i . To each
ρ of Λ we associate the function exp ρ, or more generally the symbol e ( ρ), subject
to the composition rule e ( ρ)e(σ) = e ( ρ + σ ) and the normalization e (0) = 1
(identity); and consider arbitrary finite integral linear combinations of the form
∑ρ mρ e ( ρ ), with integral coefficients mρ and summation of ρ over a finite subset
of Λ. The collection of all such sums will be conventionally called ZΛ, and the
sum defining the character χ in (7.19) is an element of ZΛ.
The Weyl group W of g acts on h∗0 , and so also on Λ and ZΛ. A Weyl reflection
w acts on ZΛ according to1 we ( ρ) = e ( w · ρ) = e ( wρ). When applied to a product,
it gives w e ( ρ )e (σ ) = e ( wρ) e(wσ ) since e ( ρ) e(σ) = e ( ρ + σ ).
1 In e(ρ)(h ), with ρ ∈ h∗0 and h ∈ h0 , ρ labels the ‘state’ while h labels the function variable. An
operator w acts either on state we(ρ)( h) = e(wρ )( h), or on function variable we(ρ )(h ) = e(ρ)(w −1 h).
In the latter view, the interpretation is ρ(w−1 h) = hρ, w−1 σ i, where σ ∈ h∗0 is the conjugate to h ∈ h0
via the Killing form. The distinction between the roles of ρ and h can be made more apparent by
246 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
Among the elements of ZΛ, there are those that have a definite symmetry un-
der the action of W: An element a is said to be symmetric if wa = a, and alternating
or skew if wa = det w a for all w in W, where det w = (−)rw and rw the number of
positive roots that w sends to negative ones. Cf. discussion on page 218.
Combinations such as ∑w we ( ρ) (summing over all w ∈ W) are symmetric for
any ρ in Λ. Another class of symmetric elements is exemplified by the character
in (7.19), χ = ∑µ nµ e ( µ ). Their symmetry stems from invariance of π under W
and nwµ = nµ for all w in W, as can be seen below:
But the key to the proof of (7.20) lies, not in the symmetric, but rather in the skew
elements of ZΛ, those that may acquire a sign under Weyl reflections w according
to the sign of the det w factors. For an arbitrary ρ ∈ Λ, define
Let B stand for any one of the equivalent expressions on the right-hand side of
(7.22). Then:
(i) The equality of the three displayed expressions stems from the equality of
δ to half the sum of positive roots: e ( δ ) = e ( ∑ α/2) = ∏ e ( α/2).
(ii) B changes sign when a Weyl reflection corresponding to a simple root αi
is applied to it. To see this, we use the first expression for B, and recall that wαi
changes the sign of αi and permutes the other positive roots. Hence wαi B = − B
for every αi, which implies B is alternating, and so must be an integral linear
combinations of the Aρ’s with strongly dominant weights ρ.
(iii) Multiplying out the product in the second expression, we see that e ( δ )
appears with coefficient 1 and has the highest weight, since all other terms cor-
respond to lower weights of the form δ − ∑ α with positive α. Moreover besides
e ( δ), there are no other contributions from the strongly dominant weights.
(iv) As B (any one of the equivalent products on the right-hand side of (7.22))
is skew, and has the term e ( δ ) with coefficient 1, it must be equal to Aδ .
where we have used the fact that detw = ± 1, and that w and s = w−1 go over the
same W.
We take the limit of χ ( ρ ) = Aλ+δ ( ρ )/Aδ( ρ ) as ρ goes to 0 along the path
ρ = tδ with t → 0. From the symmetry of Aµ ( ρ ), we have
Aλ+δ ( tδ ) A ( t (λ + δ ))
χ ( tδ) = = δ .
Aδ ( tδ ) Aδ ( tδ )
exp (h λ + δ, α i t) − 1
χ ( tδ) = exp (−hλ, δit) ∏ .
α0 exp(h δ, α it) − 1
The limiting value χ (0) as t → 0 is now well defined and gives the dimension of
the irreducible representation πλ :
h λ + δ, α i
dλ [g] = dim πλ = ∏ h δ, α i
. (7.23)
α0
A good way to calculate dλ [g] from (7.23) is to write λ and δ in terms of the
fundamental weights ω i, and every α ∈ ∆+ in terms of the simple roots αi , and
use the relation h ω i, α j i = δij | αi |2 /2 (no summation) for the dual bases. So, let
λ = ∑i mi ω i , α = ∑i k αi αi , δ = ∑i ω i, where mi ≥ 0, k αi ≥ 0, with 1 ≤ i ≤ `.
Then one obtains h α, λ + δ i = ∑i k αi ( mi + 1)| αi|2 /2 in the numerator, and h α, δ i =
∑i k αi | αi |2 /2 in the denominator. The dimension formula (7.23) then becomes
∑i k αi ( mi + 1)| αi|2
dim πλ = ∏ ∑i k αi | αi |2
. (7.24)
α0
S PECIAL CASES .
(a) If the highest weight λ of the representation equals δ, then (7.23) reduces
to dim πλ = 2 p , where p = | ∆+ | is the number of positive roots in the algebra.
(b) If all the fundamental roots have the same lengths, then (7.24) reduces to
∑i k αi ( mi + 1)
dim πλ = ∏ ∑i k αi
. (7.25)
α0
E XAMPLE 20: The algebra a1 has just one positive root equal to the single fun-
damental root, so k α1 = 1. The fundamental weight λ is specified by the non-
negative integer m1 = m. The product in (7.25) has a single factor, so that the
dimension of the irreducible representation of a1 of highest weight m is given by
dm [a1 ] = m + 1.
E XAMPLE 21: The algebra a2 has three positive roots, α1, α2 and α1 + α2, from
which we read the k αi ’s. The highest weight of the irreducible representation πλ is
λ = ( m1, m2 ). The three factors in (7.25) corresponding to the three positive roots
are (( m1 + 1)1 + 0) / (1 + 0) = m1 + 1, (0 + ( m2 + 1)1) /(0 + 1) = m2 + 1, and
(( m1 + 1)1 + ( m2 + 1)1) /(1 + 1) = ( m1 + m2 + 2) /2. This gives the dimension
of the irreducible representation of a2 of highest weight λ = ( m1, m2 ) as
1
dm1 ,m2 [a2 ] = ( m1 + 1)( m2 + 1)( m1 + m2 + 2). (7.26)
2
E XAMPLE 22: The algebra b2 has four positive roots: α1, α2, α1 + α2, and α1 + 2α2.
The simple roots α1 and α2 satisfy | α1|2 = 2| α2|2 = 1/3 and h α1, α2 i = − 1/6. In
(7.24), the numerator is 2k α1 ( m1 + 1) + k α2 ( m2 + 1), and the denominator 2k α1 + k α2 .
So we get for the four positive roots (in the same order) the factors ( m1 + 1),
( m2 + 1), (2m1 + m2 + 3) /3, and ( m1 + m2 + 2) /2, leading to the dimension for
the representation of b2 of highest weight ( m1, m2 ):
1
dm1 ,m2 [b2 ] = ( m1 + 1)( m2 + 1)( m1 + m2 + 2)(2m1 + m2 + 3). (7.27)
6
E XAMPLE 23: The algebra g2 has six positive roots, α1, α2, α1 + α2, 2α1 + α2, 3α1 +
α2, and 3α1 + 2α2, from which we read the k αi ’s. The highest weight of πλ is
λ = m1 ω 1 + m2 ω 2. Remembering now to include the different norms | α1|2 = 1
and | α2|2 = 3, we obtain the following factors in (7.24):
– For α1: m1 + 1;
– For α2: m2 + 1;
– For α1 + α2: (( m1 + 1) + 3( m2 + 1)) /(1 + 3) = ( m1 + 3m2 + 4) /4;
– For 2α1 + α2: (2( m1 + 1) + 3( m2 + 1)) / (2 + 3) = (2m1 + 3m2 + 5) /5;
– For 3α1 + α2: (3( m1 + 1) + 3( m2 + 1)) / (3 + 3) = ( m1 + m2 + 2) /2;
– For 3α1 + 2α2: (3( m1 + 1) + 6( m2 + 1)) /(3 + 6) = ( m1 + 2m2 + 3) /3.
250 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
1
dm1 ,m2 [g2] = ( m + 1)( m2 + 1)(m1 + m2 + 2)
120 1 (7.28)
· ( m1 + 2m2 + 3)( m1 + 3m2 + 4)(2m1 + 3m2 + 5).
Table 7.1: Factors in the dimension formula for the irrep with highest weight
( m1, m2, m3 , m4 ) of d4 = so(8, C )
the internal (non time-spatial) spaces of color, isospin, and hypercharge. All par-
ticles then evolve in representations of GSM on finite-dimensional Hilbert spaces
V, and since GSM and its component groups are compact Lie groups, every irre-
ducible representation (irrep) V of GSM may be decomposed to a direct product
of irreps of the form V3 ⊗ V2 ⊗ V1, where V3, V2, and V1 are respective irreps of
SU(3), SU(2), and U(1). Not only does this give us a way to organizing particles,
but also to uncovering their interactions as well.
In the Standard Model, there are three kinds of particles: the fundamental
fermions, which constitute matter; the gauge bosons, which carry the (strong and
electro-weak) interactions; and the Higgs bosons, which give masses.
Following a long line of theoretical arguments and experimental observations,
all the particles we now know are completely identified by their quantum num-
bers. Thus, we know that the fundamental fermions, of which matter is ultimately
made up, are structureless particles of spin one-half. They are of two kinds, quarks
and leptons, distinguished by the property that the former are sensitive to the
strong interaction whereas the latter are not. They come in different flavors: up
(u), down (d), electron (e), and neutrino (ν). As we will see later, there are others,
but these are the lightest ones, forming what one calls the first generation.
Each flavor of quark (u, d) comes in three different ‘strong charges’ called
colors, labeled i = 1 (red), 2 (green), and 3 (blue). Observations (e.g. the spins of
the baryons and mesons) and the spin–statistics relationship point to an internal
SU(3) symmetry of color in which the quarks of each flavor transform among
themselves in a 3 representation of SU(3). So u1, u2, u3 are basis vectors of the
standard representation of SU(3) in C3 ; and similarly d1, d2, d3 in another copy of
C3 . On the other hand, the leptons (e, ν) carry no color, and so may be considered
color neutral in this context, sitting together in C ⊕ C.
Massless quarks and leptons in motion can be characterized by their helicity
± 1/2, the component of their spin in the direction of motion. Those with he-
licity +1/2 are said to be right-handed, while those of helicity −1/2, left-handed.
The antiparticle of a right-handed particle is left-handed, and vice versa. Helicity
is a relativistically invariant quantity for massless particles, but not for massive
particles (and so these must be linear superpositions of left- and right-handed
components). At the symmetry level where the fundamental fermions carry no
mass, there may exist separate left- and right-handed quarks and leptons, uiL , uiR ,
diL , diR , e L , e R, νL , νR, and their conjugate antiquarks and antileptons. However,
both components need not exist in nature; in fact, right-handed neutrinos and
left-handed antineutrinos seem not to exist at all. So one may take for the first-
generation fundamental fermions of the SM the set of 15 particles uiL , diL , uiR , diR ,
νL , e L , and e R, although at times one may want to include νR (making it a set of 16),
for example when studying neutrino oscillations, or Higgs boson interactions.
A salient feature of the weak force (and only of the weak force) is that it vio-
lates parity, i.e. it acts non symmetrically on particles of different helicities. No
other physical law is asymmetric in left and right. The parity violation so ob-
served is maximal, such that only left-handed particles or right-handed antipar-
ticles participate in weak interactions. Thus, neutrons (n) decay into protons (p)
252 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
Table 7.2: Irreducible representations of GSM = SU (3) × SU (2) × U (1)Y for the
fermion first generation in the Standard Model. It is equivalent, and convenient,
to replace u R , d R, νR , e R with ū L , d¯L , ν̄L , e+
L in the definition of F. The values of Y
are explicitly in indicated in 1Y .
Thus, all the fundamental fermions of the first generation are defined as basis
vectors of three irreps of SU(5): F = 10 ⊕ 5 ⊕ 1.
Here the vector gauge bosons, which belong to the adjoint representation 24,
include besides the Wµi , Bµ and the QCD gluons, twelve new gauge bosons, called
256 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
Problems
7.1 (Weyl group for a2). Consider the Weyl group for a2 (see Example 1 in the
chapter). (a) Give the transformations of the positive roots and an arbitrary
weight under the Weyl group. (b) Give the orbits of the dominant weights (0, 0),
(1,0), (0,1), (1,1), (2,0), (0,2), (3,0) and (0,3).
7.2 Weyl group for b2 . (a) Find the Weyl group for the Lie algebra b2 . (b) Give the
orbits of the dominant weights (0,0), (1,0), (0,1), (1,1), (2,0), (0,2).
7.3 (Properties of Weyl reflections). Let W be the Weyl group for a Lie algebra
g. Prove the following: (a) The Weyl reflections are isometries on h∗ , i.e trans-
formations that preserve the scalar product. (b) The reflection wα on the hyper-
plane perpendicular to a simple root α permutes all other positive roots. (c) Let
δ = 1/2 ∑β∈∆+ β, i.e. half the sum of the positive roots, and w 6 = 1 be an arbitrary
Weyl reflection. δ − w · δ is a sum of distinct positive roots. (d) δ defined in (c) is
the sum of the fundamental weights, and so is a weight.
7.4 (Dual representations). We know that the NSC for an irrep πλ with highest
weight λ to be self-dual is that its lowest weight is − λ. Determine the self-duality
of the irreps of a2 , b2 , and g2 .
7.5 (Inner products of weights). Inner products for weights can be calculated in
the same way as for roots, since the two entities live on the same space, h∗0 . We
take sl (3, C ) to illustrate, and set µ = ( m1, m2 ) for any weight, where mi = µ ( hi)
and hi are the fundamental coroots. (a) Calculate h µ defined by ( hµ : h ) = µ ( h ),
where h is any vector of h0. (b) Calculate h µ, µ0i for any weights µ, µ0 ∈ h∗0 .
7.6 (Fundamental weights). Let hi with i + 1, . . . , ` be the fundamental coroots of
a simple Lie algebra, and ωi the fundamental weights, so that ωi ( h j ) = δij . Recall
that for any simple root αi, we also have αi ( h j) = a ji (the Cartan integers), so
that αi = ∑j ω j a ji . The sets Σ = { αi; i = 1, . . . , `} and Φ = { ω i; i = 1, . . . , `} are
two bases for h∗0 . In addition, there is also the canonical coordinate basis { ε i} , in
which αi = ∑i ε j c ji (where c ji are given in the text of this chapter for all simple
algebras). Find the expansion coefficients of ω i in { ε j} for the algebras b3 , d4, g2 ,
f4, e6, e7 , and e8 .
7.7 (Representations of b2 ). Find the irreducible representations of b2 with high-
est weights (0,1), (1,0),(0,2) and (2,0). Discuss the representations of the tensor
products π0,1 ⊕ π0,1, π1,0 ⊕ π1,0, and π0,1 ⊕ π1,0.
7.8 (Representations of g2 ). (a) Find the Weyl group for the Lie algebra g2 . (b)
Give the Weyl orbits of the dominant weights (1,0), (0,1), (1,1), (2,0), (0,2). (c) Give
the weights of the irreducible representation having (0,1) as the highest weight.
7.9 Find the integral expansion ∑i k i αi (where αi are simple roots) for the positive
weights of the classical algebras a` , b` , c` , and d` .
7.10 (Dimension of the irreps of a` ). Find the dimension of the irreps of a` for
arbitrary `.
Q. Ho-Kim. Group Theory: A Physicist’s Primer.
258 CHAPTER 7. LIE ALGEBRAS: REPRESENTATIONS
m1 : −2 −1 0 1 2 3
u u e e
u e e
u u e e
u e e
m2 : u u e e
α2
3
u e ω2 e
u u e ω1 u
2
u e u α1
u u u u
1
u u u
u u u u
0
u u u
u u u u
−1
u u u
u u u u
−2
Figure 7.11: A section of the weight lattice for a2 . α1 and α2 are the simple roots,
with | α1| = | α2|; ω 1 and ω 2 the fundamental weights. Weights are labeled in the
system { ω 1, ω 2 } by (m1, m2 ).
WEIGHT LATTICE 259
m1 : −2 −1 0 1 2
t t d d d
t α2 t dω 2 dω1 t
@
@
m2 : @
@
3 @
t t @d t tα1
2
t t t t t
1
t t t t t
0 −1 −2
A section of the weight lattice for b2 . α1 and α2 are the simple roots,
Figure 7.12: √
with | α | = 2| α2|; ω 1 and ω 2 the fundamental weights. Weights are labeled in
1
s s s s c c c s s
s s s s c c s s
ω2
s s s s c c s s s
s s α2 s s c ω1 s s s
s s s s s s α1 s s s
s s s s s s s s
s s s s s s s s s
s s s s s s s s
s s s s s s s s s
A section of the weight lattice for g2 . α1 and α2 are the simple roots,
Figure 7.13: √
with | α2| = 3| α1|; ω 1 and ω 2 the fundamental weights. Weights are labeled in
the system { ω 1, ω 2 } by (m1, m2 ).
Appendix A
261
262 APPENDIX A. THE SYMMETRIC GROUP
contain no proper subideals are said to be minimal. Minimal left ideals correspond to
e and are therefore objects of interest. (A right ideal is similarly
irreps of the algebra G
defined as a subalgebra invariant under right multiplication, that is the set of all
vectors h of G e such that hx is some vector of the set for all x ∈ G.)
e
6. The regular representation. As we have mentioned, the regular representation
DR of a group algebra G e results from the operators U ( x) acting on the algebra
itself. We learned from Chapter 2 that the regular representation of any group
(and therefore of any group algebra) is fully reducible and contains each distinct
irreducible representation, Dµ , a number of times equal to the dimension d µ of
that irrep. When completely reduced, it appears as a direct sum, DR = ⊕ µ dµ Dµ .
In other words, the group algebra can be decomposed into a direct sum of irre-
ducible invariant subspaces Laµ , where a = 1, 2, . . . , dµ . As subalgebras of G, e Laµ
are minimal left ideals, which we write as G e = ⊕ µ Lµ , where Lµ is a left ideal
a
completely reducible into ⊕ a Lµ . Symbolically we write
DR = ⊕ dµDµ ⇐⇒ G e = ⊕Lµ = ⊕ µ (⊕ a Laµ ).
7. The problem of finding the distinct irreducible representations reduces to iden-
e To make our discussion less abstract, let us
tifying all minimal left ideals of G.
assume that G e may be decomposed into two left ideals: G e = L1 ⊕ L2 such that
L1 ∩ L2 = { 0} , where by definition xL1 = L1 and xL2 = L2 for every x ∈ G. e
e
Then every x ∈ G is expressible in a unique way as the sum x = x1 + x2 of an
element x1 ∈ L1 and an element x2 ∈ L2. In particular, the identity e of G e (which
is also the identity element of G) is uniquely expressible as e = e1 + e2 where e1
(e2) is some element in L1 (L2). By definition of the identity, x = xe, we have
As L1 and L2 are left ideals, xe1 is in L1 and xe2 is in L2, which means
8. These ideals may or may not be reducible. If any one is not, we continue the
process of resolving the unit element into a sum of idempotents in that way, until
we can go no further: An idempotent eµ is said to be primitive if it cannot be
resolved into a sum of idempotents. If eµ is a primitive idempotent, the left ideal
Lµ = Gee µ is minimal; conversely, if Lµ is a minimal left ideal, any generating element
of Lµ is primitive.
9. To verify that an idempotent is primitive, one may use a simple criterion:
Theorem A.1 (Primitive idempotents). An idempotent eµ is primitive if and only if
e where λ x is some number, which may be zero or may depend
eµ xeµ = λ x eµ for all x ∈ G,
on x.
10. Finally, to select among primitive idempotents those which generate inequiva-
lent representations, one may use the following theorem:
Theorem A.2 (Equivalent primitive idempotents). Two primitive idempotents e1
and e2 generate equivalent irreducible representations if and only if e1 xe2 6 = 0 for some
e In other words, if e1 xe2 = 0 for all x ∈ G,
x ∈ G. e then the idempotents e1 and e2 must
generate inequivalent representations.
To generalize now, the group algebra G e associated with a finite group G can
be decomposed into left ideals Lµ and written G e = ⊕ µ Lµ , with µ running over all
inequivalent irreps of the group G. Each Lµ = Ge e µ is generated by an idempotent
eµ , which satisfies the conditions eµ eν = δµν eµ and ∑µ eµ = e. In addition, each Lµ
may be decomposed into dµ minimal left ideals Laµ with a = 1, . . . dµ corresponding
to the primitive idempotents eaµ , which obey eaµ xebµ = ceaµ δab for all x ∈ G, e where c
is a number. The idempotents eµ generate the inequivalent irreps µ of G, each of
multiplicity dµ which is the number of minimal left ideals in µ.
E XAMPLE 1: The identity representation. The algebra G e associated with any finite
group G contains the element s = | G |−1 ∑h∈G h, which satisfies gs = s and so
sgs = ss = s, for any g ∈ G. Hence s is a primitive idempotent. It generates
an irreducible representation of G on the group algebra. The basis vectors of the
representation are gs for each g ∈ G, but since gs = s (the same vector for every
g), the representation is one-dimensional, with the representation matrix given
by D ( g) = 1.
E XAMPLE 2: Algebra of C2. Group C2 has two elements a and e = a2. Any element
of the associated algebra must have the form f = αe + βa, and if it is an idempo-
tent, it must also satisfy f f = f , or α2 + β2 = α and 2αβ = β. Hence α = 1/2
and β = ±(1/2). From the previous example, we know that e1 = 12 ( e + a ) is a
primitive idempotent, which generates the identity representation. We check that
the second solution, e2 = 21 ( e − a ), is orthogonal to the first, e1 e2 = 0. In addition,
e2 ee2 = e2, e2 ae2 = − e2, and e2 ge1 = 0 with g = e, a. Hence if x = xe e + xa a is any
element of the algebra, then e2 xe2 = ( xe − xa ) e2, which according to Theorem A.1
means that e2 is also a primitive idempotent; and e2 xe1 = 0, which according to
Theorem A.2 means that e1, e2 are inequivalent.
A.1. ALGEBRAIC METHOD 265
α+ β +γ = 0, (A.7)
2
α + 2βγ − 3α = 0 , (A.8)
2
β + 2αγ − 3γ = 0 , (A.9)
2
γ + 2αβ − 3β = 0 . (A.10)
α = −( β + γ ) , (A.11)
0 = ( α − 1)( β − γ ) , (A.12)
a
β3 = γ3 = (3 − α )(3 − 2α ) . (A.13)
2
There are two cases: In the first, α = 1 and β 6 = γ; in the second, β = γ and α 6 = 1.
And so there are three solutions:
(i) α = 2, β = γ = − 1;
(ii) α = 1, β = ε, γ = ε 2;
(iii) α = 1, β = ε 2, γ = ε. √
where ε = exp(i2π/3) = − 21 + i 23 . We now check the primitivity (using Theo-
rem A.1) and the equivalence (using Theorem A.2) of the solutions.
. Solution (i): e0 = (2e − a − b ) /3
e0 e = ee0 = e0 ,
e0 a = ae0 = (− e + 2a − b ) /3,
e0 b = be0 = (− e − a + 2b ) /3,
e0 ee0 = e0 ,
e0 ae0 = (− ee0 + 2ae0 − be0 ) /3 = e0 a.
The last equation shows that e0 is not a primitive idempotent. It turns out, as seen
from the results obtained below, that e0 is decomposable, such that e0 = e2 + e3.
. Solution (ii): e2 = ( e + εa + ε 2 b ) /3
e2 e = ee2 = e2, e2 a = ae2 = ε 2 e2, e2b = be2 = εe2;
e2 ee2 = e2, e2 ae2 = ε 2e2, e2 be2 = εe2.
And so e2 is a primitive idempotent.
. Solution (iii): e3 = ( e + ε 2 a + εb ) /3.
One can check in a similar manner that e3 is also a primitive idempotent. It re-
mains to see whether e2 and e3 are equivalent or not. Applying Theorem A.2 we
find that they are not: e2 ee3 = e2 e3 = 0, e2 ae3 = ε 2 e2 e3 = 0, and e2 be3 = εe2e3 = 0.
266 APPENDIX A. THE SYMMETRIC GROUP
The left ideal L2 generated by e2 contains the basis vectors: ee2 = e2, ae2 =
e2 ε 2, and be2 = e2ε. Thus, the corresponding irreducible representation is one-
dimensional, with matrices D ( e ) = 1, D ( a ) = ε 2, and D ( b ) = ε. Similarly, the left
ideal L3 generated by e3 gives rise to another one-dimensional representation,
with matrices D ( e ) = 1, D ( a ) = ε, and D ( b ) = ε 2. To summarize we list all the
irreducible representations of the group C3:
µ : e a a2
1 : 1 1 1
2 : 1 ε2 ε
3 : 1 ε ε2
that the number of partitions of n is equal to the number of Young patterns, and
to the number of conjugacy classes of Sn , which in turn is equal to the number of
inequivalent irreducible representations of Sn . This suggests that one may associate
each Young pattern with an irreducible representation of Sn . It is important to note
that this correspondence is one-to-one. We will describe this relationship more
precisely in the next section after we have familiarized ourselves with this new
graphical tool.
E XAMPLE 4: For n = 2, 3, 4, we have the following partitions [ λ ] enumerated in
the conventional order in each case:
. n = 2: [2], [1, 1] ≡ [12]. The two partitions correspond to the graphs:
[2]: [1 2 ]:
[3]: [2,1]: [1 3 ]:
. n = 4: [4], [3,1], [2, 2] ≡ [22], [2, 1, 1] ≡ [2, 12], [1, 1, 1, 1] ≡ [14 ]. To each
partition, there corresponds a Young pattern:
14. Standard Young tableaux. A Young tableau for the symmetric group Sn (of
degree n) is a Young pattern in which n boxes are filled with n symbols (e.g. 1,
2,. . . , n) in any order, each symbol being used only once. A standard Young tableau
is a tableau in which the numbers 1, 2,. . . , n in each row appear in an increasing
order to the right, and those in each column appear increasing from top to bottom.
A normal Young tableau is a tableau in which the labels 1, 2,. . . , n appear in that
order from left to right and from the top row to the bottom row. To each pattern,
there corresponds a unique normal tableau Γ; one can obtain every other standard
tableau having the same pattern by applying an appropriate permutation p acting
on the labels in the normal tableau, producing the tableau pΓ.
This graphical tool yields a simple first result: We can find the dimension of any
irreducible representation of Sn by counting the number of standard tableaux of
the pattern corresponding to that irrep. (Another, more systematic, way is to
use the ‘hook’ formula described later.) The numbers of standard tableaux of
two patterns conjugate under the interchange of rows and columns (and so the
dimensions of the corresponding irreps of Sn ) are equal.
268 APPENDIX A. THE SYMMETRIC GROUP
15. Ordering the tableaux. It is conventional to list the tableaux of the same
pattern in the ‘dictionary order’, such that when all the numbers in the boxes
are written down on one line, one row followed by the next, the tableaux are
arranged in the increasing order of the resulting numbers. For example,
E XAMPLE 5: We list below the standard tableaux of the symmetric groups S2, S3
and S4, and give the number of standard tableaux d [λ] for each pattern [ λ ].
. S2: The partitions [2] and [1,1] of 2 result in two patterns, [2]: and
[ 12 ] : , with one standard tableau (d = 1) for each pattern. Both tableaux are
normal, and conjugates of each other.
1
[2]: 1 2 d[2] = 1; [ 12 ] : d[1,1] = 1.
2
. S3: There are three possible partitions of 3, namely [3], [2,1], and [1,1,1],
[3] : 1 2 3 d[3] = 1
1 2 1 3
[2,1] : d[2,1] = 2
3 2
1
[ 13 ] : 2 d[13 ] = 1
3
13
All the tableaux are normal, except of the pattern [2, 1], which is simply
2
standard, and can be obtained from the normal tableau of the pattern [2,1] by
applying permutation p = (23) to it. As a rule, the normal tableau for each
pattern always appears first in the dictionary ordering of all the tableaux.
A.2. GRAPHICAL METHOD 269
1 2 3 1 2 4 1 3 4
[3,1] : d[3,1] = 3
4 3 2
1 2 1 3
[ 22 ] : d[2,2] = 2
3 4 2 4
1 2 1 3 1 4
[2, 12] : 3 2 2 d[2,1,1] = 3
4 4 3
1
2
[ 14 ] : d[14 ] = 1
3
4
In each pattern, the first tableau is normal, the others are not but can be obtained
by permutations of the symbols in the normal tableau of the same pattern. For
example, the standard non-normal tableaux of pattern [3, 1] can be obtained by
applying the permutations (34) and (234) to the normal tableau. Note also that
the conjugate patterns have the same dimensions (d = 1 for the pair [4] and [14 ],
and d = 3 for the pair [3,1] and [2, 12 ]); on the other hand, [22] is self-conjugate,
with dimension d = 2. These examples illustrate the general rule.
16. Number of standard tableaux. There is a simple way of calculating the num-
ber of standard tableaux of any given pattern [ λ ], and hence the dimension of the
corresponding irrep of Sn . It is given by the ‘hook’ formula:
n!
d[λ] ( Sn ) = , (A.15)
H[λ]
where H = ∏ij hij , and hij is the hook number of the box B located on column j
and row i in the given pattern. The hook number h ij is defined as the number of
boxes to the right of B on row i, plus the boxes below B in column j, and finally,
plus one. The hook product H has n factors in it (i.e. the total number of boxes).
For example, in the Young pattern [ λ ] = [2, 1]: , there are three hooks
which produce the hook numbers h 11 = 3, h12 = h21 = 1, resulting in the hook
product H = 3 · 1 · 1 = 3, and so d [2,1] ( S3) = 3!/3 = 2. In both partitions [ n ]
and [1n ], the hook product is H = n!, and so d ( Sn) = n!/n! = 1 in both cases.
(The procedure gets its name from the calculational device of drawing a ‘hook’,
or a line passing vertically up through the bottom of column j, making a 90◦ right
turn at row i to move through the boxes to the right end of that row: h ij is the
total number of boxes through which the hook line passes.)
270 APPENDIX A. THE SYMMETRIC GROUP
sλ = ∑ h λ , (symmetrizer of Γλ ), (A.16)
h
aλ = ∑ δv vλ , (antisymmetrizer of Γλ ), (A.17)
v
For a given tableau Γλ , the horizontal operator, vertical operator, and Young op-
erator are uniquely defined (but one may also take aλ sλ ). In practice, to calculate
the sλ of tableau Γλ , one takes the sum of all permutations of the symbols in row i
of Γλ and then forms the product of such sums for all rows: sλ = ∏i ∑ p p(i) . Sim-
ilarly, to calculate aλ , one first takes the sum of all permutations of the symbols in
column j, weighted by factors δ p, then the product of such sums for all columns:
aλ = ∏j ∑ p δ p p(j) . These expressions when expanded become combinations of
permutations, as shown in the equations displayed above.
As we will discuss in more details below, the Young operator, once properly
normalized, is an idempotent, and generates a left ideal which generates an irre-
ducible representation. So, the Young operators for all standard Young tableaux
of Sn completely resolve the identity, and give the complete decomposition of the
regular representation of Sn .
18. The identity representation of Sn . For the normal tableau of one-row pattern
[ n ] the set {h[n] } consists of all elements p ∈ Sn , whereas {v[n] } contains only the
group identity e. So a[n] = e, and s[n] = ∑ p p ≡ s, which is the symmetrizer of the
whole group; it follows that the Young operator for [ n ] is Y[n] = se = s.
Since sp = ps = s for all p ∈ Sn , it follows that ss = n!s, and so s is essentially
idempotent, i.e. it is idempotent up to a constant. Thus, s/n! is strictly idempotent.
In addition, sps = ss = n!s for all p ∈ Sn , and so s is a primitive idempotent.
Therefore, s (or Y[n] = s) generates an irreducible representation of Sn on the
A.3. IDEMPOTENTS AND REPRESENTATIONS 271
where H and c are ordinary numbers. It turns out that H = n!/d [λ] ( Sn ) is the
hook product for pattern [ λ ]. Using this normalization factor, we define the nor-
malized primitive idempotent for [ λ ] as
def 1 1
eλ = Yλ = sλ v λ . (A.21)
H H
As a primitive idempotent, the Young operator eλ generates a minimal left ideal on Sen ,
which corresponds to an irreducible invariant subspace for group representation.
21. Equivalent representations from the same pattern. Up to now we have used
the symbol Γλ to denote any standard Young tableau obtained from the pattern
p
[ λ ]. Now, let us replace it with a more precise notation, Γλ , such that Γ1λ ≡ Γeλ
p
denotes the normal tableau, and Γλ = pΓ1λ every one of the other standard tableaux,
which as we know, can be obtained by applying an appropriate permutation p on
p
Γ1λ . If eλ is the primitive idempotent for the normal Young tableau Γ1λ , then eλ =
peλ p−1 is the primitive idempotent for the non-normal standard Young tableau
p p
Γλ . Therefore, the irreducible representations generated by eλ and eλ are equivalent,
p q
just as are those generated by eλ and eλ for any permutations p, q ∈ Sn .
p
This can be easily seen as follows: We need to show that eλ xeλ 6 = 0 for some
p p
x ∈ Sen (cf. Theorem A.2). As eλ = peλ p−1 , we have eλ peλ = peλ p−1 peλ = peλ,
which is non-vanishing.
272 APPENDIX A. THE SYMMETRIC GROUP
j
22. The permutation R ji , which relates the Young operators eiλ and eλ , correspond-
j j
ing to the Young tableaux Γiλ and Γλ of the same pattern, so that eλ = R ji eiλ R − 1
ji ,
can be found either by inspection of the diagrams, or more systematically in the
following way: Write the labeling numerals in successive rows of tableau Γiλ on
the first line of a two-row representation of R ji , and in the same way the numerals
j
in successive rows of tableau Γλ on the second line.
1346 1347
j
Thus, for example, if Γiλ = 27 and Γλ = 25 , then the permutation
5 6
needed is R ji = (1346275
1347256) ≡ ( 567 ) .
23. Inequivalent representations from different patterns. The Young operators
for different patterns ([ λ ] 6 = [ µ ]) generate inequivalent representations.
The proof runs as follows: Let eλ and eµ be the primitive idempotents of the
patterns λ and µ 6 = λ. Let p be any element of Sn , then
p
e µ p e λ = e µ p e λ ( p −1 p ) = ( e µ e λ ) p = 0 .
It follows that
eµ x eλ = 0 for all x ∈ Sen .
By Theorem A.2, the primitive idempotents eλ and eµ ([ λ ] 6 = [ µ ]) generate in-
equivalent representations.
24. Irreducible representations of Sn . The Young operators { eλ} associated to the
normal Young tableaux generate all the inequivalent irreducible representations of the
symmetric group Sn .
A proof of this key result is based on the following observations:
(i) The number of the inequivalent irreducible representations of Sn is equal
to the number of Young patterns.
(ii) There is one Young operator eλ for each Young pattern.
(iii) Every eλ generates an inequivalent irreducible representation of the group.
25. Complete reduction of the regular representation. The left ideals generated
by the Young operators associated to the distinct standard Young tableaux are lin-
early independent, and they span the group-algebra space Sen . The group identity
element is resolved into the sum of the Young operators associated to the stan-
dard tableaux.
p
Let { eλ } be the set of dλ primitive idempotents for the distinct standard Young
tableaux having the pattern [ λ ], including the normal tableaux. Then the state-
ment above tells us that the decomposition of the group algebra Sen and the reso-
lution of its identity element e are expressed in the form
p p p
Sen = ∑ ∑ Lλ , where Lλ = Sen eλ , (A.22)
λ p
p p p
e = ∑ ∑ eλ , where ( eλ )2 = eλ . (A.23)
λ p
A.3. IDEMPOTENTS AND REPRESENTATIONS 273
E XAMPLE 6: Group S2. There are two partitions of 2, namely [2] and [1,1]. The
corresponding Young patterns produce two standard tableaux, both normal:
. Pattern [2]: 1 2 s[2] = e + (12), a[2] = e, and Y[2] = e + (12). With
H = 2, we define the normalized primitive idempotent e[2] = [ e + (12)]/2. This
is a special case of the general situation discussed in §18 above.
1
. Pattern [1,1]: s[1,1] = e, a[1,1] = e − (12), and Y[1,1] = e − (12). With
2
H = 2, we define the normalized primitive idempotent e[1,1] = [ e − (12)]/2. This
is a special case of the situation discussed in §19 above.
So, the resolution of the identity on the group algebra Se2 is e = e[2] + e[1,1] .
E XAMPLE 7: Group S3. There are three partitions of 3, namely [3], [2,1] and [13 ].
The corresponding Young patterns produce four standard tableaux, all normal
except one.
. [3]: 1 2 3 This is again a special case of the situation discussed in §18.
All permutations are horizontal permutations, the only vertical permutation is e.
So, s(3) = ∑ p p = s (the total symmetrizer), a(3) = e, and Y(3) = s. With constant
H = 3! = 6, the normalized primitive idempotent for [3] is e[3] = s/6.
1 2
. [2,1]: e and (12) are horizontal, e and (13) are vertical. Hence
3
s[2,1] = e + (12), a[2,1] = e − (13), and Y[2,1] = s[2,1] a[2,1] = e + (12) − (13) − (321).
With H = 3, the normalized idempotent of this tableau is e[2,1] = Y[2,1] /3.
1 3
. [2, 1]0: This non-normal standard tableau can be obtained from
2
the normal tableau [2,1] by a (23) permutation. We have s[2,1]0 = e + (13), a[2,1]0 =
e − (12), and Y[2,1]0 = s[2,1]0 a[2,1]0 = e + (13) − (12) − (123), and the normalized
idempotent e[2,1]0 = Y[2,1]0 /3.
1
. [ 13 ] : 2 This is another special case of the general situation discussed
3
in §19 above. All permutations are vertical, the only horizontal being e. Therefore,
s[13 ] = e, a[13 ] = ∑p δ p p = a (the antisymmetrizer of the full group). The primitive
idempotent is Y[13 ] = a, which becomes normalized, e[13 ] = a/6, by including the
constant H = 3!.
And so, the resolution of the identity in Se3 is
The dimension of Se3 matches the sum of the dimensions of the component ideals:
3! = 6 = 1 + 2 + 2 + 1, as it should.
274 APPENDIX A. THE SYMMETRIC GROUP
It transforms a rank-r tensor ψ such that the tensor components ψ I are mapped
p p
to ψ I = D I J ( p )ψJ = ψ Ip . For example, if p = (110 220 ) then ψi ,i ,i = ψi10 ,i20 ,i3 .
1 2 3
29. The representations of the Lie group Gn and the symmetric group Sr on tensor
space Cn×r have a property, crucial to our discussion, of commuting with each
other, which can be proved as follows. First, note that as the order of the factors
gij in D ( g) is not important, we have:
The RHS of the two equations may also be rewritten, after changing the repeated
indices, as | J i D Jp I ( g) and | J i D J I −1 ( g). But D J I −1 ( g) = D Jp I ( g), and so the LHS
p p
of the two equations are equal. Since | I i is any of the basis vectors of Cn×r , we
have the commutation relation on the tensor space
φ ( p ) π ( g ) | ψ i = π ( g ) φ ( p ) | ψ i, (A.28)
for any ψ ∈ Cn×r , g ∈ Gn , and p ∈ Sr. So those tensors of rank r with a particular
symmetry under Sr will be transformed among themselves by transformations
g ∈ Gn . The decomposition of Cn×r into symmetry classes of Sr automatically
gives a decomposition of the tensor space into subspaces invariant under the
general linear group Gn . The subspaces spanned by tensors of definite symmetry
classes are irreducible invariant under the finite group Sr , and provide simulta-
neously the spaces for the irreducible representations of the Lie group Gn .
30. To construct an irreducible representation of the group Gn accessible through
tensorial analysis, it suffices to provide it with a basis in the form of irreducible
tensors of rank r in Cn×r possessing a definite symmetry under the symmetric
group Sr, and this we know how to do—by using Young’s graphical method.
To each Young pattern [ λ1, . . . , λk ], with ∑ki=1 λi = r, corresponds a symmetry
type. To obtain the class of tensors of rank r having that symmetry type, we begin
by assigning the indices i1 , i2 , . . . , ir in this order to the boxes of each successive
row from the top down. Next, we apply the appropriate Young operator Y to the
Young tableau so obtained, that is to say, we symmetrize all the indices in each
row, and then antisymmetrize all the indices in each column.
E XAMPLE 8: Rank-3 tensors in Cn . There are three Young patterns, namely [3],
[2,1] and [13 ], producing four standard tableaux, all normal except one.
. [3]: i1 i2 i3 s[3] = s (the total symmetrizer), a[3] = e, and Y[3] = s. Then we
apply Y on a generic rank-3 tensor ψi1 i2 i3 , resulting in the set of totally symmetric
tensors Y[3] ψi1 i2 i3 = ψ{i1 i2 i3 } which form the basis of an irrep subspace describable
by rank-3 tensors.
276 APPENDIX A. THE SYMMETRIC GROUP
i1 i2
. [2,1]: The projection operator for this tableau is Y[2,1] = a[2,1] s[2,1] ,
i3
where s[2,1] = e + (12), a[2,1] = e − (13). The tensors have a mixed symmetry:
ψi1 i2 i3 + ψi2 i1 i3 − ψi3 i2 i1 − ψi2 i3 i1 , and are the basis of a subspace of the symmetry
class [2,1] of symmetry [123].
i1 i3
. [2,1]’: (Here we fill the boxes with indices in an order different
i2
from the standard order.) The projection operator for this Young tableau [2, 1]0
is Y = [1 − (12)][e + (13)], so that the corresponding tensors are of the form
ψi1 i2 i3 − ψi2i1 i3 + ψi3 i2 i1 − ψi3 i1 i2 . They span an invariant subspace also belonging
to the symmetry class [2,1], but of symmetry [132], different from [123]. The two
sets of tensors are distinct but equivalent, being in the same symmetry class.
i1
. [ 13 ] : i2 The Young operator for [13] is Y[13 ] = ae = a, where s[13 ] = e
i3
and a[13 ] = ∑p δ p p = a (the antisymmetrizer of the full group). The basis func-
tions are given by the tensors ψ[i1 i2 i3 ] , totally antisymmetric in its indices.
C OMMENTS .
(a) By construction, the irreducible tensors in a given tensor subspace satisfy
certain linear relations, such as in [2, 1] subspace:
a b c b a b b a a c
(i) =− and (ii) = +
c a c c b
(To check, apply Y[2,1] = a[2,1] s[2,1] on both sides of each equation.)
(b) For a group Gn and tensors of fixed rank r, what are the allowed Young
patterns? The number of columns must be no greater than r (since ∑i λi = r), and
the number of rows must be no greater than n (because an antisymmetric tensor
component with at least one repeated index is equal to zero). Then, every pattern
with n rows or less, and r columns or less is realizable, giving the characteristic
symmetry type of a set of nonzero tensors of rank r.
(c) The Young pattern [1n ] has one standard tableau and the corresponding
irrep of Gn is one-dimensional. The associated tensor has one component S =
ε i1,...,in ψi1 ,...,in where ε i1,...,in is the invariant unit antisymmetric tensor of rank n.
Under Gn transformations, S is multiplied by det D ( g). (Note that while the
pattern [1n ] gives a one-dimensional (pseudo-scalar) representation in Cn×n , the
pattern [1] corresponds to an n-dimensional (vector) representation in Cn .) Sim-
ilarly, the pattern [2n ] also has one standard tableau, corresponding to a one-
dimensional irrep of Gn . The single tensor associated to it is multiplied by the
factor (det D ( g))2 upon a group transformation. Generally, for each [ sn ], there
exists a one-dimensional irrep S0 , distinguished by the transformation multiplica-
tive factor (det D ( g))s in a Gn transformation, S0 7 → (det D ( g))s S0 . It can also be
seen that the Young patterns [ λ1, . . . , λn ] and [ λ1 + s, . . . , λn + s ] (where s is any
positive integer) have the same standard tableaux, but differ by a scalar factor of
(det D ( g))s in their transformations under Gn .
A.4. IRREDUCIBLE TENSORS 277
[ λ 1, . . . , λ n − 1 ] ∗ = [ λ 1, λ 1 − λ n − 1 , . . . , λ 1 − λ 2 ] . (A.29)
For example, if r is an integer less than n, then [1]∗ = [1n−1 ], [1r ]∗ = [1n−r ],
and [ r ]∗ = [ rn−1 ]. In particular, if n = 2, we have [ r ]∗ = [ r ], or all the irreps of
SU (2) are self-conjugate, or real.
λ1 -
a b 6a
? b
33. The notation [ λ1, . . . , λn−1 ] gives the lengths of the rows in a pattern. One may
also use a notation that gives the heights (numbers of boxes) of successive columns
in decreasing order, ( n − 1)λn−1 , ( n − 2)λn−2 −λn−1 , . . . , 2λ2 −λ3 , 1λ1 −λ2 . This tells us
that the pattern [ λ1, . . . , λn−1 ] contains λn−1 columns of height n − 1; λn−2 − λn−1
columns of height n − 2; and so on.
34. Irreducible representations of a Lie group or a Lie algebra are identified by their
highest weights, usually written ( m1, m2, . . . ) in the Dynkin basis (consisting of the
fundamental weights). Consider some irrep of SU ( n ) identified with the Young
pattern [λ1, . . . , λn−1 ], or alternately with the highest weight ( m1, m2 , . . . , mn−1 ).
1 This should not be confused with the conjugation of Young tableaux that interchanges rows and
The correspondence between the two notations is given by a simple rule: the
Dynkin coordinate mi is equal to the number of columns in the Young pattern
containing i boxes, that is to say, mi = λi − λi+1 (taking λn = 0). So, for any irrep
and its complex conjugate, the correspondence is:
[λ1, . . . , λn−1] 7 → ( λ1 − λ2, λ2 − λ3, . . . , λn−2 − λn−1 , λn−1 ), (A.30)
[λ1, . . . , λn−1 ]∗ 7 → ( λn−1, λn−2 − λn−1 , . . . , λ2 − λ3, λ1 − λ2 ). (A.31)
Conversely, given a dominant weight ( m1, . . . , mn−1 ) of SU ( n) in the Dynkin
basis, one can draw the corresponding Young pattern formed by mi columns
of height i, with i = n − 1, n − 2, . . . , 2, 1. The lengths of the rows are then
given by λ1 = m1 + · · · + mn−1 ; λ2 = m2 + · · · + mn−1 ; . . . . . . ; λn−1 = mn−1 .
Thus, the fundamental representations ( . . . , 0, 1, 0, . . .) are 1-column patterns [1r ],
whereas symmetric tensor product representations ( r, 0, . . .) and ( r − s, s, 0, . . .)
correspond to the Young patterns [ r ] and [ r, s ].
35. The dimension of the irreps of SU ( n) can be calculated with a formula similar
to the ‘hook’ formula (A.15) devised for the Sn irreps, with modifications made
necessary by the fact that the permuted symbols are all distinct whereas here the
tensor indices may take any allowed (even equal) values. It is given by
Fn,λ
d[λ] ( SU (n)) = . (A.32)
H[λ]
The denominator H[λ] , which depends only on the pattern, is the same as for
(A.15), while the numerator Fn,λ is calculated as follows. Put the number n in
the box in the upper left corner of the tableau, and put integers in all other boxes
increasing by 1 each time moving to the right, and decreasing by 1 each time
moving down. Fn,λ is the product of all these numbers.
E XAMPLE 9: This example and all the examples that follow concern Lie group
SU ( n). For [2, 1]: , there are three hooks which produce the hook numbers
h11 = 3, h21 = h12 = 1, resulting in the hook product H = 3 · 1 · 1 = 3. The
numerator is F = n ( n + 1)( n − 1). And so the dimension of the irrep [2, 1] of
SU ( n) is F/H = n ( n2 − 1) /3. Note that [2, 1] = (1, 1, 0, . . . ) in the Dynkin basis.
V
E XAMPLE 10: [1r ] = r Cn (r boxes in one column): H = 1 · 2 · · · r = r! and
F = n ( n − 1) · · · ( n − r + 1), and so the dimension of the corresponding irrep is
n n!
d [ 1r ] = = .
r r! (n − r ) !
• [1] = Cn : d = n; in Dynkin basis [1] = (1, 0, . . . ).
• [ 1n − 1 ] = [1]∗ = Cn ∗ : d = n; in Dynkin basis [1n−1 ] = ( . . . , 0, 1).
Vn n
n
• [1 ] = C : d = 1; in Dynkin basis [1n ] = (0, . . . , 0).
E XAMPLE 11: [ r ] = Symr Cn (r boxes in one row): F = n ( n + 1) · · · ( n + r − 1) and
H = r! . The dimension of the corresponding irrep is
n+r−1 ( n + r − 1) !
d[r] = = .
r r! (n − 1) !
A.5. PRODUCTS OF REPRESENTATIONS 279
For example, for [3], d = n ( n + 1)( n + 2) /6, so that d = 10 for SU (3), and d = 56
for SU (6). Note that [ r ] = ( r, 0, . . . ) in the Dynkin basis.
E XAMPLE 12: [2, 1r−1 ] with r > 1 (two-column pattern: one column of r boxes
plus one of one box) H = ( r + 1) !/r and F = n ( n + 1)(n − 1) · · · ( n − r + 1) =
( n + 1) !/(n − r ) !. This gives the dimension
r−1 r ( n + 1) ! n+1
d [2, 1 ] = = r· .
( r + 1) !(n − r ) ! r+1
all possible ways to form new tableaux. Next, take the boxes labeled 2 and add
them to every enlarged tableau, again each box in a different column. Repeat
with the remaining boxes of [ µ ]. The tableaux one finally obtains must satisfy the
following conditions:
(i) Every new diagram is a valid Young pattern.
(ii) In each column, an integer-label appears only once.
(iii) Reading the boxes from right to left and the top down, the number of
boxes carrying an integer (say, 1) must be greater than or equal to the number
of boxes labeled with a larger integer (say, 2) at every step.
The diagrams formed in this way correspond to the irreps of Sm+n in [ λ ] ⊗ [µ],
and the number of ways a Young pattern [ ν ] appears is the multiplicity a ( λ, µ, ν)
of the irrep [ ν ] of Sm+n . This automatically gives the decomposition (A.33) for the
irreps of the Lie group Gn .
E XAMPLE 13: [2, 1] ⊗ [2, 1]. Insert digits j = 1, 2 in the boxes of the second pattern:
1 1
⊗
2
Following the rules, add the boxes labeled 1 in columns of the first pattern:
1
1 1 1
( a) : (b) : (c) : (d) : 1
1
1 1
Next, to each of the new diagram, add the remaining and last box labeled 2:
1 1 1 1 2
1 1
( a) :
2
2 6∈
1
1
(b) : 1
1 2
2
1
1
(c) : 2
1
1
2
1
(d) : 1
1
1 2
2
The diagram marked by 6 ∈ is not valid because it violates rule (iii), and the two
[3, 2, 1]’s are different; so that we obtain finally 8 patterns, leading to the following
decomposition:
A.5. PRODUCTS OF REPRESENTATIONS 281
⊗ = + + +
+ + + +
where the dimensions of the irreps of the symmetric group are given by the hook
formula (A.15). The combinatorial factor on the LHS reflects the different ways
in which r different objects are selected from r + s objects to form the set acted on
by Sr, with the remaining s objects assigned to the second set, with the symmetry
of Ss . And so, in the above example, (6!/3!3! )2 × 2 = 80 for the product, and
9 + 10 + 5 + 2 × 16 + 10 + 5 + 9 = 80 for the sum. (The combinatorial factor
mentioned above is not necessary for the representations of Gn or its subgroups.)
E XAMPLE 14: More examples:
⊗ = +
⊗ = +
⊗ = + +
⊗ = +
⊗ = + +
282 APPENDIX A. THE SYMMETRIC GROUP
The results for these simple cases together with the distributive and associative
properties of outer products can be used to calculate more complicated products
and their decomposition into direct sums containing irreps of higher dimensions.
39. The procedure just described also applies to representations of SU ( n ) pro-
vided that in the final results we remove the patterns having more than n rows,
and remember that columns having exactly n boxes in any patterns are equivalent
to one-dimensional representations. So, for example, if we consider the decom-
position of the product [2, 1] ⊗ [2, 1] for SU (3), we have
⊗ = + +
+ + +
or equivalently in terms of partitions (Eq. (A.35) minus the last two terms):
Clifford Algebras
and Spin Representations
283
284 APPENDIX B. CLIFFORD ALGEBRAS
and then taking the quotient of T ( V ) by the ideal in T ( V ) generated by all ele-
ments v ⊗ v − B ( v, v)1 with v ∈ V, where B is the quadratic form associated with
a symmetric bilinear map ϕ : V × V 7 → F. This construction is similar to that
V
of the exterior tensor algebra ∗ V, which is just the quotient of T ( V ) by the ideal
V
generated by all elements v ⊗ v in V ⊗2. In fact, the exterior tensor algebra ∗ V is
V∗
the situation in which B = 0, which we write as V = Cl ( V, B = 0).
We will assume that B is non-degenerate. Non-degenerate bilinear forms over
a real vector space of dimension n can be put, by a change of basis, in a canonical
form with + 1 (p times) and − 1 (q = n − p times) on the diagonal. The case (+ +
+ , −) corresponds to Minkowski space, and is of special interest to relativistic
physics. But here, for our present purpose, we shall need to consider just the
positive definite (p = n, q = 0) or negative definite (p = 0, q = n) forms.
4. In the following examples, we assume F = R and V = R n for some positive
integer n, and so, the sum in (B.2) has n + 1 summands, and call C ( n, B ) the Clif-
ford algebra Cl(R n , B ). Given an orthonormal basis { e1, e2, . . . , en } of R n , C ( n, B)
is generated by the ei’s when these are subject to the condition ei e j + e j ei = 2 B δij ,
with B becoming now the signature of the algebra, either B = + 1 (positive defi-
nite), or B = − 1 (negative definite) for every i, j = 1, . . . , n.
E XAMPLE 1: n = 1. With the single basis vector e1 of V = R, C (1, B) is spanned
by 1 and e1, so that dim C (1, B ) = 2. Any element of the algebra is of the form
x = x0 1 + x1 e1, where xi ∈ R. For B = + , C (1, +) ∼
= R ⊕ R. For B = − , meaning
e21 = − 1, we may identify the basis vector e1 with the imaginary unit i, and so,
C (1, −) is just the algebra of the complex numbers: C (1, −) ∼
= C.
E XAMPLE 2: n = 2. Here, V = R 2 = { x1 e1 + x2 e2 | x1, x2 ∈ R } , and C (2, B) is
generated by 1 and the two basis vectors e1, e2 ∈ V, so that any of its elements is
a real linear combination x0 1 + x1 e1 + x2 e2 + x3 e1 e2 with xi ∈ R. It is now a non-
commutative associative algebra of dimension 4. Now, for B = + when e2i = + 1,
if x, y ∈ V, then the product xy is a scalar dot product plus a cross product of
x, y; on the other hand, if x, y ∈ C (2, +), then xy is again an element of C (2, +).
As for B = − , that is, e21 = − 1, e22 = − 1, we may make the identification with
Hamilton’s numbers e1 = i, e2 = j, and e1 e2 = k, and recognize that C (2, −) is just
the algebra of the quaternions: C (2, −) ∼= H.
E XAMPLE 3: n = 3. Given a basis e1, e2, e3 of V = R 3 , obeying the multiplication
rules e2i = B and ei e j = − e jei (i 6 = j), Clifford algebra C (3, B ) is spanned by the
following basis elements:
where k is the number of factors ei present in each basis element, and cn,k is the
number of elements in C ( n, B) for a given k. The dimension of this Clifford al-
gebra is simply the sum of cn,k over k (n fixed at 3), so dim C (3, B) = 8. With
B = + , C (3, +) may be interpreted as a enlargement of the three-dimensional
SPIN LIE ALGEBRA 285
Euclidean space to include a scalar subspace, and oriented subspaces of one, two,
and three dimensions. With signature B = − (so that e2i = − 1 for i = 1, 2, 3),
the four elements 1, e1, e2, e1 e2 span a subalgebra isomorphic to C (2, −) ∼ = H, and
the remaining four are (1, e1, e2, e1 e2 ) e3 ∼
= C (2, −) · e3. Since e3 ∼
= i and iH ∼
= H,
we have C (3, −) ∼ = H (1 + i), or C (3, −) ∼ = H ⊕ H. (Note, however, that the
octonions are not a Clifford algebra since they are non-associative.)
Calculations (e.g. in [LM]) have identified all the Clifford algebras, as listed
in the following table only for signature B = − , where C ( n ) ≡ C ( n, −), and A [ n ]
means n × n matrices with entries in the algebra A; i.e. A [ n ] ≡ M( n, A ).
n C ( n ) ≡ C ( n, −) n C ( n ) ≡ C ( n, −)
0 R 4 H [2]
1 C 5 C [4]
2 H 6 R [8]
3 H ⊕H 7 R [8] ⊕ R [8]
n+8 C ( n ) ⊗ R [16]
The last line means that C ( n + 8) consists of 16 × 16 matrices with entries in
the algebra C ( n). This period-8 behavior was discovered by É. Cartan in 1908.
5. Generally, if the vector space V = R n generating Clifford algebra Cl ( V, B) is
provided with an orthonormal basis { e1, e2 , . . . , en } , then e0 = 1 together with the
elements ε I = ei1 ei2 . . . eik with I = { 1 ≤ i1 < i2 < · · · < ik ≤ n } form a basis
for Cl ( V, B ) = C ( n, B). Hence, C ( n, B ) has dimension 2n . In fact, for a given k
there are cn,k = n!/k!(n − k ) ! products ε I (similar to exterior products of degree
k). Including e0 for which cn,0 = 1, we have the total number of basis elements
given by the sum ∑nk=0 cn,k = 2n , which is the dimension of C ( n, B).
The quadratic dependence (B.1) of the ideal v ⊗ v − B ( v, v)1 gives a Z/2Z
grading to the quotient space C ( n, B), with B = ± , splitting it into two parts
C ( n, B) = C + ( n, B ) ⊕ C − ( n, B ), (B.3)
where C + is spanned by e0 and the products of an even number of vectors in V,
and C − is spanned by the products of an odd number. C − cannot form an algebra
since C − C − ⊂ C + ; but C + is a subalgebra of C ( n, B ), of dimension 2n−1 , it is this
subalgebra that defines the spin group.
6. Clifford algebras hold a special interest for us by the fact that they contain
as subalgebras the spin Lie algebras spin( n ) associated to the simply-connected
spin Lie groups Spin( n ) which provide the double covers of the (non simply-
connected) special orthogonal Lie groups SO ( n); and so to understand the rep-
resentations of the latter group, it is useful to study those of the former. In order
to see this, consider the Clifford algebra C ( n) of signature B = − generated by
e1, . . . , en satisfying the (anti-commutation) relations
ei e j + e j ei = − 2 δij; i, j = 1, . . . , n. (B.4)
The quadratic products eij ≡ ei e j with i 6 = j obey the commutation relations
1 1 1 1 1 1
eij , ekl = eik δjl − eil δjk − e jk δil + e jl δik . (B.5)
2 2 2 2 2 2
286 APPENDIX B. CLIFFORD ALGEBRAS
Therefore, the set { eij = ei e j | i 6 = j } is closed under the bracket operation, and
forms a vector subspace of C ( n ), which is identified with the spin Lie algebra
spin( n ). Note also that eij acts on any ek to give
1
eij , ek = δik e j − δjk ei. (B.6)
2
On the other hand, we know that the special orthogonal Lie algebra so( n )
admits as a basis n × n real matrices Lij , with i < j, which generate rotations in
the i-j plane, and satisfy commutation relations equivalent to (B.5)—see Sec. 3.6.5
of Chapter 3. So, we have isomorphism of the Lie algebras spin( n ) and so( n )
under 21 eij ↔ Lij . To get the simply-connected spin Lie group Spin( n ), we just
need to exponentiate spin( n ). In particular, for the subalgebra spanned by a given
1/2 e , we may use the relation 1/4 e 2 = − 1/4 to obtain the exponential exp e in
ij ij ij
a closed form (e0 = 1):
1 θ θ
exp θ eij = e0 cos + eij sin . (B.7)
2 2 2
With θ ranging over [0, 4π ], this gives a U(1) subgroup of the spin group Spin( n ).
Noting that the RHS of this equation is of the form a e0 + b ei e j with a2 + b2 = 1,
and is an element in C + ( n ), and that exp ·spin( n ) can contain only even powers of
the ei’s, we have the general definition for the spin group:
Spin( n ) = s ∈ C + ( n ) | svs−1 ∈ R n for all v ∈ R n ; s s̄ = 1 , (B.8)
where s̄ is the conjugate of s. The conjugation rules for ei’s are given by
ē0 = e0,
ēi = − ei,
ei1 · · · eik = (−)k eik · · · ei1 , 1 ≤ i1 < · · · ik ≤ n.
rv = ( ax − by ) e1 + ( bx + ay ) e2
= [(cos θ ) x − (sin θ ) y ] e1 + [(sin θ ) x + (cos θ ) y ] e2. (B.9)
the two nodes that branch off one end of the Dynkin diagram D` . The irrep ϕ`−1
has highest weight ω `−1 = 1/2( ε 1 + ε 2 + · · · + ε `−1 − ε ` ), with its general weights
given by 1/2 ∑`1 ηi ε i , where ηi = ± 1 and η ≡ ∏`1 ηi = − 1. On the other hand, ϕ` ,
which corresponds to the fundamental weight ω ` = 1/2( ε 1 + ε2 + · · · + ε`−1 + ε` ),
has its weights of the same form, but with η ≡ ∏`1 ηi = + 1. One can see then that
both ϕ`−1 and ϕ` are of dimension 2`−1 , to be known as S− and S+ .
We will now consider these special representations (called spin representations,
or, more exactly, fundamental spin representations ) of so( n, C ) on complex vector
spaces from the point of view of (complexified) Clifford algebra and spin Lie
algebra.
B` D`
f −
f f··· f f v f f··· f f S
H
ϕ1 ∧2 ∧`−2 ∧`−1 S ϕ1 ∧2 ∧`−3 ∧ `−2H f +
S
Figure B.1: Fundamental representations of the orthogonal Lie algebras b` =
V V
so (2` + 1, C ) and d` = so (2`, C ). Notation: k = k ϕ1 .
8. Consider a complex 2` -dimensional vector space, the type of spaces where the
spin representations of orthogonal Lie algebras live. It is useful to treat it as a
tensor product of ` two-dimensional spaces: C2⊗` = C2 ⊗ C2 ⊗ · · · ⊗ C2 , with
typical vectors v = | a, b, . . . i = | a i ⊗ | b i ⊗ . . . . An arbitrary matrix on C2⊗` can
be formed from the 2 × 2 identity matrix σ0 = diag [1, 1] and the Pauli matrices
j
σi (with i = 1, 2, 3 and σ3 = diagonal [1, − 1] by convention). We will call σi
the Pauli matrix σi that acts on the j th subspace C2 in C2⊗` . Then, define the
following tensor products of ` matrices σµ (µ = 0, 1, 2, 3) (cf. [Ge] p. 210):
γ1 ≡ σ1 ⊗ 1 · · · ⊗ 1 = σ11 ,
γ2 ≡ σ2 ⊗ 1 · · · ⊗ 1 = σ21 ,
γ3 ≡ σ3 ⊗ σ1 ⊗ 1 · · · ⊗ 1 = σ31 σ12 ,
γ4 ≡ σ3 ⊗ σ2 ⊗ 1 · · · ⊗ 1 = σ31 σ22 ,
... = ... ... ,
γ2`−3 ≡ σ3 ⊗ σ3 · · · ⊗ σ3 ⊗ σ1 ⊗ 1 = σ31 · · · σ3`−2 σ1`−1 ,
γ2`−2 ≡ σ3 ⊗ σ3 · · · ⊗ σ3 ⊗ σ2 ⊗ 1 = σ31 · · · σ3`−2 σ2`−1 ,
γ2`−1 ≡ σ3 ⊗ σ3 · · · ⊗ σ3 ⊗ σ1 = σ31 · · · σ3`−1 σ1` ,
γ2` ≡ σ3 ⊗ σ3 · · · ⊗ σ3 ⊗ σ2 = σ31 · · · σ3`−1 σ2` ,
γ2`+1 ≡ σ3 ⊗ σ3 · · · ⊗ σ3 ⊗ σ3 = σ31 · · · σ3`−1 σ3` .
Making use the defining conditions of the Pauli matrices, σi σj = δij + iσk ε kij ,
with i, j, k ranging over the values 1, 2, 3, one can prove that the γi ’s satisfy the
SPIN REPRESENTATIONS 289
anticommutation relations:
γi γj + γj γi = 2 δij, i, j = 1, 2, . . . , 2` + 1. (B.11)
And, keeping in mind (B.4), we see (from ei 7 → iγi ) that the γi ’s are the generators
of a complexified Clifford algebra which yields the spin representations we are
considering.
Eq. (B.11) indicates that they generalize the matrices Dirac introduced in rel-
ativistic space-time quantum mechanics to higher dimensions in which γ2`+1 =
(− i)`γ1 γ2 · · · γ2`−1 γ2` plays a role comparable to that of Dirac’s γ5 . We shall now
see that these matrices γi ’s, defined on C2⊗` , give just the natural framework for
studying the spin representations of the orthogonal Lie groups.
Now, given n γi -matrices subject to the anticommutation relations (B.11), we
define the antisymmetric quadratic products
1 1
Mij ≡ γi γj = γi , γj , ( i, j = 1, . . . , n ) (B.12)
2i 4i
(of which n ( n − 1)/2 are independent) which obey the following bracket relations
characteristic of SO (n):
Mij , Mkm = i ( δik Mjm − δjk Mim − δim Mjk + δjm Mik ) . (B.13)
mean that the γi ’s are a set of tensors that transform as the n-dimensional vector
representation of SO ( n).
9. Spin representation S of b` = so (2` + 1, C ). The Lie algebra b` is provided
with a basis of `(2` + 1) elements Mij = 1/2i γi γj , with 1 ≤ i < j ≤ 2` + 1.
Noting that (B.13) implies [ Mij , Mkm ] = 0 for i 6 = j 6 = k 6 = m, we choose the
Cartan subalgebra h as { M1,2, M3,4, . . . , M2`−1, 2` } ≡ { H1, H2, . . . , H` } . Explicitly,
we have
1 1
Hi ≡ M2i−1, 2i = γ γ = σ3i , 1 ≤ i ≤ `. (B.15)
2i 2i−1 2i 2
The set of equations Hi vi = µ ( Hi) vi on C2 admit as solution vi = | 1/2ηi i with
eigenvalues µ ( Hi) = 1/2 ηi and ηi = ± 1. So the weight vectors in the irrep S
on C2⊗` wrt CSA h are tensor products of the form v = | 1/2η1 , 1/2η2 , . . . , 1/2η` i
with the weight ( 1/2η1 , 1/2η2 , . . . , 1/2η` ), where ηi = + 1 or − 1. Clearly, S is 2` -
dimensional and has highest weight ( 1/2, 1/2, . . . , 1/2) in the given CSA basis.
10. Spin representation S± of d` = so(2`, C ). Here, we shall take as a basis
the `(2` − 1) matrices Mij , with 1 ≤ i < j ≤ 2`, satisfying the commutation
relation (B.13). In the usual Killing-Cartan analysis, we next specify a Cartan
subalgebra by the set of elements { H1, H2, . . . , H` } , where Hj = M2j−1, 2j =
1 j
2 σ3 and j = 1, . . . , `. And so the 2` -dimensional representation we obtain con-
sists of the joint eigenvectors of { H1, H2, . . . , H` } , namely, the weight vectors
290 APPENDIX B. CLIFFORD ALGEBRAS
v = | 1/2η1 , 1/2η2 , . . . , 1/2η` i with weights (± 1/2, ± 1/2, . . . , ± 1/2) (with ` terms)
wrt the chosen CSA basis. But this time there is a non trivial element, namely,
γ2`+1 = σ31 · · · σ3`−1 σ3` (which does not participate in the construction of the set
Mij ), which by (B.14) commutes with all other elements, [ Mij , γ2`+1 ] = 0 for
all Mij ∈ so(2`, C ). It acts on any weight vector v = | 1/2η1 , 1/2η2 , . . . , 1/2η` i
as a multiplication by a scalar η = η1 · · · η` , where ηi = ± 1. In other words,
we have γ2`+1 v = η v, where η is equal to − 1 if the number of minus signs
is odd, and to + 1 if it is even. (γ2`+1 is comparable to the γ5 of Dirac alge-
bra, and η to chirality.) It follows that this 2` representation we have obtained
is reducible, and its space may be split into disjoint subspaces of equal sizes via
v = 1/2(1 − γ2`+1) v + 1/2(1 + γ2`+1) v, corresponding to the decomposition of the
2` representation into two irreducible representations, which are the fundamental
spin representations S− and S+ of so(2`, C ), each of dimension 2`−1 .
E XAMPLE 6: so (5) (` = 2). The 5 generators of the Clifford algebra are γ1 = σ11 ,
γ2 = σ21 , γ3 = σ31 σ12, γ4 = σ31 σ22 , and γ5 = σ31 σ32 . The basis elements of Lie algebra
so (5) are { Mij, 1 ≤ i < j ≤ 5} , and for its CSA, we take H1 = M12 = 1/2σ31 and
H2 = M34 = 1/2σ32 . Call |±i the eigenvectors of σ3 on C2 with eigenvalues ± 1.
Then the fundamental spin representation of so (5) is 4-dimensional, consisting of
the eigenvectors of ( H1, H2 ) on C2⊗2 given by | + +i, | + −i, | − +i, and | − −i,
with respective weights ( 1/2, 1/2), ( 1/2, − 1/2), (− 1/2, 1/2), and (− 1/2, − 1/2).
E XAMPLE 7: so (6) (` = 3). The generators of the Clifford algebra are γ1 = σ11 ,
γ2 = σ21 , γ3 = σ31 σ12 , γ4 = σ31 σ22 , γ5 = σ31 σ32 σ13, and γ6 = σ31 σ32 σ23 . They serve
to determine the 15 basis elements of so(6), namely, { M12, . . . , M16, . . . , M56 } , of
these, Hi = M2i−1, 2i = 1/2σ3i (i = 1, 2, 3) are selected to define the CSA. Then
the resulting representation on C2⊗3 consists of 23 = 8 weight vectors |± , ±, ±i
of weights (± 1/2, ± 1/2, ± 1/2), divided into 4 vectors | + ++i, | + −−i, | − +−i,
and | − −+i of chirality η = + 1 (representation S+ ); and 4 vectors | + +−i,
| + −+i, | − ++i, and | − −−i of chirality η = − 1 (representation S− ).
Manifolds
291
292 APPENDIX C. MANIFOLDS
φα−1 ◦ φβ : φ− 1 −1
β ( Oα ∩ Oβ ) → φα ( Oα ∩ Oβ ) ,
18. Just as there exist in general different equivalent ways to parameterize a sur-
face, so there are different equivalent charts on space X. The maps φα : Vα → Oα
and φβ : Vβ → Oβ are said to be compatible charts (1) if both φα−1 (Oα ∩ Oβ ) and
φ− 1 n
β ( Oα ∩ Oβ ) in R are open sets; and (2) if the mutually inverse changes of coor-
dinates ( x1α , . . . , xnα ) ↔ ( x1β , . . . , xnβ ) are given by smooth functions. Thus, one can
say briefly that a smooth atlas is a collection of pairwise compatible charts. (Note that we
require only continuity for a chart, but need C ∞-differentiability via the condition
of ‘compatible charts’ to define ‘smooth atlas’.) This equivalence relation extends
to atlases: Two n-dimensional smooth atlases A 1 and A 2 on the same topological
space X are said to be equivalent if any chart from A 1 and any chart from A 2 are
compatible.
297
Figure C.2: Stereographic projection of the sphere from the north pole N to the equatorial
plane. Points Z on the sphere are projected to points z on the plane.
298 APPENDIX C. MANIFOLDS
(3) Unit sphere S2 in stereographic coordinates u, v (Fig. C.2). The unit sphere S2
in R 3 is defined by the set S2 = {( x, y, z ) ∈ R 3 | x2 + y2 + z2 = 1} . Just as in the
previous example for S1, mapping the two-dimensional surface S2 requires two
maps, φ1 and φ2, using the local coordinates ( u, v) ∈ R 2 and ( u0, v0 ) ∈ R 2:
φ1 : R 2 → S2 \ N ⊂ R 3
( u, v) 7 → (2u/s, 2v/s, ( s − 2) /s)
φ2 : R 2 → S2 \ S ⊂ R 3
( u0, v0 ) 7 → 2u0 /s0 , 2v0 /s0 , (2 − s0 ) /s0 ,
φ1 : R n → Sn \ N ⊂ R n+1
u 7 → (2u/s, ( s − 2) /s )
φ2 : R n → Sn \ S ⊂ R n+1
u0 7 → 2u0 /s0, (2 − s0 ) /s0 ,
by a smooth function of the entries (involving their products and sums), and so
is a smooth map. Its restriction to G × G is also smooth. As for the inversion, the
map A 7 → A−1 is smooth inv : G → G because, by virtue of Cramer’s formula,
the entries of A−1 are given by rational functions of the entries of A and det A (a
continuous function). So G = GL( n, R ) is a Lie group, of dimension n2. Note the
special case GL(1, R ) = R × .
If the matrices are defined over the complex field C, the corresponding group
GL( n, C ) = { A ∈ M( n, C )| det A 6 = 0} is a an open subset of the vector space of
n × n matrices over C, and hence is a smooth manifold of dimension 2 n2. So,
GL( n, C ) is a Lie group, of dimension 2 n2 over R (or, equivalently, n2 over C). Note
that GL(1, C ) = C× .
Most Lie groups we consider are ‘contained’ in the general linear group; they
can be generated as subgroups of GL ( n, F ) for F ∈ { R, C } , and are Lie groups
by virtue of the following result: A subgroup of Lie group GL( n, F ) that is also a
manifold in R N , where N = n2 (or 2 n2) for F = R (or C), is a Lie group. This
is so because the multiplication map in a subgroup H of the Lie group G is the
restriction m H of the multiplication map mG on G, which is itself smooth. And
so, the subgroup layered with a compatible manifold is a Lie group.
When the subgroup H is specified by a system of simultaneous equations on
R N that are equivalent or reducible to k independent conditions, it is a manifold in
R N of dimension N − k, and therefore is a Lie group; cf. §22 (5).
26. E XAMPLES of Lie groups that are subgroups of GL( n, R ) and GL ( n, C ):
(1) The subgroup GL+ ( n, R ) of GL ( n, R ) of invertible matrices of positive de-
terminant. This is indeed a subgroup because det ( AB) = det A · det B > 0 and
det I = 1 > 0. Moreover, by continuity of the determinant map, det : A 7 → R > , it
is an open set, and so is a manifold. Therefore GL+ ( n, R ) is a Lie group.
(2) The special linear groups SL( n, R ) and SL ( n, C ) are the subgroups of the
general linear group GL( n, F ) with determinant 1:
U ( n )) = U ( n, C ) = { A ∈ M( n, C )| A† A = I }
(A† is the hermitian adjoint of A, so that A†ij = A∗ji , with ∗ denoting complex conju-
gation). Just as for O ( n ), but now we have complex entries in the matrices. The
total number of (real) coordinates for every A ∈ U ( n ) is N = 2 n2 subjected to
k = n + 2 ( n2 − n ) /2 = n2 independent conditions, so the space of solutions for
A† A = I, and thus the manifold, has dimension N − k = 2 n2 − n2 = n2. The
group U ( n ) is a Lie group of dimension n2 (over R).
(4) The special orthogonal group SO ( n ). The defining relation for any A ∈
O ( n ) says that AT A = I, which implies (det A )2 = 1, or det A = ± 1, which
may be written as the determinant map det : O ( n ) → {± 1}. This means that
O ( n ) is the disjoint union of two components, only one of which, with det A = 1,
is connected to the identity and forms a subgroup, called the special orthogonal
group SO( n ).
In the complex case, for any A ∈ U ( n ) one has | det A | = 1, and so det A = eiθ
for any real θ, in other words, det : U ( n ) → S1. The set of unitary matrices with
determinant one is the special unitary group SU( n ). The group SO( n ) is a Lie
group of the same dimension, n ( n − 1) /2, as O ( n ); the group SU ( n ) is also a Lie
group, but its dimension (n2 − 1) is one less than that of U ( n ). This reflects the
difference in the determinant mapping in the real and complex cases.
(5) The defining condition for every A ∈ O ( n ) is AT I A = I, where Iij = δij .
This symmetric metric Iij may be replaced with a non-positive definite matrix, in
particular the non-degenerate antisymmetric matrix of even dimension
0 I
J=
−I 0
where I is the identity n × n matrix. The set of all real 2n × 2n matrices A that
satisfy AT J A = J forms a group, called the symplectic group over R:
Sp( n, R ) = { A ∈ M(2n, R )| AT J A = J } .
Arguments similar to those for O ( n ) show that Sp( n, R ) is a Lie group. Its di-
mension is given by N − k, where N = (2n )2 is the number of entries in any A ∈
Sp( n, R ), and k is the number of the independent conditions on them. This num-
ber can be calculated from the system of equations AT J A = J: there are n ( n − 1)
independent equations from the two n × n diagonal blocks, and n2 from the two
n × n off-diagonal blocks, leading to k = n2 + n ( n − 1) = 2n2 − n. So the space
of solutions (or the manifold of the group) has the dimension N − k = n (2n + 1),
which is also the dimension of Sp ( n, R ). The same arguments also apply to the
complex group Sp ( n, C ) = { A ∈ M(2n, C )| AT J A = J } , where one should note
the presence of the transpose AT and not the Hermitian adjoint A† . It has the
same dimension as Sp( n, R ).
304 APPENDIX C. MANIFOLDS
[As] M. Ashbacher, The Status of the Classification of the Finite Simple Groups.
Notices of the Amer. Math. Soc. 51, 736–740, 2004.
[Ba] J. C. Baez, The Octonions. Bull. Amer. Math. Soc. 39, 145–205, 2002.
[Bal] L. Ballentine, Quantum Mechanics. World Scientific Publishing Co., Singa-
pore, 1998.
[Be] W. Berry, Structural Functions in Music. Prentice Hall, Englewood Cliffs,
N.J. (1976). Repr. by Dover Publications, Mineola, N.Y. (1987).
[Ca] R. N. Cahn, Semi-Simple Lie Algebras and Their Representations. Dover Pub-
lications, Inc. New-York, N.Y., 2006.
[CS] J. H. Conway and D. A. Smith, On Quaternions and Octonions: Their Geom-
etry, Arithmetic and Symmetry. A. K. Peters, Ltd., 2004.
[Dy1] E. B. Dynkin, The structure of semisimple Lie algebras. Uspehi Mat. Nauk 2,
59–127, 1947. Am. Math. Soc. Trans 1, 1–143, 1950.
[Dy2] E. B. Dynkin, Semisimple subalgebras of semisimple Lie algebras. Mat. Sbornik
30, 349–462, 1952. Am. Math. Soc. Trans Ser. 2, 6, 111–244, 1957.
[Fa] G. Fano, Mathematical Methods of Quantum Mechanics. McGraw-Hill Book
Company, New-York, N.Y., 1971.
[FH] W. Fulton and J. Harris, Representation Theory. Springer-Verlag New York,
Inc. New-York, N.Y., 1991.
[Ge] H. Georgi, Lie Algebras in Particle Physics. Benjamin/Cumming, Reading,
MA., 1982.
[Gi] R. Gilmore, Lie Groups, Physics, and Geometry. Cambridge U. Press, Cam-
bridge, U.K., 2008.
[Go] D. Gorenstein, Finite Simple Groups: An Introduction to their Classification.
Plenum, New York, N.Y., 1992
[GLS] D. Gorenstein, R. Lyons, and R. Solomon, The Classification of the Finite
Simple Groups. Amer. Math. Soc., Providence , RI., 2002.
[Ha] B. C. Hall, Lie Groups, Lie Algebras, and Representations. Springer-Verlag
New York, Inc. New-York, N.Y., 2003.
[Ham] M. Hamermesh, Group Theory and its Applications to Physical Problems.
Addison-Wesley Publishing Company, Reading, MA., 1962.
305
306 REFERENCES
307
308 INDEX
H (quaternion algebra), 91, 284, see also Q8 of dimension three, 95, 101
Haar measure, 111 of dimension two, 94, 101
Hausdorff topology, 294 of rank one, 187
Heisenberg group, algebra, 80, 95, 114 of rank two, 187
helicity, 251 representation, 105
Higgs boson, 251, 254 subalgebra, 91
homeomorphism, 294 Lie bracket, 91
homomorphism, 7, 9, 38, 54, 294 Lie group, 75, 301
hook formula, 269, 278 character, 111, 141, 245
hypercharge, 252 compact, 80, 111, 116
complex, 75
ideal (Lie subalgebra), 94, 102, 178, 263 connected component, 82
idempotent, 263, 270 dimension, 75
primitive, 264, 271 generator, 86
image (of a map), 9 of finite dimension, 9
index of a subgroup, 12 of infinite dimension, 80
index of representation, 242 of transformation, 75
infinitesimal operator, 85 real, 75
injection, 291 representation, 106
inner product, 35, 220 Lie subalgebra, 91, 94
inner-product space, 35 Lie subgroup, 76, 108, 302
integration measure, 81, 111 normal, 82, 102
invariant subgroup, 16 Lie third theorem, 103
invariant subspace, 36 Lie, Sophus, 73
irreducible representation linear group, 76
see representation – irreducible linear vector space, 33–37
isometry, 79 Lorentz group, 79, 119
isomorphism, 9, 38, 94, 109
isospin, 252 manifold, 74, 291, 299
differentiable, 74
Jacobi identity, 88–89 dimensionality, 74, 299
examples, 299–300
kernel, 7, 9, 18 product, 299
Killing form, 114–116, 160, 178–180 smooth, 74
Killing, Wilhelm, 177 matrix
Klein bottle, 304 cyclic, 99
Klein group, 5, 19, 20, 21, 61, see also V diagonal, 37
nilpotent, 99
Lagrange theorem, 12 orthogonal, 9, 56, 77
lattice, 219 representation, 34, 38, 106
of dominant weights (Λd ), 219 unimodular, 3, 99
lepton, 251 unitary, 9, 36, 37
Lie algebra, 91 metric space, 76, 81, 292, 293
automorphism, 94 Mobius strip, 304
center, 94 module, 38
complexification, 95, 97 morphism, 101, 109
derived algebra, 94 multiplication table, 4, 157
homomorphism, 94
isomorphism, 94 neighborhood, 74, 292, 294
of dimension one, 93, 101 non-isotropy of root, 181
310 INDEX