Professional Documents
Culture Documents
Lie Algebras
DRAFT, Release 1.27
October 4, 2022
2
I have prepared these notes for the students of my lectures and my seminar on Lie
algebras. The lecture has been repeated several times since the summer
semester 2017 at the mathematical department of LMU
(Ludwig-Maximilians-Universität) at Munich. I thank the participants of the
seminar for making some simplifications and corrections.
Compared to the oral lecture in class these written notes contain some additional
material.
wehler@math.lmu.de
Release notes:
• Release 1.27: Chapter 1: Minor revision.
• Release 1.26: Chapter 1 and Chapter 2: General revision.
...
• Release 1.1: General revision Chapter 1.
Contents
1 Matrix functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.1 Power series of matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 3
1.2 Jordan decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.3 The exponential map of matrices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
v
vi Contents
7 Classification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
7.1 The root system of a complex semisimple Lie algebra . . . . . . . . . . . . 189
7.2 Root systems of the A, B,C, D-series . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
7.3 Outlook . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 224
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 265
Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 269
Part I
General Lie Algebra Theory
Chapter 1
Matrix functions
The paradigm of a Lie algebra is the vector space of matrices with the commutator
of two matrices as Lie bracket. These concrete examples even cover all abstract
finite dimensional Lie algebras. They are the focus of these notes. Nevertheless it is
useful to consider Lie algebras from an abstract viewpoint as a separate algebraic
structure like groups or rings.
If not stated otherwise, we denote by K either the field R of real numbers or the
field C of complex numbers. Both fields have characteristic 0. The relevant differ-
ence is the fact that C is algebraically closed, i.e. each polynomial with coefficients
from K of degree n has exactly n complex roots. This result allows to transform
matrices over C to certain standard forms by transformations which make use of the
eigenvalues of the matrix.
If not stated otherwise all vector spaces in this chapter are assumed finite-
dimensional K-vector spaces.
This formula holds for all real numbers x, y ∈ R. Then exponentiation defines a
group morphism
exp : (R, +) → (R∗ , ·)
Those students who attended a class on complex analysis will remember that
exponentiation is defined also for complex numbers. Hence exponentiation extends
to a map
exp : (C, +) → (C∗ , ·)
which also satisfies the functional equation above.
3
4 1 Matrix functions
The seamless transition from the real field R to the complex field C is due to the
fact that the exponential map is defined by a power series
∞ 1
exp(z) = ∑ ν! · zν
ν=0
The series converges not only for real numbers but for all complex numbers too.
In order to generalize the exponential map one step further we now exponentiate
strictly upper triangular matrices as arguments.
the K-algebra of strictly upper triangular matrices. Matrices from n(3, K) can be
added and multiplied with each other, and they can be multiplied by scalars from
the field K. In particular, we consider two matrices of the form
0p0 000
P = 0 0 0 , Q = 0 0 q ∈ n(3, K).
000 000
We compute
∞ 1 ν 1 p0
exp(P) := ∑ · P = 1 + P = 0 1 0
ν=0 ν! 0 01
∞ 1 ν 1 00
exp(Q) := ∑ · Q = 1 + Q = 0 1 q .
ν=0 ν! 0 01
Note: For all ν ≥ 2
Pν = Qν = 0
Hence we do not need to care about questions of convergence.
exp(P + Q).
Ei j ∈ M(n × n, K)
having value = 1 at position with index i j and value = 0 at all other positions. Then
Ei j · Ekm = δ jk · Eim
We have
0 0 pq
(P + Q)2 = (pE12 + qE23 ) · (pE12 + qE23 ) = pqE13 = 0 0 0 = PQ
0 0 0
and
(P + Q)ν = 0, ν ≥ 3.
As a consequence
1 p (pq)/2
exp(P+Q) = 1 +P+Q+(1/2)·(P+Q)2 = 1 +P+Q+(1/2)·PQ = 0 1 q .
0 0 1
ii) Fortunately, the defect can be fixed by introducing a correcting term which
derives from the commutator of the matrices P and Q
[P, Q] := PQ − QP ∈ n(3, K)
We set
C := (1/2) · [P, Q]
Computing
we obtain
pq
[P, Q] = pq · E13 , C = · E13
2
(P+Q+C)2 = (pE12 +qE23 +(pq/2)·E13 )·(pE12 +qE23 +(pq/2)·E13 ) = pq·E13 = 2C
and
(P + Q +C)ν = 0, ν ≥ 3.
6 1 Matrix functions
Hence
exp(P + Q + (1/2) · [P, Q]) =
exp(P + Q +C) = 1 + (P + Q +C) + (1/2)(P + Q +C)2
0 p pq
= 1 + P + Q + 2C = 1 + P + Q + PQ = 0 0 q =
00 0
= exp(P) · exp(Q).
iii) More general one can show: The exponential map can be defined on the vector
space of strict upper triangular matrices as a map
The goal of the present section is to extend the exponential map to all matrices
from M(n × n, K). Therefore, we first need a concept of convergence for sequences
and series of matrices. The basic ingredience is the operator norm of a matrix.
{x ∈ Kn : ∥x∥ ≤ 1}.
Intuitively, the operator norm of A is the volume to which the linear map
determined by A blows-up or blows-down the unit ball of Kn .
The operator norm ∥A∥ depends only on the linear map represented by A. Hence,
all matrices which represent the same linear map with repect to different bases
have the same operator norm.
1.1 Power series of matrices 7
A · B ∈ M(n × n, K)
because
(A · B) ·C = A · (B ·C)
for matrices A, B,C ∈ M(n × n, K).
Proposition 1.3 (Normed associative matrix algebra). The matrix algebra (M(n × n, K), ∥∥)
is a normed associative algebra:
1. ∥A∥ = 0 iff A = 0
2. ∥A + B∥ ≤ ∥A∥ + ∥B∥
3. ∥λ · A∥ = |λ | · ∥A∥ with λ ∈ R
4. ∥A · B∥ ≤ ∥A∥ · ∥B∥
Proof. ad 6) For j = 1, ..., n denote by e j the j-th canonical basis vector of Kn . Then
for all i = 1, ..., n s
n
∥A∥ ≥ ∥Ae j ∥ = ∑ |ak j |2 ≥ |ai j |,
k=1
hence
∥A∥ ≥ ∥A∥sup .
Concerning the other estimate, the inequality of Cauchy Schwarz
hence
∥A∥ ≤ n · ∥A∥sup
⊔
⊓
Aν ∈ M(n × n, K), ν ∈ N,
Because (M(n ×n, K), ∥∥) is a normed algebra according to Proposition 1.3, con-
cepts from analysis like convergence, Cauchy sequence, continuous function, and
power series also apply to matrices. In particular, the commutator is a continuous
map: For A, B ∈ M(n × n, K)
Then
• The series is absolutely convergent in ∆ (R), i.e. ∑∞
ν=0 |cν | · |z| is convergent
ν
for z ∈ ∆ (R).
• The series converges compact in ∆ (R), i.e. the convergence is uniform on each
compact subset of ∆ (R).
be the open ball in M(n × n, K) around zero with radius R. For a matrix A ∈ B(R)
then:
• The series
!
∞ n
ν ν
f (A) := ∑ cν · A := lim
n→∞
∑ cν · A ∈ M(n × n, K)
ν=0 ν=0
• The function
f : B(R) → M(n × n, K), A 7→ f (A),
is continuous.
exists. The remaining statements about convergence follow from the estimate
∞
∥ f (A)∥ ≤ ∑ |cν | · ∥A∥ν
ν=0
ii) The equation [A, f (A)] = 0 about the commutator follows: We take the limit lim
n→∞
and employ the continuity of the commutator
" # " #
∞ n n
A, ∑ cν · Aν = lim A, ∑ cν · Aν = lim ∑ cν [A, ·Aν ] = 0
n→∞ n→∞
ν=0 ν=0 ν=0
A(t0 + h) − A(t0 )
lim =: A′ (t0 ) ∈ M(n × n, K)
h→0 h
exists. In this case we employ the notation
dA
(t0 ) := A′ (t0 ).
dt
A : I → B(R) ⊂ M(n × n, K)
1.1 Power series of matrices 11
A, B : I →
− M(n × n, K).
d
A(t) · B(t) = A′ (t) · B(t) + A(t) · B′ (t).
dt
Proof. i) The series
∞
f (A(t)) = ∑ cν · Aν (t)
ν=0
and
∞
f ′ (A(t)) = ∑ ν · cν · Aν−1 (t)
ν=1
Due to the compact convergence we may differentiate f (A(t)) term by term. The
proof of the lemma reduces to proving
d ν
A (t) = ν · Aν−1 (t) · A′ (t) = ν · A′ (t) · Aν−1 (t).
dt
Here the proof goes by induction on ν ∈ N. For the induction step assume the claim
to hold for ν ∈ N. Analogous to the proof of the product rule from calculus by clever
insert of a suitable term:
12 1 Matrix functions
⊔
⊓
• If [
V λ ( f ) := ker( f − λ )n ̸= {0}
n∈N
f (V λ ( f )) ⊂ V λ ( f ),
because
[( f − λ )n , f ] = 0.
2. Every non-zero generalized eigenspace V λ ( f ) contains at least one eigenvector v
of f with eigenvalue λ : Take a non-zero vector v0 ∈ V λ ( f ) and choose n ∈ N
maximal with
v := ( f − λ )n (v0 ) ̸= 0.
Hence knowing a generalized eigenspace of f allows to find an actual eigenvector
of f .
Vi ⊊ Vi+1 , i = 0, ..., n − 1
satisfying
f (Vi ) ⊂ Vi , i = 0, ..., n
14 1 Matrix functions
det(T · 1 − A) ∈ K[T ]
µ(pchar ; λ ) ∈ N
2. The characteristic polynomial pchar splits over K into linear factors, and for all
eigenvalues λ of f the algebraic multiplicity equals the geometric multiplicity,
i.e.
µ(pchar ; λ ) = dim Vλ ( f )(Geometric multiplicity).
3. The vector space V splits as direct sum of eigenspaces
M
V= Vλ ( f ).
λ eigenvalue
f = lim fν .
ν→∞
N = (ai j ), ai j = 0 if i ≥ j.
and
{λi + aiν : ν ∈ N}
are disjoint. For each ν ∈ N define
λi + aiν , i = 1, ..., n
and for each fixed ν ∈ N all eigenvalues of Aν are pairwise distinct. Hence Aν
represents a diagonalizable endomorphism fν due to Proposition 1.13. Moreover
16 1 Matrix functions
A = lim Aν ,
ν→∞
i.e.
f = lim fν
ν→∞
⊔
⊓
with suitable k ∈ N∗ .
1. Because the ring R := K[T ] is a principal ideal domain, the ideal of all polyno-
mials which annihilate f
f = fs + fn (Jordan decomposition)
[ fs , fn ] = 0.
1.2 Jordan decomposition 17
ps (T ), pn (T ) ∈ C[T ]
fs = ps ( f ) and fn = pn ( f ).
[ fs , g] = [ fn , g] = 0.
The proof uses the fact that the field C is algebraically closed. Hence the minimal
polynomial pmin of f splits completely into linear factors. If pmin has the roots
λi , i = 1, ..., r,
then the linear factors of pmin induce a family of polynomials without a common
factor. Because R := C[T ] is a principal domain, one obtains a partition of unity in
R. It induces a partition of the identity id ∈ End(V ). The corresponding summands
form a family of pairwise commuting projectors
Ei : V →
− V, i = 1, ..., r
Setting
Vi := im Ei ⊂ V, i = 1, ..., r,
decomposes V as the direct sum
r
M
V= Vi ,
i=1
is semisimple and
18 1 Matrix functions
f − fs =: fn
is the searched nilpotent summand of f . The main tool for the verification is the
concept of the generalized eigenspaces of f .
with λi ̸= λ j if i ̸= j.
pmin (T )
pi := ∈ R,
(T − λi )mi
The only factor common to all pi , i = 1, ..., r, are the units of R, hence w.l.o.g we
have g = 1. Therefore, for i = 1, ..., r polynomials ri ∈ R exist with
r
1 = ∑ ri · pi ∈ R.
i=1
Ei := ri ( f ) · pi ( f ) ∈ End(V ).
They satisfy:
•
n
idV = ∑ Ei
i=1
(ri · pi ) · (r j · p j ) ∈ R.
Ei · E j = 0 ∈ End(V ).
Hence the family (Ei )i=1,...,r splits the identity idV as a sum of pairwise commuting
projectors. We define the corresponding subspaces
Vi := im Ei ⊂ V, i = 1, ..., r
Vi , i = 1, ..., r,
Vi = Vλi ( fs ), i = 1, ..., r.
( f −λi )m (v) = ( f −λi )m (Ei (v)) = ( f −λi )m (ri ( f )◦ pi ( f ))(v) = (ri ( f )◦ pmin ( f ))(v) = 0.
iii) The nilpotent summand: To obtain the nilpotent summand of f we consider the
polynomial
pn (T ) := T − ps (T ) ∈ C[T ]
and define the corresponding endomorphism
fn := pn ( f ) = f − fs ∈ End(V ).
Then
f = fs + fn .
The two definitions
fs := ps ( f ) and fn := pn ( f )
imply
[ f , fs ] = [ f , fn ] = 0.
In order to prove the nilpotency of fn we consider an arbitrary index i ∈ 1, ..., r and
the restriction of fn to Vi . Set
m := mi = dim Vi
On Vi
fn = f − fs = ( f − λi ) − ( fs − λi ).
Using [ fs , f ] = 0 the binomial theorem implies
m
m
f nm = ∑ (−1)m−µ ( f − λi )µ ◦ ( fs − λi )m−µ .
µ=0 µ
( f − λi )µ
( f − λi )n (v j ) ∈ V j .
( f − λi )n (v j ) = 0,
v j ∈ V λi ( f ).
Due to Remark 1.9 the endomorphism f has also an eigenvector w ∈ Vλi ( f ) with
eigenvalue λi . The f -stableness of V j implies w ∈ V j . We obtain for large m
0 = ( f − λ j )m (w) = (λi − λ j )m · w.
w ∈ Vj ⊂ V λ j ( f )
with the last inclusion proven in part iii). And the right equality is due to
w ∈ Vλi ( f ).
The equality
0 = (λi − λ j )m · w
implies j = i. As a consequence v = vi ∈ Vi , which finishes the proof of the
inclusion
22 1 Matrix functions
V λi ( f ) ⊂ Vi
Together with the opposite inclusion from part iii) we obtain
Vi = V λi ( f )
and
r r r
V∼ Vi ∼ V λi ( f ) ∼
M M M
= = = Vλi ( fs ).
i=1 i=1 i=1
v) Both polynomials pmin and pchar have the same roots: The roots of the
characteristic polynomial pchar are the eigenvalues of f . For each root λ of the
minimal polynomial pmin (T ) of f the generalized eigenspace V λ ( f ) contains an
eigenvector of f with eigenvalue λ , see Remark 1.9. Therefore any root of pmin (T )
is an eigenvalue of f .
For the oppposite inclusion we consider an eigenvalue λ of f with corresponding
eigenvector v ∈ V :
f (v) = λ · v.
Because pmin ( f ) = 0 ∈ End(V ) we have in particular
pmin ( f )(v) = 0 ∈ V.
Hence
0 = pmin ( f )(v) = pmin (λ ) · v ∈ V.
Because v ̸= 0 we conclude
pmin (λ ) = 0 ∈ C,
i.e. the eigenvalue λ is a root of the minimal polynomial.
f = fs′ + fn′
[ fs′ , fn′ ] = 0
follows
[ f , fn′ ] = [ fs′ , fn′ ] + [ fn′ , fn′ ] = [ fn′ , fn′ ] = 0.
Part iii) with
fn = pn ( f )
shows
[ fn , fn′ ] = 0.
Analogously, one proves
1.2 Jordan decomposition 23
[ fs , fs′ ] = 0.
As a consequence
fs − fs′ = fn′ − fn ∈ End(V )
is an endomorphism which is both semisimple - being the sum of two commuting
semsisimple endomorphisms - and nilpotent - being the sum of two commuting
nilpotent endomorphisms. Hence the endomorphism is zero.
vii) Killing the constant terms: We prove that the polynomials ps and pn can be
choosen with
ps (0) = pn (0) = 0 :
• If λi = 0 for one index i = 1, ..., r then
V 0 ( f ) ̸= {0}
ps (T ) = a0 + a1 T + ... + ak T k ∈ C[T ]
implies
fs = ps ( f ) = a0 · id + a1 · f + ... + ak · f k End(V ).
Choosing a non-zero vector v ∈ ker f shows a0 = 0. Hence ps (T ) ∈ C[T ] has no
constant term, i.e. ps (0) = 0. From the definition pn (T ) := T − ps (T ) follows
• If λi ̸= 0 for all indices i = 1, ..., r then we claim pmin (0) ̸= 0. The proof is
indirect: Otherwise by factoring out
q( f ) ̸= 0,
w := q( f )(v) ̸= 0
As a consequence
0 = pmin ( f ) = f (w)
implies that λ = 0 is an eigenvalue, a contradiction.
24 1 Matrix functions
ps (0)
p̃s := ps − · pmin
pmin (0)
and
pn (0)
p̃n := pn − · pmin .
pmin (0)
Then
p̃s (0) = p̃n (0)
without changing the polynomial representations
pchar ( f ) = 0 ∈ End(V ).
Note that both polynomials pmin (T ) and pchar (T ) have the same roots due to
part v) in the proof of Theorem 1.17.
Proof. Denote by
f = fs + fn
the Jordan decomposition of f according to Theorem 1.17 and take over the notation
from its proof. In particular
1.2 Jordan decomposition 25
r
pmin (T ) = ∏ (T − λ j )m j
j=1
λ := λ j , m := m j , W := V λ ( f ), and k := dim W.
fs′ = λ · idW .
A ∈ M(k × k, C)
with all diagonal elements equal to λ . The characteristic polynomial of f ′ has the
multiple root λ . It is obtained by applying the determinant formula for upper trian-
gular matrices as
pmin, f ′ (T ) = (T − λ )m ∈ C[T ]
implies
m ≤ k.
Summing up these considerations for all indices j = 1, ..., r, the characteristic poly-
nomial of f is
r λj
(f)
pchar (T ) = ∏ (T − λ j )dim V
j=1
m j ≤ dim V λ j ( f ).
In particular,
26 1 Matrix functions
pmin ( f ) = 0 =⇒ pchar ( f ) = 0
⊔
⊓
f ⊗ id
of the complexification V ⊗ R C.
has convergence radius R = ∞. We now generalize the exponential series from com-
plex numbers z ∈ C as argument to matrices A ∈ M(n × n, K) as argument.
A : I → M(n × n, K)
d
(exp A(t)) = A′ (t) · exp A(t) = exp A(t) · A′ (t).
dt
1.3 The exponential map of matrices 27
ad 2: Taking lim of
N→∞
!⊤
N
Aν N
(A⊤ )ν
∑ ν! = ∑
ν=0 ν=0 ν!
28 1 Matrix functions
ad 3: Taking lim of
N→∞
!
N
Aν N
S · Aν · S−1 N
(S · A · S−1 )ν
S· ∑ · S−1 = ∑ = ∑
ν=0 ν! ν=0 ν! ν=0 ν!
0 λn
and conclude
∞
! eλ1 ∗
Aν ..
det(exp A) = det = det .
∑ ν! .
ν=0
0 eλn
Hence
n
det(exp A) = ∏ eλi = eλ1 +...+λn = exp(tr A).
i=1
The inverse of exponentiation is taking the logarithm. But even in the case of
complex numbers the exponential map is not injective because
Anyhow the exponential map of complex numbers is locally invertible. And this
property continues to hold for the exponential of matrices. Therefore recall the com-
plex power series of the logarithm
∞ (−1)ν+1 ν
log(1 + z) = ∑ · z , z ∈ C,
ν=1 ν
1.3 The exponential map of matrices 29
B : I → M(n × n, K)
d
(log B(t)) = B(t)−1 · B′ (t) = B′ (t) · B(t)−1 .
dt
Proof. ad 1) For each n ∈ N
n n n+1
(1 − A) · ∑ Aν = ∑ Aν − ∑ Aν = 1 − An+1 .
ν=0 ν=0 ν=1
ad 2) For t ∈ I define
30 1 Matrix functions
A := 1 − B(t).
Because ∥A∥ = ∥1 − B(t)∥ < 1 apply part 1) to A:
B(t) = 1 − A
In addition,
log(B(t)) = log(1 + (B(t) − 1))
is well-defined. The chain rule from Lemma 1.7 with the power series of the loga-
rithm
∞ (−1)ν+1
f (1 + z) = ∑ · zν
ν=1 ν
and its derivative
∞
f ′ (1 + z) = ∑ (−z)ν
ν=0
implies
!
d d ∞
(log B(t)) = log(1 + (B(t) − 1)) = ∑ (1 − B(t))ν · B′ (t) =
dt dt ν=0
exp(log A) = A
Hence
log(exp A) = log(1 + (exp(A) − 1))
is a well-defined convergent power series.
In order to prove the equality
1.3 The exponential map of matrices 31
log(exp A) = A
we first consider a diagonal matrix
λ1 0
A=
.. .
.
0 λn
Then
log(eλ1 ) 0 λ1 0
log(exp(A)) =
.. =
.. = A.
. .
0 log(eλn ) 0 λn
Here we use that the series of the logarithm log(eλi ), i = 1, ..., n, computes the main
value of the logarithm.
S · A · S−1 = B
Sν · Bν · Sν−1 = ∆ν
a diagonal matrix. Because ∥A∥ = ∥B∥ we may assume: For ν ∈ N sufficiently large
also
∥∆ν ∥ = ∥Bν ∥ < log 2
and
Bν = Sν−1 · ∆ν · Sν = Sν−1 · log(exp ∆ν ) · Sν =
= log(exp(Sν−1 · ∆ν · Sν )) = log(exp Bν ).
Taking the limit and using the continuity of the functions log and exp gives
B = log(exp(B))
and
ad 2) By assumption
exp(log A) = B
goes along the same line as the proof of part 1): First, one proves the equality for
diagonal matrices. Secondly, one approximates the triagonizable matrix A by a se-
quence of diagonalizable matrices. And finally one uses the continuity of functions
which are given by power series. ⊔ ⊓
A = 2π i · 1 ∈ M(n × n, C)
exp(A) = e2π i · 1 = 1
hence
log(exp(A)) = log(1) = 0 ̸= A.
is surjective.
Proof. Consider a fixed but arbitrary matrix B ∈ GL(n, C). According to Theorem 1.17
the vector space Cn splits into the sum of the generalized eigenspaces with respect
to the eigenvalues λ of B
Cn ∼
M
= V λ (B)
λ
N := B − λ · 1
is nilpotent and
1
B = λ · 1 + N = λ (1 + · N).
λ
Moreover
∞
1 (−1)ν+1 N ν
A1 := log(1 + · N) = ∑ · ν
λ ν=1 ν λ
is well-defined because the sum is finite. We have
1 1
exp A1 = exp(log(1 + · N)) = 1 + · N
λ λ
according to Proposition 1.25. After choosing a complex logarithm µ ∈ C with
eµ = λ
and setting
A := µ · 1 + A1 ,
Theorem 1.21 implies
1
exp A = exp(µ · 1) · exp A1 = (λ · 1) · exp A1 = λ (1 + · N) = B.
λ
⊔
⊓
sl(2, C) := {A ∈ M(2 × 2, C) : tr A = 0}
and
SL(2, C) := {B ∈ GL(2, C) : det B = 1}
and consider
exp : sl(2, C) → SL(2, C).
The map is well-defined due to Theorem 1.21. The matrix
−1 b
B= ∈ SL(2, C), b ∈ C∗ ,
0 −1
and consider
exp : M(2 × 2, R) → GL+ (2, R).
The matrix
−1 b
B= ∈ GL+ (2, R), b ∈ R∗ ,
0 −1
has no inverse image.
Aµ · v = λ µ · v, µ ∈ N,
imply ! !
n n
Aµ λµ
∑ µ! ·v = ∑ ν! · v ∈ Kn .
µ=0 µ=0
exp A = B.
0 = tr A = λ1 + λ2 .
We obtain
λ
−1 −1 −1 e 1 0
S·B·S = S · (exp A) · S = exp(S · A · S )=
0 eλ2
exp A = B.
The real matrix A has two complex eigenvalues λi , i = 1, 2, which are conjugate
to each other.
λ1 = λ2 =: λ ∈ R
and exp A has an eigenvalue eλ > 0, while B has the single eigenvalue −1, a
contradiction.
For later application in Proposition 2.14 we provide the following useful formula:
Proposition 1.29 (Lie product formula). For any two matrices A, B ∈ M(n × n, C)
the exponential satisfies
!ν
A B
exp(A + B) = lim exp · exp .
ν→∞ ν ν
In the proof we will use for a power series f the standard notation
f = O(1/ν 2 )
A A
exp = 1 + + O(1/ν 2 )
ν ν
B B
exp = 1 + + O(1/ν 2 )
ν ν
and
36 1 Matrix functions
A B A B
exp · exp = 1 + + + O(1/ν 2 ).
ν ν ν ν
For large ν ∈ N we have
! !
A B
exp · exp − 1
< log 2.
ν ν
We get !
A B A B 2
exp · exp = exp + + O(1/ν )
ν ν ν ν
and
!ν !!ν
A B A B 2
exp · exp = exp + + O(1/ν ) = exp(A + B + O(1/ν)).
ν ν ν ν
Taking lim and using lim O(1/ν) = 0 and the continuity of the exponential proves
ν→∞ ν→∞
the claim !ν
A B
lim exp · exp = exp(A + B)
ν→∞ ν ν
⊔
⊓
exp : M(n × n, K) →
− GL(n, K)
is a well-defined map according to Corollary 1.22. But the map is no(!) group
morphism
1.3 The exponential map of matrices 37
H1 (X,Y ) = X +Y
H2 (X,Y ) = (1/2)[X,Y ]
H3 (X,Y ) = (1/12)[[X,Y ],Y ] − [[X,Y ], X])
H4 (X,Y ) = −(1/24)[Y, [X, [X,Y ]]]
is absolute and compact convergent in U ×U. It satisfies for all X,Y ∈ U the
functional equation
exp(Hn (X,Y )) = exp X · exp Y
Chapter 2
Fundamentals of Lie algebra theory
In general, in the associative algebra M(n × n, K) the product of two matrices is not
commutative:
AB ̸= BA, A, B ∈ M(n × n, K).
The commutator
[A, B] := AB − BA ∈ M(n × n, K)
is a measure for the degree of non-commutativity. The commutator depends K-linearly
on both matrices A and B. In addition, it satisfies the following rules:
1. Permutation of two matrices: [A, B] + [B, A] = 0.
2. Cyclic permutation of three matrices: [A, [B,C]] + [B, [C, A]] + [C, [A, B]] = 0.
These properties make
gl(n, K) := (M(n × n, K), [−, −])
the prototype of a Lie algebra.
Definition 2.1 (Lie algebra).
1. A K-Lie algebra is a K-vector space L together with a K-bilinear map
such that
• [x, x] = 0 for all x ∈ L and
39
40 2 Fundamentals of Lie algebra theory
• [x, [y, z]] + [y, [z, x]] + [z, [x, y]] = 0 for all x, y, z ∈ L (Jacobi identity).
We have just seen that the Lie algebra gl(n, K) derives from the associative al-
gebra M(n × n, K). Actually, any finite dimensional K-Lie algebra derives from a
matrix algebra, see Theorem ??. But the theory becomes more transparent when
considering Lie algebras and their morphisms as abstract mathematical objects.
A Lie algebra is a vector space with the Lie bracket as additional operator. This
additional algebraic structure resembles the multiplication within a group or a ring.
The Lie algebra concept of the commutator is taken from group theory while the
concept of an ideal comes from ring theory.
Definition 2.2 (Basic algebraic concepts). Consider a K-Lie algebra L. One de-
fines:
• A vector subspace M ⊂ L is a subalgebra of L iff
[M, M] ⊂ M,
[L, I] ⊂ I,
i.e. if I is L-invariant.
NL (M) := {x ∈ L : [x, M] ⊂ M} ⊂ L,
The center collects those elements which commute with all elements from L.
For a morphism
f : L1 → L2
between to K-Lie algebras the kernel ker( f ) ∈ L1 is an ideal. Apparently, for each
Lie algebra L the derived algebra [L, L] ⊂ L is an ideal.
If one compares these basic concepts from Lie algebra theory with concepts from
group theory then subalgebras correspond to subgroups, while ideals correspond to
normal subgroups.
with
[ f , g] := f ◦ g − g ◦ f .
In particular
gl(n, K) := gl(Kn ).
Subalgebras of the Lie algebra gl(n, K) are named matrix algebras.
All three vector spaces of matrices are stable with respect to the commutator of
matrices. Therefore they are subalgebras of gl(n, K).
Moreover for n ≥ 2
n(n, K) ⊊ t(n, K) ⊊ gl(n, K)
and
d(n, K) ⊊ t(n, K).
We will see in Chapter 3.1 and Chapter 3.2 that n(n, K) and t(n, K) are the prototype
of the important classes of respectively nilpotent and solvable Lie algebras.
Notation 2.4 (Restricting scalars from C to R). Consider a complex Lie algebra L.
By restricting scalars from C to R the elements of L also form a real Lie algebra
which is denoted LR .
The fundamental tool for studying Lie algebras are their representations. To rep-
resent an abstract Lie algebra L means to define a map from L to a Lie algebra of
matrices. The concept of a representation has also many applications in physics.
Some examples we will see in in later chapters. The scope of the concept of a
representation is not restricted to Lie algebra theory.
which satisfies
x.(y.v) − y.(x.v).
The representation ρ is faithful iff ρ is injective, i.e. ρ embeds L into a Lie algebra
of matrices.
2. A linear map
f :V →
−W
between two L-modules is a morphism of L-modules if for all x ∈ L, v ∈ V :
f (x.v) = x. f (v).
ad : L → gl(L), x 7→ ad x,
defined as
ad x : L → L, y 7→ (ad x)(y) := [x, y].
Note. One often speaks about an L-module V without naming explicitly the
representation
ρ : L → gl(V ).
ad [x, y] = [ad x, ad y] : L → L.
44 2 Fundamentals of Lie algebra theory
ad[x, y](z) = [[x, y], z] = −[[y, z], x] − [[z, x], y] (Jacobi identity)
[ad x, ad y](z) := (ad x ◦ ad y − ad y ◦ ad x)(z) = [x, [y, z]] − [y, [x, z]].
The adjoint representation maps any abstract Lie algebra to a matrix algebra.
But in general the adjoint representation is not injective. The kernel of the adjoint
representation of a Lie algebra L is the center Z(L). Nevertheless Theorem ?? will
show that any finite dimensional Lie algebra has a faithful representation, hence is
a subalgebra of a Lie algebra of matrices.
The characteristic feature of a Lie algebra L is the Lie bracket. It refines the
underlying vector space of L. In order to investigate the Lie bracket of L one studies
how each given element x ∈ L acts on L as the endomorphism ad x.
This procedure is similar to the study of number fields Q ⊂ K ⊂ C. The multi-
plication on K refines the underlying Q-vector space structure. And one studies the
multiplication by considering the Q-endomorphisms which result from the multipli-
cation by all elements x ∈ K. Norm and trace of these endomorphisms are important
parameters in algebraic number theory.
For all x ∈ L the element ad x is not only an endomorphism of L, but also a deriva-
tion of L: With respect to the Lie bracket it satisfies a rule similar to the Leibniz rule
for the derivation of the product of two functions.
Definition 2.7 (Derivation of a Lie algebra). Let L be a Lie algebra. A derivation
of L is an endomorphism
D:L→L
which satisfies the product rule
D([y, z]) := (ad x)([y, z]) = [x, [y, z]] = −[y, [z, x]] − [z, [x, y]] = [y, [x, z]] + [[x, y], z] =
Derivations arising from the adjoint representation are named inner derivations,
all other derivations are outer derivations.
Lemma 2.9 (Algebra of derivations). Let L be a Lie algebra. The set of all deriva-
tions of L
Der(L) := {D ∈ End(L) : D derivation}
is a subalgebra Der(L) ⊂ gl(L).
An important class of Lie algebras are the Lie algebras attached to the classical
groups. The classical groups are groups of matrices, and the corresponding Lie al-
gebras are the Lie algebras of their infinitesimal generators.
with derivation
d
exp(t · X) (0) = X
dt
from Proposition 1.20 is named the 1-parameter group of GL(n, C) with infinitesi-
mal generator X.
G ⊂ GL(n, C).
46 2 Fundamentals of Lie algebra theory
2
Here GL(n, C) ⊂ Cn is equipped with the subspace topology of the Euclidean
space.
fX (t) ∈ G.
G ⊂ GL(n, C)
is closed iff for any sequence (Aν )ν∈N of matrices Aν ∈ H, ν ∈ N, which con-
verges in GL(n, C), i.e. with
2. In the general context of a Lie group G one defines a Lie subgroup as a subgroup
which is also submanifold of G. One proves that any Lie subgroup of G is closed
and a Lie group itself. Therefore Definition 2.11 requires a matrix group to be a
closed subgroup.
f : R → G.
But one can show: All continuous 1-parameter subgroups of a Lie group G have
the form
f (t) = exp(t · X)
with an element X ∈ Lie G, the Lie algebra of G, and with respect to the expo-
nential map
exp : Lie G →
− G.
In particular, every continuous 1-parameter subgroup depends on the parameter t
in a differentiable - even analytical - way.
2.2 Lie algebras of the classical groups 47
Proposition 2.13 (The classical groups and the Lie algebra of their infinitesimal
generators). Consider a parameter r ∈ N and a number m = m(r) ∈ N. We use the
shorthand
G := GL(m, K)
for the matrix group and
M := gl(m, K)
for the Lie algebra. Then the classical groups and the Lie algebra of their infinitesimal
generators are defined as:
sl(m, K) := {X ∈ M : tr X = 0}.
SO(m, K) := {g ∈ G : g · g⊤ = 1, det g = 1}
so(m, K) := {X ∈ M : X + X ⊤ = 0}.
Sp(m, K) := {g ∈ G : g⊤ · σ · g = σ }
with
0 1 0 −1
σ :=
−1 0
, σ −1 =
1 0 = −σ , 1 ∈ GL(r, K) unit matrix,
sp(m, K) := {X ∈ M : X ⊤ · σ + σ · X = 0}.
SO(m, K) := {g ∈ G : g · g⊤ = 1, det g = 1}
so(m, K) := {X ∈ M : X + X ⊤ = 0}.
Proof. We show for all cases that any infinitesimal generator of a 1-parameter
subgroup of the matrix group belongs to the Lie algebra in question. Vice versa,
Theorem 1.21 implies that any element of the Lie algebra is an infinitesimal
generator of a 1-parameter subgroup of the matrix group.
d
i) Taking the derivative on both sides of the equation
dt
1 = det(exp(t · X)) = exp(tr (t · X))
gives
0 = (tr X) · exp(tr (t · X)).
Hence for t = 0:
0 = tr X.
In the opposite direction: If tr X = 0 then
Moreover
tr
sl(m, K) = ker [gl(m, K) −
→ K]
is the kernel of the trace morphism with K considered as an Abelian Lie algebra.
Hence
sl(m, K) ⊂ gl(m, K)
is even an ideal and a posteriori a subalgebra.
Hence for t = 0:
X + X ⊤ = 0.
In the opposite direction:
X + X⊤ = 0
implies
exp(tX) · exp(tX ⊤ ) = exp(t(X + X ⊤ )) = e0 = 1
and tr X = 0 implies
det(exp(tX)) = etr X = 1.
Moreover for X,Y ∈ so(m, K)
X(Y +Y ⊤ ) + (Y ⊤ +Y )X ⊤ = 0
using X = −X ⊤ . Hence so(m, K) is closed with respect to taking the commutator.
gives
Hence for t = 0:
0 = X ⊤ · σ + σ · X.
In the opposite direction:
0 = X ⊤ + σ · X · σ −1
implies
(exp(tX))⊤ · σ · exp(tX) · σ −1 = 1.
And for X,Y ∈ sp(2n, K)
after inserting
X ⊤ · σ ·Y − X ⊤ · σ ·Y
and
Y ⊤ · σ · X −Y ⊤ · σ · X
and using
X ⊤ · σ = −σ · X.
v) Taking the derivative of
50 2 Fundamentals of Lie algebra theory
Hence for t = 0
X + X ∗ = 0.
Moreover det(exp(tX)) = 1 for all t ∈ K implies
tr X = 0.
And
tr X = 0
implies
det(exp(tX)) = etr X = 1.
One checks that su(m) is a real Lie algebra, i.e. a Lie algebra over the base field R.
The reason for introducing the series in Proposition 2.13 and for the distinction
between the two series Br and Dr will become clear later in Chapter 7. The lower
bound for the parameter r ∈ N has been choosen to avoid duplicates or direct sum
decompositions. Otherwise we would have for the base field K = C and the corre-
sponding Lie algebras:
A1 = B1 = C1 , C2 = A1 ⊕ A1 , D3 = A3 .
• Elements of the complex matrix groups of the series Br , Dr preserve the K-bilinear
form on Km
m
(z, w) = ∑ z j · w j.
j=1
2.2 Lie algebras of the classical groups 51
• Elements of the complex matrix groups of the series Cr preserve the K-bilinear
form on Km
r
(x, y) = ∑ xi · yr+i − xr+i · yi .
i=1
• Elements of the special unitary group SU(m) preserve the sesquilinear (Hermi-
tian) form on Cm
m
(z, w) = ∑ z j · w j.
j=1
The bilinear form (z, w) is C-linear with respect to the first component and C-
antilinear with respect to the second component.
Over the complex numbers quadratic forms always transform to these normal
forms by a change of basis. Therefore one can restrict the investigation to the normal
forms. The situation changes over the real numbers. Here one has to discriminate
between different normal forms, depending on the signature of the quadratic form
in question.
The examples from Proposition 2.13 illustrate the general fact that the infinites-
imal generators of all 1-parameter subgroups of a matrix group form a real Lie
algebra.
Proposition 2.14 (Infinitesimal generators of a matrix group). For a matrix
group
G ⊂ GL(n, C)
the set of infinitesimal generators of all 1-parameter subgroups of G
defined by the function fX from Definition 2.11, is a real subalgebra of gl(n, C)R ,
using the Notation 2.4.
Proof. i) Scalar multiplication: If X ∈ L and s ∈ R then for all t ∈ R also
exp(t(X +Y )) ∈ G.
iii) Lie bracket: First, for all X ∈ L and all A ∈ G also the conjugate
52 2 Fundamentals of Lie algebra theory
AXA−1 ∈ L :
f :R→
− GL(n, C), f (t) := exp(tX) ·Y · exp(−tX).
d d
f (t) = (exp(tX) ·Y · exp(−tX)) = X · exp(tX) ·Y · exp(−tX) − exp(tX) ·Y · X
dt dt
Hence
df
(0) = XY −Y X = [X,Y ].
dt
On the other hand
df f (t) − f (0) exp(tX) ·Y · exp(−tX) −Y
(0) = lim = lim
dt t→0 t t→0 t
Because each fraction belongs to L, also in the limit
df
(0) = [X,Y ] ∈ L.
dt
⊔
⊓
2. A real form of the complex Lie algebra L is a real subalgebra M ⊂ LR such that
the complex linear map
M ⊗R C → L, m ⊗ 1 7→ m, m ⊗ i 7→ i · m,
i) Real forms: Apparently, sl(2, R) is a real form of sl(2, C). The C-linear map
su(2) ⊗ R C →
− sl(2, C), X ⊗ 1 7→ X, X ⊗ i 7→ iX,
sl(2, C) →
− su(2) ⊗ R C, Z = X + iY 7→ (−iX) ⊗ i + iY ⊗ 1
with
Z + Z∗ Z − Z∗
X := , Y := .
2 2i
ii) 2-dimensional subalgebra of sl(2, R): The real Lie algebra sl(2, R) contains
the 2-dimensional subalgebra
1 0 01
spanR < , >.
0 −1 00
we obtain the basis (i · σ j ) j=1,2,3 of the real Lie algebra su(2) of skew-Hermitian
trace zero matrices. The Pauli matrices have the non-zero commutator
[σ1 , σ2 ] = 2i · σ3 ,
defined by
f (i · σ3 ) := 2Z, f (i · σ2 ) := 2Y, f (i · σ1 ) := 2X.
54 2 Fundamentals of Lie algebra theory
The Lie algebra so(3, R) is isomorphic to the Lie algebra vector product
But for any two linear independent vectors x1 , x2 ∈ Vect the product x1 × x2 is
orthogonal to the plane generated by x1 and x2 . Hence
2. The real Lie algebra su(2) is the Lie algebra of the compact matrix group SU(2).
The real form sl(2, R) is the Lie algebra of the matrix group SL(2, R) which is
not compact.
The present sections illustrates some topological properties ofclassical matrix groups.
The investigation of matrix groups often reduces to the study of simply connected
matrix groups. And the latter can be studied by their Lie algebras. The following
proofs are special cases from general Lie group theory.
S 3 × R3 .
Proof. 1. We show that the matrix group SO(3, R) is path-connected: For a given
rotation matrix A we choose an orthonormal basis of R3 with the rotation axis as
third basis element. Then we may assume
cos δ sin δ 0
A = −sin δ cos δ 0 ∈ SO(3, R), δ ∈ [0, 2π[.
0 0 1
The path
cos (t · δ ) sin (t · δ ) 0
γ : [0, 1] → SO(3, R),t 7→ −sin (t · δ ) cos (t · δ ) 0 ∈ SO(3, R)
0 0 1
As a consequence, both Lie groups SO(3, R) and O(3, R) have the same Lie al-
gebra
so(3, R) = o(3, R).
2. Consider a matrix
a z
A= ∈ SU(2).
cw
Then
−1 w −z ∗ a z
A = and A = .
−c a c w
Hence
SU := {A ∈ GL(2, C) : A · A∗ = 1, det A = 1} =
w z 2 2
: z, w ∈ C, |z| + |w| = 1 ≃
−z w
≃ S 3 ⊂ C 2 ≃ R4
is homeomorphic to the 3-dimensional unit sphere. The unit sphere S3 is com-
pact, connected and simply connected. Simply connected means the vanishing of
the fundamental group
π1 (S3 , ∗) = 0
56 2 Fundamentals of Lie algebra theory
3. Set B := C2 \ {0}. On one hand, projecting a matrix onto its last column defines
the differentiable map
a z z
p : SL(2, C) → B, A = 7→ .
cw w
On the other hand, one considers the open covering U = (U1 ,U2 ) of the base B
with
U1 := {(z, w) ∈ B : z ̸= 0}, U2 := {(z, w) ∈ B : w ̸= 0}
One obtains a well-defined map
f : SL(2, C) → B × C
(
(p(A), (a − w · r−2 ) · z−1 ) if p(A) ∈ U1
a z
A= 7→
cw (p(A), (c + z · r−2 ) · w−1 ) if p(A) ∈ U2
with
r2 := |z|2 + |w|2 .
Note: If p(A) ∈ U1 ∩U2 then
SL(2, C) ≃ B × C
Several relations exists between the classical groups. These relations can be made
explicit by differentiable group morphisms. Hence it is advantegeous not to study
each group in isolation. Instead, one should focus on the relations between different
2.3 Topology of the classical groups 57
classical groups and study those properties, which the groups have in common. We
give two examples in low dimension.
Recall Remark 2.16 which started from the two different real Lie algebras
sl(2, R) ̸= su(2).
Nevertheless, both have the same complexification
are not equal. But Example 2.18 constructs a 2-fold covering projection
Φ : SU(2) →
− SO(3, R).
Due to Proposition 2.17 the group SU(2) is simply connected. The existence of
the 2-fold covering projection Φ implies that SO(3, R) is not simply connected. In
order to obtain the morphism Φ we have to find out first: How does a complex
matrix U ∈ SU(2) act on the 3-dimensional real space R3 ?
The family of Pauli matrices (σ j ) j=1,2,3 , see Remark 2.16, is a basis of Herm0 (2).
We define the map
3
3 x3 x1 − i · x2
α : R → Herm0 (2), x = (x1 , x2 , x3 ) 7→ X := ∑ xj ·σj = .
j=1
x1 + i · x2 −x3
Correspondingly, on the real vector space Herm0 (2) we consider the quadratic
form
58 2 Fundamentals of Lie algebra theory
qH (α(x)) = qE (x), x ∈ R3 .
Φ : SU(2) → SO(H), B 7→ ΦB ,
by setting
ΦB : Herm0 (2) → Herm0 (2), X 7→ B · X · B−1 .
Note B · X · B−1 ∈ Herm0 (2), because B−1 = B∗ and X ∗ = X imply
(B · X · B−1 )∗ = (B · X · B∗ )∗ = B · X · B∗ = B · X · B−1 .
Because
SO(H) ≃ SO(3, R)
is connected due to Proposition 2.17, we even have
ΦB ∈ SO(H).
3. Tangent map of Φ: We use from the theory of Lie groups without proof:
• The underlying vector space of the Lie algebra of a matrix group G is the
tangent space Lie G = Te G at the neutral element e ∈ G.
2.3 Topology of the classical groups 59
Ψ : G1 → G2
of Lie algebras.
It is common, to interpret the tangent map of the differentiable map Φ at e ∈ G1
as the linearization of Ψ at e.
Φ : SU(2) → SO(H), B 7→ ΦB ,
φ : su(2) →
− so(H), A 7→ φA .
We compute for
ϕA : H →
− H, X 7→ ϕA (X) = [A, X].
ϕiσ1 : H → H
Φ : SU(2) → SO(3, R)
Φ(SU(2)) ⊂ SO(3, R)
Comparing on both sides the components and using the determinant condition
gives
1 = |z|2 − |w|2 , w = 0, and z2 = 1.
Hence z = ±1. As a consequence
B = ±1.
vi) Universal covering space: We use from Lie group theory without proof that
a surjective morphism between Lie groups with discrete kernel is a covering
projection. The group SU(2) is simply connected according to Proposition 2.17.
Hence the map
Φ : SU(2) → SO(3, R)
is the univeral covering projection of SO(3, R), and SU(2) is a double cover
of SO(3, R).
π1 (SO(3, R), ∗) = Z2 .
Nearly the same method as in Example 2.18 allows to compute the universal cov-
ering space of the connected component of the neutral element of the Lorentz group.
More specific, the covering projection Φ of SO(3, R) from Example 2.18 extends to
a covering projection Ψ of the orthochronous Lorentz group, see Proposition 2.20:
Figure 2.1 shows Minkowski space with the embedded light cone. Points of
Minkowski space are named events. Referring to the event e at the origin, the closure
of the interior of the light cone, the set with qM (x) ≤ 0, splits into the disjoint union
of e, the future and the past cone of e. This splitting refers to the causal structure
of the world: Events from the past may have influenced e, while e may influence
events in its future. All other events, qM (x) > 0, have no causal relation to e. They
are the presence of the event e.
Here x0 denotes the time coordinate while x j , j = 1, 2, 3, are the space coordi-
nates. Using these conventions Euclidean space embeds into Minkowski space in
a natural way.
2. Lorentz group: The Lorentz group is the matrix group of isometries of the
Minkowski space
then
B ∈ O(3, 1) ⇐⇒ η = B⊤ · η · B.
Hence for B ∈ O(3, 1):
which implies
det B = ±1.
For B = (b jk )0≤ j,k≤3 ∈ O(3, 1) the time like vector
e0 := (1, 0, 0, 0)⊤
satisfies
−1 = qM (e0 ) = qM (B · e0 ) = −b200 + b210 + b220 + b230 .
Hence
b200 = 1 + b210 + b220 + b230 ≥ 1.
The subgroup
↑
L+ := {B = (b jk )0≤ j,k≤3 ∈ O(3, 1) : det B = 1, b00 ≥ 1}
↑
is named the proper orthochronous Lorentz group. It is connected because L+ is
closed and also open:
↑
L+ = {B = (b jk )0≤ j,k≤3 ∈ O(3, 1) : det B > 0 and b00 > 0}.
↑
4. Connected components of the Lorentz group: Besides L+ ⊂ O(3, 1) the other
three connected components of O(3, 1) are the neben-classes
↑
L− := {B = (b jk )0≤ j,k≤3 ∈ O(3, 1) : det B = −1, b00 ≥ 1}
↓
L+ := {B = (b jk )0≤ j,k≤3 ∈ O(3, 1) : det B = 1, b00 ≤ −1}
64 2 Fundamentals of Lie algebra theory
↓
L− := {B = (b jk )0≤ j,k≤3 ∈ O(3, 1) : det B = −1, b00 ≤ −1}
SO(3, R)
Φ
SU(2)
Ψ
SL(2, C) O(3, 1)
H := (Herm(2), qH )
Herm(2) := {X ∈ M(2 × 2, C) : X = X ∗ }
qH : Herm(2) → R, X 7→ −det X,
i.e.
qH (X) = −(a · d) + |b|2
for
ab
X= , a, d ∈ R, b ∈ C.
bd
Set σ0 := 1 ∈ Herm(2). Then the family (σ j ) j=0,...,3 is a basis of the vector
space Herm(2). The map
3
x0 + x3 x1 − i · x2
β : M → H, x = (x0 , ..., x3 ) 7→ X := ∑ xj ·σj =
j=0
x1 + i · x2 x0 − x3
2.3 Topology of the classical groups 65
qH (β (x)) = qM (x), x ∈ R4 .
Ψ : SL(2, C) → O(H), B 7→ ΨB ,
with
ΨB : H → H, X 7→ B · X · B∗ ,
is a well-defined morphism of real matrix groups. We have
↑
Ψ (SL(2C)) ⊂ L+ (H)
ψA j , j = 1, ..., 6,
shows: The family (ψA j ) j=1,...,6 is linearly independent in the vector space End(Herm(2)).
is open because its tangent map 1 ∈ SL(2, C) is a local isomorphism. The image
↑
Ψ (SL(2, C)) ⊂ L+ (H)
is surjective.
ΨB (X) = X
for all B ⊂ SL(2, C) and the matrices X ∈ Herm(2) from the basis (σ j ) j=0,...,3
of Herm(2).
For general dimensions we have the following topological results, e.g. see [25, Chap. 10, §2]
and [27, Chap. 17]:
2. Series Ar , r ≥ 1, m = r + 1:
For K ∈ {R, C} the group SL(m, K) is connected. The groups SL(m, C) are sim-
ply connected. One has
(
Z m=2
π1 (SL(m, R), ∗) =
Z/2 m ≥ 3
Spin(3, R) = SU(2),
68 2 Fundamentals of Lie algebra theory
No complex group from this list is compact, because any complex, connected
and compact Lie group is Abelian, see [27, Prop. 15.3.7].
Proposition 2.22 (Compact real form). Each complex Lie algebra from Proposition 2.13
has a compact real form, i.e. the real form can be choosen as the Lie algebra of a
compact matrix group:
1. General linear group, m ≥ 1: The real Lie algebra
is a compact real form of gl(m, C). It is the Lie algebra of infinitesimal genera-
tors of the unitary group U(m) which is a compact matrix group.
3. Series Br , m = 2r: The real Lie algebra so(m, R) is a compact real form
of so(m, C). It is the Lie algebra of infinitesimal generators of the compact real
matrix group SO(m, R).
sp(m) = {X ∈ gl(m, R) : g⊤ σ − σ · g∗ = 0}
is a compact real form of sp(m, C). It is the Lie algebra of infinitesimal genera-
tors of the compact symplectic group
Remark 2.23 (Exponential map of classical matrix groups). On one hand, in general
the exponential map
exp : Lie G → G
is not surjective for a connected classical matrix group G: According to Example 1.28
the matrix
−1 b
B= ∈ SL(2, R), b ∈ R∗ ,
0 −1
has not the form B = exp A with a matrix A ∈ sl(2, R).
exp : Lie G → G
If not stated otherwise, all Lie algebras and vector spaces in this chapter will be
assumed finite dimensional over the base field K = C or K = R.
ad x : L → L, y 7→ [x, y],
is nilpotent.
71
72 3 Nilpotent Lie algebras and solvable Lie algebras
ad x : L → L
is nilpotent.
The content of Lemma 3.2 can be parafrased as: Nilpotency implies ad-nilpotency.
Proof. The endomorphism x ∈ End(V ) acts on End(V ) by left composition and
right composition
l : End(V ) → End(V ), y 7→ x ◦ y,
r : End(V ) → End(V ), y 7→ y ◦ x.
Then
ad x = l − r
because for all y ∈ End(V )
l N = rN = 0
[l, r](y) = x ◦ (y ◦ x) − (x ◦ y) ◦ x = 0.
An important relation between the elements of End(V ) and the elements of the
Lie algebra gl(End V ) derives from the Jordan decomposition of endomorphisms
according to Theorem 1.17.
ad f ∈ L := gl(End V ),
defined as
3.1 Engel’s theorem for nilpotent Lie algebras 73
ad f : End(V ) →
− End(V ), g 7→ [ f , g],
is
ad f = ad fs + ad fn ∈ L.
In particular, ad fs is semisimple, ad fn is nilpotent and
[ad fs , ad fn ] = 0.
Ei j (vk ) = δ jk vi .
For 1 ≤ i, j, k ≤ n:
λi − λ j , 1 ≤ i, j ≤ n.
iii) Because
ad : gl(V ) → L
is a morphism of Lie algebras:
[ad fs , ad fn ] = ad([ fs , fn ]) = 0.
⊔
⊓
proves the existence of a common eigenvector for a whole Lie algebra of nilpotent
endomorphisms.
x(v) = 0 f or all x ∈ L.
Proof. The proof goes by induction on dim L, while the dimension of the
finite-dimensional vector space V is left arbitrary.
For the induction step assume dim L > 1 and assume that the theorem is true for all
Lie algebras of less dimension.
i) Existence of an ideal of codimension 1: By assumption all x ∈ L are nilpotent
endomorphisms of V . Hence by Lemma 3.2 each action
ad x : L → L, x ∈ L,
is nilpotent.
Choose
I∈A
as a subalgebra of L having maximal dimension with respect to all Lie algebras
from A . In particular
dim I < dim L
(ad x)(I) ⊂ I
y + I ∈ L/I
which has for all x ∈ I the eigenvalue zero with respect to ad x. Eigenvectors are
non-zero vectors, hence
y∈
/ I and [x, y] ∈ I for all x ∈ I.
As a consequence
y ∈ NL (I) \ I
i.e. I is properly contained in its normalizer. Due to the maximality of the
dimension of I we have NL (I) ∈ / A , hence
NL (I) = L,
π : L → L/I.
ii) Constructing a common eigenvector: The construction from part i) implies the
vector space decomposition
L = I + K · x0
with a suitable element x0 ∈ L \ I. By induction assumption the elements of the Lie
algebra I annihilate a common non-zero eigenvector, i.e.
x0 (w) ∈ W :
For all x ∈ I
x(x0 (w)) = x0 (x(w)) − [x0 , x](w) = 0.
Here the first summand vanishes because w ∈ W and x ∈ I. The second summand
vanishes because I ⊂ L is an ideal and therefore
76 3 Nilpotent Lie algebras and solvable Lie algebras
The restriction
x0 |W : W → W
is a nilpotent endomorphism and has an eigenvector v0 ∈ W with eigenvalue 0.
From the decomposition
L = I + K · x0
follows x(v0 ) = 0 for all x ∈ L. This finishes the induction step. ⊔
⊓
The fact, that any Lie algebra L ⊂ gl(V ) of nilpotent endomorphisms annihilates
a common eigenvector, allows the simultaneous triagonalization of all endomor-
phisms of L to obtain strict upper triangular matrices.
2. A flag (Vi )i=0,...,n of subspaces of V exists such that L(Vi ) ⊂ Vi−1 for all i = 1, ..., n.
W := V /(K · v1 )
x
V V
π π
x
W W
Now define
Vi := π −1 (Wi−1 ), i = 1, ..., n.
Then
x(Wi ) ⊂ Wi−1
implies
x(Vi ) ⊂ Vi−1 , i = 1, ..., n.
2) =⇒ 3) For the proof one extends successively a basis (v1 ) of V1 to a basis (v j ) j=1,...,i
of Vi , i = 1, ..., n.
A Lie algebra is Abelian when the commutator of any two elements vanishes.
Nilpotent Lie algebras are a first step to generalize this property: A Lie algebra is
named nilpotent when a number N ∈ N exists such that all N-fold commutators
vanish.
Definition 3.7 (Descending central series and nilpotent Lie algebra). Consider a
Lie algebra L.
1. The descending central series of L is the sequence (Ci L)i∈N of ideals Ci L ⊂ L,
inductively defined as
2. The Lie algebra L is nilpotent iff an index i0 ∈ N exists with Ci0 L = 0. The
smallest index with this property is named the length of the descending central
series.
Apparently, any subalgebra and any quotient algebra of a nilpotent Lie algebra L is
nilpotent too: The respective descending central series arise from the descending
central series of L. Concerning the reverse implication consider Lemma 3.8.
The Lie algebra n(n, K) of strictly upper triangular matrices and all subalgebras
of n(n, K) are nilpotent Lie algebras.
Lemma 3.8 (Central extension of a nilpotent Lie algebra). Consider a Lie al-
gebra L and an ideal I ⊂ L which is contained in the center, i.e. I ⊂ Z(L). If the
quotient L/I is nilpotent then L is nilpotent too.
Ci0 L ⊂ I ⊂ Z(L).
Ci0 +1 L = [L,Ci0 L] = 0.
⊔
⊓
In Lemma 3.8 one must not drop the assumption that the ideal I ⊂ L belongs to
the center.
Example 3.9 (Counter example for nilpotent extension). The descending central se-
ries of the Lie algebra
∗∗
L := t(2, K) = ∈ gl(2, K)
0∗
is
01
C0 L = [L, L] = n(2, K) = K ·
00
and
C1 L = [L, n(2, K)] = n(2, K) = C0 L ̸= 0.
Hence L is not nilpotent. Moreover
I := n(2, K) ⊂ L
3.1 Engel’s theorem for nilpotent Lie algebras 79
L/I ≃ d(2, K)
is also nilpotent.
We now prove Engel’s theorem. It is the main result about nilpotent Lie alge-
bras. It characterizes the nilpotency of a Lie algebra by the ad-nilpotency of all its
elements. Engel’s theorem follows as a corollary from Theorem 3.5.
Theorem 3.10 (Engel’s theorem for nilpotent Lie algebras). A Lie algebra L is
nilpotent if and only if every element x ∈ L is ad-nilpotent.
ad x : L → L, x ∈ L,
L/Z(L)
From a categorical point of view the concept of a short exact sequence is a useful
tool to handle injective or surjective morphisms and respectively their kernels and
cokernels.
i f
2. A chain complex of Lie algebra morphisms (Li − → Li+1 )i∈Z is exact or an exact
sequence of Lie algebra morphisms if for all i ∈ Z
fi−1 i f
im[Li−1 −−→ Li ] = ker[Li −
→ Li+1 ].
Using the concept of exact sequences we restate the relation between the nilpo-
tency of a Lie algebra L, an ideal of L and a quotient of L as follows:
Proposition 3.12 (Nilpotency and short exact sequences). Consider a short exact
sequence of Lie algebra morphisms
π
0 → L0 → L1 −
→ L2 → 0.
For a Lie algebra L one obtains for any i ∈ N a short exact sequence
By definition of the descending central series, the Lie algebra on the left-hand
side is contained in the center of the Lie algebra in the middle, i.e.
For a nilpotent Lie algebra L and minimal index i0 with Ci0 L = 0 the exact sequence
3.1 Engel’s theorem for nilpotent Lie algebras 81
0 → Ci0 −1 L → L → L/Ci0 −1 L
Z(L) ∩ I ̸= {0}.
In particular
CL (I) ̸= {0}.
ad x : L → L, x ∈ L,
(ad x)|I : I → I, x ∈ L,
[L, x0 ] = 0,
M = M +Ci0 L.
M ⊊ M +Ci L.
Then
M = M +Ci+1 L.
82 3 Nilpotent Lie algebras and solvable Lie algebras
Therefore
We obtain
M ⊊ M +Ci L ⊂ NL (M), q.e.d.
Solvability generalizes nilpotency. Solvable Lie algebras relate to nilpotent Lie al-
gebras like upper triangular matrices to strictly upper triangular matrices.
Definition 3.15 (Derived series and solvable Lie algebra). Consider a Lie algebra L.
1. The derived series of L is the sequence (Di L)i∈N inductively defined as
2. The Lie algebra L is solvable iff an index i0 ∈ N exists with Di0 L = 0. The small-
est index with this property is named the length of the derived series.
Lemma 3.16 (Solvability and short exact sequences). Consider an exact sequence
of Lie algebra morphisms
π
0 → L0 → L1 −
→ L2 → 0.
If any two terms of the sequence are solvable then the third term is solvable too.
Di L2 = Di π(L1 ) = π(Di L1 ).
Di0 + j0 L1 ⊂ D j0 L0 = 0.
The proof of the other direction uses the following relations between the derived
series:
Di L0 ⊂ Di L1 , π(Di L1 ) = Di L2 , q.e.d.
Corollary 3.17 (Solvable ideals). Consider a Lie algebra L. The sum I + J of two
solvable ideals I, J ⊂ L is solvable.
Proof. We apply Lemma 3.16: The quotient I/(I ∩ J) of the solvable Lie algebra I
is solvable too. By means of the canonical isomorphy
(I + J)/J ≃ I/(I ∩ J)
Definition 3.18 (Radical of a Lie algebra). The unique solvable ideal of L which
is maximal with respect to all solvable ideals of L is denoted rad L, the radical of L.
To verify that the concept is well-defined take rad L as a solvable ideal not con-
tained in any distinct solvable ideal. For an arbitrary solvable ideal I ⊂ L also
I + rad L
rad L ⊂ I + rad L
rad L = I + rad L.
Hence I ⊂ rad L.
The following Lemma 3.19 prepares the proof of Theorem 3.20 and therefore
also of Lie’s theorem about solvable Lie algebras.
84 3 Nilpotent Lie algebras and solvable Lie algebras
x(v) = λ (x) · v, x ∈ I,
λ ([x, y]) = 0
for all x ∈ I, y ∈ L.
The induction assumption applied to yi−1 (v) ∈ Wi proves: For the first summand
because yi−1 (v) ∈ Wi and [x, y] ∈ I, and for the second summand
x|W : W → W, x ∈ I,
3.2 Lie’s theorem for solvable Lie algebras 85
0 ∗
ii) Diagonal elements of the upper triangular matrices: We claim that all diagonal
elements of x|W are equal to λ (x), i.e.
λ (x) ∗
x|W =
.. , x ∈ I.
.
0 λ (x)
x(v) = λ (x) · v.
trace(x|W ) = n · λ (x).
λ :V →C
The proof of this theorem will imitate the proof of Theorem 3.5.
Proof. Without restriction we may assume L ̸= {0}. Then the proof goes by induc-
tion on dim L. The claim trivially holds for dim L = 0. For the induction step we
assume dim L > 0 and suppose that the claim is true for all solvable Lie algebras of
smaller dimension.
π : L → L/[L, L]
be the canonical projection of Lie algebras. The Lie algebra L/[L, L] is Abelian,
hence solvable. Therefore any arbitrary choosen vector subspace D ⊂ L/[L, L] of
codimension 1 is even an ideal D ⊂ L/[L, L]. Then I := π −1 (D) is an ideal of L with
The latter formula on the codimension is a general statement from the theory of
vector spaces:
dim I = dim π −1 (D) = dim D + dim [L, L]
implies
The subspace W ⊂ V is stable under the action of L, i.e. for v ∈ W and y ∈ L holds
y(v) ∈ W : For arbitrary x ∈ I, v ∈ W, y ∈ L we have
x(y(v)) = [x, y](v) + y(x(v)) = [x, y](v) + λ (x) · y(v) = λ ([x, y]) · y(v) + λ (x) · y(v).
Here
[x, y](v) = λ ([x, y]) · y(v)
because [L, I] ⊂ I. Lemma 3.19, applied to I ⊂ gl(W ), gives λ ([x, y]) = 0. Hence
and y(v) ∈ W .
L = I ⊕ C · x0 .
By part ii) the subspace W is stable under the action of the restricted endomor-
phism x0 |W . Because the field C is algebraic closed an eigenvector v0 ∈ W of x0
with eigenvalue λ ′ exists:
x0 (v0 ) = λ ′ · v0 .
The linear functional λ : I → C extends to a linear functional e
λ : L → C by
λ (x0 ) := λ ′ . Then
defining e
x : V → V, x ∈ L.
x(v) = λ (x) · v
V1 := C · v1 .
With respect to the basis B all endomorphism x ∈ L are represented by upper trian-
gular matrices, q.e.d.
ii) Assume that L solvable and dim L = n. Then also ad L ⊂ gl(L) is solvable
according to Lemma 3.16. Theorem 3.21 implies
ad L ⊂ t(n, C).
If x ∈ DL = [L, L] then
By Engel’s theorem, see Theorem 3.10, the Lie algebra DL is nilpotent, q.e.d.
Note that Corollary 3.22 is also valid for an R-Lie algebra L, because with respect
to complexification
Di L ⊗R C = Di (L ⊗R C),
and similarly for the descending central series.
3.3 Semidirect product of Lie algebras 89
The semidirect product of two Lie algebras is a Lie algebra structure on the product
vector space of the two Lie algebras. The most simple case of a semidirect product
is the direct product. It is obtained by taking the Lie bracket in each component
separately. But using certain derivations one can define a Lie bracket which mixes
the Lie brackets of the components. The semidirect product is an important tool to
construct new Lie algebras, and also for splitting Lie algebras into factors.
Definition 3.23 (Semidirect product). Consider two Lie algebras I and M with Lie
brackets respectively [−, −]I and [−, −]M , together with a morphism of Lie algebras
to the Lie algebra of derivations
θ : M → Der(I).
[−, −] : L × L → L
Proposition 3.24 (Semidirect product). Consider two Lie algebras I and M, and a
morphism of Lie algebras
θ :M→− Der(I).
1. The semidirect product
I ⋊θ M
is a Lie algebra.
2. The semidirect product fits into the exact sequence of Lie algebras
j π
0→
− I→
− I ⋊θ M −
→M→
− 0
3. The exact sequence has a section against π, i.e. a morphism of Lie algebras exists
s
s:M→
− I ⋊θ M
satisfying
π ◦ s = idM .
90 3 Nilpotent Lie algebras and solvable Lie algebras
4. We have
I ≃ j(I) ⊂ I ⋊θ M
an ideal, and
M ≃ s(M) ⊂ I ⋊θ M
a subalgebra.
• If x, y ∈ I and z ∈ M then
[x, y] = ([x1 , y1 ], 0); [[x, y], z] = ([[x1 , y1 ], 0]−θ (z2 )([x1 , y1 ]), 0) = (−θ (z2 )([x1 , y1 ]), 0)
[y, z] = ([y1 , z1 ]−θ (z2 )(y1 ), 0) = ([−θ (z2 )(y1 ), 0), [[y, z], x] = (−[θ (z2 )(y1 ), x1 ], 0)
[z, x] = ([z1 , x1 ] + θ (z2 )(x1 ), 0) = (θ (z2 )(x1 ), 0), [[z, x], y] = ([θ (z2 )(x1 ), y1 ], 0)
Because I is a Lie algebra, the claim reduces to
• If x ∈ I and y, z ∈ M then
[x, y] = (−θ (y2 )(x1 ), 0); [[x, y], z] = (θ (z2 )(θ (y2 )(x1 )), 0)
θ (z2 )(θ (y2 )(x1 )) + θ ([y2 , z2 ])(x1 ) − θ (y2 )(θ (z2 )(x1 )) = 0.
im j = j(I) = ker π.
ker π = im j = j(I) ≃ I.
π ◦ s = idM
s(M) ≃ M
is a subalgebra of I ⋊θ M.
I ⋊θ M ≃ I × M
s(M) ⊂ I ⋊θ M
being an ideal.
[P, Q] = −iℏ · 1
with P the momentum operator, Q the position operator, the unit operator 1, and
Planck’s constant
h
ℏ= = 1.054 × 10−27 erg · sec.
2π
This relation has been introduced by Born and Jordan and termed strict quantum
condition (German: verschärfte Quantenbedingung), see [3, Equation (38)].
is defined as
• Pj = E1,1+ j , the matrix with the only non-zero entry p j = 1,
• Q j = E1+ j,n+2 , the matrix with the only non-zero entry q j = 1, and
• I = E1,n+2 , the matrix with the only non-zero entry c = 1
has the commutators
Proof. Similarly to the case n = 1 one checks that heis(n) ⊂ n(n + 2, R) is closed
with respect to the Lie bracket.
The Heisenberg algebra is nilpotent, being a subalgebra of the nilpotent Lie al-
gebra n(n + 2, R).
(H := 1),
θ :R→
− Der(heis(n))
with
j(heis(n)) ⊂ heis(n) ⋊θ R
Proof. i) We have to show that θ maps to derivations. Using the shorthand D := θ (H):
• Left-hand side:
D([Pj , Pk )]) = D(0) = 0
Right-hand side:
• Left-hand side:
D([Q j , Qk )]) = D(0) = 0
Right-hand side:
• Left-hand side:
D([Pj , Qk )]) = D(δ jk E) = 0
Right-hand side:
• Left-hand side:
D([I, Pj ]) = D(0) = 0
Right-hand side:
• Left-hand side:
D([I, Q j ]) = D(0) = 0
Right-hand side:
D1 quant(n) = heis(n).
and
C2 L = [L, heis(n)] = C1 L.
Hence for all i ∈ N
Ci L ̸= {0},
which shows that quant(n) is not nilpotent. Corollary 3.22 and the subsequent note
imply that L is solvable, q.e.d.
2. The temporal development of the system, i.e. its dynamics, is governed by the
Hamiltonian H of the system according to the Schrödinger equation, the ordinary
differential equation
1 d
φ (t) + H(φ (t)) = 0, normalization: ℏ = 1.
i dt
The Hamiltonian H is the observable of the total energy of the system. The
Schrödinger picture considers the temporal development of the states and fixes
the observables. For conservative systems - and only these will be considered in
the sequel - the Hamiltonian has no explicit time dependency.
3. The Heisenberg picture takes the opposite approach: It fixes the states and con-
siders the temporal development of the observables. This approach resembles the
Hamiltonian approach of classical mechanics using canonical coordinates. In the
Heisenberg picture, the temporal development of an observable Ω is governed
by the commutator equation
1 . ∂
Ω (t) = [H, Ω (t)] + Ω (t).
i ∂t
∂
Here the partial derivation Ω vanishes in case Ω does not explicitly depend
∂t
on the time t. In particular, for a set of canonical observables P j , Q j , j = 1, ..., n:
1 .
Q (t) = [H, Q j (t)] = −P(t)
i j
96 3 Nilpotent Lie algebras and solvable Lie algebras
1 .
P j (t) = [H, P j (t)] = Q j (t).
i
A simple example of a 1-dimensional quantum system is the 1-dimensional os-
cillator. It has the Hamiltonian
1 2 mω 2 2
H= P + Q
2m 2
with respect to the momentum observable P and the position observable Q. The
real number ω denotes the oscillator frequency, the real number m denotes the
oscillator mass.
Note. There is no finite dimensional, i.e. m-dimensional dimensional representation ρ
of heis(n), m ̸= 0, with
because
tr ρ([P, Q]) = 0, but tr(λ · (−i · 1m ) ̸= 0.
The main result on the representation theory of self-adjoint operators P and Q satis-
fying the Born-Jordan commutation rules
[P, Q] = −iℏ · 1
is the Stone-von Neumann theorem, see [21, Theor. 14.8 and Chap. 9.8].
All vector spaces are assumed finite-dimensional if not stated otherwise. The com-
position f2 ◦ f1 of two endomorphisms will be denoted as product f2 · f1 or also f2 f1 .
The present section introduces a new tool for the study of Lie algebras: Important
properties of a Lie algebra L express themselves by a bilinear form, which derives
from the trace of the endomorphisms of a suitable representation of L.
Lemma 4.1 (Simple properties of the trace). Consider a vector space V and en-
domorphisms x, y, z ∈ End(V ).
1. For nilpotent x holds
tr x = 0.
2. With respect to two endomorphisms the trace is symmetric, i.e.
tr (xy) = tr (yx).
In particular, tr[x, y] = 0.
tr (xyz) = tr (yzx)
97
98 4 Killing form and semisimple Lie algebras
4) We have
[x, y]z = xyz − yxz and x[y, z] = xyz − xzy.
The ordering yxz results from the ordering xzy by cyclic permutation. Hence part 3)
implies
tr(yxz) = tr(yxz)
which proves the claim.
ρ : L → gl(V )
β : L × L → K, β (x, y) := tr (ρ(x)ρ(y)).
κ : L × L → K, κ(x, y) := tr (ad(x)ad(y))
xs = 0
we will show E = {0} or E ∗ = {0} for the dual space of Q-linear functionals
f : E → Q.
ad x : End(V ) →
− End(V )
ad x = ad xs + ad xn .
Therefore a polynomial
ps (T ) ∈ C[T ], ps (0) = 0,
exists with
ad xs = ps (ad x).
Consider a Q-linear functional f : E → Q. The endomorphism
y:V →V
is semisimple by construction.
According to the proof of Proposition 3.3 both ad xs and ad y act on End(V) as
diagonal matrices with respect to the standard basis (Ei j )1≤i, j≤n of End(V ) corre-
100 4 Killing form and semisimple Lie algebras
sponding to B:
ad y = q(ad xs ) : End(V ) →
− End(V ),
ad y = (q ◦ ps )(ad x)
xn · y ∈ End(V )
Therefore, all f (λi ), i = 1, ..., n, vanish. Hence f = 0 because the family (λi )i=1,...,n
spans E. As a consequence xs = 0, q.e.d.
Theorem 4.4 (Cartan’s trace criterion for solvability). Consider a complex vec-
tor space V and a subalgebra L ⊂ gl(V ). The following properties are equivalent:
• L is solvable
4.1 Traces of endomorphisms 101
x : V → V, x ∈ [L, L] ⊂ gl(V ),
ii) Trace check for nilpotency: Consider a fixed but arbitrary element
x0 ∈ [L, L] ⊂ gl(V ).
To show nilpotency of x0 we apply the trace check from Lemma 4.3 with A := [L, L], B := L
and
M := {x ∈ End(V ) : [x, L] ⊂ [L, L]}.
Apparently x0 ∈ [L, L] ⊂ L ⊂ M. It remains to show for all endomorphisms y ∈ M:
tr (xy) = 0.
Because [L, L] has generators [u, v], u, v ∈ L, we may assume x0 = [u, v]. According
to Lemma 4.1, part iv)
tr (x0 y) = 0.
2) Assume L solvable.
According to Lie’s theorem, see Theorem 3.21, the Lie algebra L is isomorphic
to a subalgebra of upper triangular matrices:
Denote by
(Vi )i=0,...,n , n = dim V,
102 4 Killing form and semisimple Lie algebras
and
x[u, v](Vi ) ⊂ x(Vi−1 ) ⊂ Vi−1 , i = 1, ..., n.
Therefore x · [u, v] is nilpotent and
Definition 4.5 (Simple and semisimple Lie algebras). Consider a Lie algebra L.
1. L is semisimple iff L has no Abelian ideal I ̸= 0.
2. L is simple iff L is not Abelian and L has no ideal other than 0 and L.
These concepts apply also to an ideal I ⊂ L when the ideal is considered a Lie
algebra.
Note: The trivial Lie algebra L = {0} is semisimple, but not simple. A semisimple
Lie algebra L ̸= {0} is not Abelian.
If L has a solvable Ideal I ̸= {0} then L has an Abelian ideal ̸= {0} too: Let i ∈ N
be the largest index from the derived series of I with Di I ̸= {0}. Then Di I ⊂ L is an
Abelian ideal because Di+1 I = [Di I, Di I] = {0}. The reverse implication is obvious:
Each Abelian ideal is solvable. As a consequence:
L semisimple ⇐⇒ rad(L) = 0.
Proof. According to Proposition 2.9 the adjoint representations maps to the Lie
algebra of derivations. The kernel is the center
Proposition 4.7 (Semisimple Lie algebra sl(n, K)). The Lie algebra
is semisimple.
R ⊂ t(n, C) ∩ L.
A direct conclusion from the definition of solvability is the equivalence
X ∈ R ⇐⇒ X ⊤ ∈ R.
R ⊂ d(n, C) ∩ L.
[L, R] ⊂ R.
X = λ · 1, λ ∈ C,
X = 0.
ii) Real base field: A real Lie algebra M is semisimple iff its complexification
M ⊗ RC
Later we will see in Theorem 7.11 that the Lie algebras sl(n, C), so(n, C), sp(2n, C)
are semisimple and - under the dimensional restrictions from Proposition 2.13 - even
simple Lie algebras. The most basic example from these series is the simple Lie al-
gebra sl(2, C).
Any Lie algebra becomes semisimple after dividing out its radical.
Proposition 4.8 (Semisimpleness after dividing out the radical). For any Lie
algebra L the quotient L/rad(L) is semisimple.
π :L→
− L/rad(L).
Any ideal
I ⊂ L/rad(L)
has the form
I = J/rad(L)
with the ideal J := π −1 (I) ⊂ L. If I is an Abelian ideal, then [I, I] = 0, hence [J, J] ⊂ rad(L).
Therefore the derived algebra
D1 J = [J, J]
is solvable. The latter property implies J solvable because
I = π(J) = 0, q.e.d.
4.2 Fundamentals of semisimple Lie algebras 105
According to Proposition 4.8 any Lie algebra L fits into an exact sequence
0→
− I→
− L→
− M→
− 0
with I = rad(L), a solvable subalgebra, and M = L/I semisimple. The Levi de-
composition, Theorem ??, will show that one can even obtain L as the semi-direct
product
L ≃ I ⋊θ M.
Definition 4.9 (Orthogonal space of the trace form). Consider a K-Lie algebra L
and a symmetric bilinear map
β : L × L → K.
Lemma 4.10 (Orthogonal space of ideals). Consider a vector space V , a Lie alge-
bra L ⊂ gl(V ) and the trace form
β : L × L → C, β (x, y) := tr(xy).
S := L⊥ = {x ∈ L : tr(xy) = 0 f or all y ∈ L}
tr([x, y] ◦ z) = 0
because z ∈ S. Cartan’s trace criterion, see Theorem 4.4, applied to the Lie algebra
S, shows that S is solvable. Hence
S⊂L
The principal theorem of the present section is the following Cartan criterion for
semisimplicity. It derives from Cartan’s trace criterion for solvability.
Theorem 4.12 (Cartan criterion for semisimplicity). For a Lie algebra L the fol-
lowing properties are equivalent:
• L is semisimple
ad L ⊂ gl(L).
S := L⊥ = 0.
ad(x) ◦ ad(y) : L → I ⊂ L
Hence
4.2 Fundamentals of semisimple Lie algebras 107
Lemma 4.13 (Killing form of an ideal). Consider a Lie algebra L. For any ideal I ⊂ L
the Killing form of I is the restriction of the Killing form of L to I
κI = κ|I × I.
and
A of the restriction (ad x)|I ◦ (ad y)|I : I → I
show
tr((ad x) ◦ (ad y)) = tr A = tr((ad x)|I ◦ (ad y)|I).
As a consequence
The first step on the way to split a semisimple Lie algebra as a direct sum of
simple Lie algebras is the following Proposition 4.14.
L = I ⊕ I⊥.
[I, J] ⊂ I ∩ J = {0}.
L ≃ I × J.
Proof. 1) Dimension formula: First we consider the underlying vector spaces. The
Killing form κ of L induces a map of vector spaces
F : L −→ L∗ , x 7→ κ(x, −).
The map is injective because the Killing form is nondegenerate according to Car-
tan’s criterion for semisimplicity, Theorem 4.12. It is also surjective because
dim L = dim L∗ .
We obtain
i.e.
dim L = dim I + dim I ⊥ .
Secondly, we show I ∩ I ⊥ = {0}: According to Lemma 4.10 also the orthogo-
nal space I ⊥ ⊂ L is an ideal. We apply Cartan’s trace criterion for solvability,
Theorem 4.4, to the ideal
J := I ∩ I ⊥
considered as a Lie algebra. Denote by κ the Killing form of L and by κJ the Killing
form of J. According to Lemma 4.13 the Killing form κJ is the restriction of κ. We
have
κJ ([J, J], J) = κ([J, J], J) = 0
4.2 Fundamentals of semisimple Lie algebras 109
J = {0}.
L = I ⊕ I⊥.
of simple ideals Iα ⊂ L.
2. Assume L semisimple.
i) Each ideal I ⊂ L is a finite direct sum of simple ideals. In particular, for each
ideal I ⊂ L an ideal J ⊂ L exists with
L = I ⊕ J.
L = D1 L = [L, L].
110 4 Killing form and semisimple Lie algebras
Proof. 1. i) Assume L semisimple. In case L = {0} take the empty sum with A = 0.
/
Otherwise L ̸= {0} and L is not Abelian.
If L has no proper ideals, then L itself is simple. Otherwise we choose a proper
ideal
{0} ⊊ I1 ⊊ L
of minimal dimension. Proposition 4.14 implies a direct sum representation
L = I1 ⊕ I1⊥ .
L = I1 ⊕ I2 ⊕ ... ⊕ In
does not contain any proper ideal. The decomposition stops after finitely many
steps because each steps decreases the dimension of the ideals in question.
I ∩ Iα
I ∩ Iα = {0}
or
I ∩ Iα = Iα .
The latter case is excluded because the simple Lie algebra Iα is not Abelian. As
a consequence I = 0.
pα : L →
− Iα ,
and M
idL = pα .
α∈A
If I ̸= {0} then M
{0} ̸= I = p(I) = pα (I).
α∈A
pα (I) ̸= {0}.
Simpleness of Iα implies
pα (I) = Iα .
Then
pα ([Iα , I]) = [pα (Iα ), p(I)] = [Iα , Iα ] ̸= {0}
because the simple ideal Iα is not Abelian. Hence
{0} ̸= [Iα , I] ⊂ Iα ,
which implies
[Iα , I] = Iα
and eventually
Iα = [Iα , I] ⊂ I.
If I is simple, then
Iα = I.
For a general ideal I ⊂ L we just proved
pα (I) ̸= {0} =⇒ Iα ⊂ I.
Hence
Iα , A′ = {α ∈ A : pα (I) ̸= {0}}.
M
I=
α∈A′
iii) Derived algebra: The splitting of L implies the direct sum decomposition
M M M M
[L, L] = [ Iα , Iβ ] = [Iα , Iβ ] = [Iα , Iα ].
α∈A β ∈A α,β ∈A α∈A
For each α ∈ A either [Iα , Iα ] = {0} or [Iα , Iα ] = Iα because the simple Lie alge-
bra Iα has no proper ideals. The case [Iα , Iα ] = {0} is excluded because Iα is not
Abelian. As a consequence
M
[L, L] = Iα = L, q.e.d.
α∈A
Alike to splitting a semisimple Lie algebra into a direct sum of simple Lie alge-
bras Weyl’s Theorem 4.24 splits an arbitrary finite-dimensional representation of a
semisimple Lie algebra L as a direct sum of irreducible representations of L.
ρ : L → gl(V ).
ρ(L)(W ) ⊂ W.
2. An L-module V is irreducible iff V has exactly the two L-submodules V and {0}.
Otherwise V is reducible. Notably, the zero-module is reducible.
V = W ⊕W ′
The constructions from part 2) - 4) generalize to products of more than two fac-
tors. The notations emphasize the tight relationship to similar constructions in Com-
mutative Algebra for modules over rings.
Left-hand side:
Right-hand side:
is an isomorphism of L-modules.
L. f = 0 ⇐⇒ f : V → W is L-linear.
4.3 Weyl’s theorem on complete reducibility of representations 115
2. If K = C and
ρ : L → gl(V )
is an irreducible complex representation, then any endomorphism f ∈ End(V )
commuting with all elements of ρ is a multiple of the identity, i.e. if for all x ∈ L
[ f , ρ(x)] = 0
f = µ · id.
3. If K = C, and
f1 , f2 : V → W
are two morphisms of irreducible L-modules with f2 ̸= 0. Then
f1 = c · f2
f (w) = c · w
f := f1 ◦ f2−1 : W →
− W.
x. f (w) = f (x.w),
f = c · idW
which proves
f1 = c · f2 .
The last claim about HomL (V,W ) follows from part 1), q.e.d.
β (xi , y j ) = δi j .
Note: One checks that the Casimir element does not depend on the choice of the
basis (xi )i=1,...,n .
L = ker ρ ⊕ L′
with L′ := (ker ρ)⊥ ⊂ L. The Lie algebra L′ is semisimple according to Proposition 4.14.
The restricted representation ρ|L′ → gl(V ) is faithful.
4.3 Weyl’s theorem on complete reducibility of representations 117
Theorem 4.22 (Properties of the Casimir element). The Casimir element cρ ∈ End(V )
of a faithful representation ρ of a complex semisimple Lie algebra L has the follow-
ing properties:
• Commutation: The Casimir element commutes with all elements of the represen-
tation:
[cρ , ρ(L)] = 0
• Trace: tr(cρ ) = dim L
dim L
cρ = · idV .
dim V
Proof. Set n = dim L and
n
cρ = ∑ ρ(xi )ρ(yi ) ∈ End(V )
i=1
with bases (xi )i=1,...,n and (y j ) j=1,...,n of L, which are dual with respect to the trace
form β of ρ.
Because the families (x j )1≤ j≤n and (yk )1≤k≤n are dual bases with respect to β
n
ai j = β ( ∑ aik · xk , y j ) = β ([x, xi ], y j ) = −β ([xi , x], y j ) = −β (xi , [x, y j ]) =
k=1
n
= −β (xi , ∑ b jk · yk ) = −b ji .
k=1
Here we made use of the associativity of the trace form according to Lemma 4.1. To
compute
n
[ρ(x), cρ ] = ∑ [ρ(x), ρ(xi )ρ(yi )]
i=1
and therefore
n
[ρ(x), cρ ] = ∑ ([ρ(x), ρ(xi )]ρ(yi ) + ρ(xi )[ρ(x), ρ(yi )]) =
i=1
n n n
= ∑ (ρ([x, xi ])ρ(yi )+ρ(xi )ρ([x, yi ]) = ∑ ai j ·ρ(x j )ρ(yi )+ ∑ bik ·ρ(xi )ρ(yk ) =
i=1 i, j=1 i,k=1
n n
= ∑ ai j · ρ(x j )ρ(yi ) + ∑ b ji · ρ(x j )ρ(yi ) =
i, j=1 i, j=1
n
= ∑ (ai j + b ji ) · (ρ(x j )ρ(yi )) = 0.
i, j=1
Here we have changed in the second sum the summation indices (i, k) 7→ ( j, i).
iii) Scalar: For an irreducible representation ρ we get with part i) and the Lemma
of Schur, Theorem 4.19,
cρ = µ · idV
and with part ii)
tr(cρ ) = µ · dim V = dim L, q.e.d.
Proof. Due to semisimpleness we have L = [L, L], see Theorem 4.15. If x = [u, v] ∈ L
then
tr(ρ(x)) = tr(ρ([u, v])) = tr([ρ(u), ρ(v)]) = 0
according to Lemma 4.1. For 1-dimensional V , i.e. V = C, and for all x ∈ L
Note: The irreducible direct summands of V are uniquely determined, not only their
isomorphy classes.
2. codimV W arbitrary.
1. Codimension = 1: Consider pairs (V,W ) with an L-module V and a submodule
W ⊂ V with codimV W = 1.
0→W →V →C→0
We construct a complement of W by induction on dim W . The induction step em-
ploys one of two alternative subcases. Subcase 1a) uses the induction hypothesis,
while subcase 1b) uses the Casimir element.
{0} ⊊ W ′ ⊊ W,
0 → W /W ′ → V /W ′ → C → 0
Because
dim (W /W ′ ) < dim W
the pair (V /W ′ ,W /W ′ ) satisfies the induction assumption. We obtain a submod-
ule W ′ ⊂ W̃ ⊂ V and a splitting of L-modules
V /W ′ = W /W ′ ⊕ W̃ /W ′ .
0 → W ′ → W̃ → W̃ /W ′ ≃ C → 0
The estimates
codimW̃ W ′ = 1 and codimW ′ W ≥ 1
imply
dim W ′ < dim W
also the pair (W̃ ,W ′ ) satisfies the induction assumption. We obtain a submodule X ⊂ W̃
and a splitting of L-modules
W̃ = W ′ ⊕ X.
We claim
V = W ⊕ X.
For the proof, on one hand the two splittings imply
Hence
dim V = dim W + dim X.
On the other hand,
X ⊂ W̃
(W ∩ X) ⊂ (W ∩ W̃ ).
The first splitting implies
{0} = (W /W ′ ) ∩ (W̃ /W ′ ) = (W ∩ W̃ ) ⊂ W ′ ,
and therefore
(W ∩ X) ⊂ (W ∩ W̃ ) ⊂ W ′ .
As a consequence
W ∩ X = (W ∩ X) ∩W ′ = W ∩ (X ∩W ′ ).
X ∩W ′ = {0}.
Hence
W ∩ X = {0}.
Therefore
V =W ⊕X
which finishes the induction step for reducible W .
ρ : L → gl(V )
X := ker cρ ⊂ V
is an L-submodule.
ρ(L)(V ) ⊂ W.
cρ (W ) ⊂ W.
cρ |W = µ · idW .
From
tr cρ = dim L
follows µ ̸= 0, and in particular
W ∩ X = {0}.
X = ker cρ ̸= {0}.
Because codimV W = 1 we get
V =W ⊕X
{0} ⊊ W ⊊ V.
→ V,
W ,−
f0 : V → W
HomC (V,W )
W := { f ∈ V : f |W = 0}.
In order to prove that V is a L-module, consider f ∈ HomC (V,W ) with f |W = λ · idW
and x ∈ L, w ∈ W :
codimV W = 1
4.3 Weyl’s theorem on complete reducibility of representations 123
C · f0 , f0 ∈ V ⊂ HomC (V,W ),
i.e.
V = W ⊕ C · f0 .
Without restriction
f0 |W = idW .
The L-module C · f0 is 1-dimensional, hence trivial according to Lemma 4.23.
According to Remark 4.18 the equality L. f0 = 0 implies the L-linearity of
f0 : V → W.
Therefore
X := ker f0 ⊂ V
is a submodule. Due to our considerations at the beginning of case 2)
V = W ⊕ X, q.e.d.
x :V →V
x = xs + xn ∈ End(V ),
Proof. The proof has two separate parts. The first part deals with the Lie algebra L
and considers L ⊂ gl(V ) as a matrix algebra. The second part deals with the Lie
algebra L when considered an abstract semisimple Lie algebra.
x :V →V
ad x = ad xs + ad xn
N := Ngl(V ) L ⊂ gl(V ).
xs , xn ∈ N.
L ⊂ L̃ ⊂ N
and
tr((z.y)|W ) = tr([z, y]|W ) = 0.
We define \ \
L̃ := N ∩ LW = (N ∩ LW ).
W ⊂V W ⊂V
xs (W ) ⊂ W and xn (W ) ⊂ W.
tr (xn |W ) = 0 and xn ∈ LW .
As a consequence also
xs = x − xn ∈ LW .
We obtain
xs , xn ∈ L̃
because xs , xn ∈ N and xs , xn ∈ LW for all L-submodules W ⊂ V .
L̃ = L ⊕ M.
[L, y] = 0.
y = µ · idW .
y|W = 0.
L = L̃, q.e.d.
The previous Theorem 4.25 assures: For a semisimple Lie algebra L the Jordan
decomposition of elements from ad L induces a decomposition of elements from L.
ad x : L → L, y 7→ [x, y]
ad x = fs + fn ∈ End(L)
with
fs ∈ ad(L) and fn ∈ ad(L),
see Theorem 4.25. Defining
s := ad −1 ( fs ) ∈ L and n := ad −1 ( fn ) ∈ L
provides the decomposition
x = s+n
with s ∈ L ad-semisimple, n ∈ L ad-nilpotent and [s, n] = 0. This decomposition is
named the abstract Jordan decomposition of x ∈ L.
ad x : L →
− L
In the abstract case, neither x ∈ L nor its components s, n ∈ L from the abstract
Jordan decomposition are defined as endomorphisms of a vector space. But Corol-
lary 4.27 will show: For any representation
ρ : L → gl(V )
ρ : L → gl(V )
ad s : L → L
is semisimple, i.e. the eigenvectors (vi )i∈I of ad s ∈ End(L) span the vector space L.
Hence the vectors from ρ(L)
(ρ(vi ))i∈I
span the vector space F. They are eigenvectors of ad(ρ(s)) because
implies
ad ρ(s) ∈ End(F)
128 4 Killing form and semisimple Lie algebras
ad n ∈ End(L)
ad ρ(n) ∈ End(F)
is nilpotent because
(ad ρ(n))k = ρ ◦ (ad n)k .
Apparently,
[ρ(s), ρ(n)] = ρ([s, n]) = 0
which implies [ad ρ(s), ad ρ(n)] = 0. Summing up:
ad ρ(x) ∈ ad(F).
By definition
ρ(x) = ρ(s) + ρ(n) ∈ F
is the abstract Jordan decomposition in the Lie algebra F of the element ρ(x) ∈ F.
fs , fn ∈ ρ(L).
ρ(x) = fs + fn
is the abstract Jordan decomposition of ρ(x) ∈ F. From the uniqueness of the ab-
stract Jordan decomposition in F derives
Which are the fundamental results from the theory of complex semisimpleLie
algebras?
(Lα ⊕ L−α )
M
L=H⊕
α∈Φ +
Lα = C · xα , L−α = C · yα
satisfying
[xα , yα ] = hα ∈ H.
For each root α ∈ Φ+ the Lie algebra
2. Integrality of the roots: For the first time the importance of integer numbers for
complex Lie algebras shows up as the integer = 2 in the commutator relations
Via the adjoint representation the element h ∈ sl(2, C) acts as diagonal endomor-
phism on sl(2, C), and the corresponding eigenvalues, are
0, 2, −2.
In the general case, under the adjoint representation the Cartan algebra H ⊂ L
acts in a diagonal way on L, and the corresponding eigenvalues, the roots from
the root set Φ = Φ + ∪ Φ − ,
α :H → − C
satisfy
< α1 , α2 >:= α1 (hα2 ) ∈ Z (Cartan integer),
see Proposition 7.4. The proof relies on the existence of the subalgebras
Sα ≃ sl(2, C) ⊂ L
131
3. Root system: The integrality of the Cartan integers proves that the root set Φ
of L is a root system, see Definition 6.2: First, the non-degenerateness of the
Killing-form κ provides an identification of the Cartan algebra H ⊂ L with its
dual space
≃
H− → H ∗ , tα 7→ α,
and introduces on H ∗ the non-degenerate bilinear form
Secondly, the root set Φ spans the dual space H ∗ . After choosing a basis of H ∗
formed by roots, any other root is a rational linear combination of the basis roots.
The restriction of the form (−, −) to the rational vector space
VQ := spanQ Φ
(v, α)
σα : V →
− V, v 7→ v− < v, α > ·α, < v, α >:= 2 · = v(hα ),
(α, α)
To generalize this result to L-modules one introduces for any Lie algebra A the
universal enveloping algebra UA of A as a quotient of the tensor algebra of A,
see Definition 8.1. To obtain an irreducible UL-module V of arbitrary highest
weight λ ∈ H ∗ one takes an arbitrary non-zero vector e and sets
and
x.e = 0 for all α ∈ Φ + and x ∈ Lα .
The 1-dimensional representation Ce of the Borel subalgebra
M
B := H ⊕ Lα ⊂ L
α∈Φ +
UL ⊗ UB Ce.
see Theorem 8.26. The fundamental weights are obtained from a base
∆ = {α1 , ..., αr }
of the root system of L by applying the Cartan matrix, see Proposition 8.25.
Chapter 5
Cartan decomposition
sl(2, C) = C · h ⊕ (C · x ⊕ C · y)
The base field in this chapter is K = C, the field of complex numbers. All Lie
algebras are complex Lie algebras unless stated otherwise.
ad x : L →
− L, y 7→ [x, y],
A second result, Theorem 5.17 from Section 5.3, will show that no other elements
of L commute with a maximal toral subalgebra T ⊂ L, i.e. CL (T ) = T .
133
134 5 Cartan decomposition
We recall from Linear Algebra the following Lemma 5.2: The restriction of a
diagonalizable endomorphism to an invariant subspace stays diagonalizable.
x :V →V
x|W : W → W
we show: If
w = v1 + ... + vk ∈ W
with eigenvectors vi of x with pairwise distinct eigenvalues λi , then vi ∈ W for
all i = 1, ..., k.
The claim holds for k = 1. For the induction step k − 1 7→ k consider
w = v1 + ... + vk ∈ W
with
5.1 Toral subalgebra 135
x(w) = λ1 · v1 + ... + λk · vk ∈ W
and apply the induction assumption to
w = v1 + ... + vk ∈ W
implies v1 ∈ W , q.e.d.
x = s+n
C·s ⊂ L
is a toral subalgebra.
[x, y] = λ · y.
ad(y) : L →
− L
adT (y) : T →
− T
We obtain
α j · λ j ̸= 0.
The 3-dimensional complex Lie algebra sl(2, C) is the prototype of complex semisim-
ple Lie algebras. Also its representation theory is of fundamental importance:
• The representation theory of sl(2, C) is the means to clarify the structure of gen-
eral complex semisimple Lie algebras, see Proposition 7.3 about the Cartan de-
composition. Proposition 5.4 presents the structure of sl(2, C) in a form which
generalizes to arbitrary complex semisimple Lie algebras.
L := sl(2, C)
ad : L → gl(L)
are
0 0 0 0 01 0 −1 0
ad h = 0 2 0 , ad x = −2 0 0 , ad y = 0 0 0
0 0 −2 0 00 2 0 0
• The element h ∈ L is ad-semisimple: ad h is a diagonal matrix. It has the eigen-
values
0, α := 2, −α
with eigenspaces respectively
H := L0 = C · h, Lα = C · x, L−α = C · y.
In particular
L = H ⊕ L2 ⊕ L−2
as a direct sum of complex vector spaces.
• With respect to the basis B the Killing form of L, see Definition 4.2, has the
symmetric matrix with integer coefficients
2 0 0
4 · 0 0 1 ∈ M(3 × 3, Z).
0 1 0
• The restriction
κH := κ|H × H
of the Killing form is positive definite: κH (h, h) = 8. The scalar product κH in-
duces the isomorphism
∼
→ H ∗ , h 7→ κH (h, −) = 8 · h∗ .
H−
138 5 Cartan decomposition
h
tα = .
4
It satisfies
2 · tα
h= .
κ(tα ,tα )
Proof. Only the following issues need a separate proof:
i) Simpleness: Due to the commutator relations L is not Abelian. Assume the
existence of a proper ideal I ⊂ L:
{0} ⊊ I ⊊ L.
The splitting
L = H ⊕ L2 ⊕ L−2
implies the splitting of the ideal I
I = (I ∩ H) ⊕ (I ∩ L2 ) ⊕ (I ∩ L−2 )
according to Lemma 5.2. We now use once more the fact [I, L] ⊂ I.
• First, h ∈
/ I: Otherwise 2x = [h, x] ∈ I and − 2y = [h, y] ∈ I, which implies I = L,
a contradiction.
• Secondly, x ∈
/ I: Otherwise h = [x, y] ∈ I, contradicting the first part.
• Thirdly, y ∈
/ I: Otherwise −h = [y, x] ∈ I, contradicting the first part.
Each of the three summands in the splitting of I has dimension at most = 1.
Hence I = {0}, a contradiction, proving the non-existence of I. Therefore L is
simple.
Remark 5.5 (Basis of Pauli matrices). A different basis of sl(2, C) is the family (σ j ) j = 1, 2, 3
of Pauli matrices, see Remark 2.16: We have
σ1 + iσ2 σ1 − iσ2
h = σ3 , x = , y=
2 2
5.2 sl(2, C)-modules 139
The element
h ∈ L := sl(2, C)
is ad-semisimple by definition because ad h ∈ End(L) is semisimple. Therefore h ∈ L
coincides with its semisimple component in the abstract Jordan decomposition of L.
Corollary 4.27 shows the far reaching consequences: The element h acts as semisim-
ple endomorphism on any L-module V . Hence any L-module decomposes as a direct
sum of eigenspaces with respect to the action of h.
Definition 5.6 (Weight, weightspace and primitive element). Consider an sl(2, C)-module V ,
not necessarily finite-dimensional.
1. Denote by
V λ = {v ∈ V : h.v = λ · v}, λ ∈ C,
the eigenspaces with respect to the action of h. If
V λ ̸= {0}
then V λ is named a weight space of V , its non-zero elements are named weight
vectors, and the corresponding eigenvalue λ ∈ C is a weight of V .
For any sl(2, C)-module V not only the action of h ∈ sl(2, C), but also the action
of the two other elements of the standard basis
B = (h, x, y)
1 i
ei := · y .e, i ≥ 0, e−1 := 0,
i!
satisfy for all i ≥ 0:
1. Weight: h.ei = (λ − 2i) · ei .
λ =: m ∈ N,
the family (ei )i=0,...,m is linearly independent, and ei = 0 for all i > m.
1 1
y.ei = y.(yi .e) = (yi+1 .ei+1 ) = (i + 1) ei+1 .
i! i!
5.2 sl(2, C)-modules 141
= (λ − 2i − 2) (y.ei ) = (λ − 2i − 2) (i + 1)ei+1 ,
hence
h.ei+1 = (λ − 2i − 2) ei+1 .
ad 3) With the convention e−1 := 0 the formula
x.ei = (λ − i + 1) · ei−1
[x, y] = h ∈ sl(2, C)
0 = x.em+1 = (λ − m)em
is solvable. According to Lie’s theorem, see Theorem 3.20, the sl(2, C)-operation
on V has a common eigenvector e ∈ V . Consider the eigenvalues λh , λx defined
by
h.e = λh · e and x.e = λx · e.
The commutator [h, x] = 2x implies
λ =: m ∈ N.
and successive application of y lowers the weight up to the value −m. Therefore λ ≥ 0.
Chapter 8 will generalize Proposition 5.7 and Corollary 5.8 from sl(2, C) to ar-
bitrary semisimple Lie algebras L. The solvable Lie algebra B will be called a Borel
subalgebra of L. The submodule generated by a primitive element will be named a
standard cyclic module, cf. Definition 8.16.
The construction of the sl(2, C)-module V (λ ) from Proposition 5.7 generates all
irreducible sl(2, C)-modules.
λ
V λ −2i .
M
V=
i=0
is bijective.
L ×V →
−V
V (λ1 ) ≃ V (λ2 ),
144 5 Cartan decomposition
in particular
V (0) = V 0 ≃ C.
→ gl(C2 ).
ρ : L = sl(2, C) ,−
V (1) = V 1 ⊕V −1
with
1 1 −1 0
V = C· ,V = C ·
0 1
and primitive element
1
e= .
0
5.2 sl(2, C)-modules 145
ad : L → gl(L).
V (2) = L = L2 ⊕ L0 ⊕ L−2
P(u, v) ∈ C[u, v]
of degree = n.
The vector space C[u, v] of polynomials in two variables u and v has a basis
of monomials (uµ · vν )µ,ν∈N . A homogeneous polynomial of degree n ∈ N is an
element
n
P(u, v) = ∑ aµ · uµ · vn−µ ∈ C[u, v], aµ ∈ C.
µ=0
Denote by
Pol n ⊂ C[u, v]
the subspace of homogeneoups polynomials of degree n. One has
dim Pol n = n + 1
Pol 1 ≃ C · u ⊕ C · v
∂ ∂
Dz := (z.u) + (z.v) ,
∂u ∂v
notably for un−i · vi ∈ Pol n
h.(un−i ·vi ) := Dh (un−i ·vi ) = u·(n−i)·un−i−1 ·vi −i·v·un−i ·vi−1 = (n−2i)·un−i ·vi
e := un ∈ Pol n
We obtain
h.ei = (n − 2i) · ei
n n−i i n
y.ei = y. ·u ·v = · (n − i) · un−i−1 · vi+1 =
i i
n
= (i + 1) · · un−i−1 · vi+1 = (i + 1) · ei+1
i+1
n n−i i n
x.ei = x. ·u ·v = i· · un−i+1 · vi−1 =
i i
n
(n − i + 1) · · un−i+1 · vi−1 = (n − i + 1) · ei−1 ,
i−1
which proves Pol n ≃ V (n) with primitive element e := un ∈ Pol n .
induces an isomorphy of L-modules between Pol λ and the symmetric power with
exponent λ of the tautological L-module.
5.3 Cartan subalgebra and root space decomposition 147
For a semisimple Lie algebra L we choose an arbitrary, but fixed maximal toral
subalgebra T ⊂ L. According to Proposition 5.3 any toral subalgebra is Abelian.
Hence all endomorphisms
ad h : L → L, h ∈ T,
are simultaneously diagonizable, and the whole Lie algebra L splits as a direct sum
of the common eigenspaces of T . This decomposition is called the root space de-
composition of L with respect to T .
We recall from Lemma 4.1 that the Killing form of L is “associative”
Definition 5.12 (Root space decomposition). Consider a pair (L, T ) with a semisim-
ple Lie algebra L and a maximal toral subalgebra T ⊂ L.
1. For a complex linear functional
α :T →C
set
Lα := {x ∈ L : [h, x] = α(h) · x f or all h ∈ T }.
If α ̸= 0 and Lα ̸= {0} then α is a root of (L, T ) and Lα is the corresponding root
space. Any non-zero vector v ∈ V λ is named a root vector.
Φ := {α : T → C : α ̸= 0, Lα ̸= {0}}.
• The centralizer CL (T ) contains with each element also its ad-semisimple and
its ad-nilpotent summand. Therefore one can consider both types of elements
separately.
• The case of ad-semisimple elements is trivial due to the maximality of T with
respect to ad-semisimple elements.
• If non-zero ad-nilpotent elements of CL (T ) exist, they cannot belong to T . A
posteriori, as a subset of T the centralizer CL (T ) cannot contain any non-zero ad-
nilpotent elements because all elements of a toral subalgebra are ad-semisimple.
• Because the subalgebra CL (T ) is Abelian all its ad-nilpotent elements belong to
the null space of the Killing form restricted to CL (T ).
• The Killing form is nondegenerate on CL (T ).
The following Lemmata 5.13 and 5.15 as well as Proposition 5.16 prepare the
proof of Theorem 5.17.
Lemma 5.13 (Orthogonality of root spaces). Consider a pair (L, T ) with a semisim-
ple Lie algebra L and a maximal toral subalgebra T ⊂ L. Consider the correspond-
ing root space decomposition from Definition 5.12
M
L = L0 Lα .
α∈Φ
[h, [x, y]] = −([x, [y, h]] + [y, [h, x]]) = [x, β (h) · y] − [y, α(h) · x]
Corollary 5.14 (Negative of a root). For a semisimple Lie algebra L the negative (−α)
of a root α ∈ Φ is again a root of L.
Proof. Assume on the contrary that −α is not a root. Then for all roots β ∈ Φ
α + β ̸= 0.
implies
κ(Lα , L) = 0.
Theorem 4.12 on the non-degeneratedness of the Killling form implies Lα = 0, a
contradiction, q.e.d.
s, n ∈ CL (T )
ad x = ad s + ad n ∈ End(L)
Hence s, n ∈ C.
ad(T + C · x) ⊂ ad(L)
T + C · x ⊂ T,
i.e. x ∈ T , q.e.d.
The strategy to prove Theorem 5.17 uses the fact that the Killing form restricted
to a maximal toral subalgebra is non-degenerate. The following Proposition 5.16
proves this result in two steps. Nevertheless, Theorem 5.17 will show a posteriori
that both steps coincide.
κT := κ|T × T : T × T → C
are non-degenerate.
κ(x, L) = 0.
is nilpotent. Hence
κC (h, n) = tr((ad h) ◦ (ad n)) = 0
by Corollary 4.1. For a general element x ∈ C with abstract Jordan decomposition
x = s+n
we first have s ∈ C due to Lemma 5.15, and then s ∈ T due to the same Lemma.
Hence
κC (h, x) = κC (h, s) + κC (h, n) = κT (h, s) + κC (h, n) = 0
which implies h = 0, q.e.d.
T = CL (T ).
CL (T ) ⊂ T.
For the proof set C := CL (T ). Due to Lemma 5.15 the main step to be proved will
be: Any ad-nilpotent element n ∈ C belongs to T . For proving that step we will use
the non-degenerateness of the restricted Killing form, see Proposition 5.16.
ii) The centralizer C is nilpotent: According to Engel’s theorem, see Theorem 3.10,
it suffices to show that for each element x ∈ C the endomorphism
adC x ∈ End(C)
152 5 Cartan decomposition
x = s + n.
adC x = adC n
is nilpotent.
{0} ̸= I := [C,C] ⊂ C
T ∩ [C,C] = {0}
x = s+n
x = s + n.
is nilpotent and
κC (n, y) = trace(adC (n) ◦ adC (y)) = 0.
Eventually, Proposition 5.16 implies n = 0, a contradiction, q.e.d.
Consider a pair (L, T ) with a complex semisimple Lie algebra L and a maximal
toral subalgebra T ∈ L. Theorem 5.17 allows to replace in the root space decom-
position of L from Definition 5.12 the eigenspace L0 = CL (T ) by T . Moreover,
Corollary 5.14 allows to choose from each pair (α, −α) of roots of L exactly one
member. The set of chosen members is denoted Φ + ⊂ Φ. How to make a canonical
choice will be explained in Chapter 6 and 7.
(Lα ⊕ L−α )
M M
L=T⊕ Lα = T ⊕
α∈Φ α∈Φ +
Here we have applied Corollary 5.14: For any root α ∈ Φ also (−α) ∈ Φ. Hence
we may choose from each pair (α, −α) of roots of L exactly one member. The set
of chosen members is denoted by Φ + ⊂ Φ. How to make a canonical choice will be
explained in Chapter 6 and 7.
T = L0
x = xT + ∑ xα , xT ∈ T, xα ∈ Lα .
α∈Φ
For all h ∈ T
xα = 0.
NL (T ) = T, q.e.d.
In the following we will employ the standard notation H for Cartan subalgebras,
and denote by the letter H a fixed toral subalgebra of a semisimple Lie algebra L,
i.e. a fixed Cartan subalgebra of L.
Chapter 6
Root systems
The base field in the present chapter is R, all vector spaces are real and finite
dimensional.
The ambient space of an abstract root system is a real vector space V . One may con-
ceive of V as the vector space spanned by the root set of a semisimple Lie algebra L.
The root set is a distinguished set of linear functionals on the Cartan algebra H ⊂ L.
When following this intuition roots are conceived as linear functionals.
σα : V → V
with
155
156 6 Root systems
1. σα (α) = −α
σα (x) = x − α ∗ (x) · α :
If
x = µ1 · α + µ2 · v, v ∈ Hα ,
then
σα (x) = −µ1 · α + µ2 · v = x − 2µ1 · α.
Hence
α ∗ (x) = −2 · µ1 .
The linear functional α ∗ satisfies α ∗ (α) = 2 and ker α ∗ = Hα .
σα (x) := x − α ∗ (x) · α, x ∈ V,
In the following we define a root system Φ. It has the decisive property that for
all α ∈ Φ the linear functional α ∗ ∈ V ∗ takes on integral values on all elements β ∈ Φ.
• (R3) Cartan integers: For all α, β ∈ Φ, integer values < β , α >∈ Z exist with
The integers
< β , α >∈ Z, α, β ∈ Φ,
are named the Cartan integers of Φ.
• (R4) Reducedness: For each α ∈ Φ the only roots proportional to α are α itself
and −α, i.e.
(R · α) ∩ Φ = {α, −α}.
The dimension of V is the rank of the root system, the elements of Φ are the roots
of V .
Note that the Cartan integers < β , α > are linear in the first argument β but not
necessarily in the second argument α. For any α ∈ Φ:
< α, α >= 2.
σi (Φ) ⊂ Φ, i = 1, 2.
u := σ2 ◦ σ1 .
It satisfies
u(α) = α and u(Φ) = Φ.
Moreover, u induces the identity on the quotient V /Rα because
u(x) = σ2 (σ1 (x)) = σ2 (x − α1∗ (x) · α) = x − α2∗ (x) · α − α1∗ (x) · α + α2∗ (α1∗ (x) · α) · α
implies
u(x) ∈ x + Rα.
As a consequence, u has the single eigenvalue λ = 1. Hence the characteristic poly-
nomial pchar (T ) of u has the only root λ = 1. The minimal polynomial pmin (T )
158 6 Root systems
(u|Φ)n = id|Φ.
pmin (T ) = T − 1
Definition 6.4 (Weyl group). The Weyl group W of a root system Φ of V is the
subgroup of GL(V ) generated by all symmetries σα , α ∈ Φ.
The Weyl group permutes the elements of the finite set Φ, hence W is a finite
group. As a consequence, one may average an arbitrary scalar product over the ele-
ments of the Weyl group, obtaining an invariant scalar product with respect to W .
Lemma 6.5 (Invariant scalar product). Let Φ be a root system of a vector space V .
Then a scalar product (−, −) exists on V which is invariant under the Weyl group W
of Φ. With respect to any invariant scalar product the Cartan integers satisfy
(β , α)
< β , α >= 2 · , α, β ∈ Φ.
(α, α)
Proof. Take an arbitrary scalar product B on V and define the bilinear form (−, −)
as
(x, y) := ∑ B(w(x), w(y)), x, y ∈ V,
w∈W
as the average over the Weyl group, which is a finite group. Apparently, (−, −)
is positive definite, hence a scalar product. By construction, the scalar product
is W -invariant. For two roots α, β ∈ Φ the corresponding Cartan integer < β , α >
is defined by the equation
σα (β ) = β − < β , α > α.
Because
σα = σα−1 , σα (α) = −α,
and due to the invariance of the scalar product:
which proves
(α, β ) (β , α)
< β , α >= 2 · = 2· , q.e.d.
(α, α) (α, α)
Now that we have an Euclidean space (V, (−, −)) of the root system Φ, we can
define the length of a root and the angle between two roots. We choose a fixed
invariant scalar product.
Definition 6.6 (Length of roots and angle between roots). Consider a root system Φ
and the corresponding Euclidean vector space (V, (−, −)).
1. The length of a root α ∈ Φ is defined as
p
∥α∥ := (α, α).
2. The angle
0 ≤ θ := ∢(α, β ) ≤ π
included between two roots α, β ∈ Φ is defined as
(α, β )
cos(θ ) := .
∥α∥ · ∥β ∥
Lemma 6.7 (Possible angles and length ratio of two roots). Consider a root sys-
tem Φ and two non proportional roots α, β ∈ Φ. Table 6.1 displays the only possible
angles ∢(α, β ) included by α and β and their ratios of length.
The last column of the table indicates the type of the root system of rank = 2 with
base ∆ = {α, β }, see Definition 6.10 and Theorem 6.26.
Proof. Employing the Cartan integers < α, β >, < β , α > ∈ Z we have
(α, β ) (β , α)
4 > 4 · cos2 θ = 2 · 2
·2· = < α, β > · < β , α > ≥ 0.
∥β ∥ ∥α∥2
Both Cartan integers have equal sign and we may assume ∥β ∥ ≥ ∥α∥, i.e.
because
∥β ∥2 < β,α > | < β,α > |
2
= = , q.e.d.
∥α∥ < α, β > | < α, β > |
The formulas from the proof of Lemma 6.7 indicate : For two non-propotional
roots α, β ∈ Φ the product of their Cartan integers determines the included angle θ = ∢(α, β ),
while the quotient of the Cartan integers determines their length ratio ∥β ∥/∥α∥.
Moreover, Table 6.1 shows that the angle determines the length ratio with the only
exception of the orthogonal case θ = π/2.
• Rank = 2: Figure 6.1 displays all root systems of rank = 2, see also Theorem 6.26.
Lemma 6.9 (Roots with acute angle). If two non-proportional roots α, β ∈ Φ in-
clude an acute angle, i.e. (α, β ) > 0, then also α − β ∈ Φ.
Proof. According to Table 6.1 the assumption (α, β ) > 0 implies
Definition 6.10 (Base of a root system, positive and negative roots). Consider a
root system Φ of a vector space V .
Theorem 6.11 (Existence of a base). Every root system Φ of a vector space V has
a base ∆ .
Proof. The construction of a candidate for ∆ is straightforward. But the proof that
the candidate is indeed a base, will take several steps.
α = α1 + α2 with α1 , α2 ∈ Φt+
∆ := {α ⊂ Φt+ : α indecomposable}
is a base of Φ.
ii) Representation of elements from Φt+ : Each β ∈ Φt+ has the form
β= ∑ kα · α
α∈∆
with all kα ≥ 0: Otherwise consider the non-empty subset C ⊂ Φt+ of all elements
which lack such a representation.
Choose an element β ∈ C with t(β ) > 0 minimal. By construction β is decom-
posable. Hence
β = β1 + β2
with β1 , β2 ∈ Φ + and β1 ∈ C or β2 ∈ C. We get
6.1 Abstract root system 163
which implies
0 < t(β1 ),t(β2 ) < t(β ),
a contradiction to the minimality of β .
iii) Angle between elements from ∆ : We claim that two different roots α ̸= β
from ∆ are either orthogonal or include an obtuse angle, i.e. (α, β ) ≤ 0.
Otherwise (α, β ) > 0 and we obtain from Lemma 6.9 the root
γ := α − β ∈ Φ.
0= ∑ nα · α
α∈A
∑ kα · α = ∑ kα · α =: v ∈ V
α∈A1 α∈A2
(v, v) = ∑ kα · kβ · (α, β ) ≤ 0.
α∈A1 ,β ∈A2
Hence v = 0. Now
0 = t(v) = ∑ kα · t(α)
α∈A1
with t(α) > 0 for all α ∈ A1 implies kα = 0 for all α ∈ A1 . Similarly kα = 0 for
all α ∈ A2 .
The sequence of all steps i) until iv) proves the claim of the theorem, q.e.d.
164 6 Root systems
The proof of Theorem 6.11 constructs a base ∆ by starting from a certain func-
tional t ∈ V ∗ . We show that any base can be obtained in this way: Any base of a root
system Φ is the set of indecomposable elements of Φ from the positive half-space
of a suitable linear functional.
Lemma 6.12 (Base and a determining linear functional). Consider a base ∆ of a
root system Φ of a vector space V .
and
∆ = {α ∈ Φt+ : α indecomposable}.
Proof. The base ∆ defines the decomposition
˙ −.
Φ = Φ + ∪Φ
Because (α)α∈∆ is a basis of the vector space V , a functional t ∈ V ∗ exists with t(α) > 0
for all α ∈ ∆ . Set
From
∆ ⊂ Φt+
follows
Φ + ⊂ Φt+ and Φ − ⊂ Φt− .
And the decomposition
˙ − = Φ = Φt+ ∪Φ
Φ + ∪Φ ˙ t−
implies
Φ + = Φt+ and Φ − = Φt− .
As a consequence, the indecomposable elements from both sets Φ + and Φt+ are
equal, i.e. ∆ = ∆ ′ , q.e.d.
Part iii) in the proof of Theorem 6.11 shows (α, β ) ≤ 0 because all elements from ∆
belong to the same half-space.
6.2 Action of the Weyl group 165
Our aim in the present chapter is the classification of all possible root systems. This
result derives from two facts.
The first is the integrality of the Cartan integers of a root system Φ of a real
vector space V . This fact restricts the angle and the relative length of two roots, see
Lemma 6.7. The second fact are the different group operations of the Weyl group, in
particular the transitive operation on the set of bases of a given root system. Besides
the action of the Weyl group W on Φ
W × Φ → Φ, (w, α) 7→ w(α),
G×X →
− X, (g, x) 7→ g.x
and
(g1 · g2 ).x = g1 .(g2 .x) for all g1 , g2 ∈ G and all x ∈ X.
The group acts transitive if for all x ∈ X the induced map
G→
− X, g 7→ g.x,
is surjective.
Lemma 6.15 (Action of the Weyl group on Cartan integers and on symmetries).
Consider a root system Φ of a vector space V with Weyl group W .
1. The Cartan integers are linear in the first argument, i.e. for all α ∈ Φ the Cartan
functional
< −, α >: V → − C
is C-linear.
166 6 Root systems
2. Denote by
A := {< −, α >∈ V ∗ : α ∈ Φ}
the set of all Cartan functionals. The map
is a group operation.
3. Denote by
B := {sα : V →
− V |α ∈ Φ}
the set of all symmetries. The map
− B, (w, σα ) 7→ w ◦ σα ◦ w−1 ,
W ×B →
is a group action.
4. Denote by
C := {∆ ⊂ Φ : ∆ a base of Φ}
the set of all bases of Φ. The map
W ×C →
− C, (w, ∆ ) 7→ {w(α) : α ∈ ∆ },
is a group action.
Proof. 1. According to Lemma 6.5 a scalar product (−, −) on V exists which is
invariant with respect to W . The formula
(−, α)
< −, α >= 2 · , α ∈ Φ,
(α, α)
Hence the map W × A → − A is well-defined. One checks that it satisfies the two
properties of a group action.
sβ = w ◦ σα ◦ w−1 :
6.2 Action of the Weyl group 167
σβ (v) = v− < v, β > ·β = (w ◦ w−1 )(v)− < w−1 (v), α > ·w(α) =
(w ◦ σα ◦ w−1 )(v) = w(σα (w−1 (v))) = w(w−1 (v)− < w−1 (v), α > ·α).
Hence the map W × B → − B is well-defined. One checks that it satisfies the two
properties of a group action.
In accordance with Lemma 6.15 about the action on the Cartan integers one defines
the action of W on the dual space V ∗ :
w(∆ ) = ∆ ′ .
3. Any root extends to a base: For any root α ∈ Φ an element w ∈ W exists with
w(α) ∈ ∆ .
i) The symmetry σα of any root α ∈ ∆ leaves the set Φ + \ {α} invariant: Assume
that ∆ comprises at least two roots. Consider an element β ∈ Φ + \ {α}. It has a
representation
β = ∑ kγ · γ, kγ ≥ 0, for all γ ∈ ∆ .
γ∈∆
Hence also σα (β ) is not proportional to α and has at least one coefficient kγ > 0.
According to Definition 6.10, part i), all coefficients are non-negative and
σα (β ) ∈ Φ + \ {α}.
ii) Consider the distinguished element ρ ∈ V defined as half the sum of all posi-
tive roots
ρ := (1/2) · ∑ β .
β ∈Φ +
σα (ρ) = ρ − α.
We now show that the first three claims can be satisfied already by taking ele-
ments from W∆ .
α = ρ − σα (ρ)
or
t(w(α)) ≥ 0.
6.2 Action of the Weyl group 169
iv) Also part 2) of the proposition is satisfied even with an element w ∈ W∆ : First,
Lemma 6.12 characterizes the base ∆ ′ by a functional t ′ ∈ V ∗ with
t ′ (α ′ ) > 0
t := t ′ ◦ w ∈ V ∗
t ′ (w(α)) = 0,
|t(β )| > ε
but
t(α) = ε.
Denote by ∆ (t) the base of Φ induced by t according to Lemma 6.12. By part iv)
an element w ∈ W∆ exists with
w(∆ (t)) = ∆ .
σα = w−1 ◦ σβ ◦ w ∈ W∆ , q.e.d.
Definition 6.17 (Cartan matrix). Consider a root system Φ of a vector space V and
a base ∆ = {α1 , ..., αr } of Φ.
The Cartan matrix of ∆ is the matrix of the Cartan integers of the roots from ∆
Here the index i denotes the row and the index j denotes the column.
Note that the Cartan matrix is not necessarily symmetric. All diagonal elements
of the Cartan matrix have the value
< αi , αi >= 2.
The Cartan matrix is defined with reference to a base ∆ and with reference to
a numbering of its elements. Lemma 6.18 shows that for any two bases of Φ any
numbering of the elements of the first base induces a numbering of the elements of
the second base such that the respective Cartan matrices are equal.
Lemma 6.18 (Independence of the Cartan matrix from the choosen base). Con-
sider a root system Φ of rank r = rank Φ. Any two bases ∆ , ∆ ′ of Φ have the same
Cartan matrix. More specifically:
Cartan(∆ ) = (< αi , α j >1≤i, j≤r ) = Cartan(∆ ′ ) = (< w(αi ), w(α j ) >1≤i, j≤r ).
6.2 Action of the Weyl group 171
which implies
< w(αi ), w(α j ) >=< w−1 (w(αi )), α j >=< αi , α j >, q.e.d.
The Cartan matrix encodes the full information of the root system Φ, notably the
dimension of its ambient space V :
Theorem 6.19 (The Cartan matrix characterizes the root system). Consider a
root system Φ of an Euclidean space V and a base
∆ = {α1 , ..., αr }
f : ∆ → ∆′
exists with
F : V → V′
exists with
F(Φ) = Φ ′ and F|∆ = f .
Proof. i) Determination of F: Because the elements from ∆ form a basis of the
vector space V we may define
F : V → V′
as the uniquely determined linear extension of f . And because ∆ and ∆ ′ have the
same cardinality the linear map F is an isomorphism.
ii) We show that the Weyl groups W and W ′ are conjugate via F, i.e.
W ′ = F ◦ W ◦ F −1 or W ′ ◦ F = F ◦ W :
172 6 Root systems
σ f (α) ◦ F = F ◦ σα .
W ′ ◦F = F ◦W .
β = w(α).
Due to Theorem 6.19 a root system Φ is completely determined by its Cartan matrix
with respect to a base ∆ of Φ. Hence the classification of root systems reduces to the
classification of possible Cartan matrices. Thanks to Lemma 6.7 and Corollary 6.13
the Cartan integers satisfy a set of restrictions. The present section shows that the
language of Cartan matrices translates to a data structure from Discrete Mathemat-
ics. The data structure is a weighted graph called the Coxeter graph of the root
system.
Therefore, we first define the Coxeter graph of ∆ , and then classify all possible
Coxeter graphs. This classification is achieved by a second language translation: We
translate the language of Coxeter graphs to the language of Euclidean vector spaces,
and apply same basic results from Linear Algebra. The result is Theorem 6.21.
6.3 Coxeter graph and Dynkin diagram 173
(α, β )2
< α, β >< β , α >= 4 · .
∥α∥2 · ∥β ∥2
Definition 6.20 (Coxeter graph). Consider a root system Φ of a vector space V and
a base ∆ of Φ. The Coxeter graph of Φ with repect to ∆ is the undirected weighted
graph
Coxeter(Φ) = (N, E)
with
• vertex set N := ∆
2π
• Series Dr , r ≥ 4: The first r − 3 subsequent pairs of roots include the angle 3 ,
root αr−2 includes with each of the two roots αr−1 and αr the angle 2π 3 .
5π
• Exceptional type G2 : ∆ = (α1 , α2 ). The two roots include the angle 6 .
• Exceptional type F4 : ∆ = (α1 , ..., α4 ). The pairs (α1 , α2 ) and (α3 , α4 ) include
the angle 2π 3π
3 , the pair (α2 , α3 ) includes the angle 4 .
174 6 Root systems
The range of r ∈ N for the types Ar , Br ,Cr , Dr has been choosen to avoid dublets
with low value of r.
Note. The Coxeter graph employing the product of Cartan integers as its weights
does not encode the relative length of two roots, i.e. their length ratio. The length
ratio derives from the quotient of the Cartan integers
2
∥β ∥ < β,α >
= .
∥α∥ < α, β >
Knowing the product of the Cartan integers is not sufficient to reconstruct the Car-
tan matrix. Hence the Coxeter graph does not encode the full information about
the root system. We will see that a base of the root systems belonging to the
types Br ,Cr , G2 , F4 is made up by roots with different lengths. As a consequence,
after complementing the Coxeter graph by the information about the length ratio the
series Br and Cr of root systems will differ for r ≥ 3, see Theorem 6.26.
6.3 Coxeter graph and Dynkin diagram 175
But in any case, the (absolute) length of a root is not defined because an invariant
scalar product of a root system is not uniquely determined.
For the proof of Theorem 6.21 we will capture the characteristic properties of
the Coxeter graph in the following definition which embeds the graph into a finite-
dimensional Euclidean space.
Definition 6.22 (Admissible graph). Consider an Euclidean space (V, (−, −)). An
undirected weighted graph (N, E) with vertex set N ⊂ V is admissible if
• its vertex set satisfies
N = {v1 , ..., vr } ⊂ V
with a linearly independent family (vi )i=1,...,r of unit vectors satisfying
(vi , v j ) ≤ 0 for 1 ≤ i ̸= j ≤ r
Proof (of Theorem 6.21). The proof will classify all connected admissible graphs (N, E)
arising from root systems Φ of rank r in an Euclidean space (V, (−, −)).
ii) An admissible graph has less edges than vertices, i.e. |E| < |N|:
follows
2 · (vi , v j ) ≤ −1,
176 6 Root systems
therefore
0 < r − |E| i.e. |E| < |N|.
iii) An admissible graph has no cycles: A cycle is an admissible graph according
to part i). But for a cycle the number of vertices equalizes the number of edges, in
contradiction to part ii).
iv) For any vertex of an admissible graph the weighted sum of incident edges is at
most = 3:
Denote by
Inc(v) := {(e, m) ∈ E : e incident with v}
the set of weighted edges incident to a vertex v. We have to show
3≥ ∑ m.
(e,m)∈Inc(v)
Denote by {w1 , ..., wk } the set of vertices adjacent to v. Because an admissible graph
(wi , w j ) = 0
for 1 ≤ i ̸= j ≤ k, i.e. the family (wi )i=1,...,k is orthonormal in (V, (−, −)). The family
(v, w1 , ..., wk )
k
v = ∑ (v, wi ) · wi .
i=0
6.3 Coxeter graph and Dynkin diagram 177
The case (v, w0 ) = 0 is excluded, because v is linearly independent from (wi )i=1,...,k .
Hence
(v, w0 ) ̸= 0
and we obtain
k
1 = (v, v) = ∑ (v, wi )2
i=0
and
k
∑ (v, wi )2 < 1.
i=1
As a consequence
k
∑ m = ∑ 4 · (v, wi )2 < 4.
(e,m)∈Inc(v) i=1
v) Blowing down a simple path, i.e. all its edges have weight = 1, results in a new
admissible graph (N ′ , E ′ ):
Denote by C = {w1 , ..., wn } ⊂ N the vertices of the path. By assumption for i = 1, ..., n − 1
The graph (N ′ , E ′ ), resulting from blowing down the original path, has
• vertex set
N ′ = (N \C) ∪ {w0 }, w0 := ∑ w,
w∈C
n−1
n + 2 · ∑ (wi , wi+1 ) = n − (n − 1) = 1.
i=1
In (N, E) any vertex w from N \ C is adjacent to at most one vertex from the path,
because an admissible graph is cycle free according to part iii). Hence
• either (w, w0 ) = 0
• or exactly one index i = 1, ..., n exists with 0 ̸= (w, wi ).
In either case holds
4 · (w, w0 )2 ∈ {0, 1, 2, 3}.
vi) A connected admissible graph does not contain any subgraph from Figure 6.5:
c) One edge with weight = 2 and one vertex with three adjacent vertices.
In any of the subgraphs b)-d) it would be possible to blow down a path to a vertex
with weighted sum of incident edges ≥ 4, which contradicts part v) and iv).
vii) Any conncected admissible graph belongs to one of the types from Figure 6.6:
d) All edges have weight = 1, a single vertex has 3 incident edges, each with
weight = 1.
The result follows from excluding the subgraphs from part vi).
viii) All connected admissible graphs from type vii,a) belong to series Ar . Obvious.
ix) All connected admissible graphs from type vii,b) belong to series Br or F4 :
2 · (ui , ui+1 ) = −1
we compute
p p−1
(u, u) = ∑ i · j · (ui , u j ) = ∑ i2 · (ui , ui ) + 2 · ∑ i(i + 1) · (ui , ui+1 ) =
1≤i, j≤p i=1 i=1
180 6 Root systems
(p, q) = (1, ≥ 2), (p, q) = (≥ 2, 1), (p, q) = (2, 2), (p, q) = (1, 1).
The first two give the same Coxeter graph. It has type Br , r ≥ 3, equal to type Cr .
The third possibility gives F4 . The fourth possibility is type B2 .
xi) All connected admissible graphs having type vii,d) belong to series Dr , r ≥ 4,
or series Er , r ∈ {6, 7, 8}:
Note r, p, q, ≥ 2. The three vectors (u, v, w) are pairwise orthogonal and the four vec-
tors (u, v, w, x) are linearly independent. Denote by θ1 , θ2 , θ3 the respective angles
between the vectors x and u, v, w. Similar to the proof of part iv) we obtain
3
1 > ∑ cos2 θi .
i=1
or
1/r + 1/p + 1/q > 1.
W.l.o.g we may assume
q≤ p≤r
and q ≥ 2 because q = 1 reproduces type An . Hence
which implies
3 > q ≥ 2, i.e. q = 2.
We obtain
1 < 1/r + 1/p + 1/2, i.e. 1/2 < 1/r + 1/p,
and
1/2 < 2/p,
hence
2 ≤ p < 4.
In case p = 3 we have r < 6. In case p = 2 the parameter r may have any value ≥ 2.
Summing up: When r ≥ p ≥ q then the only possibilities for (r, p, q) are
6.3 Coxeter graph and Dynkin diagram 183
˙ 2 , Φ1 ̸= 0,
Φ = Φ1 ∪Φ / Φ2 ̸= 0,
/
exists with
(Φ1 , Φ2 ) = 0.
Otherwise Φ is irreducible.
2. Analogously defined are the terms reducible and irreducible for a base ∆ of Φ.
˙ 2 ; Φ1 ̸= 0,
Φ = Φ1 ∪Φ / Φ2 ̸= 0.
/
Define ∆i := Φi ∩ ∆ , i = 1, 2. Then
˙ 2
∆ = ∆1 ∪∆
and (∆1 , ∆2 ) = 0.
(Φ1 , ∆2 ) ⊂ (Φ1 , Φ2 ) = 0.
(Φ1 ,V ) = (Φ1 , ∆2 ) = 0.
Therefore
184 6 Root systems
Φ1 = 0,
which is excluded. As a consequence:
∆1 ̸= 0/ and similarly ∆2 ̸= 0.
/
The decomposition
˙ 2
∆ = ∆1 ∪∆
proves the reducibility of ∆ .
˙ 2.
∆ = ∆1 ∪∆
Φi := W (∆i ), i = 1, 2.
According to Lemma 6.16 any root β ∈ Φ has the form β = w(α) for suitable w ∈ W
and α ∈ ∆ . Therefore
Φ = Φ1 ∪ Φ2 .
The Weyl group is generated by the symmetries σα , α ∈ ∆ . Explicit calculation
shows:
• The orthogonality (∆1 , ∆2 ) = 0 implies that for two roots αi ∈ ∆i , i = 1, 2, the
corresponding symmetries σα1 and σα2 commute.
• If α1 ∈ ∆1 then for a root α ∈ span ∆1 also σα1 (α) ∈ span ∆1 .
• If α2 ∈ ∆2 then σα2 (α) = α for any root α ∈ span ∆1 .
As a consequence W (∆1 ) ⊂ span ∆1 and similarly W (∆2 ) ⊂ span ∆2 . The orthog-
onality (∆1 , ∆2 ) = 0 implies the orthogonality
(Φ1 , Φ2 ) = 0, notably Φ1 ∩ Φ2 = 0.
/
The Dynkin diagram of a root system complements the Coxeter graph by the
information about the length ratio of the elements from a base. This information can
be encoded by an orientation of the edges pointing from the long root to the short
root in case two non-orthogonal roots have different length.
6.3 Coxeter graph and Dynkin diagram 185
Definition 6.25 (Dynkin diagram of a root system). Consider a root system Φ and
its Coxeter graph Coxeter(Φ) = (N, E). The Dynkin diagram of Φ is the directed
weighted graph
Dynkin(Φ) := (N, ED )
with
• Vertex set: N = ∆ .
Note: The Coxeter graph and the Dynkin diagram of a root system have the same
set of vertices. In the Dynkin diagram an oriented edge
e = (α, β )
is an ordered pair; the edge is considered oriented from the long root α to the short
root β , ∥α∥ ≥ ∥β ∥. If both roots have the same length, then also the edge with
the inverse orientation belongs to the Dynkin diagram. In this case the two ori-
ented edges from the Dynkin diagram are equivalent to the corresponding single
undirected edge from the Coxeter graph. Therefore, often the figure of the Dynkin
diagram suppresses the arrows of two opposite orientations.
• Series Br , r ≥ 2
• Series Cr , r ≥ 3
• Series Dr , r ≥ 4
• Exceptional type G2
• Exceptional type F4
Notably, the two types Br and Cr are distinguished by the length ratio of their
roots.
2. If one dismisses the restriction of the rank r then one has the following repetitions
of low rank:
C1 = B1 = A1 and C2 = B2 and D3 = A3 .
Proof. The statement follows from the classification of Coxeter graphs according
to Theorem 6.21 and the restriction of the length ratio of two roots from a base
6.3 Coxeter graph and Dynkin diagram 187
according to Lemma 6.7: A weighted edge (e, m) from the Coxeter graph links two
roots with length ratio ̸= 1 if and only if m ∈ {2, 3}. Therefore the Dynkin diagrams
distinguish between the two series Br and Cr , q.e.d.
Dynkin(Φ) = (N, ED ).
< α, α >:= 2.
< α, β >:= 0.
Remember that an edge without orientation from the figure of a Dynkin diagram
replaces two oriented edges of the diagram.
Proof. The proof evaluates for each pair of roots the relation between the included
angle and the length ratio of the roots, see Lemma 6.7, q.e.d.
Chapter 7
Classification
The objective of the present chapter is to classify complex semisimple Lie algebras
by their Dynkin diagram, more precisely the Dynkin diagram of its root system.
The result is one of the highlights of Lie algebra theory. It completely encodes the
structure of these Lie algebras by a certain finite graph, a data structure from discrete
mathematics.
We consider a fixed pair (L, H) with a complex semisimple Lie algebra L and a
Cartan subalgebra H ⊂ L. We denote by Φ the root set corresponding to the Cartan
decomposition of (L, H) M
L=H⊕ Lα .
α∈Φ
From the Cartan criterion we know that a complex Lie algebra L is semisimple if
and only if its Killing form is non-degenerate. We even know from Proposition 5.16:
Also the restriction of the Killing form to a Cartan subalgebra H
κH := κ|H × H → C
is non-degenerate.
189
190 7 Classification
The map
∼
→ H ∗ ,tλ 7→ λ ,
H−
is a linear isomorphism of complex vector spaces. In particular, for each root α ∈ Φ ⊂ H ∗
of L its inverse image under this isomorphism is denoted tα ∈ H, i.e.
α = κH (tα , −).
This formula relates roots α ∈ Φ, being elements from the dual space H ∗ , to ele-
ments tα ∈ H of the Cartan subalgebra.
The objective of this section is to show that the root set Φ of (L, H) satisfies the
axioms for a root system of a real vector space V in the sense of Definition 6.2. We
consider the real vector space
V := spanR Φ.
We will show that the root set Φ of (L, H) has the following properties:
/ Φ, dimR V = dimC H ∗ .
• (R1) Finite and generating: |Φ| < ∞, 0 ∈
• (R3) Cartan integers: The coefficients < β , α >, α, β ∈ Φ, from the representa-
tion
σα (β ) = β − < β , α > ·α
are integers, i.e. < β , α >∈ Z.
7.1 The root system of a complex semisimple Lie algebra 191
To motivate the separate steps in deriving the root system of the Lie algebra L we
first consider the example of sl(3, C).
Example 7.2 (The root system of sl(3, C)).
Set L := sl(3, C).
i) Cartan subalgebra: The subalgebra
(Lα ⊕ L−α ).
M
L=H⊕
α∈{α1 ,α2 ,α3 }
All root spaces are 1-dimensional. The root set Φ = {±α1 , ±α2 , ±α3 } satisfies:
• Lα1 = span < x1 := E12 >, L−α 1 = span < y1 := E21 >
α1 : H → C, α1 (h1 ) = 2, α1 (h2 ) = −1
• Lα2 = span < x2 := E23 >, L−α 2 = span < y2 := E32 >
• Lα3 = span < x3 := E13 >, L−α 3 = span < y3 := E31 >
α3 : H → C, α3 (h1 ) = 1, α3 (h2 ) = 1
Therefore α3 = α1 + α2 .
iii) Killing form: To compute the Killing form κ and its restriction
κH : H × H → C
Therefore
2 −1
(κH (hi , h j )1≤i, j≤2 ) = 6 · .
−1 2
192 7 Classification
we get
tα1 = (1/6) · h1 ,tα2 = (1/6) · h2 .
The family (α1 , α2 ) is linearly independent and a basis of H ∗ . According to
Definition 7.1, on H ∗ the induced bilinear form
(−, −) : H ∗ × H ∗ → C
Apparently
dimC V = dimC H ∗ = 2.
The restriction of the bilinear form (−, −) to V is a real positive definite form, hence
a scalar product. We now prove by explicit computation: The root set Φ of (L, H) is
a root system of the real vector space V in the sense of Definition 6.2.
v) Distinguished symmetries: Consider the Euclidean space (V, (−, −)). For
for i = 1, 2, 3 define the maps
σi : V → V, x 7→ x − 6 · (x, αi ) · αi .
σi (αi ) = −αi .
Moreover
σ2 (α1 ) = α3 , σ2 (α3 ) = α1
7.1 The root system of a complex semisimple Lie algebra 193
vi) Cartan integers: From the symmetries of part v) one reads off the Cartan
numbers
viii) Base: Apparently the set ∆ := {α1 , α2 } is a base of Φ. The Cartan matrix with
respect to ∆ and the given numbering is
2 −1
Cartan(∆ ) = .
−1 2
According to Lemma 6.7 both roots from ∆ have the same lenght and include the
angle (2/3)π. Therefore Coxeter graph and Dynkin diagram of Φ contain the same
information. According to the classification from Theorem 6.26 the root system
of sl(3, C) has type A2 . The scalar product (−, −) induced from the Killing form is
invariant with respect to the Weyl group W = span < σα1 , σα2 >.
The following two propositions collect the main properties of the root set Φ of a
semisimple complex Lie algebra L. They generalize the result from Proposition 5.4
about the structure of sl(2, C).
Proposition 7.3 considers the complex linear structure of L and its canonical
subalgebras sl(2, C) ⊂ L. While Proposition 7.4 considers the integrality proper-
ties of the root set Φ with respect to the bilinear form induced by the Killing form,
see Definition 7.1. These properties assure that Φ is a root system in the sense of
194 7 Classification
Definition 6.2. They allow to apply the classification of respectively Coxeter graphs
from Theorem 6.21 and Dynkin diagrams from Theorem 6.26 to obtain the classifi-
cation of complex semisimple Lie algebras in Proposition 7.7 - 7.10 and Section 7.3.
Recall from Corollary 5.14 that for any root α ∈ Φ of L also the negative −α is
a root of L.
Proposition 7.3 (Semisimple Lie algebras as sl(2, C)-modules). Consider a pair (L, H)
with L a complex semisimple Lie algebra and H ⊂ L a Cartan subalgebra. Denote
by (−, −) the nondegenerate bilinear form on H ∗ derived from the restricted Killing
form κH on H according to Definition 7.1. The root set Φ of the Cartan decomposi-
tion M
L=H⊕ Lα
α∈Φ
2. Duality of root spaces: For each α ∈ Φ the vector spaces Lα and L−α are dual
with respect to the Killing form, i.e. the bilinear map
[x, y] = κ(x, y) · tα
[Lα , L−α ] = C · tα .
α : [Lα , L−α ] →
− C, [x, y] 7→ κ(x, y),
[xα , yα ] = hα .
[hα , xα ] = 2 · xα , [hα , yα ] = −2 · yα .
7.1 The root system of a complex semisimple Lie algebra 195
h ̸= 0 and α(h) = 0
for all α ∈ Φ. We obtain [h, Lα ] = 0 for all α ∈ Φ. In addition [h, H] = 0 because the
Cartan subalgebra H is Abelian, see Proposition 5.3. As a consequence [h, L] = 0,
i.e.
h ∈ Z(L).
We obtain h = 0 because the center of the semisimple Lie algebra L is trivial. This
contradiction proves span Φ = H ∗ .
In order to prove the remaining issues we recall from Lemma 5.13 for two linear
functionals λ , µ ∈ H ∗ with λ + µ ̸= 0 the orthogonality
κ(Lλ , Lµ ) = 0.
ad 2)
is injective. Interchanging the role of the two roots α and −α and using (Lα )∗∗ ≃ Lα
shows that the map is also surjective.
• We have
[Lα , L−α ] ⊂ L0 = H
according to Lemma 5.13. The Killing form is associative according to Lemma 4.1.
Hence for all h ∈ H, x ∈ Lα , y ∈ L−α
[x, y] = κ(x, y) · tα .
→ gl(L)
ad : L ,−
S ≃ ad(S) ⊂ gl(L).
According to Lie’s theorem, see Theorem 3.21, the solvable subalgebra ad(S)
embeds into the subalgebra of upper triangular matrices. The element
Hence
2
hα = · tα .
(α, α)
• For an arbitrary element xα ∈ Lα , xα ̸= 0, according to part 2) a unique
element yα ∈ L−α exists with
[xα , yα ] = hα .
We obtain
[hα , xα ] = α(hα ) · xα = 2 · xα
[h, yα ] = −α(hα ) · yα = −2 · yα
[xα , yα ] = hα .
Therefore the subalgebra Sα :=< xα , yα , hα > is isomorphic to sl(2, C), q.e.d.
It derives from the relation between hα and tα from Proposition 7.3 and from the
defining relation κH (tβ , −) = β . The numbers β (hα ) will turn out as the Cartan
integers < β , α > of the root system. Notably they are integers.
198 7 Classification
Proposition 7.4 (Integrality and rationality properties of the root set). Consider
a pair (L, H) with L a complex semisimple Lie algebra and H ⊂ L a Cartan subal-
gebra. The root set Φ from the Cartan decomposition
M
L=H⊕ Lα
α∈Φ
3. Rationality and scalar product: Consider a basis B = (α1 , ..., αr ) of the complex
vector space H ∗ made up from roots αi ∈ Φ, i = 1, ..., r. Then any root β ∈ Φ is
a rational combination of elements from B, i.e.
and
dimQ VQ = dimC H.
The bilinear form (−, −) restricts from H ∗ to a rational form on the subspace VQ ⊂ H ∗
and its restriction
(−, −)Q : VQ ×VQ → Q
is positive definite, i.e. a scalar product.
V := R ⊗ QVQ = spanR Φ ⊂ H ∗
and the scalar product (−, −)R on V . For each root α ∈ Φ the map
2 · (β , α)R
σα : V → V, β 7→ σα (β ) := β − · α,
(α, α)R
β − p · α ∈ Φ and β + q · α ∈ Φ.
β + k · α ∈ H ∗ , −p ≤ k ≤ q,
are roots.
[Lα , Lβ ] = Lα+β .
Proof. ad 1) According to Proposition 7.3, part 2) the root spaces Lα and L−α are
dual with respect to the Killing form κ.
In order to show dim Lα = 1 we assume on the contrary
α(hα ) = 2.
The linear functional κ(xα , −) on L−α is non-zero. Because dim L−α > 1 the func-
tional has a non-trivial kernel, i.e. an element y ∈ L−α , y ̸= 0, exists with κ(xα , y) = 0.
The latter formula implies [xα , y] = 0, see Proposition 7.3, part 2). As a consequence y ∈ L
is a primitive element of an irreducible Sα -submodule of L with weight
(−α)(hα ) = −2 < 0.
The latter property contradicts the fact that all irreducible sl(2, C)-modules have a
non-negative highest weight, see Theorem 5.9.
200 7 Classification
z ∈ Lβ −p·α
we get β − β (hα ) · α ∈ Φ.
ad 3)
2
(α j , α j )
and applying the bilinear form (−, α j ) to the resulting equation gives a linear
system of equations
b = A·c
for the vector of indeterminates c := (c1 , ..., cr )⊤ . The left-hand side is the vector
⊤
2 · (β , α j
b=
(α j , α j )
The system is defined over the ring Z because for two roots γ, δ ∈ Φ
7.1 The root system of a complex semisimple Lie algebra 201
(γ, δ )
2· = γ(hδ ) ∈ Z
(δ , δ )
A ∈ M(r × r, Z)
by the non-zero scalar 2/(α j , α j ). And the latter matrix defines the bilinear
form (−, −) on H ∗ , which is non-degenerate because it corresponds to the
Killing form κH (−, −) on H. As a consequence, the solution of the linear system
of equations is unique and is already defined over the base field Q, i.e.
• We claim that the bilinear form (−, −)Q is rational, i.e. it is defined over the
field Q, and positive definite. For any two roots α, β ∈ Φ we have to show
In order to evaluate the trace we employ the defining property of a root space: If
z ∈ Lγ then
(ad tµ )(z) = γ(tµ ) · z
and
(ad tλ ◦ ad tµ )(z) = γ(tλ ) · γ(tµ ) · z.
Using the Cartan decomposition
M
L=H⊕ Lγ
γ∈Φ
2 · (γ, α) γ(tα )
= 2· = γ(hα ) =:< γ, α >∈ Z
(α, α) (α, α)
202 7 Classification
We obtain
In particular
and
(β , α) = (1/2)(α, α)· < β , α >∈ Q.
ad 4) Due to the characterization α(hα ) = 2 from Proposition 7.3 the map σα is a
symmetry of V with vector α. Due to part 2) its Cartan numbers are integers
2 · (β , α)
< β , α >= = β (hα ) ∈ Z.
(α, α)
The inclusion σα (Φ) ⊂ Φ has been proven in part 2). In order to prove the
invariance of the scalar product with respect to the Weyl group it suffices to
consider three roots α, β , γ ∈ Φ:
we obtain
(σα (β ), σα (γ)) = (β , γ).
ad 5) Assume on the contrary the existence of a pair of proportional roots different
from ±α. W.l.o.g. we may assume α ∈ Φ and β = t · α ∈ Φ, 0 < t < 1,t ∈ R.
From β (hα ) ∈ Z and
follows
7.1 The root system of a complex semisimple Lie algebra 203
2t ∈ Z.
Hence the two roots from any pair of proportional roots belong to a set
{±α, ±(1/2)α}, α ∈ Φ.
[hα , z] = 2α(hα ) · z = 4 · z.
Because 3α ∈
/ Φ we have xα .z = 0.
With respect to the sl(2, C)-module structure of L induced from the action of
Sα =< xα , yα , hα >,
• xα .z ∈ L3α . But 3α ∈
/ Φ, hence xα .z = 0.
m + 1 = p + q + 1.
xα : Lβ → Lβ +α , v 7→ xα .v,
is an isomorphism, q.e.d.
Theorem 7.5 (Root system of a complex simple Lie algebra). Consider a pair (L, H)
with L a complex semisimple Lie algebra and H ⊂ L a Cartan subalgebra.
is a root system in the sense of Definition 6.2 of the real vector space
V := spanR Φ ⊂ (H ∗ )R .
It is named the root system of the Lie algebra L. The root system attaches to
each α ∈ Φ the symmetry
σα : V → V, β 7→ β − β (hα ) · α
Proof. The proof collects the results from Proposition 7.4, q.e.d.
We will show in the present section that the root systems of type Ar , Br ,Cr , Dr are
the root systems of the complex Lie algebras belonging to the classical groups of
the correponding types, see Proposition 2.13. We show the simpleness of these Lie
algebras as a consequence of their Cartan decomposition. We follow [22, Chap. 7.7]
and [25, Chap. X, §3].
The Lie algebras of the classical groups are subalgebras of the Lie algebra gl(n, C).
We introduce the following notation for the elements of the canonical basis of the
vector space M(n × n, C):
7.2 Root systems of the A, B,C, D-series 205
Ei, j ∈ M(n × n, C)
is the matrix with entry = 1 at place (i, j) and entry = 0 for all other places. Our
matrix computations are based on the formulas
The family
(Ei,i )1≤i≤n
is a basis of the subspace of diagonal matrices d(n, C). Denote the elements of the
dual base by
ε̂i := Ei,i ∗ ∈ d(n, C)∗ , i = 1, ..., n.
L ⊂ sl(n, C).
We have to show rad(L) = {0}. Due to its definition the Lie algebra L contains with
any matrix also its transpose:
A ∈ L =⇒ A⊤ ∈ L.
rad(L) ⊂ t(n, C) ∩ L.
Because
rad(L)⊤ := {A⊤ : A ∈ rad(L)}
is also solvable and rad(L) is maximal solvable, we have
rad(L)⊤ ⊂ rad(L),
rad(L) ⊂ rad(L)⊤ .
As a consequence
206 7 Classification
rad(L) = rad(L)⊤
is also contained in the set of lower triangular matrices. Therefore rad(L) is con-
tained in the diagonal matrices
rad(L) ⊂ d(n, C) ∩ L.
( j0 , k0 ), 1 ≤ j0 ̸= k0 ≤ n,
we choose an element
n
B= ∑ b j,k · E j,k ∈ L, b j,k ∈ C for 1 ≤ j, k ≤ n,
j,k=1
with b j0 ,k0 ̸= 0. Such choice is possible, cf. the normal forms of the matrices from L
in [28, Ch. I.2]. Then
n n
[A, B] = ∑ ai · b j,k · [Ei,i , E j,k ] = ∑ (a j − ak )b j,k · E j,k ∈ d(n, C) ∩ L.
i, j,k=1 j,k=1
As a consequence
(a j0 − ak0 )b j0 ,k0 = 0
and a j0 = ak0 . Varying now the indices j0 ̸= k0 we obtain
a1 = ... = an
Proposition 7.7 (Typ Ar ). The Lie algebra L := sl(r +1, C), r ≥ 1, has the following
characteristics:
1. dim L = (r + 1)2 − 1.
2. The subalgebra
H := d(r + 1, C) ∩ L
is a Cartan subalgebra with dim H = r.
hi := Ei,i − Ei+1,i+1
is a basis of H.
εi − ε j , 1 ≤ i ̸= j ≤ r + 1.
Ei j , 1 ≤ i ̸= j ≤ r + 1.
αi := εi − εi+1 .
The positive roots are the elements of Φ + = {εi − ε j : i < j}. They have the
representation
j−1
εi − ε j = ∑ αk ∈ Φ + .
k=i
Sα ≃ sl(2, C)
All roots α ∈ ∆ have the same length. The only pairs of roots from ∆ , which are
not orthogonal, are
(αi , αi+1 ), i = 1, ..., r − 1.
2π
Each pair includes the angle . In particular the Dynkin diagram of the root
3
system Φ has type Ar from Theorem 6.26, see Figure 7.2.
[h, Ei, j ] = h ◦ Ei, j − Ei, j ◦ h = εi (h) · Ei, j − ε j (h) · Ei, j = (εi (h) − ε j (h)) · Ei, j
If α ̸= β and < α, β >, < β , α ≯= 0 then the symmetry of the Cartan matrix implies
< β , α, > ∥β ∥2
1= =
< α, β > ∥α∥2
Hence all simple roots have the same length. The angles between two simple roots
can be read off from the Cartan matrix. Hence the root system Φ has the Dynkin
diagram Ar from Theorem 6.26, q.e.d.
In dealing with the Lie algebra sp(2r, C) we note that X ∈ sp(2r, C) iff
7.2 Root systems of the A, B,C, D-series 209
σ · X⊤ · σ = X
with
0 1
σ := and σ −1 = −σ .
−1 0
This condition is equivalent to
A B
X=
C −A⊤
Proposition 7.8 (Typ Cr ). The Lie algebra L := sp(2r, C), r ≥ 3, has the following
characteristics:
1. dim L = r(2r + 1).
2. The subalgebra
D 0
H := ∈ L : D ∈ d(r, C)
0 −D
is a Cartan subalgebra with dim H = r.
3. The family
(hi := Ei,i − Er+i,r+i )1≤i≤r
is a basis of H.
entry < αi , α j > at position (row, column) = ( j, i). Note the distinguished entry −2
in the last row: The Cartan matrix is not symmetric.
All roots α j ∈ ∆ , 1 ≤ j ≤ r − 1 have equal length, they are the short roots. The
root αr is the long root:
√ ∥αr ∥
2= , 1 ≤ j ≤ r − 1.
∥α j ∥
The only pairs of simple roots with are not orthogonal are
2π
For i = 1, ..., r − 2 these pairs include the angle , while the pair (αr−1 , αr )
3
3π
inludes the angle . In particular the Dynkin diagram of the root system Φ has
4
type Cr according to Theorem 6.26, see Figure 7.2.
r2 − r
dim L = r2 + 2 · ( + r) = 2r2 + r = r(2r + 1).
2
4) Note εi (h) = −εr+i (h) for all h ∈ H. For h ∈ H the commutators are:
[h, Ei, j − Er+ j,r+i ] = h · (Ei, j − Er+ j,r+i ) − (Ei, j − Er+ j,r+i ) · h =
= εi (h)Ei, j − εr+ j (h)Er+ j,r+i − ε j (h)Ei, j + εr+i (h)Er+ j,r+i =
= (εi (h) − ε j (h))Ei, j − εr+ j − εr+i )Er+ j,r+i =
= (εi (h) − ε j (h))Ei, j + ε j (h) − εi (h))Er+ j,r+i =
(εi (h) − ε j (h))(Ei, j − Er+ j,r+i )
212 7 Classification
[h, Ei,r+ j + E j,r+i ] = εi (h)Ei,r+ j + ε j (h)E j,r+i − εr+ j (h)Ei,r+ j − εr+i (h)E j,r+i =
[h, Er+i, j + Er+ j,i ] = εr+i (h)Er+i, j + εr+ j (h)Er+ j,i − ε j (h)Er+i, j − εi (h)Er+ j,i =
r−1
2 · εi = αr + ∑ 2 · αk , 1 ≤ i ≤ r − 1.
k=i
Part 6) The Cartan matrix has the entries β (hα ), α, β ∈ ∆ . We have to consider the
roots
• βk = εk − εk+1 , 1 ≤ k ≤ r − 1, and
• βr = 2εr
and the roots
• α j = ε j − ε j+1 , 1 ≤ j ≤ r − 1, with elements hα j = h j − h j+1 and
• αr = 2εr with element hαr = hr .
Accordingly, we calculate the cases:
• For 1 ≤ j, k ≤ r − 1:
βk (hα j ) = (εk − εk+1 )(E j, j − E j+1, j+1 ) = δk, j − δk, j+1 − δk+1, j + δk+1, j+1 =
• For 1 ≤ k ≤ r − 1, j = r:
• For j = k = r:
βr (hαr ) = 2 · εr (Er,r ) = 2.
β (hα ) ∥α∥2
=
α(hβ ) ∥β ∥2
Hence
− 2 αr (hαr−1 ) ∥αr−1 ∥2
2= = = , q.e.d.
−1 αr−1 (hαr ) ∥αr ∥2
In dealing with the Lie algebra so(2r, C) it is useful to consider a matrix M ∈ so(2r, C)
as a scheme having 2×2-matrices as entries. Hence we introduce the non-commutative
ring
R := M(2 × 2, C)
and consider the matrices
with entries
a jk ∈ R, 1 ≤ j, k ≤ r.
For each 1 ≤ j, k ≤ r we introduce the matrix
E j,k ∈ M(r × r, R)
with onyl one nonzero entry, namely 1 ∈ R at place ( j, k). We distinguish the fol-
lowing elements from R, which can be expressed by using the Pauli matrices, see
Remark 2.16:
0 −i
h := σ2 =
i 0
1 i −1 1 i 1
s := = (1/2) · (iσ3 − σ1 ), t := ) = (1/2) · (iσ3 + σ1 )
2 −1 −i 2 1 −i
1 i 1 1 i −1
u := = (1/2) · (σ0 − iσ2 ) v := = (1/2) · i(σ0 − σ2 )
2 −1 i 2 1 i
satisfying
• s = s⊤ , h · s = s, s · h = −s
• t = t ⊤ , h · t = −t,t · h = t
• u⊤ = v, h · u = u · h = u
214 7 Classification
• v⊤ = u, h · v = v · h = −v
• [s,t] = −h
• u2 − v2 = −h
E.g. the matrix h · Ei,i ∈ M(r × r, R) is the block matrix with the single block h ∈ R
at the diagonal place with index (i, i).
Proposition 7.9 (Typ Dr ). The Lie algebra L := so(2r, C), r ≥ 4, has the following
characteristics:
1. dim L = r(2r − 1).
2. The subalgebra
for short
Sα ≃ sl(2, C)
Sα ≃ sl(2, C)
Note the distinguished entries in the last row and in the last column.
All roots α ∈ ∆ have the same length. The only pairs of simple roots with are not
orthogonal are
2π
These pairs include the angle . In particular the Dynkin diagram of the root
3
system Φ has type Dr from Theorem 6.26, see Figure 7.2.
Proof. 3) It suffices to consider only pairs of indices (k, l) with k < l because the
generators belonging to (k, l) and (l, k) differ only by sign.
[z, s·(E j,k −Ek, j )] = ε j (z)·hs·E j,k −εk (z)·hs·Ek, j −εk (z)·sh·E j,k +ε j (z)·sh·Ek, j =
[z,t · (E j,k − Ek, j )] = ε j (z) · ht · E j,k − εk (z) · ht · Ek, j − εk (z) ·th · E j,k + ε j (z) ·th · Ek, j =
[z, u·E j,k −v·Ek, j ] = ε j (z)·hu·E j,k −εk (z)·hv·Ek, j −εk (z)·uh·E j,k +ε j (z)·vh·Ek, j =
[z, v·E j,k −u·Ek, j ] = ε j (z)·hv·E j,k −εk (z)·hu·Ek, j −εk (z)·vh·E j,k +ε j (z)·uh·Ek, j =
r−2 r
εi + ε j = ∑ αk + ∑ αk , 1 ≤ i < j ≤ r.
k=i k= j
7.2 Root systems of the A, B,C, D-series 217
[hα , xα ] = [h·(E j, j +Ek,k ), s·(E j,k −Ek, j )] = hs·(E j, j +Ek,k )(E j,k −Ek, j )−sh·(E j,k −Ek, j )(E j, j +Ek,k ) =
[hα , xα ] = [h·(E j, j −Ek,k ), u·E j,k −v·Ek, j ) = hu·E j,k +hv·Ek, j +uh·E j,k +vh·Ek, j =
(εk −εk+1 )(h·(E j, j −E j+1, j+1 )) = δk j −δk, j+1 −δk+1, j +δk+1, j+1 = 2δk j −δk, j+1 −δk+1, j =
2, if i = j
= −1, if |i − j| = 1
0, if |i − j| ≥ 2
If k = r and 1 ≤ j ≤ r − 1 then
(εr−1 + εr )(h · (E j, j − E j+1, j+1 )) = δr−1, j − δr−1, j+1 + δr, j − δr, j+1 = −δr, j+2
(
−1, if j = r − 2
=
0, if j ̸= r − 2
If 1 ≤ k ≤ r − 1 and j = r then
218 7 Classification
(εk − εk+1 )(h · (Er−1,r−1 + Er,r )) = δk,r−1 + δk,r − δk+1,r−1 − δk+1,r = −δk,r−2
(
−1, if k = r − 2
=
0, if k ̸= r − 2
If k = r = 2 then
< α, β > ∥β ∥2
= , q.e.d.
< β , α > ∥α∥2
In order to investigate the Lie algebra L := so(2r + 1, C) we use the same no-
tations and employ matrices similar to those introduced to study so(2r, C). More
specific: For 1 ≤ j, k ≤ r we extend the matrix
by zero in the last row and the last column to the matrix
Like E j,k the matrix Ẽ j,k has non-zero entries only at the two places (2 j − 1, 2k − 1)
and (2 j, 2k), both entries with value = 1.
Proposition 7.10 (Typ Br ). The Lie algebra L := so(2r + 1, C), r ≥ 2, has the fol-
lowing characteristics:
1. dim L = r(2r + 1).
2. The subalgebra
3. For 1 ≤ j ≤ r denote by
ε j := (h · Ẽ j, j )∗ ⊂ H ∗
the dual functionals. For any pair 1 ≤ j < k ≤ r each of the four elements
s · (Ẽ j,k − Ẽk, j ),t · (Ẽ j,k − Ẽk, j ), u · Ẽ j,k − v · Ẽk, j , v · Ẽ j,k − u · Ẽk, j
And in addition, for each index j = 1, ..., r a 1-dimensional root space is gener-
ated by the matrix
Y j ∈ so(2r + 1, C)
with exactly four non-zero entries: The vector
i
B2 := ∈ M(2 × 1, C)
−1
α = ±ε j .
∆ = {α1 , ..., αr }
with - note the different form of the last root -
• α j := ε j − ε j+1 , 1 ≤ j ≤ r − 1,
• αr := εr .
• If α = ε j − εk , 1 ≤ j < k ≤ r, then
• If α = ε j , 1 ≤ j ≤ r, then
hα := 2h · E j, j , xα = X j , yα = Y j .
Note the distinguished entries in the last row and the last column.
The roots α j , j = 1, ..., r − 1, have equal length; they are the long roots. The
root αr is the single short root. The length ratio is
∥α j ∥ √
= 2, j = 1, ..., r − 1.
∥αr ∥
(α j , α j+1 ), j = 1, ..., r − 1.
7.2 Root systems of the A, B,C, D-series 221
2π
Each pair (α j , α j+1 ), j = 1, ..., r − 2, encloses the angle , while the last
3
3π
pair (αr−1 , αr ) encloses the angle . The Dynkin diagram of Φ has type Br
4
from Theorem 6.26, see Figure 7.2.
→ so(2l + 1, C)
so(2l, C) ,−
In addition to the action of H on the root space elements from Proposition 7.9 one
has the action on the additional elements X j and Y j : For 1 ≤ j ≤ r the element z ∈ H
acts according to
[z, X j ] = ε j (z) · X j , [z,Y j ] = −ε j (z) ·Y j .
4) The positive roots have the base representation:
r
εj = ∑ αk , 1 ≤ j ≤ r,
k= j
j−1
εi − ε j = ∑ αk , 1 ≤ i < j ≤ r.
k=i
5) Note:
B1 · B⊤ ⊤
2 − B2 · B1 = 2 · h ∈ M(2 × 2, C).
Theorem 7.11 (The classical complex Lie algebras are simple). The Lie algebras
of the complex classical groups belonging to types
Ar , r ≥ 1; Br , r ≥ 2;Cr , r ≥ 3; Dr , r ≥ 4,
are simple.
Proof. We know from Proposition 7.6 that any classical Lie algebra L is semisimple.
According to Theorem 4.15 the semisimple Lie algebra L splits into a direct sum of
222 7 Classification
simple Lie algebras. Elements from different summands commute pairwise. There-
fore the direct sum reduces to one single summand, because the Dynkin diagram of
L is connected. Hence L is simple, q.e.d.
Remark 7.12 (Semisimplicity of the classical Lie algebras). Apparently, one can for-
mulate the content of Proposition 7.7 until Proposition 7.10 without using that the
classical Lie algebras L are semisimple. Then these propositions give a second proof
for the semisimplicity of L, besides Proposition 7.6. One needs the decomposition
of L as a direct sum of H and 1-dimensional eigenspaces Lα
M
L=H⊕ Lα ,
α∈Φ
Sα ≃ sl(2, C) ⊂ L
h∈R⊂H
α(h) ̸= 0.
[h, x] = α(h) · x.
7.2 Root systems of the A, B,C, D-series 223
[h, x] = −[x, h] ∈ R ⊂ H
x = h+ ∑ xα
α∈Φ
xα ̸= 0.
hα = [xα , yα ] ∈ R
and
yα = −(1/2) · [hα , yα ] ∈ R.
Therefore
sl(2, C) ≃ spanC < xα , yα , hα >⊂ R,
a contradition, because R is solvable but sl(2, C) is not.
a contradiction.
Both cases i) and ii) lead to a contradition. This fact shows R = {0} and proves
that L is semisimple, q.e.d.
224 7 Classification
Remark 7.13 (Real simple Lie algebras). Also the Lie algebras of the real classical
groups belonging to types
Ar , r ≥ 1; Br , r ≥ 2;Cr , r ≥ 3; Dr , r ≥ 4;
are simple.
7.3 Outlook
According to Theorem 7.5 the root set of any complex semisimple Lie alge-
bra is a root system. The previous section shows that all Dynkin diagrams from
the A, B,C, D-series, see Theorem 6.26, are realized by the root system of a simple
Lie algebra.
At least the following related issues remain:
• Show that also the exceptional Dynkin diagrams of type F4 , G2 , Er with r = 6, 7, 8
result from root systems, see [28, Chap. 12.1].
• Show that also the exceptional root systems can be obtained from simple Lie
algebras.
• Show that a semisimple Lie algebra is uniquely determined by its root system.
Chapter 8
Irreducible Representations
In the present section L denotes an arbitrary K-Lie algebra, not necessaritly semisim-
ple. The Lie algebra and the vector spaces may have infinite dimension if not stated
otherwise.
Before studying the structure of L-modules we make a disgression from Lie al-
gebras to associative algebras. We assume that all associative algebras A have a unit
element 1 ∈ A and all morphism ρ between associative algebras respect the unit
element.
[x, y] := x · y − y · x, x, y ∈ A.
225
226 8 Irreducible Representations
Definition 8.1 (Tensor algebra, symmetric algebra and universal enveloping al-
gebra).
1. Consider a K-vector space M. The tensor algebra T M of M is the pair (T M, j)
with
• the associative K-algebra
T k M, T k M := M ⊗k ,
M
T M :=
k∈N
j : M = T 1 M ,→ T M.
SM := T M/I
and denote by
pSM : T M →
− SM
the canonical projection. The pair
(SM, pSM ◦ j)
with
pUL : T L → UL
the canonical projection.
xy ∈ UL or x · y ∈ UL for x, y ∈ UL.
Moreover we consider the map j : L ,−→ T L as an identification. The restriction
on K = T 0 L
pUL |T 0 L : T 0 L →
− UL
is injective, because
L⊗m .
M
I⊂
m≥1
i:L→
− UL
u ⊗ v − v ⊗ u with u, v ∈ T M,
ρUL : UL → A
UL
i
L ρUL
ρ
A
T := T L, T ⊗m := T Lm , S := SL, U := UL.
Definition 8.3 (Filtration and graded algebra of the universal enveloping alge-
bra). Consider an arbitrary but fixed K-Lie algebra L with universal enveloping
algebra U.
1. For m ∈ N denote by
the K-subspace of the universal enveloping algebra formed by all products with
at most m factors from L. In particular,
Then
FU := (FmU)m∈N
is a filtration of the universal enveloping algebra U:
satisfying
FnU · FmU ⊂ Fn+mU.
2. The successive quotient vector spaces of the filtration
GmU := FmU/Fm+1U, m ∈ N,
Then
pU,GU : U →
− GU
3. In general the Lie algebra L is not Abelian. For non-Abelian L the ideal
(F mU := pU (Tm ))m∈N
pU,GU : U →
− GU.
4. The notation FmU denotes the filtration by lower indices, while the notation GmU
denotes the graduation by upper indices.
x · y ∈ Fn+m−1U.
pGU := [pU,GU : U →
− GU] ◦ [pU : T →
− U] : T →
− GU
Hence
pGU (I) ⊂ {0} = F1U/F1U ⊂ F2U/F1U.
The inclusion I ⊂ ker pGU implies the existence of a morphism
φ : SL →
− GU
pGU
T GU
pS φ
S
The main theorem about the universal enveloping algebra UL is due to Poincaré-
Birkhoff-Witt, see Theorem 8.8. It states that φ is an isomorphism of associative
8.1 The universal enveloping algebra 231
algebras. Hence the graded algebra GUL of the universal enveloping algebra UL
is the algebra of polynomials in finitely many variables, a well-understood alge-
braic object. The proof of the injectivity of φ in requires some preparations, see
Notation 8.5, Proposition 8.6 and Lemma 8.7.
SL ≃ K[(zλ )λ ∈Ω ].
Sm L := S0 L ⊕ ... ⊕ Sm L ⊂ SL, m ∈ N,
Σ = (λ1 , ..., λm )
we set
xΣ := xλ1 ⊗ ... ⊗ xλm ∈ T m L
and
zΣ := zλ1 · ... · zλm ∈ Sm L.
If λ1 ≤ ... ≤ λm then the sequence Σ is increasing.
ρ :L→
− gl(SL)
The algebraic proof of the proposition follows a standard scheme, see [28, § 17.4],
[32, Ch. II.6], [46, Part I, Ch. III, No. 4] .
Proof. Because the K-algebra SL is Abelian we may assume all m-tuples in the
proof
Σ = (λ1 , ..., λm )
as increasing.
Due to the filtration of SL the existence of the linear map f is equivalent to a com-
patible family ( fm )m∈N of linear maps
fm : L ⊗ Sm L →
− GS,
fm (xλ ⊗ zΣ ) = zλ · zΣ ∈ Sm+1 L
fm (xλ ⊗ zΣ ) ≡ zλ · zΣ mod Sk L
Here condition Am considers for given Σ only indices λ ∈ Ω less or equal than
each index from Σ . Condition Bm considers arbitrary indices λ and allows a correc-
tion term of lower order in comparison to the result from Am . While condition Cm
captures the requirement for commutators from L.
f0 : L ⊗ S0 L →
− SG, f0 (xλ ⊗ 1) := zλ ∈ S1 L.
Apparently f0 is the only linear map which satisfies condition A0 . It also satisfies
condition B0 and C0 .
fm (xλ ⊗ zΣ ) ∈ SL
fm (xλ ⊗ zΣ ) := zλ · zΣ .
λ > µ := λ1 .
fm−1 (xµ ⊗ zT ) = zµ · zT = zΣ .
or
fm (xλ ⊗ zΣ ) = fm (xµ ⊗ fm (xλ ⊗ zT )) + fm ([xλ , xµ ] ⊗ zT ).
Concerning the terms on the right-hand side of the last equation:
• First term: The compatibility of fm with fm−1 and the induction
hypothesis Bm−1 imply
or
fm (xλ ⊗ zT ) = zλ · zT + y,
with
y = fm−1 (xλ ⊗ zT ) − zλ · zT ∈ Sm−1 L.
234 8 Irreducible Representations
fm : L ⊗ Sm L →
− Sm+1 L
µ < λ and µ ≤ T.
λ < µ and λ ≤ T
because
[xµ , xλ ] = −[xλ , xµ ].
Apparently, condition Cm holds for λ = µ. Hence we are left with the case:
neither λ ≤ T nor µ ≤ T.
x.z := fm (x ⊗ z)
We obtain
xµ .zT = xµ .(zν · zΨ )
8.1 The universal enveloping algebra 235
using condition Am−2 for the last equality. Condition Cm−1 applied with the
indices µ and ν implies
hence
xµ .zT = xν .(xµ .zΨ ) + [xµ , xν ].zΨ .
Condition Bm−2 , a consequence of Bm−1 , implies the existence of a
suitable w ∈ Sm−2 L such that
xµ .zΨ = zµ · zΨ + w.
Hence
xµ .zT = xν .(zµ · zΨ ) + xν .w + [xµ , xn u].zΨ .
We apply xλ to both sides of the equation:
The trick of the proof is now to apply the induction assumption and the part of
condition Cm , which has already been proven, to each of the three summands
separately. And afterwards to add the results and to remove the intermediate term w
from the result. Using the shorthand T1 , T2 , T3 for the three summands on the
right-hand side of the equation, we transform each of the summands as
• T1 : Condition Cm applies because ν ≤ (µ,Ψ ), hence
xλ .(xµ .zT ) = xν .(xλ .(xµ .zΨ ))+[xλ , xν ].(xµ .zΨ )+[xµ , xν ].(xλ .zΨ )+[xλ , [xµ , xν ]].zΨ
We now interchange the role of λ and µ and subtract the resulting equation from
the first equation. Due to the symmetry the underlined terms from both equations
236 8 Irreducible Representations
= xν .(xλ .(xµ .zΨ )) − xν .(xµ .(xλ .zΨ )) + [xλ , [xµ , xν ]].zΨ − [xµ , [xλ , xν ]].zΨ =
= xν .([xλ , xµ ].zΨ ) + [xλ , [xµ , xν ]].zΨ + [xµ , [xν , xλ ]].zΨ =
= [xλ , xµ ].(xν .zΨ ) + [xν , [xλ , xµ ]].zΨ + [xλ , [xµ , xν ]].zΨ + [xµ , [xν , xλ ]].zΨ =
= [xλ , xµ ].zT + ([xν , [xλ , xµ ]] + [xλ , [xµ , xν ]] + [xµ , [xν , xλ ]]).zΨ =
= [xλ , xµ ].zT
Here the last equality is due to the Jacobi identity. The end of the induction step
finishes the proof of the Proposition, q.e.d.
I := ker [pSL : T L →
− SL], J := ker [pUL : T L →
− UL].
tm ∈ T m L ∩ I.
The meaning of Lemma 8.7 can be easily seen in the case m = 2 by the element
t = v1 ⊗ v2 − v2 ⊗ v1 − [v1 , v2 ] ∈ T 2 L ∩ J :
t2 = v1 ⊗ v2 − v2 ⊗ v1 ∈ I.
ρ :L→
− gl(SL)
ρ :L→
− End(SL)
satisfying
ρ([x, y]) = ρ(x) ◦ ρ(y) − ρ(y) ◦ ρ(x).
Due to the universal property of the universal enveloping algebra UL we obtain a
unique morphism of associative algebras
8.1 The universal enveloping algebra 237
ρUL : UL →
− End(SL)
L
ρ
j
TL End(SL)
pUL
ρUL
UL
ρ := ρUL ◦ pUL : T L →
− End(SL)
SL ≃ UL.
238 8 Irreducible Representations
The following Theorem 8.8 shows how to generalize this fact to arbitrary Lie
algebras.
2. Injectivity: We set
I := ker pSL ⊂ T L
in the commutative diagram
pGUL
TL GUL
pSL φ
SL
pUL : T L →
− UL
and set
J := ker pUL .
We show for any m ∈ N: If t ∈ T m L satisfies
pGUL (t) = 0
i.e.
pUL (t) ∈ Fm−1UL
then t ∈ I, hence pSL (t) = 0. The particular case m = 1 proves the injectivity of φ .
i.e.
pUL (t − t ′ ) = 0 or t − t ′ ∈ J.
Lemma 8.7, applied to the element
t − t ′ ∈ T mL ∩ J
The following Proposition 8.9 follows from Theorem 8.8. It states in particular
that any Lie algebra L embeds into its universal enveloping algebra UL.
Proposition 8.9 (Basis of the universal enveloping algebra). Consider a finite-
dimensional Lie algebra L with tensor algebra T , symmetric algebra S, universal
enveloping algebra U, and graded universal enveloping algebra GU.
1. If W ⊂ T m is a subspace which under the restriction of the canonical map
pS : T →
− S
pU (W ) ⊂ FmU
i = pU ◦ j : L →
−U
is injective.
together with the element 1 is a basis of U. Note that the factors are arranged
according to increasing indices of the basis elements.
of L. Hereby all elements from BH are to be placed before the elements from BL/H .
Then the injection H ,→ L induces an injection
UH ,→ UL,
pU
T U
pU,GU
φ
S GU
the map φ is an isomorphism due to theorem 8.8. The diagram induces by re-
striction for each index m ∈ N the commutative diagram
Tm FmU
≃
Sm Gm U
→ Tm →
W ,− − Sm →
− GmU
→ Tm →
W ,− − Gm U
− FmU →
Because
pU (W ) ∩ Fm−1U = {0}
we obtain
pU (W ) ≃ FmU/Fm−1U
or
FmU ≃ pU (W ) ⊕ Fm−1U.
2. We apply the result from part 1) with index m := 1 and W := j(L). Then
which implies
i(L) = pU ( j(L)) ≃ L.
8.1 The universal enveloping algebra 241
3. We apply the result of part 1) for a given index m ∈ N and the subspace
FmU = Fm−1U ⊕ pU (W ).
All statements from Proposition 8.9 hold also for Lie algebras with a countable
infinite dimension. The proof goes through along the same lines.
Example 8.10 (Basis of the universal enveloping algebra). Consider the Lie algebra
L = sl(2, C).
hi x j yk , i, j, k ∈ N.
The following table shows the product formulas which allow to compute inductively
the product of two basis elements. The entries of the table show the product with an
element from the left column with an element from the upper row.
h x y
h h2 hx hy
x hx − 2x x2 xy
y hy + 2y xy − h y2
U : Lie →
− Ass,
i.e. for any pair of a K-Lie algebra L and an associative K-algebra A there is a
natural isomophism of K-vector spaces
Corollary 8.12 (Equivalence of categories). For any Lie algebra L the category L-mod
of L-modules and the category UL-mod of UL-modules are equivalent.
The present section generalizes the representation theory of the Lie algebra sl(2, C)
from Chapter 5.2 to the representation theory of an arbitrary complex semisimple
Lie algebra L. The L-modules M under consideration are not necessarily finite di-
mensional.
We fix
• a pair (L, H) with a semisimple complex Lie algebra L and a Cartan subalgebra H ⊂ L,
• and a base
∆ = {α1 , ..., αr }
of the roots Φ of (L, H).
˙ Φ −,
Φ = Φ+ ∪
δ = (1/2) · ∑ α,
α∈Φ +
(Lα ⊕ L−α ).
M M
L=H⊕ Lα = H ⊕
α∈Φ α∈Φ +
8.2 General weight theory 243
The subalgebra of L M
B := H ⊕ Lα
α∈Φ +
which comprises besides H only the root spaces of positive roots, is named the
corresponding Borel subalgebra. Moreover we consider the subalgebras of L of
root spaces belonging to respectively positive and negative roots
N + := Lα , N − :=
M M
Lα .
α∈Φ + α∈Φ −
xα ∈ Lα , yα ∈ L−α , hα = [xα , yα ] ∈ H
satisfying
spanC < hα , xα , yα >≃ sl(2, C).
Proof. The algebra N + is nilpotent because its descending central series converges
to zero: One uses the formula
Lα = N +
M
[B, B] ⊂
α∈Φ +
is nilpotent, q.e.d.
Both, the terms in Definition 5.6 from the representation theory of sl(2, C)-modules
and the terms in Definition 5.12 for the adjoint representation of semisimple com-
plex Lie algebras generalize to representations of arbitray semisimple complex Lie
algebras.
Definition 8.14 (Weight and weight space). Consider a semisimple Lie algebra L
and an L-module V .
244 8 Irreducible Representations
µ(hα ) ∈ Z,
µ(hα ) ≥ 0.
Lemma 8.15 (Action of the root spaces). Consider a semisimple Lie algebra L and
an L-module V .
1. For any root α of L and any weight µ of V we have
Lα .V µ ⊂ V α+µ .
2. The sum
V ′ := ∑ Vµ
µ weight
is direct, i.e.
V′ =
M
V µ.
µ weight
It is a submodule of V .
Consider a vector space V . Then Lemma 8.2 applies to the associative algebra A = gl(V ):
Any representation
ρ : L → gl(V )
extends uniquely to a morphism of associative algebras
ρUL : UL → gl(V ).
8.2 General weight theory 245
for u ∈ UL and v ∈ V.
V = UL.e
x.e = λ (x) · e.
If α ∈ Φ + and
xα ∈ L α ⊂ N +
we consider
Sα := span < hα , xα , yα >≃ sl(2, C).
Then
2xα .e = [hα , xα ].e = λ (hα ) · λ (xα ) · e − λ (xα ) · λ (hα ) · e = 0.
Hence
N + .e = 0
and e is a primitive element, q.e.d.
The following Proposition 8.18 clarifies the structure of standard cyclic modules;
compare also Proposition 5.7.
246 8 Irreducible Representations
3. Finite dimensionality of weight spaces: For each weight µ of V the weight space
is finite-dimensional, i.e.
dim V µ < ∞.
In particular
dim V λ = 1.
4. The L-module V does not decompose as a direct sum of proper L-submodules,
i.e. there is no representation
V = V1 ⊕V2
Proof. Note that V is not necessarily finite dimensional: In general there is no lower
bound for the weights µ.
1. Proposition 8.9, part 4) applied to the vector space decomposition
L = N− ⊕ B
BN − = (yα )α∈Φ +
by the basis of B
BB = (hα )α∈∆ ∪ (xα )α∈Φ +
to a basis of L represents UL as a free UN − -module with basis BB . Therefore
Lα .V µ ⊂ V µ+α
and the fact, that any negative root α ∈ Φ − is a linear combination of elements
from ∆ with negative integer coefficients.
is
1 = y0α1 · ... · y0αr ,
hence dim V λ = 1.
V λ = V1 λ ⊕V2 λ .
λi , i = 1, 2,
dim V λ = 1
according to part 1), the two primitive elements e and e′ are scalar multiples.
3. i) Consider two irreducible L-modules V1 and V2 with the same heighest weight λ .
We choose primitive elements e1 ∈ V1 and e2 ∈ V2 and define the L-module
V := V1 ⊕V2
pi : V → Vi , i = 1, 2.
p2 |E : E →
− V2
N2 = E ∩V1 .
p1 : E →
− V1
is an isomorphism. Therefore V1 ≃ V2 .
ei ∈ Viλi , i = 1, 2.
We may assume φ (e1 ) = e2 . For arbitrary elements h ∈ H from the Cartan sub-
algebra H ⊂ L we have
λ2 (h) · e2 = h.e2 = h.φ (e1 ) = φ (h.e1 ) = φ (λ1 (h) · e1 ) = λ1 (h) · φ (e1 ) = λ1 (h) · e2 ,
λ1 = λ2 ∈ H ∗ , q.e.d.
Not any irreducible L-module has a primitive element as the following example
shows.
Example 8.20 (An irreducible sl(2, C)-module without primitive elements). Con-
sider the Lie algebra
250 8 Irreducible Representations
with the basis (en )n∈2Z . We define the following action of L using x as raising oper-
ator and y as lowering operator:
h.en := n · en
x.en := an · en+2
y.en := bn · en−2 .
i) The sequences (an )n∈N , (bn )n∈N have to be choosen in a suitable way: Three
conditions have to be satisfied, in order that these definitions determine an L-module
structure on V . The first two conditions are fulfilled for any choice: First,
and
2x.en = 2 · an · en+2 .
Secondly
[h, y].en := h.(y.en ) − y.(h.en ) = 2 · bn · en−2
and
2y.en = 2 · bn · en−2 .
The third condition, the equality of
[x, y].en = x.(y.en ) − y.(x.en ) = x.(bn · en−2 ) − y.(an · en+2 ) = (an−2 · bn − an · bn+2 )en
and
h.en = n · en ,
requires for all n ∈ 2Z
n = an−2 · bn − an · bn+2 .
That condition can be satisfied by choosing the two sequences (an )n∈N , (bn )n∈N as
being linear in the indeterminate n. We choose
µ +n µ −n
an := , bn :=
2 2
with an arbitrary µ ∈ C \ Z.
8.2 General weight theory 251
ii) The choice of µ assures that V contains no proper L-submodule, i.e. that V
is irreducible: Assume the existence of a non-zero submodule W ⊂ V and choose a
non-zero vector w ∈ W . The vector w is a linear combination of finitely many weight
vectors of V . Hence w is contained in a finite direct sum V ′ of weight spaces. As a
consequence
w ∈ W ∩V ′ ,
and W ∩ V ′ is a non-zero, h-stable, and finite dimensional vector space. Hence the
endomorphism induced by the action of h has an eigenvector. In particular, W ∩V ′
and a posteriori W contains a weight vector of V . Then W = V due to the definition
of V .
λ :H →
− C
Note that λ is not required to satisfy any condition with respect to integrality or
positivity. The proof of the theorem relies heavily on the existence and the properties
of the universal enveloping algebra of L.
Proof. We first define a representation of the Borel-algebra B ⊂ L: We choose
a 1-dimensional vector space
V λ := C · eB
and define
Vλ := UL ⊗UB V λ .
Here UL acts on the first component of the elementary tensors. Then Vλ is generated
by the element
252 8 Irreducible Representations
e := 1 ⊗ eB ∈ Vλ .
The latter element is not-zero because UL is a free UB-module having a basis which
extends the unit 1 ∈ UL, see Proposition 8.9, part 4). The UL-module Vλ , sometimes
named Verma module, can be considered also as an L-module.
In general the L-module Vλ is not irreducible. Therefore, the final step will factor
out the proper L-submodules of Vλ : Consider a proper submodule
V ′ ⊊ Vλ .
V′ = (V ′ ) ⊂
µ
∑ ∑ (Vλ )µ
µ weight of V ′ µ weight of Vλ
V ′ = Vλ ,
V′ ⊂ ∑ (Vλ )µ
µ weight of Vλ
µ̸=λ
• And on the other hand, Theorem 8.23 implies: If V has an integral and non-
negative highest weight, then V is finite-dimensional.
We continue using the notation which has been fixed in the beginning of Section 8.2.
All Lie algebras are assumed finite-dimensional.
8.3 Finite-dimensional irreducible representations 253
For an arbitrary but fixed positive root α ∈ Φ + the vector space V is a finite-dimensional sl(2, C)-module
with respect to
sl(2, C) ≃ spanC < hα , xα , yα >⊂ L.
According to Theorem 5.9 the numbers µ(hα ) are integers.
2. According to Theorem 3.20 the action of the Borel subalgebra B ⊂ L on the finite-
dimensional vector space V has a common eigenvector. The latter is a primitive
element due to Lemma 8.17.
3. According to Proposition 8.18 the standard cyclic L-module V does not split as
the direct sum of distinct L-modules. Thanks to Weyl’s theorem on complete
reducibility, Theorem 4.24, the module V is irreducible, q.e.d.
µ(hα ) ∈ Z, α ∈ Φ +
254 8 Irreducible Representations
Set
mα := λ (hα ).
Ba assumption mα ∈ Z and mα ≥ 0. Consider the Sα -submodule of V generated by
the operation of Sα on a primitive element e ∈ V λ of the L-module V , and in
particular the distinguished element
α +1
eα := ym
α .e ∈ V.
[xα , yβ ] = 0
which implies
α +1 α +1
xβ .eα = xβ .(ym
α .e) = ym
α .(xβ .e) = 0,
because e is a primitive element. On the other hand, the formulas from
Proposition 5.7 concerning the generated sl(2, C)-module show
xα .eα = 0.
λ − (mα + 1)α.
But according to Corollary 8.19 it must have the same weight λ as the primitive
element e. Therefore eα = 0. Then the vector space
is a finite-dimensional Sα -module.
Sα .(L.F) ⊂ L.F,
8.3 Finite-dimensional irreducible representations 255
zα .(z.F) ⊂ L.F
zα .z = z.zα + [zα , z]
hence
zα .(z.F) ⊂ z.(zα .F) + [zα , z].F ⊂ L.F
because zα .F ⊂ F by assumption. As a consequence, also L.F is a
finite-dimensional Sα -module, i.e.
L.F ∈ Tα .
ii) The Weyl group acts on the weight set: We show that the Weyl group permutes
the weights of V . Consider a weight µ and a non-zero element v ∈ V µ . According
to part i) for any positive root α ∈ Φ + exists a finite-dimensional Sα -module Vα
with v ∈ Vα . Consider the integer
pα := µ(hα ) ∈ Z,
and set pα
(yα ) .v
if pα ≥ 0
x :=
(xα )−pα .v
if pα ≤ 0
iii) Finite weight set: To show that the weight set of V is finite consider an arbitrary
but fixed weight µ. According to Corollary 8.19 it has the form
We shall show that the coefficients pα , α ∈ ∆ , are bounded; hence there are only
finitely many.
w(∆ ) = −∆ .
According to part ii) also w(µ) is a weight of V . Using again Corollary 8.19 the
weight has the form
Then
µ = w−1 (w(µ)) = w−1 (λ ) − ∑ qα · w−1 (α) =
α∈∆
w−1 (∆ ) = −∆ .
Note that
{qα : α ∈ ∆ } = {rα : α ∈ ∆ },
but for α ∈ ∆ possibly qα ̸= rα because not necessarily w(α) = −α. We obtain
! !
λ − w−1 (λ ) = µ+ ∑ pα · α − µ− ∑ rα · α = ∑ (pα + rα ) · α.
α∈∆ α∈∆ α∈∆
λ − w−1 (λ ) = ∑ cα · α
α∈∆
with coefficients
cα := pα + rα with pα , rα ∈ N.
We obtain
0 ≤ pα = cα − rα ≤ cα .
The estimate shows: All coefficients (pα )α∈∆ of the weight µ are bounded by the
constant
max{cα : α ∈ ∆ }
8.3 Finite-dimensional irreducible representations 257
which is independent from µ. Hence there exist only finitely many weights.
iv) Weight space decomposition: The module V decomposes into finitely many
weight spaces according to part iii). Each weight space is finite-dimensional due to
Proposition 8.18. Hence V is finite-dimensional, q.e.d.
(λα )α∈∆
are the fundamental weights of L with respect to (H, ∆ ). The uniquely de-
termined irreducible L-modules Vα with highest weight λα , α ∈ ∆ , are the
fundamental L-modules or fundamental representations of L.
2. The dominant weights are the linear combinations of the fundamental weights
with non-negative integer coefficents
λ= ∑ nα · λα
α∈∆
∆ = {α1 , ..., αr }
of its roots. Then the Cartan matrix of L relates the elements of ∆ and the funda-
mental weights according to
Proof. In order to prove the claim one has to show for i = 1, ..., r, the equality of
linear functionals on H
r
αi = ∑ < αi , α j > ·λ j .
j=1
• Right-hand side:
!
r
∑ < αi , α j > ·λ j (hk ) =< αi , αk >, q.e.d.
j=1
Because all entries of the Cartan matrix of a semisimple Lie algebra L are inte-
gers, Proposition 8.25 shows: The r-dimensional root lattice Λroot of L is a sublattice
of the lattice Λ with basis the fundamental weights.
The dominant weights are the highest weights of the finite-dimensional, irre-
ducible L-modules. Hence the dominant weights characterize the finite-dimensional
irreducible L-modules. The result is a direct consequence of Theorem 8.23.
µ= ∑ nα · λα ∈ Λ , nα ∈ Z for all α ∈ ∆ .
α∈∆
λ= ∑ nα · λα , nα ∈ C.
α∈∆
8.3 Finite-dimensional irreducible representations 259
λ (hα ) = nα ∈ N
µ(hα ) ∈ Z, q.e.d.
Hence
µ= ∑ µ(hα ) · λα , q.e.d.
α∈∆
Generalizing the notation for irreducible sl(2, C)-modules from Theorem 5.9 we
introduce the notation:
is denoted
V (n1 , ..., nr ).
H := {h ∈ d(r + 1, C) : tr h = 0},
Then we have
260 8 Irreducible Representations
• the basis of H
(hi := hαi := Ei,i − Ei+1,i+1 )i=1,...,r
• and its dual basis (λi )i=1,...,r of fundamental weights
λi = ε1 + ... + εi , i = 1, ..., r.
or
λ1 = (1/3)(2α1 + α2 ), λ2 = (1/3)(α1 + 2α2 ),
see Figure 8.1.
8.3 Finite-dimensional irreducible representations 261
Fig. 8.1 Lattice of fundamental weights of sl(3, C), monoid of dominant weights shaded
List of results
263
264 List of results
...
FINIS
References
In addition,
• References with focus on mathematics:
[4], [5], [7], [25], [26], [27], [28], [50], [51]
• References with focus on physics:
[1], [3], [17], [20], [21], [24], [37], [42]
• References with focus on both mathematics and physics:
[2], [18], [21], [47]
1. Alex, Arne; Kalus, Matthias; Huckleberry, Alan; von Delft, Jan: A numerical algorithm
for the explicit calculation of SU(N) and SL(N, C) Clebsch–Gordan coefficients. Journal of
Mathematical Physics 52, 023507 (2011)
2. Böhm, Manfred: Lie-Gruppen und Lie-Algebren in der Physik. Eine Einführung in die math-
ematischen Grundlagen. Springer, Berlin (2011)
3. Born, Max, Jordan, Pascual: Zur Quantenmechanik. Zeitschrift für Physik 34, 858-888 (1925)
4. Bourbaki, Nicolas: Éléments de mathématique. Groupes et Algèbres de Lie. Chapitre I. Dif-
fusion C.C.L.S., Paris (without year)
5. Bourbaki, Nicolas: Éléments de mathématique. Groupes et Algèbres de Lie. Chapitre VII,
VIII. Algèbres de Lie. Diffusion C.C.L.S., Paris (without year)
6. Bourbaki, Nicolas: Éléments de mathématique. Groupes et Algèbres de Lie. Chapitre II, III.
Algèbres de Lie. Diffusion C.C.L.S., Paris (without year)
7. Bourbaki, Nicolas: Elements of Mathematics. Algebra II. Chapters 4-7. Springer, Berlin
(2003)
8. Bourbaki, Nicolas: Elements of Mathematics. General Topology. Chapters 1-4. Springer,
Berlin (1989)
9. Bröcker, Theodor; tom Dieck, Tammo: Representations of Compact Lie groups. Springer,
New York (1985)
10. Dugundji, James: Topology. Allyn and Bacon, Boston (1966)
11. Duistermaat, Johannes J.; Kolk, Johan A.C.: Lie Groups. Springer (2013)
12. Fischer, Gerd: Lineare Algebra. Eine Einführung für Studienanfänger. Springer Spektrum,
16. Aufl. (2008)
265
266 References
13. Forster, Otto: Analysis 1. Differential- und Integralrechnung einer Veränderlichen. Vieweg,
Reinbek bei Hamburg (1976)
14. Forster, Otto: Komplexe Analysis. Universität Regensburg. Regensburg (1973)
15. Forster, Otto: Riemannsche Flächen. Springer, Berlin (1977)
16. Gasiorowicz, Stephen: Elementary Particle Physics. John Wiley & Sons, New York (1966)
17. Georgi, Howard: Lie Algebras in Particle Physics. Westview, 2nd ed. (1999)
18. Gilmore, Robert: Lie Groups, Lie Algebras, and some of their Applications. Dover Publica-
tions, Mineola (2005)
19. Gunning, Robert; Rossi, Hugo: Analytic Functions of Several Complex Variables. Prentice-
Hall, Englewood Cliffs, N.J. (1965)
20. Grawert, Gerald: Quantenmechanik. Akademische Verlagsanstalt, Wiesbaden (1977)
21. Hall, Brian: Quantum Theory for Mathematicians. Springer, New York (2013)
22. Hall, Brian: Lie Groups, Lie Algebras, and Representations. An Elementary Introduction.
Springer, Heidelberg (2ed 2015)
23. Hatcher, Allen: Algebraic Topology. Cambridge University Press, Cambridge (2002). Down-
load https://www.math.cornell.edu/ hatcher/AT/AT.pdf
24. Heisenberg, Werner: Die Kopenhagener Deutung der Quantentheorie. In ders.: Physik und
Philosophie. Ullstein, Frankfurt/Main (1959)
https://www.marxists.org/reference/subject/philosophy/works/ge/heisenb3.htm
Download 6.6.2018
25. Helgason, Sigurdur: Differential Geometry, Lie Groups and Symmetric Spaces. Academic
Press, New York (1978)
26. Hilgert, Joachim, Neeb, Karl-Hermann: Lie Gruppen und Lie Algebren. Braunschweig (1991)
27. Hilgert, Joachim, Neeb, Karl-Hermann: Structure and Geometry of Lie Groups. New York
(2012)
28. Humphreys, James E.: Introduction to Lie Algebras and Representation Theory. Springer,
New York (1972)
29. Iachello, Francesco: Lie Algebras and Application. Lecture Notes in Physics 708, Springer,
Berlin (2006)
30. Ireland, Kenneth, Rosen, Michael: A Classical Introduction to Modern Number Theory.
Springer, New York (1982)
31. Kac, Victor, in Introduction to Lie Algebras, http://math.mit.edu/classes/18.745/index.html
Cited 17 September 2018
32. Knapp, Anthony: Lie Groups, Lie Algebras, and Cohomology. Princeton University Press,
Princeton (1988)
33. Knapp, Anthony: Lie Groups Beyond an Introduction. Birkhäuser, Boston 2nd ed. (2005)
34. Lang, Serge: Algebra. Addison-Wesley, Reading (1965)
35. Lang, Serge: SL(2,R). Springer, New York (1985)
36. Lipkin, Harry J.: Lie Groups for Pedestrians (Deutsch: Anwendung von Lieschen Gruppen in
der Physik). North-Holland Publishing. Amsterdam. 2nd ed. (1966)
37. Messiah, Albert: Quantum Mechanics. Volume 1. North-Holland Publishing Company, Am-
sterdam (1970)
38. Montgomery, Deane; Zippin, Leo: Topological Transformation Groups. Interscience Publish-
ers, New York (1955)
39. Narasimhan, Raghavan: Several Complex Variables. The University of Chicago Press,
Chicago (1971)
40. Neukirch, Jürgen: Algebraische Zahlentheorie. Springer, Berlin (1992)
41. Rollnik, Horst: Teilchenphysik II. Innere Symmetrien der Teilchen. Bibliographisches Insti-
tut. Mannheim (1971)
42. Roman, Paul: Advanced Quantum Theory. An outline of the fundamental ideas. Addison-
Wesley, Reading Mass. (1965)
43. Rim, Donsub: An elementary proof that symplectic matrices have determinant one.
https://arxiv.org/abs/1505.04240} Download 9.5.2018
44. Scharlau, Winfried; Opolka Hans: Von Fermat bis Minkowski. Eine Vorlesung über Zahlen-
theorie. Springer, Berlin (1980)
References 267
45. Serre, Jean-Pierre: Complex Semisimple Lie Algebras. Reprint 1987 edition, Springer, Berlin
(2001)
46. Serre, Jean-Pierre: Lie Algebras and Lie Groups. 1964 Lectures given at Harvard University.
2nd edition, Springer, Berlin (2006)
47. Schottenloher, Martin: Geometrie und Symmetrie in der Physik. Leitmotiv der Mathematis-
chen Physik. Vieweg, Braunschweig (1995)
48. Stöcker, Ralph; Zieschang, Heiner: Algebraische Topologie. Eine Einführung. Teubner,
Stuttgart (1988)
49. Spanier, Edward: Algebraic Topology. Tata McGraw-Hill, New Delhi (Repr. 1976)
50. Varadarajan, Veeravalli S.: Lie Groups, Lie Algebras, and their Representations. Springer,
New York (1984)
51. Weibel, Charles A.: An Introduction to Homological Algebra. Cambridge University Press,
Cambridge (1994)
Index
Killing form, 98
derivation, 44
derivation L-module, 43
inner, 45 Lie algebra
outer, 45 definition, 39
derived algebra, 41 nilpotent, 77
derived series, 82 of skew-Hermitian matrices, 48
descending central series, 77 of skew-symmetric matrices, 47
direct sum decomposition, 107 of traceless matrices, 47
dominant weight, 257 semisimple, 102
dynamical Lie algebra of quantum mechanics, simple, 102
93 solvable, 82
logarithm, 28
eigenspace, 13
exponential, 26 matrix
269
270 Index