2
E
6
:

E
7
:

E
8
:
The graphs listed in the theorem are called (simply laced) Dynkin diagrams. These graphs
arise in a multitude of classication problems in mathematics, such as classication of simple Lie
algebras, singularities, platonic solids, reection groups, etc. In fact, if we needed to make contact
with an alien civilization and show them how sophisticated our civilization is, perhaps showing
them Dynkin diagrams would be the best choice!
As a nal example consider the representation theory of nite groups, which is one of the most
fascinating chapters of representation theory. In this theory, one considers representations of the
group algebra A = C[G] of a nite group G the algebra with basis a
g
, g G and multiplication
law a
g
a
h
= a
gh
. We will show that any nite dimensional representation of A is a direct sum of
irreducible representations, i.e. the notions of an irreducible and indecomposable representation
are the same for A (Maschkes theorem). Another striking result discussed below is the Frobenius
divisibility theorem: the dimension of any irreducible representation of A divides the order of G.
Finally, we will show how to use representation theory of nite groups to prove Burnsides theorem:
any nite group of order p
a
q
b
, where p, q are primes, is solvable. Note that this theorem does not
mention representations, which are used only in its proof; a purely grouptheoretical proof of this
theorem (not using representations) exists but is much more dicult!
This text is based on a minicourse given by Pavel Etingof at the 2004 Clay Mathematics Insti
tute Research Academy. The remaining authors, who were participants of the Academy, improved
and extended the initial lecture notes, and added solutions of homework problems.
The goal of the text is not to provide a systematic introduction to representation theory, but
rather to convey to the reader the spirit of this fascinating subject, and to highlight its beauty by
discussing a few striking results. In other words, the authors would like to share with the reader
the fun they had during the days of the Academy!
1
.
The text contains many problems, whose solutions are given in the last section. These problems
were given as homework during the CMI Research Academy. Sometimes they are designed to
illustrate a notion or result from the main text, but often contain new material. The problems are
a very important part of the text, and we recommend the reader to try to solve them after reading
the appropriate portions.
The only serious prerequisite for reading this text is a good familiarity with linear algebra, and
some prociency in basic abstract algebra (groups, elds, polynomials etc.) The necessary material
is discussed in the rst seven chapters of Artins Algebra textbook. Perhaps the only basic notion
from linear algebra well need which is not contained in standard texts is that of tensor product of
1
For more about the subject, we recommend the reader the excellent textbook of Fulton and Harris Representation
theory
3
vector spaces; we recommend the reader to solve the preparatory problems below to attain a better
familiarity with this notion.
Acknowledgments. The authors are very grateful to the Clay Mathematics Institute (and
personally to David Ellwood and Vida Salahi) for hospitality and wonderful working conditions.
They also are very indebted to the Academy fellows Josh NicholsBarrer and Victor Ostrik, whose
mathematical insights and devotion were crucial in making the Academy a success and in creating
this text.
1.1 Preparatory problems on tensor products
Recall that the tensor product V W of vector spaces V and W over a eld k is the quotient of
the space V W whose basis is given by formal symbols v w, v V , w W, by the subspace
spanned by the elements
(v
1
+v
2
)wv
1
wv
2
w, v(w
1
+w
2
)vw
1
vw
2
, (av)wa(vw), v(aw)a(vw),
where v V, w W, a k.
Problem 1.3. (a) Let U be any kvector space. Construct a natural bijection between bilinear
maps V W U and linear maps V W U.
(b) Show that if v
i
is a basis of V and w
j
is a basis of W then v
i
w
j
is a basis of
V W.
(c) Construct a natural isomorphism V
W Hom(V, W) in the case when V is nite
dimensional.
(d) Let V be a vector space over C. Let S
n
V be the subspace of V
n
(nfold tensor product of
V ) which consists of the tensors that are symmetric under permutation of components. Let
n
V be
the subspace of V
n
which consists of the tensors which are antisymmetric, i.e., s
ij
T = T, where
s
ij
is the permutation of i and j. (These spaces are called the nth symmetric, respectively exterior,
power of V ). If v
i
is a basis of V , can you construct a basis of S
n
V,
n
V ? If dimV = m, what
are their dimensions?
(e) Let A : V W be a linear operator. Then we have A
n
: V
n
W
n
, and its restrictions
S
n
A : S
n
V S
n
W,
n
A :
n
V
n
W. Suppose V = W and has dimension N, and assume that
the eigenvalues of A are
1
, ...,
N
. Find Tr(S
n
A), Tr(
n
A). In particular, show that Tr(
N
A) =
det(A).
Problem 1.4. Let J
N
be the linear operator on C
N
given in the standard basis by the formula
J
N
e
i
= e
i1
for i > 1, J
N
e
1
= 0. Thus J
N
is a Jordan block of size N.
Find the Jordan normal form of the operator B = J
N
1
M
+1
N
J
M
on C
N
C
M
, where 1
L
denotes the identity operator on C
L
.
Hint. Compute dimensions of kernels of B
j
for all j.
Problem 1.5. Hilberts problem. It is known that if A and B are two polygons of the same
area then A can be cut by nitely many straight cuts into pieces from which one can make B. David
Hilbert asked in 1900 whether it is true for polyhedra in 3 dimensions. In particular, is it true for
a cube and a regular tetrahedron of the same volume?
The answer is no, as was found by Dehn in 1901. The proof is very beautiful. Namely, to any
polyhedron A let us attach its Dehn invariant D(A) in
4
V = 1(1/) (the tensor product of vector spaces). Namely,
D(A) =
a
l(a)
(a)
,
where a runs over edges of A and l(a), (a) are the length of a and the angle at a.
(a) Show that if you cut A into B and C by a straight cut, then D(A) = D(B) +D(C).
(b) Show that = arccos(1/3)/ is not a rational number.
Hint. Let p
n
, n 1 be the sequence dened by the recursive equation p
n+1
= p
2
n
2, p
0
= 2/3.
Show that p
n
= 2cos2
n
. On the other hand, show that p
n
must take innitely many dierent
values. From this, derive that cannot be rational.
(c) Using (a) and (b), show that the answer to Hilberts question is negative. (Compute the
Dehn invariant of the regular tetrahedron and the cube).
2 Basic notions of representation theory
2.1 Algebras
Let k be a eld. Unless stated otherwise, we will assume that k is algebraically closed, i.e. any
nonconstant polynomial with coecients in k has a root in k. The main example is the eld of
complex numbers C, but we will also consider elds of characteristic p, such as the algebraic closure
F
p
of the nite eld F
p
of p elements.
Denition 2.1. An associative algebra over k is a vector space A over k together with a bilinear
map A A A, (a, b) ab, such that (ab)c = a(bc).
Denition 2.2. A unit in an associative algebra A is an element 1 A such that 1a = a1 = a.
Proposition 2.3. If a unit exists, it is unique.
Proof. Let 1, 1
= 1
.
From now on, by an algebra A we will mean an associative algebra with a unit. We will also
assume that A ,= 0.
Example 2.4. Here are some examples of algebras over k:
1. A = k.
2. A = k[x
1
, ..., x
n
] the algebra of polynomials in variables x
1
, ..., x
n
.
3. A = EndV the algebra of endomorphisms of a vector space V over k (i.e., linear maps from
V to itself). The multiplication is given by composition of operators.
4. The free algebra A = kx
1
, ..., x
n
). The basis of this algebra consists of words in letters
x
1
, ..., x
n
, and multiplication is simply concatenation of words.
5. The group algebra A = k[G] of a group G. Its basis is a
g
, g G, with multiplication law
a
g
a
h
= a
gh
.
Denition 2.5. An algebra A is commutative if ab = ba for all a, b A.
5
For instance, in the above examples, A is commutative in cases 1 and 2, but not commutative in
cases 3 (if dimV > 1), and 4 (if n > 1). In case 5, A is commutative if and only if G is commutative.
Denition 2.6. A homomorphism of algebra f : A B is a linear map such that f(xy) = f(x)f(y)
for all x, y A, and f(1) = 1.
2.2 Representations
Denition 2.7. A representation of an algebra A (also called a left Amodule) is a vector space
V together with a homomorphism of algebras : A EndV .
Similarly, a right Amodule is a space V equipped with an antihomomorphism : A EndV ;
i.e., satises (ab) = (b)(a) and (1) = 1.
The usual abbreviated notation for (a)v is av for a left module and va for the right module.
Then the property that is an (anti)homomorphism can be written as a kind of associativity law:
(ab)v = a(bv) for left modules, and (va)b = v(ab) for right modules.
Example 2.8. 1. V = 0.
2. V = A, and : A EndA is dened as follows: (a) is the operator of left multiplication by
a, so that (a)b = ab (the usual product). This representation is called the regular representation
of A. Similarly, one can equip A with a structure of a right Amodule by setting (a)b := ba.
3. A = k. Then a representation of A is simply a vector space over k.
4. A = kx
1
, ..., x
n
). Then a representation of A is just a vector space V over k with a collection
of arbitrary linear operators (x
1
), ..., (x
n
) : V V (explain why!).
Denition 2.9. A subrepresentation of a representation V of an algebra A is a subspace W V
which is invariant under all operators (a) : V V , a A.
For instance, 0 and V are always subrepresentations.
Denition 2.10. A representation V ,= 0 of A is irreducible (or simple) if the only subrepresenta
tions of V are 0 and V .
Denition 2.11. Let V
1
, V
2
be two representations over an algebra A. A homomorphism (or
intertwining operator) : V
1
V
2
is a linear operator which commutes with the action of A, i.e.
(av) = a(v) for any v V
1
. A homomorphism is said to be an isomorphism of representations
if it is an isomorphism of vector spaces.
Note that if a linear operator : V
1
V
2
is an isomorphism of representations then so is the
lienar operator
1
: V
2
V
1
(check it!).
Two representations between which there exists an isomorphism are said to be isomorphic. For
practical purposes, two isomorphic representations may be regarded as the same, although there
could be subtleties related to the fact that an isomorphism between two representations, when it
exists, is not unique.
Denition 2.12. Let V = V
1
, V
2
be representations of an algebra A. Then the space V
1
V
2
has
an obvious structure of a representation of A, given by a(v
1
v
2
) = av
1
av
2
.
Denition 2.13. A representation V of an algebra A is said to be indecomposable if it is not
isomorphic to a direct sum of two nonzero representations.
6
It is obvious that an irreducible representation is indecomposable. On the other hand, we will
see below that the converse statement is false in general.
One of the main problems of representation theory is to classify irredicible and indecomposable
representations of a given algebra up to isomorphism. This problem is usually hard and often can
be solved only partially (say, for nite dimensional representations). Below we will see a number
of examples in which this problem is partially or fully solved for specic algebras.
We will now prove our rst result Schurs lemma. Although it is very easy to prove, it is
fundamental in the whole subject of representation theory.
Proposition 2.14. (Schurs lemma) Let V
1
, V
2
be irreducible representations of an algebra A over
any eld F. Let : V
1
V
2
be a nonzero homomorphism of representations. Then is an
isomorphism.
Proof. The kernel K of is a subrepresentation of V
1
. Since ,= 0, this subrepresentation cannot
be V
1
. So by irreducibility of V
1
we have K = 0. The image I of is a subrepresentation of V
2
.
Since ,= 0, this subrepresentation cannot be 0. So by irreducibility of V
2
we have I = V
2
. Thus
is an isomorphism.
Corollary 2.15. (Schurs lemma for algebraically closed elds) Let V be a nite dimensional
irreducible representation of an algebra A over an algebraically closed eld k, and : V V is an
intertwining operator. Then = Id (the scalar operator).
Proof. Let be an eigenvalue of (a root of the characteristic polynomial of ). It exists since
k is an algebraically closed eld. Then the operator Id is an intertwining operator V V ,
which is not an isomorphism (since its determinant is zero). Thus by Schurs lemma this operator
is zero, hence the result.
Corollary 2.16. Let A be a commutative algebra. Then every irreducible nite dimensional rep
resentation V of A is 1dimensional.
Remark. Note that a 1dimensional representation of any algebra is automatically irreducible.
Proof. For any element a A, the operator (a) : V V is an intertwining operator. Indeed,
(a)(b)v = (ab)v = (ba)v = (b)(a)v
(the second equality is true since the algebra is commutative). Thus, by Schurs lemma, (a) is a
scalar operator for any a A. Hence every subspace of V is a subrepresentation. So 0 and V are
the only subspaces of V . This means that dimV = 1 (since V ,= 0).
Example 2.17. 1. A = k. Since representations of A are simply vector spaces, V = A is the only
irreducible and the only indecomposable representation.
2. A = k[x]. Since this algebra is commutative, the irreducible representations of A are its
1dimensional representations. As we discussed above, they are dened by a single operator (x).
In the 1dimensional case, this is just a number from k. So all the irreducible representations of A
are V
= spanas and S)
r
= spansb, the left, respectively right, ideal generated by S.
2.4 Quotients
Let A be an algebra and I a twosided ideal in A. Then A/I is the set of (additive) cosets of I.
Let : A A/I be the quotient map. We can dene multiplication in A/I by (a) (b) := (ab).
This is welldened because if (a) = (a
) then
(a
b) = (ab + (a
a)b) = (ab)
because (a
) then
(ab
) = (ab +a(b
b)) = (ab)
because a(b
, y
+
, y
= x
x
+
= y
+
y
= y
y
+
= 1. One customarily writes x
+
as x, x
as x
1
, y
+
as
y, and y
as y
1
.
Problem 2.24. Let A be the Weyl algebra, generated over an algebraically closed eld k by two
generators x, y with the relation yx xy = 1.
(a) If chark = 0, what are the nite dimensional representations of A? What are the twosided
ideals in A? .
(b) Suppose for the rest of the problem that chark = p. What is the center of A?
(c) Find all irreducible nite dimensional representations of A.
Problem 2.25. Let q be a nonzero complex number, and A be the algebra over C generated by
X
1
and Y
1
with dening relations XX
1
= X
1
X = 1, Y Y
1
= Y
1
Y = 1, and XY = qY X.
(a) What is the center of A for dierent q? If q is not a root of unity, what are the twosided
ideals in A?
(b) For which q does this algebra have nite dimensional representations?
(c) Find all nite dimensional irreducible representations of A for such q.
2.7 Quivers
Denition 2.26. A quiver Q is a directed graph, possibly with selfloops and/or multiple arrows
between two vertices.
Example 2.27.
//
oo
OO
10
We denote the set of vertices of the quiver Q as I, and the set of edges as E. For an edge h E,
let h
, h
h
h
//
3. p
h
a
h
= a
h
, p
i
a
h
= 0 for i ,= h
1
,= h
2
.
This algebra is called the path algebra because a basis of P
Q
is formed by elements a
, where is
a path in Q (possibly of length 0). If = h
n
h
2
h
1
= a
hn
a
h
2
a
h
1
.
If is a path of length 0, starting and ending at point i, then a
is dened to be p
i
. Clearly
a
2
a
1
= a
1
(where
2
1
is the concatenation of paths
2
and
1
) if the nal point of
1
equals
the initial point of
2
and a
2
a
1
= 0 otherwise.
We now justify our statement that a representation of a quiver is the same as a representation
of the path algebra of a quiver.
Let Vbe a representation of the path algebra P
Q
. From this representation of the algebra P
Q
, we
can construct a representation of Q as follows: let V
i
= p
i
V and let x
h
= a
h
[
p
h
V
: p
h
V p
h
V.
Similarly, let (V
i
, x
h
) be a representation of a quiver Q. From this representation, we can
construct a representation of the path algebra P
Q
: let V =
i
V
i
, let p
i
: V V
i
V be the
projection onto V
i
, and let a
h
= i
h
x
h
p
h
: V V
h
V
h
V where i
h
: V
h
V is the
inclusion map.
It is clear that the above assignments V (p
i
V) and (V
i
)
i
V
i
are inverses of each other.
Thus, we have a bijection between isomorphism classes of representations of the algebra P
Q
and of
the quiver Q.
Remark 2.31. In practice, it is generally easier to consider a representation of a quiver as in
Denition 2.28. The above serves to show, as stated before, that the theory of representations of
quivers is a part of the larger theory of representations of algebras.
We lastly dene several previous concepts in the context of quivers representations.
Denition 2.32. A subrepresentation of a representation (V
i
, x
h
) of a quiver Q is a representation
(W
i
, x
h
) where W
i
V
i
for all i I and where x
h
(W
h
) W
h
and x
h
= x
h
[
W
h
: W
h
W
h
for
all h E.
11
Denition 2.33. The direct sum of two representations (V
i
, x
h
) and (W
i
, y
h
) is the representation
(V
i
W
i
, x
h
y
h
).
As with representations of algebras, a representation (V
i
) of a quiver Q is said to be irreducible
if its only subrepresentations are (0) and (V
i
) itself, and indecomposable if it is not isomorphic to
a direct sum of two nonzero representations.
Denition 2.34. Let (V
i
, x
h
) and (W
i
, y
h
) be representations of the quiver Q. A homomorphism
: (V
i
) (W
i
) of quiver representations is a collection of maps
i
: V
i
W
i
such that
y
h
h
=
h
x
h
for all h E.
Problem 2.35. Let A be a Z
+
graded algebra, i.e., A =
n0
A[n], and A[n] A[m] A[n + m].
If A[n] is nite dimensional, it is useful to consider the Hilbert series h
A
(t) =
dimA[n]t
n
(the
generating function of dimensions of A[n]). Often this series converges to a rational function, and
the answer is written in the form of such function. For example, if A = k[x] and deg(x
n
) = n then
h
A
(t) = 1 +t +t
2
+... +t
n
+... =
1
1 t
Find the Hilbert series of:
(a) A = k[x
1
, ..., x
m
] (where the grading is by degree of polynomials);
(b) A = k < x
1
, ..., x
m
> (the grading is by length of words);
(c) A is the exterior algebra
k
[x
1
, ..., x
m
], generated by x
1
, ..., x
m
with the dening relations
x
i
x
j
+x
j
x
i
= 0 for all i, j (grading is by degree).
(d) A is the path algebra P
Q
of a quiver Q as dened in the lectures.
Hint. The closed answer is written in terms of the adjacency matrix M
Q
of Q.
2.8 Lie algebras
Let g be a vector space over a eld k, and let [ , ] : g g g be a skewsymmetric bilinear map.
(So [a, b] = [b, a].) If k is of characteristic 2, we also require that [x, x] = 0 for all x (a requirement
equivalent to [a, b] = [b, a] in elds of other characteristics).
Denition 2.36. (g, [ , ]) is a Lie algebra if [ , ] satises the Jacobi identity
_
[a, b] , c
+
_
[b, c] , a
+
_
[c, a] , b
= 0. (2)
Example 2.37. Some examples of Lie algebras are:
1. 1
3
with [u, v] = u v, the crossproduct of u and v
2. Any space g with [ , ] = 0 (abelian Lie algebra)
3. Any associative algebra A with [a, b] = ab ba
4. Any subspace U of an associative algebra A such that [a, b] U for all a, b U
5. sl(n), the set of n n matrices with trace 0
For example, sl(2) has the basis
e =
_
0 1
0 0
_
f =
_
0 0
1 0
_
h =
_
1 0
0 1
_
with relations [e, f] = h, [h, f] = 2f, [h, e] = 2e.
12
6. The Heisenberg Lie algebra H of matrices
_
0
0 0
0 0 0
_
It has the basis
x =
_
_
0 0 0
0 0 1
0 0 0
_
_
y =
_
_
0 1 0
0 0 0
0 0 0
_
_
c =
_
_
0 0 1
0 0 0
0 0 0
_
_
with relations [y, x] = c and [y, c] = [x, c] = 0.
7. The algebra a(1) of matrices (
0 0
)
Its basis consists of X = (
1 0
0 0
) and Y = (
0 1
0 0
), with [X, Y ] = Y .
8. so(n), the space of skewsymmetric n n matrices, with [a, b] = ab ba
Denition 2.38. Let g
1
, g
2
be Lie algebras. A homomorphism : g
1
g
2
of Lie algebras is a
linear map such that ([a, b]) = [(a), (b)].
Denition 2.39. A representation of a Lie algebra g is a vector space V with a homomorphism
of Lie algebras : g EndV .
Example 2.40. Some examples of representations of Lie algebras are:
1. V=0
2. Any vector space V with = 0
3. Adjoint representation V = g with (a)(b) = [a, b]
def
= ab ba
That this is a representation follows from Equation (2).
It turns out that a representation of a Lie algebra g is the same as a representation of a certain
associative algebra (g). Thus, as with quivers, we can view the theory of representations of Lie
algebras as part of the theory of representations of associative algebras.
Denition 2.41. Let g be a Lie algebra with basis x
i
and [ , ] dened by [x
i
, x
j
] =
k
c
k
ij
x
k
. The
universal enveloping algebra (g) is the associative algebra generated by the x
i
s with the
relations x
i
x
j
x
j
x
i
=
k
c
k
ij
x
k
.
Example 2.42. The associative algebra (sl(2)) is the algebra generated by e, f, h with relations
he eh = 2e hf fh = 2f ef fe = h.
Example 2.43. The algebra (H), where H is the Heisenberg Lie algebra of Example 2.37.6, is
the algebra generated by x, y, c with the relations
yx xy = c yc cy = 0 xc cx = 0.
The Weyl algebra is the quotient of (H) by the relation c = 1.
Finally, let us dene the important notion of tensor product of representations.
Denition 2.44. The tensor product of two representations V, W of a Lie algebra g is the space
V W with
V W
(x) =
V
(x) Id +Id
W
(x).
It is easy to check that this is indeed a representation.
13
Problem 2.45. Representations of sl(2). According to the above, a representation of sl(2) is
just a vector space V with a triple of operators E, F, H such that HE EH = 2E, HF FH =
2F, EF FE = H (the corresponding map is given by (e) = E, (f) = F, (h) = H.
Let V be a nite dimensional representation of sl(2) (the ground eld in this problem is C).
(a) Take eigenvalues of H and pick one with the biggest real part. Call it . Let
V () be the
generalized eigenspace corresponding to . Show that E[
V ()
= 0.
(b) Let W be any representation of sl(2) and w W be a nonzero vector such that Ew = 0.
For any k > 0 nd a polynomial P
k
(x) of degree k such that E
k
F
k
w = P
k
(H)w. (First compute
EF
k
w, then use induction in k).
(c) Let v
V () be a generalized eigenvector of H with eigenvalue . Show that there exists
N > 0 such that F
N
v = 0.
(d) Show that H is diagonalizable on
V (). (Take N to be such that F
N
= 0 on
V (), and
compute E
N
F
N
v, v
V (), by (b). Use the fact that P
k
(x) does not have multiple roots).
(e) Let N
v
be the smallest N satisfying (c). Show that = N
v
1.
(f ) Show that for each N > 0, there exists a unique up to isomorphism irreducible representation
of sl(2) of dimension N. Compute the matrices E, F, H in this representation using a convenient
basis. (For V nite dimensional irreducible take as in (a) and v V () an eigenvector of H.
Show that v, Fv, ..., F
.
(g) Show that the operator C = EF + FE + H
2
/2 (the socalled Casimir operator) commutes
with E, F, H and equals
(+2)
2
Id on V
.
Now it will be easy to prove the direct sum decomposition. Assume the contrary, and let V be
a representation of the smallest dimension, which is not a direct sum of smaller representations.
(h) Show that C has only one eigenvalue on V , namely
(+2)
2
for some nonnegative integer .
(use that the generalized eigenspace decomposition of C must be a decomposition of representations).
(i) Show that V has a subrepresentation W = V
v
i
). Show that V
i
are subrepresentations of V and derive a
contradiction with the fact that V cannot be decomposed.
(l) (JacobsonMorozov Lemma) Let V be a nite dimensional complex vector space and A : V
V a nilpotent operator. Show that there exists a unique, up to an isomorphism, representation of
sl(2) on V such that E = A. (Use the classication of the representations and Jordan normal form
theorem)
(m) (ClebschGordan decomposition) Find the decomposition into irreducibles of the represen
tation V
of sl(2).
Hint. For a nite dimensional representation V of sl(2) it is useful to introduce the character
V
(x) = Tr(e
xH
), x C. Show that
V W
(x) =
V
(x) +
W
(x) and
V W
(x) =
V
(x)
W
(x).
14
Then compute the character of V
and of V
1
2
r
: A
r
i=1
EndV
i
is surjective.
Proof. Let
V
N
1
,...,Nr
= (V
1
V
1
)
. .
N
1
copies
(V
2
V
2
)
. .
N
2
copies
(V
r
V
r
)
. .
Nr copies
.
Let p
ij
: V
N
1
,...,Nr
V
i
be the projection onto V
ij
, the j
th
copy of V
i
.
We shall need the following two lemmas:
15
Lemma 3.2. Let W _ V = V
N
1
,...,Nr
be a subrepresentation. Then, there exists an automorphism
of V such that p
iN
i
((W)) = 0 for some i.
Proof of Lemma 3.2. We induct on N =
r
i=1
N
i
. The base case, N = 0, is clear. For the inductive
step, rst pick an irreducible nonzero subrepresentation Y of W (which clearly exists by Problem
2.18). As Y ,= 0, there exist some i, j such that p
ij
[
Y
: Y V
i
is nonzero. Assume, without loss of
generality, that j = 1. As p
i1
[
Y
is nonzero, by Schurs lemma, p
i1
[
Y
is an isomorphism, and Y
= V
i
.
As Y can be isomorphic to only one V
l
, p
lm
[
Y
= 0 for all l ,= i, m. As p
im
[
Y
(p
i1
[
Y
)
1
: V
i
Y
V
i
is a homomorphism, by Schurs lemma for algebraically closed elds, p
im
[
Y
(p
i1
[
Y
)
1
=
m
Id
and p
im
[
Y
=
m
p
i1
[
Y
for some scalar
m
.
We now dene an automorphism : V V. For v V, we write v = (v
lm
) where v
lm
=
p
lm
(v) V
l
. We now let (v) = (v
lm
), where
v
lm
= v
lm
for l ,= i,
v
i1
= v
i1
,
v
im
= v
im
m
v
i1
for 2 m N
i
.
Clearly is an automorphism. Suppose that v Y . As p
lm
[
Y
= 0 for all l ,= i, m, v
lm
= v
lm
=
p
lm
(v) = 0 for l ,= i, m. As p
im
[
Y
=
m
p
i1
[
Y
, v
im
=
m
v
i1
and v
im
= v
im
m
v
i1
= 0 for m ,= 1.
Thus, p
lm
((Y )) = v
lm
= 0 unless l = i and m = 1. Also, p
i1
((Y ))
= Y .
Consider the map
: V = V
N
1
,...,N
i
,,...,Nr
V
N
1
,...,N
i
1,...,Nr
(v
lm
) (v
lm
for (l, m) ,= (i, 1))
which has V
i1
(the rst copy of V
i
) as its kernel. But this is also (Y ), so ker = (Y ). As
(Y ) (W), ker [
(W)
= (Y ) and ((W))
= (W)/(Y )
= W/Y .
By the inductive hypothesis, there exists some : V
N
1
,...,N
i
1,...,Nr
V
N
1
,...,N
i
1,...,Nr
such
that for some l, p
lN
l
((((W)))) = 0, where N
l
= N
l
for l ,= i and N
i
= N
i
1. Dene
= (Id
V
i1
) (where Id
V
i1
is the identity on the rst copy of V
i
). As Id
V
i1
= Id
V
,
Id
V
i1
= (Id
V
i1
0) + ( ). Thus, p
lN
l
((W)) = p
lN
l
((((W)))) = 0 (as (l, N
l
) ,= (i, 1)).
This completes the inductive step.
Lemma 3.3. Let W V
N
1
,...,Nr
be a subrepresentation. Then, W
= V
M
1
,...,Mr
for some M
i
N
i
.
Proof of Lemma 3.3. We induct on N =
r
i=1
N
i
. The base case, N = 0 is clear, as then W = 0.
If W = V
N
1
,...,Nr
, then we simply have M
i
= N
i
. Otherwise, by Lemma 3.2, there exists an
automorphism of V
N
1
,...,Nr
such that p
iN
i
((W)) = 0 for some i. Thus, (W) V
N
1
,...,N
i
1,...,Nr
.
By the inductive assumption, W
= (W)
= V
M
1
,...,Mr
.
Proof of Theorem 3.1. First, by replacing A with A/ ker
1
2
r
, we can assume, without
loss of generality, that the map
1
2
r
is injective. As A is now isomorphic to its
image, we can also assume that A End V
1
EndV
r
. Thus, A is a subrepresentation of
EndV
1
EndV
r
.
Let d
i
= dimV
i
. As EndV
i
= V
i
V
i
. .
d
i
copies
, we have A V
d
1
,...,dr
. Thus, by Lemma 3.3,
A
= V
M
1
,...,Mr
for some M
i
d
i
. Thus, dimA =
r
i=1
M
i
d
i
.
16
Next, consider End
A
(A). As the V
i
s are pairwise nonisomorphic, by Schurs lemma, no copy
of V
i
in A can be mapped to a distinct V
j
. Also, by Schur, End
A
(V
i
) = k. Thus, End
A
(A)
=
i
Mat
M
i
(k), so dimEnd
A
(A) =
r
i=1
M
2
i
. By Problem 2.20, End
A
(A)
= A, so
r
i=1
M
2
i
=
dimA =
r
i=1
M
i
d
i
. Thus,
r
i=1
M
i
(d
i
M
i
) = 0. As M
i
d
i
, d
i
M
i
0. Next, as 1 A, the
map
i
: A EndV
i
is nontrivial. As EndV
i
is a direct sum of copies of V
i
, A must contain a copy
of V
i
. Thus M
i
> 0, and we must have d
i
M
i
= 0 for all i, so M
i
= d
i
, and A =
r
i=1
EndV
i
.
3.2 Representations of matrix algebras
In this section we consider representations of algebras A =
i
Mat
d
i
(k).
Theorem 3.4. Let A =
r
i=1
Mat
d
i
(k). Then the irreducible representations of A are V
1
=
k
d
1
, . . . , V
r
= k
dr
, and any nite dimensional representation of A is a direct sum of copies of
V
1
, . . . , V
r
.
In order to prove Theorem 3.4, we shall need the notion of a dual representation.
Denition 3.5. (Dual representation) Let V be a representation of A. Then the dual space V
is
a right Amodule (or equivalently, a representation of A
op
) with the action
f a = (v f(av)) ,
as (f (ab)) (v) = f((ab)v) = f(a(bv)) = (f a)(bv) = ((f a) b)(v).
Proof of Theorem 3.4. First, the given representations are clearly irreducible, as for any v, w
V
i
0, there exists a A such that av = w. Next, let X be an n dimensional representation
of A. Then, X
is an n dimensional representation of A
op
. But (Mat
d
i
(k))
op
= Mat
d
i
(k) with
isomorphism (X) = X
T
, as (AB)
T
= B
T
A
T
. Thus, A
= A
op
and X
is an n dimensional
representation of A. Dene
: A A
. .
n copies
X
by
(a
1
, . . . , a
n
) = a
1
x
1
+ +a
n
x
n
where x
i
is a basis of X
: X A
n
is
injective. But A
n
= A
n
. Hence, Im
= X is a subrepresentation of A
n
. Next, as Mat
d
i
(k) = V
d
i
i
,
A = V
d
1
,...,dr
and A
n
= V
nd
1
,...,ndr
. By Lemma 3.3, X
= V
M
1
,...,Mr
= M
1
V
1
M
r
V
r
for some
M
i
as desired.
3.3 Finite dimensional algebras
Denition 3.6. The radical I of a nite dimensional algebra A is the set of all elements of A
which act by 0 in all irreducible representations of A. It is denoted Rad(A).
Proposition 3.7. Rad(A) is a twosided ideal.
Proof. If i I, a A, v V , where V is any irreducible representation of A, then aiv = a 0 = 0
and iav = iv
= 0 where v
= av.
17
Theorem 3.8. A nite dimensional algebra A has only nitely many irreducible representations
V
i
up to isomorphism, these representations are nite dimensional, and
A/I
=
i
EndV
i
,
where I = Rad(A).
Proof. First, for any irreducible representation V of A, and for any nonzero v V , Av V is a
nite dimensional subrepresentation of V . (It is nite dimensional as A is nite dimensional.) As
V is irreducible and Av ,= 0, V = Av and V is nite dimensional.
Next, suppose that we had innitely many nonisomorphic irreducible representations. Let
V
1
, V
2
, . . . , V
r
be any r nontrivial nonisomorphic irreducible representations, with r > dimA. By
Theorem 3.1, the homomorphism
i
: A
i
EndV
i
is surjective. But this is impossible as
dimEndV
i
r > dimA. Thus, A has only nitely many
nonisomorphic irreducible representations.
Next, let V
1
, V
2
, . . . , V
r
be all nonisomorphic irreducible nite dimensional representations of
A. By Theorem 3.1, the homomorphism
i
: A
i
EndV
i
is surjective. The kernel of this map is exactly I.
Corollary 3.9.
i
(dimV
i
)
2
dimA, where the V
i
s are the irreducible representations of A.
Proof. As dimEndV
i
= (dimV
i
)
2
, Theorem 3.8 implies that dimA dimI =
i
dimEndV
i
=
i
(dimV
i
)
2
. As dimI 0,
i
(dimV
i
)
2
dimA.
Denition 3.10. A nite dimensional algebra A is said to be semisimple if Rad(A) = 0.
Proposition 3.11. For a nite dimensional algebra A over an algebraically closed eld k, the
following are equivalent:
1. A is semisimple
2.
i
(dimV
i
)
2
= dimA, where the V
i
s are the irreducible representations of A
3. A
=
i
Mat
d
i
(k) for some d
i
4. Any nite dimensional representation of A is completely reducible (that is, isomorphic to a
direct sum of irreducible representations)
5. A (as a vector space) is a completely reducible representation of A
Proof. As dimA dimI =
i
(dimV
i
)
2
, clearly dimA =
i
(dimV
i
)
2
if and only if I = 0. Thus,
(1) (2).
Next, by Theorem 3.8, if I = 0, then clearly A
=
i
Mat
d
i
(k) for d
i
= dimV
i
. Thus, (1) (3).
Conversely, if A
=
i
Mat
d
i
(k), then A
=
i
EndU
i
for some U
i
s with dimU
i
= d
i
. Clearly
18
each U
i
is irreducible (as for any u, u
U
i
0, there exists a A such that au = u
), and the
U
i
s are pairwise nonisomorphic representations. Thus, the U
i
s form a subset of the irreducible
representations V
i
of A. Thus, dimA =
i
(dimU
i
)
2
i
(dimV
i
)
2
dimA. Thus, (3) (2)(
(1)).
Next (3) (4) by Theorem 3.4. Clearly (4) (5). To see that (5) (3), let A =
i
n
i
V
i
.
Consider End
A
(A) (endomorphisms of A as a representation of A). As the V
i
s are pairwise non
isomorphic, by Schurs lemma, no copy of V
i
in A can be mapped to a distinct V
j
. Also, by Schur,
End
A
(V
i
) = k. Thus, End
A
(A)
=
i
Mat
n
i
(k). But End
A
(A)
= A
op
by Problem 2.20, so A
op
=
i
Mat
n
i
(k). Thus, A
= (
i
Mat
n
i
(k))
op
=
i
(Mat
n
i
(k))
op
. But (Mat
n
i
(k))
op
= Mat
n
i
(k) with
isomorphism (X) = X
T
, as (AB)
T
= B
T
A
T
. Thus, A
=
i
Mat
n
i
(k).
Let A be an algebra and V a nitedimensional representation of A with action . Then the
character of V is the linear function
V
: A k given by
V
(a) = tr[
V
((a)).
If [A, A] is the span of commutators [x, y] := xy yx over all x, y A, then [A, A] ker
V
. Thus,
we may view the character as a mapping
V
: A/[A, A] k.
Theorem 3.12. (1) Characters of irreducible nitedimensional representations of A are linearly
independent. (2) If A is a nitedimensional semisimple algebra, then the characters form a basis
of (A/[A, A])
.
Proof. (1) If V
1
, . . . , V
r
are nonisomorphic irreducible nitedimensional representations of A, then
V
1
Vr
: A End V
1
End V
r
is surjective by the density theorem, so
V
1
, . . . ,
Vr
are
linearly independent. (Indeed, if
V
i
(a) = 0 for all a A, then
i
Tr(M
i
) = 0 for all M
i
End
k
V
i
. But each tr(M
i
) can range independently over k, so it must be that
1
= =
r
= 0.)
(2) First we prove that [Mat
d
(k), Mat
d
(k)] = sl
d
(k), the set of all matrices with trace 0. It is
clear that [Mat
d
(k), Mat
d
(k)] sl
d
(k). If we denote by E
ij
the matrix with 1 in the ith row of
the jth column and 0s everywhere else, we have [E
ij
, E
jm
] = E
im
for i ,= m, and [E
i,i+1
, E
i+1,i
] =
E
ii
E
i+1,i+1
. Now E
im
E
ii
E
i+1,i+1
forms a basis in sl
d
(k), and indeed [Mat
d
(k), Mat
d
(k)] =
sl
d
(k), as claimed.
By semisimplicity, we can write A = Mat
d
1
k Mat
dr
k. Then [A, A] = sl
d
1
(k) sl
dr
(k),
and A/[A, A]
= k
r
. By the corollary to the density theorem, there are exactly r irreducible
representations of A (isomorphic to k
d
1
, . . . , k
dr
, respectively), and therefore r linearly independent
characters in the rdimensional vector space A/[A, A]. Thus, the characters form a basis.
3.4 JordanHolder and KrullSchmidt theorems
To conclude the discussion of associative algebras, let us state two important theorems about their
nite dimensional representations.
Let A be an algebra over an algebraically closed eld k. Let V be a representation of A. A
(nite) ltration of A is a sequence of subrepresentations 0 = V
0
V
1
... V
n
= V .
Theorem 3.13. (JordanHolder theorem). Let V be a nite dimensional representation of A,
and 0 = V
0
V
1
... V
n
= V , 0 = V
0
... V
m
= V be ltrations of V , such that the
representations W
i
:= V
i
/V
i1
and W
i
:= V
i
/V
i1
are irreducible for all i. Then n = m, and there
exists a permutation of 1, ..., n such that W
(i)
is isomorphic to W
i
.
19
Proof. First proof (for k of characteristic zero). Let I A be the annihilating ideal of V (i.e.
the set of elements that act by zero in V ). Replacing A with A/I, we may assume that A is nite
dimensional. The character of V obviously equals the sum of characters of W
i
, and also the sum of
characters of W
i
. But by Theorem 3.12, the characters of irreducible representations are linearly
independent, so the multiplicity of every irreducible representation W of A among W
i
and among
W
i
are the same. This implies the theorem.
Second proof (general). The proof is by induction on dimV . The base of induction is clear,
so let us prove the induction step. If W
1
= W
1
(as subspaces), we are done, since by the induction
assumption the theorem holds for V/W
1
. So assume W
1
,= W
1
. In this case W
1
W
1
= 0 (as
W
1
, W
1
are irreducible), so we have an embedding f : W
1
W
1
V . Let U = V/(W
1
W
1
), and
0 = U
0
U
1
... U
p
= U be a ltration of U with simple quotients Z
i
= U
i
/U
i1
. Then we see
that:
1) V/W
1
has a ltration with successive quotients W
1
, Z
1
, ..., Z
p
, and another ltration with
successive quotients W
2
, ...., W
n
.
2) V/Y has a ltration with successive quotients W
1
, Z
1
, ..., Z
p
, and another ltration with
successive quotients W
2
, ...., W
n
.
By the induction assumption, this means that the collection of irreducible modules with mul
tiplicities W
1
, W
1
, Z
1
, ..., Z
p
coincides on one hand with W
1
, ..., W
n
, and on the other hand, with
W
1
, ..., W
m
. We are done.
Theorem 3.14. (KrullSchmidt theorem) Any nite dimensional representation of A can be uniquely
(up to order of summands) decomposed into a direct sum of indecomposable representations.
Proof. It is clear that a decomposition of V into a direct sum of indecomposable representation
exists, so we just need to prove uniqueness. We will prove it by induction on dimV . Let V =
V
1
... V
m
= V
1
... V
n
. Let i
s
: V
s
V , i
s
: V
s
V , p
s
: V V
s
, p
s
: V V
s
be the natural
maps associated to these decompositions. Let
s
= p
1
i
s
p
s
i
1
: V
1
V
1
. We have
n
s=1
s
= 1. Now
we need the following lemma.
Lemma 3.15. Let W be a nite dimensional indecomposable representation of A. Then
(i) Any homomorphism : W W is either an isomorphism of nilpotent;
(ii) If
s
: W W, s = 1, ..., n are nilpotent homomorphisms, then so is :=
1
+... +
n
.
Proof. (i) Generalized eigenspaces of are subrepresentations of V , and V is their direct sum.
Thus, can have only one eigenvalue . If is zero, is nilpotent, otherwise it is an isomorphism.
(ii) The proof is by induction in n. The base is clear. To make the induction step (n 1 to n),
assume that is not nilpotent. Then by (i) is an isomorphism, so
n
i=1
i
= 1. The morphisms
i
are not isomorphisms, so they are nilpotent. Thus 1 theta
1
n
=
1
1
+ ... +
1
n1
is
an isomorphism, which is a contradiction with the induction assumption.
By the lemma, we nd that for some s,
s
must be an isomorphism; we may assume that s = 1.
In this case, V
1
= Imp
1
i
1
Ker(p
1
i
1
), so since V
1
is indecomposable, we get that f := p
1
i
1
: V
1
V
1
and g := p
1
i
1
: V
1
V
1
are isomorphisms.
Let B =
j>1
V
j
, B
=
j>1
V
j
; then we have V = V
1
B = V
1
B
attached to these
decompositions. We claim that h is an isomorphism. To show this, it suces to show that Kerh = 0
20
(as h is a map between spaces of the same dimension). Assume that v Kerh B. Then v V
1
.
On the other hand, the projection of v to V
1
is zero, so gv = 1. Since g is an isomorphism, we get
v = 0, as desired.
Now by the induction assumption, m = n, and V
j
= V
(j)
for some permutation of 2, ..., n.
The theorem is proved.
3.5 Problems
Problem 3.16. Extensions of representations.
Let A be an algebra over an algebraically closed eld k, and V, W be a pair of representations
of A. We would like to classify representations U of A such that V is a subrepresentation of U,
and U/V = W. Of course, there is an obvious example U = V W, but are there any others?
Suppose we have a representation U as above. As a vector space, it can be (nonuniquely)
identied with V W, so that for any a A the corresponding operator
U
(a) has block triangular
form
U
(a) =
_
V
(a) f(a)
0
W
(a)
_
,
where f : A Hom
k
(W, V ).
(a) What is the necessary and sucient condition on f(a) under which
U
(a) is a repre
sentation? Maps f satisfying this condition are called (1)cocycles (of A with coecients in
Hom
k
(W, V )). They form a vector space denoted Z
1
(W, V ).
(b) Let X : W V be a linear map. The coboundary of X, dX, is dened to be the function A
Hom
k
(W, V ) given by dX(a) =
V
(a)XX
W
(a). Show that dX is a cocycle, which vanishes if and
only if X is a homomorphism of representations. Thus coboundaries form a subspace B
1
(W, V )
Z
1
(W, V ), which is isomorphic to Hom
k
(W, V )/Hom
A
(W, V ). The quotient Z
1
(W, V )/B
1
(W, V ) is
denoted Ext
1
(W, V ).
(c) Show that if f, f
Z
1
(W, V ) and f f
B
1
(W, V ) then the corresponding extensions
U, U
is an isomorphism which
acts as the identity on V and projects onto Id
W
, then f f
B
1
(V, W). Such an isomorphism is
called an isomorphism of extensions. Thus, the space Ext
1
(W, V ) classies extensions of W
by V .
(d) Assume that W, V are nitedimensional irreducible representations of A. For any f
Ext
1
(W, V ), let U
f
be the corresponding extension. Show that U
f
is isomorphic to U
f
as repre
sentations if and only if f and f
n
+...,
where
i
: A End(V ) are linear maps,
0
= , and (ab) = (a) (b).
If b(t) = 1 +b
1
t +b
2
t
2
+..., where b
i
End(V ), and is a formal deformation of , then b b
1
is also a deformation of , which is said to be isomorphic to .
(a) Show that if Ext
1
(V, V ) = 0, then any deformation of is trivial, i.e. isomorphic to .
(b) Is the converse to (a) true? (consider the algebra of dual numbers A = k[x]/x
2
).
4 Representations of nite groups: basic results
4.1 Maschkes Theorem
Theorem 4.1 (Maschke). Let G be a nite group and k an algebraically closed eld whose charac
teristic does not divide [G[. Then the algebra k[G] is semisimple. In particular, k[G] =
i
EndV
i
,
where V
i
are the irreducible representation of G.
Proof. We need to show that if V is a nitedimensional representation of G and W V is any
subrepresentation, then there exists a subrepresentation W
V such that V = W W
as the
direct sum of representations.
Choose any complement
W of W in V . (Thus V = W
W as vector spaces, but not necessarily
as representations.) Let P be the projection along
W onto W, i.e., the operator on V dened by
P[
W
= Id and P[
W
= 0. Let
P :=
1
[G[
gG
(g)P(g
1
),
where (g) is the action of g on V , and let
W
= ker P.
Now P[
W
= Id and P(V ) W, so P
2
= P, so P is a projection along W
. Thus, V = W W
as
vector spaces.
Moreover, for any h G and any y W
,
P(h)y =
1
[G[
gG
(g)P(g
1
h)y =
1
[G[
G
(h)P(
1
)y = (h)Py = 0,
so (h)y ker P = W
. Thus, W
r
i=1
End V
i
, where the V
i
are irreducible representations and V
1
= k is the
trivial onedimensional representation. Then
k[G] = k
r
i=2
End V
i
= k
r
i=2
d
i
V
i
,
where d
i
= dimV
i
. By Schurs Lemma,
Hom
k[G]
(k, k[G]) = k
Hom
k[G]
(k[G], k) = k,
for nonzero homomorphisms : k[G] k and : k k[G] unique up to scaling. We can take
such that (g) = 1 for all g G, and such that (1) =
gG
g. Then
(1) =
_
gG
g
_
=
gG
1 = [G[.
If [G[ = 0, then has no left inverse, a contradiction.
4.2 Characters
If V is a nitedimensional representation of a nite group G, then its character is dened by the
formula
V
(g) = tr[
V
((g)). Obviously,
V
(g) is simply the restriction of the character
V
(a) of
V as a representation of the algebra A = k[G] to the basis G A, so it carries exactly the same
information. The character is a central or class function:
V
(g) depends only on the conjugacy
class of g; i.e.,
V
(hgh
1
) =
V
(g).
Theorem 4.3. If the characteristic of k does not divide [G[, characters of irreducible representa
tions of G form a basis in the space F
c
(G, k) of class functions on G.
Proof. By the Maschke theorem, k[G] is semisimple, so by Theorem 3.12, the characters are linearly
independent and are a basis of (A/[A, A])
= Hom
k
(k[G], k) [ gh hg ker g, h G
V Irrep G
(dimV )V,
where the direct sum is taken over all nonisomorphic irreducible representations.
23
Proof. By the Maschke theorem we can write C[G] as a direct sum of irreducible representations:
C[G]
=
V Irrep G
n
V
V.
By the corollary to Schurs lemma, an endomorphism of representations on V must be a scalar
times the identity. For any irreducible representation V
,= V , any homomorphism of representa
tions from V
to V must be 0. Thus,
dimHom
G
(V
, V ) =
_
1 if V
= V
0 else.
Comparing dimensions in
Hom
G
(C[G], V ) =
V
Irrep G
n
V
Hom
G
(V
, V )
gives dimHom
G
(C[G], V ) = n
V
. Furthermore,
Hom
G
(C[G], V ) = : G V (hg) =
V
(h)(g),
so such are of the form(h) =
V
(h)v for an arbitrary vector v V . It follows that dimHom
G
(C[G], V ) =
dimV .
Thus, n
V
= dimV , and the theorem is proved.
4.4 Examples
The following are examples of representations of nite groups over C.
1. Finite abelian groups G = Z
n
1
Z
n
k
. Let G
is an abelian group: if
1
,
2
: G C
are irreducible
representations then so are
1
(g)
2
(g) and
1
(g)
1
. G
by (m) = e
2im/n
. Then Z
n
=
k
: k = 0, . . . , n 1,
so Z
n
= Z
n
. In general,
(G
1
G
2
G
n
)
= G
1
G
2
G
n
,
so G
= G for any nite abelian group G. This isomorphism is, however, noncanonical:
the particular decomposition of G as Z
n
1
Z
n
k
is not unique as far as which elements
of G correspond to Z
n
1
, etc. is concerned. On the other hand, G
= (G
is a canonical
isomorphism, given by : G (G
.
24
The 2dimensional representation can be visualized as representing the symmetries of the
equilateral triangle with vertices 1, 2, 3 at the points (cos 120
, sin 120
), (cos 240
, sin 240
),
(1, 0) of the coordinate plane, respectively. Thus, for example,
((12)) =
_
1 0
0 1
_
, ((123)) =
_
cos 120
sin120
sin 120
cos 120
_
.
To show that this representation is irreducible, consider any subrepresentation V . V must be
the span of a subset of the eigenvectors of ((12)), which are the nonzero multiples of (1, 0)
and (0, 1). V must also be the span of a subset of the eigenvectors of ((123)), which are the
nonzero multiples of (1, i) and (i, 1). Thus, V must be either C
2
or 0.
3. The quaternion group Q
8
= 1, i, j, k, with dening relations
i = jk = kj, j = ki = ik, k = ij = ji, 1 = i
2
= j
2
= k
2
.
The 5 conjugacy classes are 1, 1, i, j, k, so there are 5 dierent irreducible
representations, the sum of the squares of whose dimensions is 8, so their dimensions must
be 1, 1, 1, 1, and 2.
The center Z(Q
8
) is 1, and Q
8
/Z(Q
8
)
= Z
2
Z
2
. The four 1dimensional irreducible
representations of Z
2
Z
2
can be pulled back to Q
8
. That is, if q : Q
8
Q
8
/Z(Q
8
) is the
quotient map, and any representation of Q
8
/Z(Q
8
), then q gives a representation of Q
8
.
The 2dimensional representation is V = C
2
, given by (1) = Id and
(i) =
_
0 1
1 0
_
, (j) =
_
1 0
0
1
_
, (k) =
_
0
1 0
_
.
These are the Pauli matrices, which arise in quantum mechanics.
4. The symmetric group S
4
. The order is 24, and there are 5 conjugacy classes: e, (12), (123), (1234), (12)(34).
Thus the sum of the squares of the dimensions of 5 irreducible representations is 24. As with
S
3
, there are two of dimension 1: the trivial and sign representations, C and C
1
. The other
three must have dimensions 2, 2, and 3. Because S
3
= S
4
/V , where V is the Viergruppe
e, (12)(34), (13)(24), (14)(23), the 2dimensional representation of S
3
can be pulled back to
the 2dimensional representation of S
4
, which we will call C
2
.
We can consider S
4
as the group of rotations of a cube (acting by permuting the interior
diagonals); this gives the 3dimensional representation C
3
+
.
The last 3dimensional representation is C
3
, the product of C
3
+
with the sign representation,
or equivalently the permutation group of a regular tetrahedron. C
3
+
and C
3
= (1)
3
= 1.
4.5 Duals and tensor products of representations
If V is a representation of a nite group G, then V
is also a representation, via
V
(g) = (
V
(g)
)
1
= (
V
(g)
1
)
=
V
(g
1
)
.
The character is
V
(g) =
V
(g
1
).
For complex representations,
V
(g) =
i
, where the
i
are the eigenvalues of g in V . These
eigenvalues must be roots of unity because (g)
G
= (g
G
) = (e) = Id. Thus
V
(g) =
V
(g
1
) =
1
i
=
i
=
i
=
V
(g).
25
In particular, V
= V
V W
(g) =
V
(g)
W
(g).
It is an interesting problem to decompose V W into the direct sum of irreducible representations.
4.6 Orthogonality of characters
We dene the Hermitian inner product form on F
c
(G, C) (the space of central functions) by
(f
1
, f
2
) =
1
[G[
gG
f
1
(g)f
2
(g).
The following theorem says that characters of irreducible representations of G form an orthonormal
basis of F
c
(G, C) under this inner product.
Theorem 4.7. For any representations V, W
(
V
,
W
) = dimHom(V, W),
and
(
V
,
W
) =
_
1, if V
= W,
0, if V W
if V, W are irreducible.
Proof. By the denition
(
V
,
W
) =
1
[G[
gG
V
(g)
W
(g) =
1
[G[
gG
V
(g)
W
(g)
=
1
[G[
gG
V W
(g) = tr [
V W
(P),
where P =
1
G
gG
g Z(C[G]). (Here Z(C[G]) denotes the center of C[G]). If X is an irreducible
representation of G then
P[
X
=
_
Id, if X = C,
0, X ,= C.
Therefore, for any representation X the operator P[
X
is the Ginvariant projector onto the subspace
X
G
of Ginvariants in X. Thus,
tr [
V W
(P) = dimHom
G
(C, V W
)
= dim(V W
)
G
= dimHom
G
(W, V ).
Here is another orthogonality formula for characters, in which summation is taken over irre
ducible representations rather than group elements.
26
Theorem 4.8. Let g, h G, and let Z
g
denote the centralizer of g in G. Then
V
(g)
V
(h) =
[Z
g
[ifgisconjugatetoh
0, otherwise.
where the summation is taken over all irreducible representations of G.
Proof. As noted above,
V
(h) =
V
(h), so the left hand side equals (using Maschkes theorem):
V
(g)
V
(h) = tr[
C[G]
(x gxh
1
).
If g and h are not conjugate, this trace is clearly zero, since the matrix of the operator x gxh
1
in the basis of group elements has zero diagonal entries. On the other hand, if g and h are in the
same conjugacy class, the trace is equal to the number of elements x such that x = gxh
1
, i.e., the
order of the centralizer Z
g
of g. We are done.
4.7 Unitary representations. Another proof of Maschkes theorem for complex
representations
Denition 4.9. A unitary nite dimensional representation of a group G is a representation of G
on a complex nite dimensional vector space V over C equipped with a Ginvariant positive denite
Hermitian form (, ), i.e. such that
V
(g) are unitary operators: (
V
(g)v,
V
(g)w) = (v, w).
Remark 4.10. Not any nite dimensional representation admits a unitary structure.
Theorem 4.11. If G is nite, then any nite dimensional representation of G has a unitary
structure. If the representation if irreducible, this structure is unique up to scaling.
Proof. Take any positively dened form B on V and dene another form B as follows:
B(v, w) =
gG
B(gv, gw)
Then B is a positive denite Hermitian form on V, and
V
(g) are unitary operators. If V is an
irreducible representation and B
1
, B
2
are two Hermitian forms on V, then B
1
(v, w) = B
2
(v, Aw)
for some homomorphism A : V V. This implies A = Id, 1.
Theorem 4.12. A nite dimensional unitary representation V of any group G is completely re
ducible.
Proof. Let W be a subrepresentation of V . Let W
is a subrepresentation of W, and V = W W
.
This implies that V is completely reducible.
4.8 Orthogonality of matrix elements
Let V be an irreducible representation of G, and v
1
, v
2
, . . . , v
n
be an orthonormal basis of V under
the Hermitian form. The matrix elements of V are T
V
ij
(x) = (v
i
,
V
(x)v
j
).
Proposition 4.13. (1) Matrix elements of nonisomorphic representations are orthogonal in F(G, C)
under the form (f, g) =
1
G
xG
f(x)g(x).
27
(2) (t
V
ij
, t
V
i
) =
ii
jj
1
dimV
Proof. Let V and W be two representations of G. Take v
i
to be an orthonormal basis of V and
w
i
to be an orthonormal basis of W. Putting P =
xG
x, we have
xG
v
i
, xv
j
)w
i
, xw
j
) =
xG
v
i
, xv
j
)w
i
, xw
j
) = v
i
w
i
, P(v
j
w
j
))
If V ,= W, this is zero, since P projects to the trivial representation. If V = W, we need to
consider v
i
v
, P(v
j
v
)). We have
V V
= C L
C = span(
v
k
v
k
)
L = span(
P
a
kl
=0
a
kl
v
k
v
l
)
The projection of v
i
v
to C C L is
1
dimV
v
k
v
ii
j
))) =
ii
jj
dimV
which nishes the proof.
4.9 Character tables, examples
The characters of all the irreducible representations of a nite group can be arranged into a char
acter table, with conjugacy classes of elements as the columns, and characters as the rows. More
specically, the rst row in a character table lists representatives of conjugacy classes, the second
one the numbers of elements in the conjugacy classes, and the other rows are the values of the char
acters on the conjugacy classes. Due to theorems 4.7 and 4.8 the rows and columns of a character
table are orthonormal with respect to the appropriate inner products.
Note that in any character table, the row corresponding to the trivial representation consists
of ones, and the column corresponding to the neutral element consists of the dimensions of the
representations.
Here is, for example, the character table of S
3
:
S
3
Id (12) (123)
# 1 3 2
C 1 1 1
C
1 1 1
C
2
2 0 1
It is obtained by explicitly computing traces in the irreducible representations.
For another example consider A
4
, the group of even permutations of 4 items. There are three
onedimensional representations. Since there are four conjugacy classes in total, there is one more
irreducible representation of dimension 3. Finally, the character table is
28
A
4
Id (123) (132) (12)(34)
# 1 4 4 3
C 1 1 1 1
C
1
2
1
C
2 1
2
1
C
3
3 0 0 1
where = exp(
2i
3
).
The last row can be computed using the orthogonality of rows. Another way to compute the
last row is to note that C
3
is the representation of A
4
by rotations of the regular tetrahedron: in
this case (123), (132) are the rotations by 120
0
and 240
0
around a perpendicular to a face of the
tetrahedron, while (12)(34) is the rotation by 180
0
around an axis perpendicular to two opposite
edges.
Example 4.14. The following three character tables are of Q
8
, S
4
, and A
5
respectively.
Q
8
1 1 i j k
# 1 1 2 2 2
C
++
1 1 1 1 1
C
+
1 1 1 1 1
C
+
1 1 1 1 1
C
1 1 1 1 1
C
2
2 2 0 0 0
S
4
Id (12) (12)(34) (123) (1234)
# 1 6 3 8 6
C 1 1 1 1 1
C
1 1 1 1 1
C
2
2 0 2 1 0
C
3
+
3 1 1 0 1
C
3
3 1 1 0 1
A
5
Id (123) (12)(34) (12345) (13245)
# 1 20 15 12 12
C 1 1 1 1 1
V
+
3
= C
3
(1)
3 0 1
1+
5
2
1
5
2
V
3
= C
3
(2)
3 0 1
1
5
2
1+
5
2
V
4
= C
4
4 1 0 1 1
V
5
= C
5
5 1 1 0 0
Indeed, the computation of the characters of the 1dimensional representations is straightfor
ward.
The character of the 2dimensional representation of Q
8
is obtained by using character orthog
onality.
For S
4
, the 2dimensional irreducible representation is obtained from the 2dimensional irre
ducible representation of S
3
via the surjective homomorphism S
4
S
3
, which allows to obtain its
character from the character table of S
3
.
The character of the 3dimensional representation C
3
+
is computed from its geometric realization
by rotations of the cube. Namely, by rotating the cube, S
4
permuts the main diagonals. Thus (12)
29
is rotaion by 180
0
around an axis that is perpendicular to two opposite edges, (12)(34) is rotation
by 180
0
around an axis that is perpendicular to two opposite faces, (123) is rotation around a main
diagonal by 120
0
, and (1234) is rotation by 90
0
around an axis that is perpendicular to two opposite
faces; this allows us to compute the traces easily, using the fact that the trace of a rottation by the
angle in 1
3
is 1 + 2 cos . Now the character of C
3
N
k
ij
V
k
, N
k
ij
= (
i
j
,
k
)
Example 4.15. The following tables represent computed tensor product multiplicities of irre
ducible representations of S
3
, S
4
, and A
4
respectively.
C C
C
2
C C C
C
2
C
C C
2
C
2
C
2
C
2
CC C
2
C
+
C
C
2
C
3
+
C
3
C
+
C
+
C
C
2
C
3
+
C
3
C
+
C
2
C
3
C
3
+
C
2
C
2
C
2
C C
C
3
C
3
+
C
3
C
3
+
C
3
C
3
+
C
3
+
C
3
C
3
+
C
3
C
2
C
+
C
3
+
C
3
C
2
C
C
3
+
C
3
C
3
C
3
C
3
+
C
3
+
C
3
C
2
C
C
3
+
C
3
C
2
C
+
C
3
+
C
3
30
C V
+
3
V
3
V
4
V
5
C C V
+
3
V
3
V
4
V
5
V
+
3
C V
5
V
+
3
V
4
V
5
V
3
V
4
V
5
V
+
3
V
3
V
4
V
5
V
3
C V
5
V
+
3
V
+
3
V
4
V
5
V
+
3
V
3
V
4
V
5
V
4
V
+
3
V
3
C V
4
V
5
V
+
3
V
3
2V
5
V
4
V
5
C V
+
3
V
3
2V
4
2V
5
4.11 Problems
Problem 4.16. Let G be the group of symmetries of a regular Ngon (it has 2N elements).
(a) Describe all irreducible complex representations of this group.
(b) Let V be the 2dimensional complex representation of G obtained by complexication of the
standard representation on the real plane (the plane of the polygon). Find the decomposition of
V V in a direct sum of irreducible representations.
Problem 4.17. Let G be the group of 3 by 3 matrices over F
p
which are upper triangular and have
1s on the diagonal, under multiplication (its order is of course p
3
). It is called the Heisenberg
group. For any complex number z such that z
p
= 1 we dene a representation of G on the space V
of complex functions on F
p
, by
(
_
_
1 1 0
0 1 0
0 0 1
_
_
f)(x) = f(x 1),
(
_
_
1 0 0
0 1 1
0 0 1
_
_
f)(x) = z
x
f(x).
(note that z
x
makes sense since z
p
= 1).
(a) Show that such a representation exists and is unique, and compute (g) for all g G.
(b) Denote this representation by R
z
. Show that R
z
is irreducible if and only if z ,= 1.
(c) Classify all 1dimensional representations of G. Show that R
1
decomposes into a direct sum
of 1dimensional representations, where each of them occurs exactly once.
(d) Use (a)(c) and the sum of squares formula to classify all irreducible representations of
G.
Problem 4.18. Let V be a nite dimensional complex vector space, and GL(V ) be the group of
invertible linear transformations of V . Then S
n
V and
m
V (m dim(V )) are representations of
GL(V ) in a natural way. Show that they are irreducible representations.
Hint: Choose a basis e
i
in V . Find a diagonal element H of GL(V ) such that (H) has
distinct eigenvalues. (where is one of the above representations). This shows that if W is a
subrepresentation, then it is spanned by a subset S of a basis of eigenvectors of (H). Use the
invariance of W under the operators (1+E
ij
) (where E
ij
is dened by E
ij
e
k
=
jk
e
i
) for all i ,= j
to show that if the subset S is nonempty, it is necessarily the entire basis.
Problem 4.19. Recall that the adjacency matrix of a graph (without multiple edges) is the matrix
in which the ijth entry is 1 if the vertices i and j are connected with an edge, and zero otherwise.
Let be a nite graph whose automorphism group is nonabelian. Show that the adjacency matrix
of must have repeated eigenvalues.
31
Problem 4.20. Let I be the set of vertices of a regular icosahedron ([I[ = 12). Let F(I) be the
space of complex functions on I. Recall that the group G = A
5
of even permutations of 5 items
acts on the icosahedron, so we have a 12dimensional representation of G on F(I).
(a) Decompose this representation in a direct sum of irreducible representations (i.e., nd the
multiplicities of occurrence of all irreducible representations).
(b) Do the same for the representation of G on the space of functions on the set of faces and
the set of edges of the icosahedron.
Problem 4.21. Let F be a nite eld with q elements, and G be the group of inhomogeneous
linear transformations, x ax + b, over F (i.e., a F
of invertible elements
of H under multiplication.
The algebra H is called the quaternion algebra.
(d) For q = a+bi+cj+dk, a, b, c, d 1, let q = abicjdk, and [[q[[
2
= q q = a
2
+b
2
+c
2
+d
2
.
Show that q
1
q
2
= q
2
q
1
, and [[q
1
q
2
[[ = [[q
1
[[ [[q
2
[[.
(e) Let G be the group of quaternions of norm 1. Show that this group is isomorphic to SU(2).
(Thus SU(2) is the 3dimensional sphere).
(f ) Consider the action of G on the space V H spanned by i, j, k, by x qxq
1
, q G,
x V . Since this action preserves the norm on V , we have a homomorphism h : SU(2) SO(3).
Show that this homomorphism is surjective and that its kernel is 1, 1.
Problem 4.23. Using the classication of nite subgroups of SO(3) (M. Artin, Algebra, p.184),
classify nite subgroups of SU(2) (use the homomorphism SU(2) SO(3)).
Problem 4.24. Find the characters and tensor products of irreducible complex representations of
the Heisenberg group from Problem 4.17.
Problem 4.25. Let G be a nite group, and V a complex representation of G which is faithful,
i.e. the corresponding map G GL(V ) is injective. Show that any irreducible representation of G
occurs inside S
n
V (and hence inside V
n
) for some n.
Problem 4.26. This problem is about an application of representation theory to physics (elasticity
theory). We rst describe the physical motivation and then state the mathematical problem.
Imagine a material which occupies a certain region U in the physical space V = 1
3
(a space
with a positive denite inner product). Suppose the material is deformed. This means, we have
32
applied a dieomorphism (=change of coordinates) g : U U
, C
2
are real. For S
4
there
are ve irreducible representations C
+
, C
, C
2
, C
3
+
, C
3
3
, V
4
, V
5
are real. As for Q
8
, its onedimensional
representations are real, and the twodimensional one is quaternionic.
Theorem 5.2. (FrobeniusSchur) The number of involutions (=elements of order 2) in G is equal
to the sum of dimensions of real representations minus the sum of dimensions of quaternionic
representations.
Proof. Let A : V V have eigenvalues
1
,
2
, . . . ,
n
. We have
Tr[
S
2
V
(AA) =
ij
j
Tr[
2
V
(AA) =
i<j
j
Thus,
Tr[
S
2
V
(A A) Tr[
2
V
(A A) =
1in
2
i
= Tr(A
2
).
Thus for g G we have
V
(g
2
) =
S
2
V
(g)
2
V
(g)
Therefore,
V
(
gG
g
2
) = [G[
_
_
_
1, if V is real
1, if V is quaternionic
0, if V is complex
Finally, the number of involutions in G equals
1
[G[
V
dimV
V
(
gG
g
2
) =
real V
dimV
quat V
dimV
5.2 Frobenius determinant
Enumerate the elements of a nite group G as follows: g
1
, g
2
, . . . , g
n
. Introduce n variables indexed
with the elements of G :
x
g
1
, x
g
2
, . . . , x
gn
.
Denition 5.3. Consider the matrix X
G
with entries a
ij
= x
g
i
g
j
. The determinant of X
G
is some
polynomial of degree n of x
g
1
, x
g
2
, . . . , x
gn
that is called Frobenius determinant.
The following theorem, discovered by Dedekind and proved by Frobenius, became the starting
point for creation of representation theory.
Theorem 5.4.
det X
G
=
r
j=1
P
j
(x)
deg P
j
for some pairwise nonproportional irreducible polynomials P
j
(x).
We will need the following rather simple lemma.
34
Lemma 5.5. Let Y be a n n matrix with entries y
ij
. Then det Y is an irreducible polynomial of
y
ij
.
Proof. Suppose that det Y = q
1
q
2
. . . q
k
, k 2 where q
i
are irreducible homogeneous polynomials.
Because det Y is linear on each column, there is exactly one q
i
that depends on any column.
Thus q
1
is a linear function on some column, say j
th
. Pick y
ij
n
i=1
so that q
1
= 0. Then any
matrix Y with such j
th
column must have det Y = 0. This is clearly false for n 2. Contradiction.
Now we are ready to proceed to the proof Theorem 5.4.
Proof. Let V = C[G] be the regular representation of G. Consider operatorvalued polynomial
L(x) =
gG
x
g
(g),
where (g) EndV is induced by g. The action of L(x) on element h G is
L(x)h =
gG
x
g
(g)h =
gG
x
g
gh =
zG
x
zh
1z
So the matrix of the linear operator L(x) in the basis g
1
, g
2
, . . . , g
n
which is X
G
with permuted
columns and, hence, the same determinant up to sign. Further, we have
det
V
L(x) =
r
i=1
(det
V
i
L(x))
dimV
i
.
We set P
i
= det
V
i
L(x). Recall that C[G] =
r
i=1
EndV
i
. Let e
im
be bases of V
i
and E
i,jk
EndV
i
be the matrix units in these bases. Then E
i,jk
is a basis of C[G] and
L(x)[
V
i
=
y
i,jk
E
i,jk
,
where y
i,jk
are new coordinates on C[G] related to x
g
by a linear transformation. Then
P
i
(x) = det [
V
i
L(x) = det(y
i,jk
)
Hence, P
i
are irreducible (by lemma 5.5) and not proportional.
5.3 Algebraic integers
We are now passing to deeper results in representation theory of nite groups. These results require
the theory of algebraic numbers, which we will now briey review.
Denition 5.6. z C is an algebraic integer if z is a root of a monic polynomial with integer
coecients.
Denition 5.7. z C is an algebraic integer if z is an eigenvalue of a matrix with integer
entries.
Proposition 5.8. Denitions (5.6) and (5.7) are equivalent.
35
Proof. To show (5.7) (5.6), notice that z is a root of the characteristic polynomial of the matrix
(a monic polynomial with integer coecients).
To show (5.6) (5.7), suppose z is a root of
p(x) = x
n
+a
1
x
n1
+. . . +a
n1
x +a
n
.
Then the characteristic polynomial of the following matrix is p(x):
_
_
_
_
_
_
_
0 0 0 . . . 0 a
n
1 0 0 . . . 0 a
n1
0 1 0 . . . 0 a
n1
.
.
.
0 0 0 . . . 1 a
1
_
_
_
_
_
_
_
.
Since z is a root of the characteristic polynomial of this matrix, it is its eigenvalue.
We will sometimes be using the symbol A to denote the set of algebraic integers.
Proposition 5.9. A is a ring.
Proof. We will be using denition (5.7). Let be an eigenvalue of
/ Mat
n
(C)
with eigenvector v, let be an eigenvalue of
B Mat
m
(C)
with eigenvector w. Then + is an eigenvalue of
/Id
m
+ Id
n
B,
and is an eigenvalue of
/B.
The corresponding eigenvector is in both cases v w.
Proposition 5.10. A = Z.
Proof. We will be using denition (5.6). Let z be a root of
p(x) = x
n
+a
1
x
n1
+. . . +a
n1
x +a
n
,
and suppose
z =
p
q
, gcd(p, q) = 1.
Notice that the leading term of p(x) will have q
n
in the denominator, whereas all the other terms
will have a lower power of q there. Thus, if
q ,= 1,
then
p(z) / Z,
a contradiction. Thus,
z A z Z.
The reverse inclusion follows because n Z is a root of x n.
36
5.4 Frobenius divisibility
Theorem 5.11. Let G be a nite group, and let V be an irreducible (necessarily nitedimensional)
representation of G over C. Then
[G[ divides dimV.
Proof. Let
C
1
, C
2
, . . . , C
n
be the conjugacy classes of G, with
C
1
= e.
Let
p
C
i
C[G]
be dened for each conjugacy class as
p
C
i
=
gC
i
g.
Since G acts transitively on each conjugacy class, every conjugate of p
C
i
is equal to itself, i.e. p
C
i
is a central element in C[G]. By Schurs lemma, p
C
i
acts on V by a scalar
i
; therefore,
[C
i
[
V
(g
C
i
) = tr (p
C
i
) = dimV
i
.
Therefore,
i
=
V
(g
C
i
)
[C
i
[
dimV
,
where g
C
i
is a representative of C
i
. Claim. The number
i
is an algebraic integer for all i.
Proof of claim. Notice that
p
C
i
p
C
j
=
gC
i
,hC
j
gh =
uG
ux,
where x is the number of ways to obtain u as gh for some g C
i
, h C
j
.
Thus,
p
C
i
p
C
j
=
i,j,k
N
k
ij
p
C
k
,
where N
k
ij
is the number of ways to obtain some element of C
k
as gh for some g C
i
, h C
j
.
Therefore,
j
=
i,j,k
N
k
ij
k
Let
=
_
_
_
_
_
2
.
.
.
n
_
_
_
_
_
,
then
N
i
(
) =
i
,
37
where N
i
is the matrix whose jk
th
entry is N
k
ij
. Since
,= 0,
i
is an eigenvalue of an integer matrix N
i
, and by denition (5.7) an algebraic integer.
Now, consider
V
(g
C
i
).
This is an algebraic integer, since
i
was just proven to be an algebraic integer, and
V
(g
C
i
) is a
sum of roots of unity (it is the sum of eigenvalues of the matrix of (g
C
i
), and since
g
G
C
i
= e
in G, the eigenvalues of (g
C
i
) are roots of unity), and A is a ring (5.9). On the other hand, from
the denition of
i
,
C
i
V
(g
C
i
) =
i
[C
i
[
V
(g
C
i
)
V
(g
C
i
)
dimV
.
Recalling that
V
is a class function, this is equivalent to
gG
V
(g)
V
(g)
dimV
=
[G[(
V
,
V
)
dimV
.
Since V was an irreducible representation,
(
V
,
V
) = 1,
so
C
i
V
(g
C
i
) =
[G[
dimV
.
Since
[G[
dimV
and
C
i
V
(g
C
i
) A,
by (5.10)
[G[
dimV
Z.
5.5 Burnsides Theorem
This famous result in group theory was proved by the British mathematician William Burnside
in the late 19th century. Here is a proof of his theorem using Representation Theory.
38
Denition 5.12. A group G is called solvable if there exists a series of nested normal subgroups
e = G
1
G
2
. . . G
n
= G
where G
i+1
/G
i
is abelian for all 1 i n 1.
Remark 5.13. These groups are called solvable because they rst arose as Galois groups of poly
nomial equations which are solvable in radicals.
Theorem 5.14 (Burnside).
Any group G of order p
a
q
b
, where p and q are prime and a, b 0, is solvable.
Before proving Burnsides theorem we will prove several other results which may be of indepen
dent interest.
Theorem 5.15. Let V be an irreducible representation of a nite group G and let C be a conjugacy
class of G with gcd([C[, dim(V )) = 1.
Then for any g C, either
V
(g) = 0 or g acts as a scalar on V .
The proof will be based on the following two lemmas.
The rst lemma is a standard fact about algebraic numbers.
Lemma 5.16. Let x, y be algebraic numbers. Then any conjugate of x + y can be expressed as
x
i
+y
j
, where x
i
and y
j
are conjugates of x and y respectively.
Proof. Recall that, by denition, a conjugate of an algebraic number is any root of its minimal
polynomial (thus any number is its own conjugate).
Let x
1
, . . . , x
m
and y
1
, . . . , y
n
be the sets of all conjugates of x and y respectively. Consider
the polynomial
g(z) =
i,j
(z x
i
y
j
) = g
0
z
mn
+g
1
z
mn1
+. . . +g
mn
.
The coecients g
k
depend polynomially on x
i
and y
j
. Moreover, these polynomials are symmetric
with respect to x
i
and with respect to y
j
and have rational coecients. Therefore all g
k
.
Notice that x + y is a root of g. As we have shown, g [z], so the minimal polynomial of
x+y divides g. Therefore, every conjugate of x+y is a root of g and can be written as x
i
+y
j
.
Lemma 5.17. If
1
,
2
. . .
n
are roots of unity such that
1
n
(
1
+
2
+. . . +
n
) is an algebraic integer,
then either
1
= . . . =
n
or
1
+. . . +
n
= 0.
Proof. Let a =
1
n
(
1
+. . . +
n
). Let q = x
m
+q
m1
x
m1
+. . . +q
1
x+q
0
be the minimal polynomial
of a, and let a
i
, i = 1, . . . , m be the set of all the conjugates of a.
By 5.16, a
i
=
1
n
(
1
+
2
+ . . . +
n
), where
i
are conjugate to
i
. Since conjugates of roots
of unity are roots of unity, [
i
[ = 1. This means that [
1
+ . . . +
n
[ n and [a
i
[ 1. Let
q = x
n
+q
1
x
n1
+.. + q
n
x be the minimal polynomial of a. Then, [q
0
[ =
n
i=1
[a
i
[ 1. However,
by our assumption, a is an algebraic integer and q
0
Z. Therefore, either q
0
= 0 or [q
0
[ = 1.
39
Assume that [q
0
[ = 1. Then [a
i
[ = 1 for all i, and in particular, [a[ = [
1
n
(
1
+ . . . +
n
)[ =
1 =
1
n
([
1
[ + [
2
[ + . . . + [
n
[). This means that all
i
have the same argument. It follows that
1
= . . . =
n
since all
i
have the same absolute value.
Otherwise, q
0
= 0. Then x[q and since q is irreducible, q = x and a = 0.
Proof of theorem 5.15.
Let dimV = n. Let
1
,
2
, . . .
n
be eigenvalues of (g). Since G is a nite group, (g) is
diagonalizable and
i
are roots of unity. We know that
1
n
([C[
V
(g)) A and that
V
(g) A. Since
GCD(n, [C[) = 1, there are integers , such that n +[C[ = 1. Therefore,
V
(g) +
[C[
V
(g)
n
=
V
(g)
n
A.
However, since
V
(g) =
1
+. . . +
n
, we get that either
1
+. . . +
n
=
V
(g) = 0 or
1
= . . . =
n
by 5.17. If
1
= . . . =
n
, then, since (g) is diagonalizable, it must be scalar. Otherwise,
V
(G) =
0.
Proposition 5.18. Let G be a nite, simple nonabelian group and let V be a nontrivial, irreducible
representation of G. Then, if g G acts by a scalar in V , g = e.
Proof. Assume that g ,= e. Let N be the set of all x G whose action in V is scalar. Clearly,
N G and g N. Since g ,= e, this means that N ,= e and N = G.
Now let K be the kernel of : G EndV . Since is a group homomorphism, K G and
because V is nontrivial, K ,= G and K = e.
This means that is an injection and G
= Im. But (x) is scalar for any x G, so G is
commutative, which is a contradiction.
We are now ready to prove another result in group theory which will later imply Burnsides
Theorem.
Theorem 5.19. Let G be a group and let C be a conjugacy class of order p
k
where p is prime and
k > 0. Then G has a proper normal subgroup.
Proof. Assume the contrary, i.e. that G is simple.
We can choose an element e ,= g C.
Let R be the regular representation of G. Since g ,= e,
R
(g) = 0. On the other hand,
R =
V X
(dimV )V , where X is the set of all irreducible representations of G. Therefore
0 =
R
(g) =
V X
dimV
V
(g).
We can divide X into three parts:
40
1. S, the set of irreducible representations whose dimension is divisible by p,
2. T, the set of nontrivial representations whose dimension is not divisible by p and
3. I, the trivial representation.
Lemma 5.20. If V T then
V
(g) = 0.
Proof. Since gcd([C[, dim(V )) = 1, by 5.15, either
1.
V
(g) = 0 or
2. g acts as a scalar in V , and by 5.18, g = e which is a contradiction.
Also, if V S, we have
1
p
dim(V )
V
(g) A, so
a =
V S
1
p
dim(V )
V
(g) A.
Therefore,
0 =
V S
dimV
V
(g) +
V T
dimV
V
(g) + dimI
I
(g) = pa + 1.
This means that a =
1
p
which is not an algebraic integer, so we have a contradiction.
Now we can nally prove Burnsides theorem.
Assume that there exists a group of order p
a
q
b
that is not solvable. We may assume that G has
the smallest order among such groups. Since [G[ ,= 1, either a or b must be nonzero.
We may assume without loss of generality that b ,= 0.
Lemma 5.21. Let G be a group as above.
(i) G is simple.
(ii) G has a trivial center (in particular, it is not abelian).
(iii) G has a conjugacy class C of order p
k
.
Proof.
(i) Assume that N is a nontrivial proper normal subgroup of G. Since [N[ divides [G[, [N[ = p
r
q
s
for some r a, s b.
41
Let H = N/G. Then [H[ = p
ar
q
bs
. By our minimality assumption, both N and H are
solvable, and there exist normal series
e = N
1
N
2
. . . N
m
= N and e = H
1
H
2
. . . H
n
= H
with abelian quotients.
Let be the canonical epimorphism G G/N = H. Then
1
(H
j
)
1
(H
j+1
) and
1
(H
j+1
)/
1
(H
j
) = H
j+1
/H
j
for any j.
Consider the normal series
e N
1
. . . N
m
= N =
1
(e) =
1
(H
1
) . . .
1
(H
n
) = G.
The quotient of any two consecutive subgroups of this series is either N
i+1
/N
i
or
1
(H
j+1
)/
1
(H
j
) =
H
j+1
/H
j
, all of which are abelian. Because of this, G is solvable, which is a contradiction.
Therefore G is simple.
(ii) The center of G, Z(G) is a normal subgroup of G. If Z(G) = G then G is abelian and e G
is a normal series with abelian factors. Therefore, since G is simple, Z(G) = e.
(iii) Assume that G does not have a conjugacy class of order p
k
. Let C be any conjugacy class.
The order of C divides [G[, so [C[ = p
i
q
j
. Because of our assumption, either j ,= 0 and
q divides [C[ or [C[ = 1 and C is central. But there is exactly one central element, e, so
[G[ =
[C[ = 1modq where the sum is taken over all the conjugacy classes C. However,
since b ,= 0, q divides [G[, which is a contradiction.
But by Theorem 5.19, this is impossible! Therefore, there are no groups of order p
a
q
b
which
are not solvable, and we have proven Burnsides Theorem.
5.6 Induced Representations
Given a representation V of a group G and a subgroup H < G, there is a natural way to construct
a representation of H. The restricted representation of V to H, Res
G
H
V is the representation given
by the vector space V and the action
Res
G
H
V
=
V
[
H
.
There is also a natural, but more complicated way to construct a representation of a group G
given a representation V of its subgroup H.
5.6.1 Denition
Denition 5.22. If G is a group, H < G, and V is a representation of H, then the induced
representation Ind
G
H
V is a representation of G with
Ind
G
H
V = f : G V [f(hx) =
V
(h)f(x)x G, h H
and the action g(f)(x) = f(xg)g G.
42
5.6.2
Let us check that this is indeed a representation:
g(f)(hx) = f(hxg) =
V
(h)f(xg) =
V
(h)g(f)(x), and g(g
(f))(x) = g
(f)(xg) = f(xgg
) =
(gg
, x G and h H.
Remark 5.23. In fact, Ind
G
H
V is naturally equivalent to Hom
H
(k[G], V ).
Remark 5.24. Notice that if we choose a representative x
).
Because of this,
dim(Ind
G
H
V ) = dimV
[G[
[H[
.
5.6.3 The character of induced representation
Let us now compute the character of Ind
G
H
V .
For a left Hcoset of G, let us dene
V
= f Ind
G
H
V [f(g) = 0 g , .
Theorem 5.25. (The Mackey formula) One has
(g) =
1
[H[
xG,xgx
1
H
V
(xgx
1
)
Proof. One has
Ind
G
H
V =
,
and so
(g) =
(g),
where
.
Since g() = g is a left Hcoset for any left Hcoset ,
(g) = 0 if ,= g.
Now assume that = g. Choose x
. Then x
g = hx
where h = x
gx
1
H. Consider
the vector space homomorphism : V
). Since f V
is uniquely
determined by f(x
), is an isomorphism. We have
(gf) = g(f)(x
) = f(x
g) = f(hx
) =
V
(h)f(x
) = h(f),
and gf =
1
h(f). This means that
(g) =
V
(h). Therefore
(g) =
G\H,g=
V
(x
gx
1
).
43
Since it does not matter which representative x
= C
2
C
.
2. Let G = S
3
, H = Z
3
. Then we obtain Ind
G
H
C = CC
, Ind
G
H
C
= Ind
G
H
C
2 = C
2
.
3. Let G = S
4
, H = S
3
. Then Ind
G
H
C = CC
3
, Ind
G
H
C
= C
C
3
+
, Ind
G
H
C
2
= C
2
C
3
C
3
+
.
5.7 Representations of S
n
Let us give, without proof, a description of the representations of the symmetric group S
n
for any
n.
Let be a partition of n, i.e. a representation of n in the form n =
1
+
2
+ ... +
p
, where
i
are positive integers, and
i
i+1
To such we will attach a Young diagram Y
, which is the
subset of the coordinate plane dened by the inequalities: 0 x p, 0 y
[x]
+
, where [x]
+
denotes the smallest integer N such that N x. Clearly, Y
in
some way (without repretitions). For example, we will consider the Young tableau T
obtained by
lling in the numbers in the increasing order, left to right, bottom to top.
We can dene two subgroups of S
n
corresponding to Y
:
1. Row subgroup P
.
2. Column subgroup P
.
Clearly, P Q = 1.
Dene the Young projectors:
p
+
:=
1
[P
gP
g,
q
:=
1
[Q
gQ
(1)
g
g,
where (1)
g
denotes teh sign of the permutation g. Set c
= p
.
The irreducible representations of S
n
are described by the following theorem.
Theorem 5.27. The subspace V
:= C[S
n
]c
of C[S
n
] is an irreducible representation of S
n
under
left multiplication. Every irreducible representation of S
n
is isomorphic to V
for a unique .
Example 5.28.
For the partition = (1, ..., 1), P
= S
n
, Q
= 1, so c
is the
trivial representation.
For the partition = (n), Q
= S
n
, P
= 1, so c
is the
sign representation.
45
n = 3. For = (2, 1), V
= C
2
.
n = 4. For = (2, 2), V
= C
2
; for = (3, 1), V
= C
3
+
; for = (2, 1, 1), V
= C
3
.
Corollary 5.29. All irreducible representations of S
n
can be given by matrices with rational entries.
Problem 5.30. Find the sum of dimensions of all irreducible representations of the symmetric
group S
n
.
Hint. Show that all irreducible representations of S
n
are real, i.e. admit a nondegenerate
invariant symmetric form. Then use the FrobeniusSchur theorem.
6 Representations of GL
2
(F
q
)
6.1 Conjugacy classes in GL
2
(F
q
)
Let F
q
be a nite eld of size q of characteristic other than 2. Then
[GL
2
(F
q
)[ = (q
2
1)(q
2
q),
since the rst column of an invertible 2 by 2 matrix must be nonzero and the second column may
not be a multiple of the rst one. Factoring,
[GL
2
(F
q
)[ = q(q + 1)(q 1)
2
.
The goal of this section is to describe the irreducible representations of GL
2
(F
q
).
To begin, let us nd the conjugacy classes in GL
2
(F
q
).
Representatives
Number of elements in a conjugacy
class
Number of classes
Scalar
_
x 0
0 x
_
1 (this is a central element)
q 1 (one for every non
zero x)
Parabolic
_
x 1
0 x
_
q
2
1 (elements that commute with
this one are of the form
_
t u
0 t
_
, t ,=
0)
q 1 (one for every non
zero x)
Hyperbolic
_
x 0
0 y
_
, y ,= x
q
2
+q (elements that commute with
this one are of the form
_
t 0
0 u
_
, t, u ,=
0)
1
2
(q 1)(q 2) (x, y ,= 0
and x ,= y)
Elliptic
_
x y
y x
_
, x F
q
, y
F
q
, F
q
F
2
q
(characteris
tic polynomial over F
q
is irre
ducible)
q
2
q (the reason will be described
below)
1
2
q(q 1) (matrices with
y and y are conjugate)
More on the conjugacy class of elliptic matrices: these are the matrices whose characteristic
polynomial is irreducible over F
q
and which therefore dont have eigenvalues in F
q
. Let A be such
a matrix, and consider a quadratic extension of F
q
,
F
q
(
), F
q
F
2
q
.
46
Over this eld, A will have eigenvalues
=
1
+
2
and
=
1
2
,
with corresponding eigenvectors
v, v (Av = v, Av = v).
Choose a basis
e
1
= v +v, e
2
=
(v v).
In this basis, the matrix A will have the form
_
1
2
2
1
_
,
justifying the description of representative elements of this conjugacy class.
In the basis v, v, matrices that commute with A will have the form
_
0
0
_
,
for all
F
q
2
,
so the number of such matrices is q
2
1.
6.2 Representations of GL
2
(F
q
)
In this section, G will denote the group GL
2
(F
q
).
6.2.1 1dimensional representations
First, we describe the 1dimensional representations of G.
Proposition 6.1. [G, G] = SL
2
(F
q
).
Proof. Clearly,
det(xyx
1
y
1
) = 1,
so
[G, G] SL
2
(F
q
).
To show the converse, let us show that every elementary matrix with determinant 1 is a commutator;
such matrices generate SL
2
(F
q
). We may restruict our attention to the matrices
_
1 1
0 1
_
or
_
1 0
1 1
_
47
Clearly, only showing that the rst matrix is a commutator suces, since if
_
1 1
0 1
_
= ABA
1
B
1
then
_
1 0
1 1
_
=
_
1 x
0 1
_
t
= (B
t
)
1
(A
t
)
1
B
t
A
t
.
To show that matrix
_
1 1
0 1
_
is a commutator, we observe that the commutator of the following two matrices is as required:
A =
_
k 0
0
1
k
_
, B =
_
1 y
0 1
_
,
where
y =
1
k
2
1
, k / 0, 1, 1
(the case of F
3
is considered below).
In the case of F
3
, for
A =
_
2 2
0 1
_
, B =
_
2 2
0 2
_
, we have ABA
1
B
1
=
_
1 1
0 1
_
.
This completes the proof.
Therefore,
G/[G, G]
= F
q
via g det(g).
The onedimensional representations of G thus have the form
(g) =
_
det(g)
_
,
where is a homomorphism
: F
q
C
;
since F
q
is cyclic, there are q 1 such representations, denoted C
.
6.2.2 Principal series representations
Let
B G, B =
_
0
_
q
F
q
48
(the isomorphism maps an element of B/[B, B] to its two eigenvalues).
Let
: B C
be a homomorphism dened by
_
a b
0 c
_
=
1
(a)
2
(c),for some pair of homomorphisms
1
,
2
: F
q
C.
Dene
V
1
,
2
= Ind
G
B
C
,
where C
1
,
2
) =
[G[
[B[
= q + 1.
Theorem 6.2. 1.
1
,=
2
V
1
,
2
is irreducible.
2.
1
=
2
= V
1
,
2
= C
, where W
= W
i = ; V
1
,
2
= V
1
,
2
i
1
,
2
=
1
,
2
(in the second case,
1
,=
2
,
1
,=
2
).
Proof. From the Mackey formula, we have
tr
V
1
,
2
(g) =
1
[B[
aG, aga
1
B
(aga
1
).
If
g =
_
x 0
0 x
_
,
the expression on the right evaluates to
2
(x)
[G[
[B[
=
1
(x)
2
(x)
_
q + 1
_
.
If
g =
_
x 1
0 x
_
,
the expression evaluates to
2
(x) 1,
since here
aga
1
B a B.
If
g =
_
x 0
0 y
_
,
the expression evaluates to
_
1
(x)
2
(y) +
1
(y)
2
(x)
_
1,
since here
aga
1
B a B or a is an element of a permutation of B.
49
If
g =
_
x y
y x
_
,
the expression on the right evaluates to 0 because matrices of this type dont have eigenvalues over
F
q
.
From the denition,
i
(x)(i = 1, 2) is a root of unity, so
[G[
V
1
,
2
,
V
1
,
2
) = (q + 1)
2
(q 1) + (q
2
1)(q 1)
+ 2(q
2
+q)
(q 1)(q 2)
2
+ (q
2
+q)
x=y
1
(x)
2
(y)
1
(y)
2
(x).
The last two summands come from the expansion
[a +b[
2
= [a[
2
+[b[
2
+ab +ab.
If
1
=
2
= ,
the last term is equal to
(q
2
+q)(q 2)(q 1),
and the total in this case is
(q + 1)(q 1)[(q + 1) + (q 1) 2q(q 2)] = (q + 1)(q 1)2q(q 1) = 2[G[,
so
1
,
2
,
V
1
,
2
) = 2.
Clearly,
C
Ind
G
B
C
,
,
since
Hom
G
(C
, Ind
G
B
C
,
) = Hom
B
(C
, C
) = C (Theorem 5.26).
Ind
G
B
C
,
= C
; W
are distinct.
If
1
,=
2
, let z = xy
1
, then the last term of the summation is
(q
2
+q)
x=y
1
(z)
2
(z) = (q
2
+q)
x;z=1
2
(z) = (q
2
+q)(q 1)
z=1
2
(z).
Since
zFq
2
(z) = 0,
being the sum of all roots of unity, the last term becomes
(q
2
+q)(q 1)
z=1
2
(1) = (q
2
+q)(q 1).
The dierence between this case and the case of
1
=
2
is equal to
(q
2
+q)[(q 2)(q 1) + (q 1)] = [G[,
50
so this is an irreducible representation.
To prove the third assertion of the theorem, we note that on hyperbolic characters the function
1
(x)
2
(y) +
1
(y)
2
(x)
determines
1
,
2
up to permutation.
6.2.3 Complimentary series representations
Let F
q
2 F
q
be a quadratic extension F
q
(
), F
q
F
2
q
. We regard this as a 2dimensional vector
space over F
q
; then GL
2
(F
q
) is the group of linear transformations of F
q
2 over F
q
. Let K GL
2
(F
q
)
be the cyclic group of multiplication by an element of F
q
,
K =
_
x y
y x
_
, [K[ = q
2
1.
For : K C
a homomorphism, let
Y
= Ind
G
K
.
This representation, of course, is very reducible. Let us compute its character, using the Mackey
formula. We get
_
x 0
0 x
_
= q(q 1)(x);
(A) = 0 for A parabolic or hyperbolic;
_
x y
y x
_
=
_
x y
y x
_
+
_
x y
y x
_
q
.
The last assertion is because if we regard the matrix as an element of F
q
2, conjugation is an
automorphism of F
q
2 over F
q
, but the only nontrivial automorphism of F
q
2 over F
q
is the q
th
power
map.
We thus have
Ind
G
K
q
= Ind
G
K
V
,
,
where is the trivial character and W
_
x 0
0 x
_
= q(q + 1)(x);
(A) = 0 for A parabolic or elliptic;
_
x 0
0 y
_
= (x) +(y).
51
Thus the virtual representation
W
V
,,
V
,
Ind
G
K
_
x 0
0 x
_
= (q 1)(x);
_
x 1
0 x
_
= (x);
_
x 0
0 y
_
= 0;
_
x y
y x
_
=
_
x y
y x
_
q
_
x y
y x
_
.
In all that follows, we will have
q
,= .
The following two lemmas will establish that the inner product of this character with itself is
equal to 1, that its value at 1 is positive, and that the above conditions imply that it is the character
of an irreducible representation of G.
Lemma 6.3. Let be the character of the virtual representation dened above. Then
, ) = 1
and
(1) > 0.
Proof.
(1) = q(q + 1) (q + 1) q(q 1) = q 1 > 0.
We now compute the inner product , ). Since is a root of unity, this will be equal to
1
(q 1)
2
q(q + 1)
_
(q1)(q1)
2
1+(q1)1(q
2
1)+
q(q 1)
2
elliptic
(()
q
())(()
q
())
Because is also a root of unity, the last term of the expression evaluates to
(())(()) + (
q
())(
q
()) +
elliptic
q1
() +
1q
().
The rst two summands here are equal to 1. Lets evaluate the last one.
Associating the elliptic elements with F
q
2 as a vector space over F
q
, we have, since F
q
2
is cyclic
and
q
,= ,
q
2
q1
() =
F
q
2
1q
() = 0.
However, we are only interested in F
q
2
F
q
, since we already considered the conjugacy classes
of scalars. Therefore,
elliptic
q1
() +
1q
() = 0
q1
() +
1q
() = 0 2(q 1) = 2(q 1)
52
since F
q
is cyclic of order q 1. Therefore,
, ) =
1
(q 1)
2
q(q + 1)
_
(q1)(q1)
2
1+(q1)1(q
2
1)+
q(q 1)
2
(2(q
2
q)2(q1))
_
= 1.
Lemma 6.4. Let
V
1
, V
2
, . . . , V
m
be (possibly reducible) representations of a nite group G over C, and let
1
,
2
, . . . ,
m
be the respective characters. Let
p
1
, p
2
, . . . p
m
Z
(not necessarily nonnegative!), let
= p
1
+p
2
+ +p
m
.
If , ) = 1 and (1) > 0, then is a character of an irreducible representation of G.
Proof. Let
W
1
, W
2
, . . . , W
s
be the irreducible representations of G, let
V
i
=
j
a
ij
W
j
(a
ij
Z
+
0).
If
j
are characters of W
j
, then
i
=
j
a
ij
j
=
i,j
p
i
a
ij
j
=
q
j
j
,
where
q
j
=
i
p
i
a
ij
Z.
Since
, ) =
j
q
2
j
= 1,
we must have q
j
= 0 except for one value of j = j
0
, with q
j
0
= 1. Thus,
=
j
0
,
and since (1) > 0 we have
=
j
0
.
53
We have now shown that for any pair (, ) with
q
,= the representation with the same
character as
W
V
,,
V
,
Ind
G
K
exists and is irreducible. This character is distinct for distinct pairs (, ) when is not a q
th
power, so there are
q(q1)
2
such representations, each of dimension q 1.
We have thus found q 1 1dimensional representations of G,
q(q1)
2
principal series represen
tations, and
q(q1)
2
complimentary series representations, for a total of q
2
1 representations, i.e.
the number of conjugacy classes in G. We can also check the sum of squares formula:
(q 1) 1
2
+ (q 1) q
2
+
(q 1)(q 2)
2
(q + 1)
2
+
q(q 1)
2
(q 1)
2
= (q 1)
2
q(q + 1) = [G[.
7 Quiver Representations
7.1 Problems
Problem 7.1. Field embeddings. Recall that k(y
1
, ..., y
m
) denotes the eld of rational functions
of y
1
, ..., y
m
over a eld k.
(a) Let f : k[x
1
, ..., x
n
] k(y
1
, ..., y
m
) be an injective homomorphism. Show that m n. (Look
at the growth of dimensions of the spaces W
N
of polynomials of degree N in x
i
and their images
under f as N ).
(b) Let f : k(x
1
, ..., x
n
) k(y
1
, ..., y
m
) be a eld embedding. Show that m n.
Problem 7.2. Some algebraic geometry.
Let k be an algebraically closed eld, and G = GL
n
(k) be the group of nondegenerate matrices
of size n over k. An algebraic representation of G is a nite dimensional representation : G
GL
N
(k) such that the matrix elements
ij
(g) are polynomials in g
pq
and 1/det(g). For example,
the standard representation V = k
n
and the dual representation V
are algebraic (show it!).
Let V be an algebraic representation of G. Show that if G has nitely many orbits on V then
dim(V ) n
2
. Namely:
(a) Let x
1
, ..., x
N
be linear coordinates on V . Let us say that a subset X of V is Zariski dense
if any polynomial f(x
1
, ..., x
N
) which vanishes on X is zero (coecientwise). Show that if G has
nitely many orbits on V then G has at least one dense orbit on V .
(b) Use (a) to construct a eld embedding k(x
1
, ..., x
N
) k(g
pq
), then use problem 1.
(c) generalize the result of this problem to the case when G = GL
n
1
(k) ... GL
nm
(k).
Problem 7.3. Dynkin diagrams.
Let be a graph, i.e. a nite set of points (vertices) connected with a certain number of edges.
We assume that is connected (any vertex can be connected to any other by a path of edges) and
has no selfloops (edges from a vertex to itself ). Suppose the vertices of are labeled by integers
1, ..., N. Then one can assign to an N N matrix R
= (r
ij
), where r
ij
is the number of edges
connecting vertices i and j. This matrix is obviously symmetric, and is called the adjacency matrix.
Dene the matrix A
= 2I R
is positive denite. Dynkin diagrams appear in many areas of mathematics (singularity theory,
54
Lie algebras, representation theory, algebraic geometry, mathematical physics, etc.) In this problem
you will get a complete classication of Dynkin diagrams. Namely, you will prove
Theorem. is a Dynkin diagram if and only if it is one on the following graphs:
A
n
:
D
n
:

E
6
:

E
7
:

E
8
:
where = A
N
, D
N
. (Use the row decomposition rule,
and write down a recusive equation for it). Deduce by Sylvester criterion that A
N
, D
N
are Dynkin
diagrams.
(b) Compute the determinants of A
for E
6
, E
7
, E
8
(use row decomposition and reduce to (a)).
Show they are Dynkin diagrams.
(c) Show that if is a Dynkin diagram, it cannot have cycles. For this, show that det(A
) = 0
for a graph below:
(show that the sum of rows is 0). Thus has to be a tree.
(d) Show that if is a Dynkin diagram, it cannot have vertices with 4 or more incoming edges.
For this, show that det(A
) = 0 for a graph below (including the case when the two nodal vertices
coincide).
(e) Using (d) show that can have no more than one vertex with 3 incoming edges.
(f ) Show that det(A
iD
x
2
i
i,jD
b
ij
x
i
x
j
.
Hint. It suces to check the result for integers: x
i
= n
i
. For this, consider the space W of
representations V of Q such that dimV
i
= n
i
. Show that the group
i
GL
n
i
(k) acts with nitely
many orbits on W k, and use problem 2 to derive the inequality.
(b) Deduce that q is a positive denite quadratic form.
(c) Show that a quiver of nite type can have no selfloops. Then, using problem 3, deduce the
theorem.
Problem 7.5. Let G be a nite subgroup of SU(2), and V be the 2dimensional representation of
G coming from its embedding into SU(2). Let V
i
, i I, be all the irreducible representations of G.
Let r
ij
be the multiplicity of V
i
in V V
j
.
(a) Show that r
ij
= r
ji
.
(b) The McKay graph of G, M(G), is the graph whose vertices are labeled by i I, and i is
connected to j by r
ij
edges. Show that M(G) is connected. (Use problem 1)
56
(c) Show that M(G) is an ane Dynkin graph. For this, show that the matrix a
ij
= 2
ij
r
ij
is positive semidenite but not denite, and use problem set 5.
Hint. Let f =
x
i
V
i
, where
V
i
be the characters of V
i
. Show directly that ((2
V
)f, f) 0.
When is it = 0? Next, show that M(G) has no selfloops, by using that if G is not cyclic then G
contains the central element Id SU(2).
(d) Which groups from problem 2 correspond to which diagrams?
(e) Using the McKay graph, nd the dimensions of irreducible representations of all nite
G SU(2). Compare with the results on subgroups of SO(3) we obtained earlier.
One central question when looking at representations of quivers is whether a certain quiver has
only nitely many indecomposable representions. We already proved that only those quives whose
underlying undirected graph is a Dynkin diagram may have this property. To see if they actually
do have this property, we rst explicitly decompose representations of certain easy quivers.
7.2 Indecomposable representations of the quivers A
1
, A
2
, A
3
Example 7.6 (A
1
). The quiver A
1
consists of a single vertex and has no edges. Since a repre
sentation of this quiver is just a single vector space, the only indecomposable representation is the
ground eld itself. Therefore the quiver A
1
has only one indecomposable representation, namely
the eld of complex numbers.
Example 7.7 (A
2
). The quiver A
2
consists of two vertices connected by a single edge.
//
V
A
//
W
To decompose this representation, we rst let V
be a complement to the kernel of A in V and
let W
V
A
//
W
=
ker V
0
//
0
V
//
A
ImA
0
0
//
//
//
or
//
oo
2
By an object of the type 1
//
0 we mean a map from a onedimensional vector space to the zero space.
Similarly, an object of the type 0
//
1 is a map from the zero space into an onedimensional space. The object
1
//
1 means an isomorphism from a onedimensional to another onedimensional space. Similarly, numbers
in such diagrams always mean the dimension of the attached spaces and the maps are the canonical maps (unless
specied otherwise).
57
1. We rst look at the orientation
//
//
.
Then a representation of this quiver looks like
V
A
//
W
B
//
Y
.
Like in 7.7 we rst split away
ker A
0
//
0
0
//
0
.
This object is a multiple of 1
//
0
//
0 . Next, let Y
0
0
//
0
0
//
A
//
W
B
// //
Y
.
Next, let X = ker(B A) and let X
be a complement of X in V . Let W
be a complement
of A(X) in W such that A(X
) W
. Then we get
A
//
W
B
// //
Y
=
X
A
//
A(X)
B
//
0
X
A
//
B
// //
Y
The rst of these summands is a multiple of 1
//
1
//
0 . Looking at the second summand,
we now have a situation where A is injective, B is surjective and furthermore ker(BA) = 0.
To simplify notation, we redene
V = X
, W = W
.
Next we let X = Im(B A) and let X
= B
1
(X
). Then W
A
//
W
B
// //
Y
=
V
A
//
A(V )
B
//
X
0
//
B
// //
//
oo
.
Very similarly to the other orientation, we can split away objects of the type
1
//
0 0
oo
, 0
//
0 1
oo
58
which results in a situation where both A and B are injective:
A
//
W
oo
B
?
_
Y
.
By identifying V and Y as subspaces of W, this leads to the problem of classifying pairs of
subspaces of a given space W up to isomorphism (the pairs of subspaces problem). To
do so, we rst choose complements V
, W
, Y
//
W
oo ?
_
Y
=
V
//
oo ?
_
V Y
//
V Y
oo
V Y
.
The second summand is a multiple of the object 1
//
1 1
oo
. We go on decomposing the
rst summand. Again, to simplify notation, we let
V = V
, W = W
, Y = Y
be a complement of V Y in W. Then
we get
//
W
oo ?
_
Y
=
V
//
V Y
oo ?
_
Y
0
//
0
oo
The second of there summands is a multiple of the indecomposable object 0
//
1 0
oo
.
The rst summand can be further decomposed as follows:
//
V Y
oo ?
_
Y
=
V
//
V
0
oo
0
//
Y
oo
These summands are multiples of
1
//
1 0
oo
, 0
//
1 1
oo
So  like in the other orientation  we get 6 indecomposable representations of A
3
:
1
//
0 0
oo
, 0
//
0 1
oo
, 1
//
1 1
oo
,
0
//
1 0
oo
, 1
//
1 0
oo
, 0
//
1 1
oo
7.3 Indecomposable representations of the quiver D
4
As a last  slightly more complicated  example we consider the quiver D
4
.
Example 7.9 (D
4
). We restrict ourselves to the orientation
//
oo
OO
.
So a representation of this quiver looks like
V
1
A
1
//
V
3
A
3
oo
V
2
A
2
OO
59
The rst thing we can do is  as usual  split away the kernels of the maps A
1
, A
2
, A
3
. More
precisely, we split away the representations
ker A
1
0
//
0
oo
0
OO
0
//
0
oo
ker A
2
0
OO
0
//
ker A
3
0
oo
0
OO
These representations are multiples of the indecomposable objects
1
0
//
0
oo
0
OO
0
//
0
oo
1
0
OO
0
//
1
0
oo
0
OO
So we get to a situation where all of the maps A
1
, A
2
, A
3
are injective.
V
1
A
1
//
V
3
?
_
A
3
oo
V
2
?
A
2
OO
As in 2, we can then identify the spaces V
1
, V
2
, V
3
with subspaces of V . So we get to the triple of
subspaces problem of classifying a triple of subspaces of a given space V .
The next step is to split away a multiple of
0
//
0
oo
0
OO
to reach a situation where
V
1
+V
2
+V
3
= V.
Now, by letting Y = V
1
V
2
V
3
and choosing complements V
1
, V
2
, V
3
, V
of Y in V
1
, V
2
, V
3
, V
respectively, we can decompose this representation into
//
3
?
_
oo
?
OO
//
oo
Y
OO
O
The last summand is a multiple of the indecomposable representation
//
oo
1
OO
O
60
So  considering the rst summand and renaming the spaces to simplify notation  we are in a
situation where
V = V
1
+V
2
+V
3
, V
1
V
2
V
3
= 0
As a next step, we let Y = V
1
V
2
and we choose complements V
1
, V
2
, V
of Y in V
1
, V
2
and V ,
such that V
3
V
V
1
//
V
3
?
_
oo
V
2
?
OO
=
//
V
3
?
_
oo
?
OO
//
0
oo
Y
OO
O
The second summand is a multiple of the indecomposable object
//
0
oo
1
OO
O
.
In the resulting situation we have V
1
V
2
= 0. Similarly we can split away multiples of
//
oo
0
OO
and
0
//
oo
1
OO
O
to reach a situation where the spaces V
1
, V
2
, V
3
do not intersect pairwise
V
1
V
2
= V
1
V
3
= V
2
V
3
= 0
Next, if V
1
_ V
2
V
3
we let Y = V
1
(V
2
V
3
). We let V
1
be a complement of Y in V
1
. Since then
V
1
(V
2
V
3
) = 0, we can select a complement V
of V
1
in V which contains V
2
V
3
. This gives
us the decomposition
V
1
//
V
3
?
_
oo
V
2
?
OO
=
//
0
oo
0
OO
//
V
3
?
_
oo
V
2
?
OO
The rst of these summands is a multiple of
//
0
oo
0
OO
By splitting these away we get to a situation where V
1
V
2
V
3
. Similarly, we can split away
objects of the type
0
//
0
oo
1
OO
O
and
0
//
oo
0
OO
to reach a situation in which the following conditions hold
61
1. V
1
+V
2
+V
3
= V
2. V
1
V
2
= 0, V
1
V
3
= 0, V
2
V
3
= 0
3. V
1
V
2
V
3
, V
2
V
1
V
3
, V
3
V
1
V
2
But this implies that
V
1
V
2
= V
1
V
3
= V
2
V
3
= V.
So we get
dimV
1
= dimV
2
= dimV
3
= n
and
dimV = 2n
Since V
3
V
1
V
2
we can write every element of V
3
in the form
x V
3
, x = (x
1
, x
2
), x
1
V
1
, x
2
V
2
We then can dene the projections
B
1
: V
3
V
1
, (x
1
, x
2
) x
1
B
2
: V
3
V
2
, (x
1
, x
2
) x
2
Since V
3
_ V
1
, V
3
_ V
2
, these maps have to be injective and therefore are isomorphisms. We then
dene the isomorphism
A = B
2
B
1
1
Let e
1
, . . . , e
n
be a basis for V
1
. Then we get
V
1
= Ce
1
Ce
2
Ce
n
V
2
= CAe
1
CAe
2
CAe
n
V
3
= C(e
1
, Ae
1
) C(e
2
, Ae
2
) C(e
n
, Ae
n
)
So we can think of V
3
as the graph of an isomorphism A : V
1
V
2
. From this we obtain the
decomposition
V
1
//
V
3
?
_
oo
V
2
?
OO
=
n
j=1
C(1, 0)
//
C
2
C(1, 1)
?
_
oo
C(0, 1)
?
OO
These correspond to the indecomposable object
1
//
1
oo
1
OO
Here, the maps coorrespond to the embeddings into the plane of the lines x = 0, y = 0, and y = x.
Thus the quiver D
4
with the selected orientation has 12 indecomposable objects. If one were
to explicitly decompose representations for the other possible orientations, one would also nd 12
indecomposable objects.
It appears as if the number of indecomposable representations does not depend on the orienta
tion of the edges, and indeed  Gabriels theorem generalizes this observation.
62
7.4 Roots
From now on, let be a xed graph of type A
n
, D
n
, E
6
, E
7
, E
8
. We denote the adjacency matrix
of by R
.
Denition 7.10 (Cartan Matrix). We dene the Cartan matrix as
A
= 2Id R
On the lattice Z
n
(or the space 1
n
) we then dene an inner product
B(x, y) = x
T
A
y
corresponding to the graph .
Lemma 7.11. 1. B is positive denite
2. B(x, x) takes only even values for x Z
n
.
Proof. 1. This follows by denition, since is a Dynkin diagram.
2. By the denition of the Cartan matrix we get
B(x, x) = x
T
Ay =
i,j
x
i
a
ij
x
j
= 2
i
x
2
i
+
i,j, i=j
x
i
a
ij
x
j
But since A is symmetric, we obtain
B(x, x) = 2
i
x
2
i
+
i,j, i=j
x
i
a
ij
x
j
= 2
i
x
2
i
+ 2
i<j
a
ij
x
i
x
j
which is even.
Denition 7.12 (Root). A root is a shortest (with respect to B), nonzero vector in Z
n
So, a root is a nonzero vector x Z
n
such that
B(x, x) = 2
Remark 7.13. There can be only nitely many roots, since all of them have to lie in a ball of
some radius.
Denition 7.14. We call vectors of the form
i
= (0, . . . ,
ith
..
1 , . . . , 0)
simple roots.
The
i
naturally form a basis of the lattice Z
n
.
Lemma 7.15. Let be a root, =
n
i=1
k
i
i
. Then either k
i
0 for all i or k
i
0 for all i.
63
Proof. Assume the contrary, i.e. k
i
> 0, k
j
< 0. Without loss of generality, we can also assume
that k
s
= 0 for all s between i and j. We can identify the indices i, j with vertices of the graph .
i
Next, let be the edge connecting i with the next vertex towards j and i
1
j
2
Then we have i
1
, j
2
. We dene
=
m
1
k
m
m
, =
m
2
k
m
m
With this choice we get
= +.
Since k
i
> 0, k
j
< 0 we know that ,= 0, ,= 0 and therefore
B(, ) 2, B(, ) 2.
Furthermore,
B(, ) = k
i
k
i
since
1
,
2
are only connected at . But this has to be a positive number, since k
i
> 0 and k
i
0
for i
2
. This yields
B(, ) = B( +, +) = B(, )
. .
2
+2 B(, )
. .
0
+B(, )
. .
2
4
But this is a contradiction, since was assumed to be a root.
Denition 7.16 (positive and negative roots). We call a root =
i
k
i
i
a positive root, if all
k
i
0. A root which k
i
0 for all i is called a negative root.
Remark 7.17. The Lemma states that every root is either positive or negative.
Example 7.18. 1. Let be of the type A
n1
. Then the lattice L = Z
n1
can be realized as a
subgroup of the lattice Z
n
of all vectors (x
1
, . . . , x
n
) such that
i
x
i
= 0.
The vectors
1
= (1, 1, 0, . . . , 0)
2
= (0, 1, 1, 0, . . . , 0)
.
.
.
n1
= (0, . . . , 0, 1, 1)
64
naturally form a basis of L. Furthermore, the standard inner product
(x, y) =
x
i
y
i
on Z
n
restricts to the inner product B given by on L, since it takes the same values on the
basis vectors:
(
i
,
i
) = 2
(
i
,
j
) =
_
1 i, j adjacent
0 otherwise
This means that vectors of the form
(0, . . . , 0, 1, 0, . . . , 0, 1, 0, . . . , 0) =
i
+
i+1
+ +
j1
and
(0, . . . , 0, 1, 0, . . . , 0, 1, 0, . . . , 0) = (
i
+
i+1
+ +
j1
)
are the roots of L. Therefore the number of positive roots in L equals
n(n 1)
2
2. As a fact we also state the number of positive roots in the other Dynkin diagrams:
D
n
n(n 1)
E
6
36 roots
E
7
63 roots
E
8
120 roots
Denition 7.19 (Root reection). Let Z
n
be a positive root. The reection s
is dened by
the formula
s
(v) = v B(v, )
We denote s
i
by s
i
and call these simple reections.
Remark 7.20. s
xes B, since
B(s
(v), s
() =
Therefore s
i
oo
OO
We call a vertex i Q a source, if all edges connected to i point away from i.
i
oo //
Denition 7.25. Let Q be any quiver and i Q be a sink (a source). Then we let Q
i
be the
quiver obtained from Q by reversing all arrows pointing into (pointing out of) i.
We are now able to dene the reection functors.
Denition 7.26. Let Q be a quiver, i Q be a sink. Let V be a representation of Q. Then we
introduce the reection functor
F
+
i
: RepQ RepQ
i
by the rule
F
+
i
(V )
k
= V
k
if k ,= i
F
+
i
(V )
i
= ker
_
_
ji
V
j
V
i
_
_
Also, all maps stay the same but those now pointing out of i; these are replaced by the obvious
projections.
66
Denition 7.27. Let Q be a quiver, i Q be a source. Let V be a representation of Q. Let be
the canonical map
: V
i
ij
V
j
Then we dene the reection functor
F
i
: RepQ RepQ
i
by the rule
F
i
(V )
k
= V
k
if k ,= i
F
i
(V )
i
= Coker () =
_
_
ij
V
j
_
_
/(Im)
Again, all maps stay the same but those now pointing into i; these are replaced by the obvious
projections.
Proposition 7.28. Let Q be a quiver, V an indecomposable representation of Q.
1. Let i Q be a sink. Then either dimV
i
= 1, dimV
j
= 0 for j ,= i or
:
ji
V
j
V
i
is surjective.
2. Let i Q be a source. Then either dimV
i
= 1, dimV
j
= 0 for j ,= i or
: V
i
ij
V
j
is injective.
Proof. 1. Choose a complement W of Im. Then we get
V =
0
//
0
oo
0
OO
V
Since V is indecomposable, one of these summands has to be zero. If the rst summand is
zero, then has to be surjective. If the second summand is zero, then the rst has to be of
the desired form, because else we could write it as a direct sum of several objects of the type
0
//
0
oo
0
OO
which is impossible, since V was supposed to be indecomposable.
2. Follows similarly by splitting away the kernel of .
67
Proposition 7.29. Let Q be a quiver, V be a representation of Q.
1. If
:
ji
V
j
V
i
is surjective, then
F
i
F
+
i
V = V
2. If
: V
i
ij
V
j
is injective, then
F
+
i
F
i
V = V
Proof. In the following proof, we will always mean by i j that i points into j in the original
quiver Q. We only prove the rst statement and we also restrict ourselves to showing that the
spaces of V and F
i
F
+
i
V are the same. It is enough to do so for the ith space. Let
:
ji
V
j
V
i
be surjective and let
K = ker .
When applying F
+
i
, the space V
i
gets replaced by K. Furthermore, let
: K
ji
V
j
After applying F
i
, K gets replaced by
K
=
_
_
ji
V
j
_
_
/(Im)
But
Im = K
and therefore
K
=
_
_
ji
V
j
_
_
/
_
_
ker
ji
V
j
V
i
_
_
= Im
ji
V
j
V
i
by homomorphism theorem. Since was assumed to be surjective, we get
K
= V
i
Proposition 7.30. Let Q be a quiver, V be an indecomposable representation. Then F
+
i
V and
F
i
V (whenever dened) are either indecomposable or 0
68
Proof. We prove the proposition for F
+
i
V  the case F
i
V follows similarly. By 7.28 it follows that
either
:
ji
V
j
V
i
is surjective or dimV
i
= 1, dimV
j
= 0, j ,= i. In the last case
F
+
i
V = 0
So we can assume that is surjective. In this case, assume that F
+
i
V is decomposable
F
+
i
V = X Y
with X, Y ,= 0. But F
+
i
V is injective at i, since the maps are canonical embeddings. Therefore X
and Y also have to be injective at i and hence (by 7.29)
F
+
i
F
i
X = X, F
+
i
F
i
Y = Y
In particular
F
i
X ,= 0, F
i
Y ,= 0.
Therefore
V = F
i
F
+
i
V = F
i
X F
i
Y
which is a contradiction, since V was assumed to be indecomposable. So we can infer that
F
+
i
V
is indecomposable.
Proposition 7.31. Let Q be a quiver and V a representation of Q.
1. Let i Q be a sink and let V be surjective at i. Then
d(F
+
i
V ) = s
i
(d(V ))
2. Let i Q be a source and let V be injective at i. Then
d(F
i
V ) = s
i
(d(V ))
Proof. We only prove the rst statement, the second one follows similarly. Let i Q be a sink and
let
:
ji
V
j
V
i
be surjective. Let K = ker . Then
dimK =
ji
dimV
j
dimV
i
Therefore we get
_
d(F
+
i
V ) d(V )
_
i
=
ji
dimV
j
2 dimV
i
= B(d(V ),
i
)
and
_
d(F
+
i
V ) d(V )
_
j
= 0, j ,= i.
This implies
d(F
+
i
V ) d(V ) = B(d(V ),
i
)
i
d(F
+
i
V ) = d(V ) B(d(V ),
i
)
i
= s
i
(d(V ))
69
7.7 Coxeter elements
Denition 7.32. Let Q be a quiver and let be the underlying graph. Fix any labeling 1, . . . , r
of the vertices of . Then the Coxeter element c of Q corresponding to this labeling is dened as
c = s
1
s
2
. . . s
r
Lemma 7.33. Let
=
i
k
i
i
with k
i
0 for all i but not all k
i
= 0. Then there is N N, such that
c
N
must have at least one strictly negative one. It is enough to show that 1 is not an eigenvalue for c,
since
(1 +c +c
2
+ +c
M1
)v = w ,= 0
cw = c
_
1 +c +c
2
+ +c
M1
_
v = (c +c
2
+c
3
+ +c
M1
+ 1)v = w
Assume the contrary, i.e. 1 is a eigenvalue of c and let v be a corresponding eigenvector.
cv = v s
1
. . . s
r
v = v
s
2
. . . s
r
v = s
1
v
But since s
i
only changes the ith coordinate of v, we get
s
1
v = v and s
2
. . . s
r
v = v
Repeating the same procedure, we get
s
i
v = v
for all i. But this means
B(v,
i
) = 0
for all i and since B is nondegenerate, we get v = 0. But this is a contradiction, since v is an
eigenvector.
70
7.8 Proof of Gabriels theorem
Let V be an indecomposable representation of Q. We introduce a xed labeling 1, . . . r on Q, such
that i < j if one can reach j from i. This is possible, since we can assign the highest label to any
sink, remove this sink from the quiver, assign the next highest label to a sink of the remaining
quiver and so on. This way we create a labeling of the desired kind.
We now consider the sequence
V
(0)
= V, V
(1)
= F
+
r
V, V
(2)
= F
+
r1
F
+
r
V, . . .
This sequence is well dened because of the selected labeling: r has to be a sink of Q, r 1 has
to be a sink of Q
r
and so on. Furthermore we note that V
(r)
is a representation of Q again, since
every arrow has been reversed twice (since we applied a reection functor to every vertex). This
implies that we can dene
V
(r+1)
= F
+
r
V
(r)
, . . .
and continue the sequence to innity.
Theorem 7.34. There is m N, such that
d
_
V
(m)
_
=
p
for some p.
Proof. If V
(i)
is surjective at the appropriate vertex k, then
d
_
V
(i+1)
_
= d
_
F
+
k
V
(i)
_
= s
k
d
_
V
(i)
_
This implies, that if V
(0)
, . . . , V
(i1)
are surjective at the appropriate vertices, then
d
_
V
(i)
_
= . . . s
r1
s
r
d(V )
By 7.33 this cannot continue indenitely  since d
_
V
(i)
_
may not have any negative entries. Let i
be the smallest number such that V
(i)
is not surjective at the appropriate vertex. Then V
(i)
is not
surjective at this vertex (which is a sink) but by 7.30 it is indecomposable. So, by 7.28, we get
d(V
(i)
) =
p
for some p.
Corollary 7.35. Let Q be a quiver, V be any indecomposable representation. Then d(V ) is a
positive root.
Proof. By 7.34
s
i
1
. . . s
in
(d(V )) =
p
.
Since the s
i
preserve B, we get
B(d(V ), d(V )) = B(
p
,
p
) = 2.
Corollary 7.36. Let V, V
be indecomposable representations of Q such that d(V ) = d(V
). Then
V and V
are isomorphic.
71
Proof. Let i be such that
d
_
V
(i)
_
=
p
.
Then we also get d
_
V
(i)
_
=
p
. So
V
(i)
= V
(i)
=: V
i
.
Furthermore we have
V
(i)
= F
+
k
. . . F
+
r1
F
+
r
V
(0)
V
(i)
= F
+
k
. . . F
+
r1
F
+
r
V
(0)
But both V
(i1)
, . . . , V
(0)
and V
(i1)
, . . . , V
(0)
have to be surjective at the appropriate vertices.
This implies
F
r
F
r1
. . . F
k
V
i
=
_
F
r
F
r1
. . . F
k
F
+
k
. . . F
+
r1
F
+
r
V
(0)
= V
(0)
= V
F
r
F
r1
. . . F
k
F
+
k
. . . F
+
r1
F
+
r
V
(0)
= V
(0)
= V
These two corollaries show that there are only nitely many indecomposable representations
(since there are only nitely many roots) and that the dimension vector of each of them is a positive
root. The last statement of Gabriels theorem follows from
Corollary 7.37. For every positive root , there is an indecomposable representation V with
d(V ) =
Proof. Consider the sequence
s
r
, s
r1
s
r
, . . .
Consider the rst element of this sequence which is a negative root (this has to happen by 7.33)
and look at one step before that, call this element . So is a positive root and s
i
is a negative
root for some i. But since the s
i
only change one coordinate, we get
=
i
and
(s
q
. . . s
r1
s
r
) =
i
.
We let C
(i)
be the representation having dimension vector
i
. Then we dene
V = F
r
F
r1
. . . F
q
C
(i)
This is an indecomposable representation and
d(V ) = .
72