Professional Documents
Culture Documents
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 65
4.22 Polynomial representations of GL(V) . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
4.23 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
4.24 Representations of GL
2
(F
q
) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
4.24.1 Conjugacy classes in GL
2
(F
q
) . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
4.24.2 1-dimensional representations . . . . . . . . . . . . . . . . . . . . . . . . . . . 68
4.24.3 Principal series representations . . . . . . . . . . . . . . . . . . . . . . . . . . 69
4.24.4 Complementary series representations . . . . . . . . . . . . . . . . . . . . . . 72
4.25 Artins theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
4.26 Representations of semidirect products . . . . . . . . . . . . . . . . . . . . . . . . . . 74
5 Quiver Representations 76
5.1 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
5.2 Indecomposable representations of the quivers A
1
, A
2
, A
3
. . . . . . . . . . . . . . . . 79
5.3 Indecomposable representations of the quiver D
4
. . . . . . . . . . . . . . . . . . . . 81
5.4 Roots . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 85
5.5 Gabriels theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 87
5.6 Reection Functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
5.7 Coxeter elements . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 91
5.8 Proof of Gabriels theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 92
5.9 Problems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 94
6 Introduction to categories 96
6.1 The denition of a category . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
6.2 Functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
6.3 Morphisms of functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
6.4 Equivalence of categories . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
6.5 Representable functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
3
6.6 Adjoint functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
6.7 Abelian categories . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
6.8 Exact functors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
7 Structure of nite dimensional algebras 103
7.1 Projective modules . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
7.2 Lifting of idempotents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
7.3 Projective covers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 104
INTRODUCTION
Veryroughlyspeaking,representationtheorystudiessymmetryinlinearspaces. Itisabeautiful
mathematicalsubjectwhichhasmanyapplications,rangingfromnumbertheoryandcombinatorics
to geometry, probability theory, quantum mechanics and quantum eld theory.
Representation theory was born in 1896 in the work of the German mathematician F. G.
Frobenius. ThisworkwastriggeredbyalettertoFrobeniusbyR.Dedekind. InthisletterDedekind
madethefollowing observation: take themultiplication tableofanitegroupGandturnitintoa
matrix X
G
by replacing every entry g of this table by a variable x
g
. Then the determinant of X
G
factors into a product of irreducible polynomials in {x
g
}, each of which occurs with multiplicity
equaltoitsdegree. Dedekindcheckedthissurprisingfactinafewspecialcases,butcouldnotprove
it in general. So he gave this problem to Frobenius. In order to nd a solution of this problem
(which we will explain below), Frobenius created representation theory of nite groups.
1
Thepresentlecturenotesarosefromarepresentationtheorycoursegivenbytherstauthorto
the remaining six authors in March 2004 within the framework of the Clay Mathematics Institute
Research Academy for high school students, and its extended version given by the rst author to
MIT undergraduate math students in the Fall of 2008. The lectures are supplemented by many
problems and exercises, which contain a lot of additional material; the more dicult exercises are
provided with hints.
Thenotescoveranumberofstandardtopicsinrepresentationtheoryofgroups,Liealgebras,and
quivers. We mostlyfollow [FH],withtheexceptionofthesections discussingquivers,whichfollow
[BGP]. We also recommend the comprehensive textbook [CR]. The notes should be accessible to
students with a strong background in linear algebra and a basic knowledge of abstract algebra.
Acknowledgements. Theauthorsare grateful to the Clay Mathematics Institute forhosting
the rst version of this course. The rst author is very indebted to Victor Ostrik for helping him
prepare this course, and thanks Josh Nichols-Barrer and Thomas Lam for helping run the course
in2004andforusefulcomments. HeisalsoverygratefultoDarijGrinbergforverycarefulreading
ofthetext,formanyusefulcommentsandcorrections,andforsuggestingtheExercisesinSections
1.10, 2.3, 3.5, 4.9, 4.26, and6.8.
For more on the history of representation theory, see [Cu].
4
1
1 Basic notions of representation theory
1.1 What is representation theory?
Intechnicalterms,representationtheorystudiesrepresentationsofassociativealgebras. Itsgeneral
content can be very briey summarized as follows.
Anassociative algebraover aeldk isavector space Aover k equippedwithanassociative
bilinear multiplication a, bab, a, bA. We will always consider associative algebras with unit,
i.e., with an element 1 such that 1 a=a 1 =a for all aA. A basic example of an associative
algebra is the algebra EndV of linear operators from a vector space V to itself. Other important
examplesincludealgebrasdenedbygeneratorsandrelations,suchasgroupalgebrasanduniversal
enveloping algebras of Lie algebras.
A representation of an associative algebra A (also called a left A-module) is a vector space
V equippedwith a homomorphism:AEndV, i.e., a linear map preserving the multiplication
and unit.
A subrepresentation of a representation V is a subspace U V which is invariant underall
operators (a), a A. Also, if V
1
, V
2
are two representations of A then the direct sum V
1
V
2
has an obvious structure of a representation of A.
A nonzero representation V of A is said to be irreducible if its only subrepresentations are
0 and V itself, and indecomposable if it cannot be written as a direct sum of two nonzero
subrepresentations. Obviously, irreducible implies indecomposable, but not vice versa.
Typical problemsof representation theory are as follows:
1. Classify irreducible representations of a given algebra A.
2. Classify indecomposable representations of A.
3. Do 1 and 2 restricting to nite dimensional representations.
As mentioned above, the algebra A is often given to us by generators and relations. For
example, the universal enveloping algebra U of the Lie algebra sl(2) is generated by h, e, f with
dening relations
heeh= 2e, hffh=2f, ef fe=h. (1)
Thismeansthattheproblemofnding,say,N-dimensionalrepresentationsofAreducestosolving
a bunch of nonlinear algebraic equations with respect to a bunch of unknown N by N matrices,
for example system (1) with respect to unknownmatrices h, e, f.
It is really striking that such, at rst glance hopelessly complicated, systems of equations can
in fact be solved completely by methods of representation theory! For example, we will prove the
following theorem.
Theorem 1.1. Let k=C be the eld of complex numbers. Then:
(i) The algebra U has exactly one irreducible representation V
d
of each dimension, up to equiv-
alence; this representation is realized in the space of homogeneous polynomials of two variables x, y
of degree d1, and dened by the formulas
(h) =x
x
y
y
, (e) =x
y
, (f) =y
x
.
(ii)AnyindecomposablenitedimensionalrepresentationofU isirreducible. Thatis,anynite
5
|
dimensional representation of U is a direct sum of irreducible representations.
As another example consider the representation theory of quivers.
A quiver is a nite oriented graph Q. A representation of Q over a eld k is an assignment
ofak-vectorspaceV
i
toeveryvertexiofQ,andofalinearoperatorA
h
:V
i
V
j
toeverydirected
edgehgoingfromitoj (loopsandmultipleedgesareallowed). Wewillshowthatarepresentation
of a quiverQis thesame thing asa representation ofa certain algebra P
Q
called thepath algebra
of Q. Thusone may ask: what are the indecomposable nite dimensional representations of Q?
More specically, let us say that Q is of nite type if it has nitely many indecomposable
representations.
We will prove the following striking theorem, proved by P. Gabriel about 35 years ago:
Theorem 1.2. The nite type property of Q does not depend on the orientation of edges. The
connected graphs that yield quivers of nite type are given by the following list:
A
n
:
D
n
:
E
6
:
E
7
:
E
8
:
The graphs listed in the theorem are called (simply laced) Dynkin diagrams. These graphs
arise in a multitude of classication problems in mathematics, such as classication of simple Lie
algebras, singularities, platonic solids, reection groups,etc. Infact, if we neededto make contact
with an alien civilization and show them how sophisticated our civilization is, perhaps showing
them Dynkin diagrams would be the bestchoice!
Asanalexampleconsidertherepresentationtheoryofnitegroups,whichisoneofthemost
fascinating chapters of representation theory. In this theory, one considers representations of the
group algebra A=C[G] of a nite group G the algebra with basis a
g
, g G and multiplication
law a
g
a
h
= a
gh
. We will show that any nite dimensional representation of A is a direct sum of
irreducible representations, i.e., the notions of an irreducible and indecomposable representation
are the same for A (Maschkes theorem). Another striking result discussedbelow is the Frobenius
divisibility theorem: the dimension of any irreducible representation of A divides the order of G.
Finally,wewillshowhowtouserepresentationtheoryofnitegroupstoproveBurnsidestheorem:
a
any nite group of order p q
b
, where p, q are primes, is solvable. Note that this theorem does not
mention representations, which are used only in its proof; a purely group-theoretical proof of this
theorem (not using representations) exists but is much more dicult!
6
1.2 Algebras
Let us now begin a systematic discussionof representation theory.
Let k be a eld. Unless stated otherwise, we will always assume that k is algebraically closed,
i.e., any nonconstant polynomial with coecients in k has a root in k. The main example is the
eldofcomplexnumbersC,butwewillalsoconsidereldsofcharacteristicp,suchasthealgebraic
closure F
p
of the nite eld F
p
of p elements.
Denition 1.3. An associative algebra over k is a vector space A over k together with a bilinear
map AAA, (a, b)ab, such that (ab)c=a(bc).
Denition 1.4. A unit in an associative algebra A is an element 1A such that 1a=a1 =a.
Proposition 1.5. If a unit exists, it is unique.
Proof. Let 1,1
= 1
.
From now on, by an algebra A we will mean an associative algebra with a unit. We will also
assume that A=0.
Example 1.6. Here are some examples of algebras over k:
1. A=k.
2. A=k[x
1
, ..., x
n
] the algebra of polynomials in variables x
1
, ..., x
n
.
3. A=EndV the algebra of endomorphismsof a vector space V over k (i.e., linear maps, or
operators, from V to itself). The multiplication is given by composition of operators.
4. The free algebra A = k x
1
, ..., x
n
. A basis of this algebra consists of words in letters
x
1
, ..., x
n
, andmultiplication in this basis is simply concatenation of words.
5. The group algebra A=k[G] of a group G. Its basis is {a
g
, g G}, with multiplication law
a
g
a
h
=a
gh
.
Denition 1.7. An algebra A is commutative if ab=ba for all a, bA.
Forinstance,intheaboveexamples,Aiscommutativeincases1and2,butnotcommutativein
cases3(ifdimV >1),and4(ifn >1). Incase5,AiscommutativeifandonlyifGiscommutative.
Denition 1.8. A homomorphism of algebras f : A B is a linear map such that f(xy) =
f(x)f(y) for all x, yA, andf(1)=1.
1.3 Representations
Denition 1.9. A representation of an algebra A (also called a left A-module) is a vector space
V together with a homomorphismof algebras :AEndV.
Similarly, a right A-module is a space V equipped with an antihomomorphism :AEndV;
i.e., satises (ab) =(b)(a) and (1)=1.
The usual abbreviated notation for (a)v is av for a left module and va for the right module.
Thenthe propertythat is an (anti)homomorphism can bewritten asa kindof associativity law:
(ab)v=a(bv) for left modules, and (va)b=v(ab) for right modules.
Here are some examples of representations.
7
V
Example 1.10. 1. V =0.
2. V =A,and:AEndAisdenedasfollows: (a)istheoperatorofleftmultiplicationby
a, so that (a)b=ab (the usual product). This representation is called the regular representation
of A. Similarly, one can equip A with a structure of a right A-module by setting (a)b:=ba.
3. A=k. Thena representation of A is simply a vector space over k.
4. A=k x
1
, ..., x
n
. ThenarepresentationofAisjustavectorspaceV overkwithacollection
of arbitrary linear operators (x
1
), ..., (x
n
) :V V (explain why!).
Denition 1.11. AsubrepresentationofarepresentationV ofanalgebraAisasubspaceW
which is invariant underall the operators (a) :V V, aA.
For instance, 0 and V are always subrepresentations.
Denition 1.12. ArepresentationV =0ofAisirreducible(orsimple)iftheonlysubrepresenta-
tions of V are 0 and V.
Denition 1.13. Let V
1
, V
2
be two representations of an algebra A. A homomorphism (or in-
tertwining operator) : V
1
V
2
is a linear operator which commutes with the action of A, i.e.,
(av) =a(v)foranyvV
1
. Ahomomorphismissaidtobeanisomorphismofrepresentations
if it is an isomorphism of vector spaces. The set (space) of all homomorphisms of representations
V
1
V
2
is denoted by Hom
A
(V
1
, V
2
).
Note that if a linear operator : V
1
V
2
is an isomorphism of representations then so is the
linear operator
1
:V
2
V
1
(check it!).
Tworepresentationsbetweenwhichthereexistsanisomorphismaresaidtobeisomorphic. For
practical purposes,twoisomorphicrepresentationsmayberegardedasthesame,althoughthere
could be subtleties related to the fact that an isomorphism between two representations, when it
exists, is not unique.
Denition 1.14. Let V
1
, V
2
be representations of an algebra A. Then the space V
1
V
2
has an
obvious structure of a representation of A, given by a(v
1
v
2
) =av
1
av
2
.
Denition 1.15. Anonzerorepresentation V ofanalgebra Aissaidtobeindecomposableifitis
not isomorphic to a direct sumof two nonzero representations.
It is obvious that an irreducible representation is indecomposable. On the other hand, we will
see below that the converse statement is false in general.
Oneofthemainproblemsofrepresentationtheoryistoclassifyirreducibleandindecomposable
representations of a given algebra upto isomorphism. This problem is usually hardand often can
be solved only partially (say, for nite dimensional representations). Below we will see a number
of examples in which this problem is partially or fully solved for specic algebras.
We will now prove our rst result Schurs lemma. Although it is very easy to prove, it is
fundamental in the whole subject of representation theory.
Proposition 1.16. (Schurs lemma) Let V
1
, V
2
be representations of an algebra A over any eld
F (which need not be algebraically closed). Let : V
1
V
2
be a nonzero homomorphism of
representations. Then:
(i) If V
1
is irreducible, is injective;
8
(ii) If V
2
is irreducible, is surjective.
Thusm, if both V
1
and V
2
are irreducible, is an isomorphism.
Proof. (i) The kernel K of is a subrepresentation of V
1
. Since = 0, this subrepresentation
cannot be V
1
. So by irreducibility of V
1
we have K =0.
(ii) The image I of is a subrepresentation of V
2
. Since =0, this subrepresentation cannot
be 0. So by irreducibility of V
2
we have I =V
2
.
Corollary 1.17. (Schurs lemma for algebraically closed elds) Let V be a nite dimensional
irreducible representation of an algebra A over an algebraically closed eld k, and :V V is an
intertwining operator. Then = Id for some k (a scalar operator).
Remark. Note that this Corollary is false over the eld of real numbers: it suces to take
A=C(regarded as an R-algebra), and V =A.
Proof. Letbeaneigenvalueof(arootofthecharacteristicpolynomialof). Itexistssincek is
analgebraically closed eld. Thentheoperator Idisan intertwining operatorV V, which
is not an isomorphism (since its determinant is zero). Thus by Proposition 1.16 this operator is
zero, hence the result.
Corollary 1.18. Let A be a commutative algebra. Then every irreducible nite dimensional rep-
resentation V of A is 1-dimensional.
Remark. Notethata1-dimensionalrepresentationofanyalgebraisautomaticallyirreducible.
Proof. Let V beirreducible. For any element aA, the operator (a) :V V is an intertwining
operator. Indeed,
(a)(b)v =(ab)v =(ba)v=(b)(a)v
(the second equality is true since the algebra is commutative). Thus, by Schurs lemma, (a) is
a scalar operator for any a A. Hence every subspace of V is a subrepresentation. But V is
irreducible,so0andV aretheonlysubspacesofV. ThismeansthatdimV =1(sinceV =0).
Example 1.19. 1. A=k. Since representations of A are simply vector spaces, V =A isthe only
irreducible and the only indecomposable representation.
2. A = k[x]. Since this algebra is commutative, the irreducible representations of A are its
1-dimensional representations. As we discussed above, they are dened by a single operator (x).
Inthe 1-dimensional case, thisis justa numberfromk. Soall theirreduciblerepresentations of A
are V
=span{as} and S
r
=span{sb}, the left, respectively right, ideal generated by S.
10
1.5 Quotients
Let A be an algebra and I a two-sided ideal in A. Then A/I is the set of (additive) cosets of I.
Let:AA/I bethequotientmap. WecandenemultiplicationinA/I by(a) (b):=(ab).
This is well dened because if (a) =(a
) then
(a
b) =(ab+ (a
a)b) =(ab)
because (a
) then
(ab
) =(ab+a(b
b))=(ab) +(a(b
b))=(ab)
because a(b
Proof. (i)Firstletusshowthattheelementsx
i
y
j
areaspanningsetforA. Todothis,notethat
any word in x, y can be ordered to have all the x on the left of the y, at the cost of interchanging
some x and y. Since yxxy = 1, this will lead to error terms, but these terms will be sums of
monomialsthathaveasmallernumberoflettersx, y thantheoriginalword. Therefore,continuing
this process, we can order everything and represent any word as a linear combination of x
i
y
j
.
Theproofthatx
i
y
j
arelinearlyindependentisbasedonrepresentationtheory. Namely,letabe
avariable,andE =t
a
k[a][t, t
1
](heret
a
isjustaformalsymbol,soreallyE =k[a][t, t
1
]). ThenE
df d(t
a+n
)
isarepresentationofAwithactiongivenbyxf =tf andyf =
dt
(where
dt
:=(a+n)t
a+n1
).
i
Supposenow that we have a nontrivial linear relation
c
ij
x y
j
=0. Thenthe operator
j
L= c
ij
t
dt
acts by zero in E. Let us write L as
r
j
L= Q
j
(t) ,
dt
j=0
where Q
r
=0. Then we have
r
Lt
a
=
Q
j
(t)a(a1)...(aj+1)t
aj
.
j=0
This must be zero, so we have
r
Q
j
(t)a(a1)...(aj+1)t
j
= 0 in k[a][t, t
1
]. Taking the
j=0
leading term in a, we get Q
r
(t)=0, a contradiction.
(ii)Anywordinx, y, x
1
, y
1
canbeorderedatthecostofmultiplyingitbyapowerofq. This
easily implies both the spanning propertyand the linear independence.
Remark. The proof of (i) shows that the Weyl algebra A can be viewed as the algebra of
polynomial dierential operators in one variable t.
The proof of (i) also bringsup the notion of a faithful representation.
Denition. A representation :AEndV is faithful if is injective.
For example, k[t] is a faithful representation of the Weyl algebra, if k has characteristic zero
(check it!), but not in characteristic p, where (d/dt)
p
Q = 0 for any polynomial Q. However, the
representation E =t
a
k[a][t, t
1
], as weve seen, is faithful in any characteristic.
Problem 1.26. Let A be the Weyl algebra, generated by two elements x, y with the relation
yxxy1 = 0.
(a) If chark= 0, what are the nite dimensional representations of A? What are the two-sided
ideals in A?
Hint. For the rstquestion, usethe factthat for two square matrices B, C,Tr(BC) =Tr(CB).
For thesecond question, show that any nonzero two-sided ideal inA contains anonzero polynomial
in x, and use this to characterize this ideal.
Suppose for the rest of the problem that chark=p.
(b) What is the center of A?
12
WedenotethesetofverticesofthequiverQasI,andthesetofedgesasE. ForanedgehE,
let h
, h
Denition 1.30. Arepresentation of aquiver Qis anassignment to each vertex iI ofa vector
space V
i
andto each edge hE of a linear map x
h
:V
h
V
h
.
Itturnsoutthatthetheoryofrepresentationsofquiversisapartofthetheoryofrepresentations
of algebras in the sense that for each quiver Q, there exists a certain algebra P
Q
, called the path
algebra of Q, such that a representation of the quiver Q is the same as a representation of the
algebra P
Q
. We shall rst dene the path algebra of a quiver and then justify our claim that
representations of these two objects are the same.
Denition 1.31. The path algebra P
Q
of a quiver Q is the algebra whose basis is formed by
oriented paths in Q, including the trivial paths p
i
, i I, corresponding to the vertices of Q, and
multiplication is concatenation of paths: ab is the path obtained by rst tracing b and then a. If
two paths cannot be concatenated, the productis dened to bezero.
Remark 1.32. It is easy to see that for a nite quiver
p
i
=1, so P
Q
is an algebra with unit.
iI
Problem 1.33. Show that the algebra P
Q
is generated by p
i
for i I and a
h
for h E with the
dening relations:
13
1. p
2
i
=p
i
, p
i
p
j
= 0 for i= j
2. a
h
p
h
=a
h
, a
h
p
j
= 0 for j =h
3. p
h
a
h
=a
h
, p
i
a
h
= 0 for i=h
We nowjustifyourstatement thatarepresentationofaquiveristhesamethingasarepresen-
tation of the path algebra of a quiver.
LetVbearepresentationofthepathalgebraP
Q
. Fromthisrepresentation,wecanconstructa
representation of Q as follows: let V
i
=p
i
V, and forany edge h, let x
h
=a
h
|
p
h
V
:p
h
Vp
h
V
be the operator correspondingto the one-edge path h.
Similarly, let (V
i
, x
h
) be a representation of a quiver Q. From this representation, we can
construct a representation of the path algebra P
Q
: let V =
i
V
i
, let p
i
: V V
i
V be the
projectionontoV
i
,andforanypathp=h
1
...h
m
leta
p
=x
h
1
...x
hm
:V
h
V
h
bethecomposition
m
1
of the operators correspondingto the edges occurring in p.
It is clear that the above assignments V(p
i
V) and (V
i
)
V
i
are inverses of each other.
i
Thus,wehaveabijectionbetweenisomorphismclasses ofrepresentationsofthealgebra P
Q
andof
the quiver Q.
Remark 1.34. In practice, it is generally easier to consider a representation of a quiver as in
Denition 1.30.
We lastly dene several previous concepts in the context of quivers representations.
Denition 1.35. Asubrepresentationofarepresentation(V
i
, x
h
)ofaquiverQisarepresentation
(W
i
, x
h
)whereW
i
V
i
forall iI andwherex
h
(W
h
)W
h
andx
h
=x
h
|
W
h
:W
h
W
h
for
all hE.
Denition 1.36. Thedirectsumoftworepresentations(V
i
, x
h
)and(W
i
, y
h
)istherepresentation
(V
i
W
i
, x
h
y
h
).
As with representations of algebras, a nonzero representation (V
i
) of a quiver Q is said to be
irreducible if its only subrepresentations are (0) and (V
i
) itself, and indecomposable if it is not
isomorphic to a direct sum of two nonzero representations.
Denition 1.37. Let (V
i
, x
h
) and (W
i
, y
h
) be representations of the quiver Q. A homomorphism
: (V
i
) (W
i
) of quiver representations is a collection of maps
i
: V
i
W
i
such that
y
h
h
=
h
x
h
for all hE.
Problem 1.38. Let A be a Z
+
-graded algebra, i.e., A =
n0
A[n], and A[n] A[m] A[n+m].
If A[n] is nite dimensional, it is useful to consider the Hilbert series h
A
(t) =
dim
A[n]t
n
(the
generating function of dimensions of A[n]). Often this series converges to a rational function, and
the answer is written in the form of such function. For example, if A=k[x] and deg(x
n
) =n then
h
A
(t) = 1 +t+t
2
+...+t
n
+...=
1
1t
Find the Hilbert series of:
(a) A=k[x
1
, ..., x
m
] (where the grading is by degree of polynomials);
(b) A=k < x
1
, ..., x
m
> (the grading is by length of words);
14
(c) A is the exterior (=Grassmann) algebra
k
[x
1
, ..., x
m
], generated over some eld k by
x
1
, ..., x
m
with the dening relations x
i
x
j
+x
j
x
i
= 0 and x
2
i
= 0 for all i, j (the grading is by
degree).
(d) A is the path algebra P
Q
of a quiver Q (the grading is dened by deg(p
i
) = 0, deg(a
h
) = 1).
Hint. The closed answer is written in terms of the adjacency matrix M
Q
of Q.
1.9 Lie algebras
Let g be a vector space over a eld k, and let [,] :ggg be a skew-symmetric bilinear map.
(That is, [a, a]=0, andhence [a, b] =[b, a]).
Denition 1.39. (g,[,]) is a Lie algebra if [,] satises the Jacobi identity
[a, b], c + [b, c], a + [c, a], b = 0. (2)
Example 1.40. Some examples of Lie algebras are:
1. Any space g with [,]=0 (abelian Lie algebra).
2. Any associative algebra A with [a, b] =abba .
3. Any subspace U of an associative algebra A such that [a, b]U for all a, bU.
4. The space Der(A) of derivations of an algebra A, i.e. linear maps D : A A which satisfy
the Leibniz rule:
D(ab) =D(a)b+aD(b).
Remark 1.41. Derivations are important because they are the innitesimal version of auto-
morphisms. For example, assume that g(t) is a dierentiable family of automorphisms of a -
nite dimensional algebra A over R or C parametrized by t (, ) such that g(0) = Id. Then
D :=g
0 0
0
0 0 0
It has the basis
_ _ _ _ _ _
0 0 0 0 1 0 0 0 1
x=
_
0 0 1
_
y=
_
0 0 0
_
c=
_
0 0 0
_
0 0 0 0 0 0 0 0 0
with relations [y, x] =c and [y, c] = [x, c]=0.
4. The algebra a(1) of matrices (
0 0
)
Its basis consists of X = (
1 0
) and Y = (
0 1
), with [X, Y] =Y.
0 0 0 0
5. so(n), the space of skew-symmetric nn matrices, with [a, b] =abba.
Exercise. Show that Example 1 is a special case of Example 5 (for n=3).
Denition 1.42. Let g
1
,g
2
be Lie algebras. A homomorphism : g
1
g
2
of Lie algebras is a
linear map such that ([a, b])=[(a), (b)].
Denition 1.43. A representation of a Lie algebra g is a vector space V with a homomorphism
of Lie algebras :gEndV.
Example 1.44. Some examples of representations of Lie algebras are:
1. V =0.
2. Any vector space V with =0 (the trivial representation).
3. The adjoint representation V =g with (a)(b):=[a, b].
That this is a representation follows from Equation (2).
It turns out that a representation of a Lie algebra g is the same thing as a representation of a
certain associative algebra U(g). Thus, as with quivers, we can view the theory of representations
of Lie algebras as a part of the theory of representations of associative algebras.
k
Denition 1.45. Let g be a Lie algebra with basis x
i
and [,] denedby [x
i
, x
j
] =
k
c
ij
x
k
. The
universal enveloping algebra U(g) is the associative algebra generated by the x
i
s with the
dening relations x
i
x
j
x
j
x
i
=
k
c
ij
k
x
k
.
Remark. Thisisnotaverygooddenitionsinceitdependsonthechoice ofabasis. Laterwe
will give an equivalent denition which will bebasis-independent.
Example 1.46. Theassociative algebraU(sl(2))isthealgebrageneratedbye,f,hwithrelations
heeh= 2e hf fh=2f ef fe=h.
Example1.47. ThealgebraU(H),whereHistheHeisenbergLiealgebra,isthealgebragenerated
by x, y, c with the relations
yxxy=c yccy= 0 xccx= 0.
Note that the Weyl algebra is the quotient of U(H) by the relation c=1.
16
1.10 Tensor products
Inthissubsectionwerecall thenotionoftensorproductofvector spaces,whichwillbeextensively
used below.
Denition1.48. ThetensorproductVW ofvectorspacesV andW overaeldkisthequotient
of the space V W whose basis is given by formal symbols vw, v V, w W (called pure
tensors), by the subspace spannedby the elements
(v
1
+v
2
)wv
1
wv
2
w, v(w
1
+w
2
)vw
1
vw
2
, avwa(vw), vawa(vw),
where vV, wW, ak.
Exercise. Show that V W can be equivalently dened as the quotient of the free abelian
group V W generated by vw, vV, wW by the subgroupgenerated by
(v
1
+v
2
)wv
1
wv
2
w, v(w
1
+w
2
)vw
1
vw
2
, avwvaw,
where vV, wW, ak.
Thisallowsonetodenethetensorproductofanynumberofvectorspaces,V
1
...V
n
. Note
that this tensor productis associative, in the sense that (V
1
V
2
)V
3
can be naturally identied
with V
1
(V
2
V
3
).
Inparticular,peopleoftenconsidertensorproductsoftheformV
n
=V ...V (ntimes)for
agivenvectorspaceV,and,moregenerally, E :=V
n
(V
)
m
. Thisspaceiscalledthespaceof
tensors of type (m, n) on V. For instance, tensors of type (0,1) are vectors, of type (1,0) - linear
functionals(covectors), oftype(1,1)- linearoperators,oftype(2,0)- bilinearforms,oftype(2,1)
- algebra structures, etc.
If V isnite dimensional with basise
i
, i= 1, ..., N, ande
i
is the dual basisof V
, then a basis
of E is the set of vectors
e
i
1
...e
in
e
j
1
...e
jm
,
and a typical element of E is
N
T
i
1
...in j
1
jm
j
1
...jm
e
i
1
...e
in
e ...e ,
i
1
,...,in,j
1
,...,jm=1
where T is a multidimensional table of numbers.
Physicists dene a tensor as a collection of such multidimensional tables T
B
attached to every
basis B in V, which change according to a certain rule when the basis B is changed. Here it is
important to distinguish upper and lower indices, since lower indices of T correspond to V and
upper ones to V
. The physicists dont write the sum sign, but remember that one should sum
over indices that repeat twice - once as an upper index and once as lower. This convention is
called the Einstein summation, and it also stipulates that if an index appears once, then there is
no summation over it, while no index is supposedto appear more than once as an upperindex or
more than once as a lower index.
Onecanalsodenethetensorproductoflinearmaps. Namely,ifA:V V
andB :W W
arelinearmaps,thenonecandenethelinearmapAB :V W V
givenbytheformula
(AB)(vw) =AvBw (check that this is well dened!)
The most important properties of tensor productsare summarized in the following problem.
17
Problem 1.49. (a) Let U be any k-vector space. Construct a natural bijection between bilinear
maps V W U and linear maps V W U.
(b) Show that if {v
i
} is a basis of V and {w
j
} is a basis of W then {v
i
w
j
} is a basis of
V W.
(c) Construct a natural isomorphism V
1
, ...,
N
. Find Tr(S
n
A), Tr(
n
A).
(g) Show that
N
A = det(A)Id, and use this equality to give a one-line proof of the fact that
det(AB)=det(A) det(B).
Remark. Note that a similar denitionto the above can beusedto denethe tensor product
V
A
W,whereAisanyring,V isarightA-module,andW isaleftA-module. Namely, V
A
W
is the abelian group which is the quotient of the group V W freely generated by formal symbols
vw, vV, wW, modulothe relations
(v
1
+v
2
)wv
1
wv
2
w, v(w
1
+w
2
)vw
1
vw
2
, vawvaw, aA.
Exercise. Throughout this exercise, we let k be an arbitrary eld (not necessarily of charac-
teristic zero, and not necessarily algebraically closed).
If A and B are two k-algebras, then an (A, B)-bimodule will mean a k-vector space V with
both a left A-module structure and a right B-module structure which satisfy (av)b = a(vb) for
any v V, a A and b B. Note that both the notions of left A-module and right A-
moduleareparticularcasesofthenotionofbimodules;namely, aleftA-moduleisthesameasan
(A, k)-bimodule, anda right A-module is the same as a (k, A)-bimodule.
Let B beak-algebra, W aleft B-moduleandV arightB-module. We denotebyV
B
W the
k-vector space (V
k
W)/vbwvbw|vV, wW, bB. We denote the projection of
a pure tensor vw (with vV and wW) onto the space V
B
W by v
B
w. (Note that this
tensor productV
B
W is the one denedin the Remark after Problem1.49.)
If, additionally, A is another k-algebra, and if the right B-module structure on V is part of an
(A, B)-bimodule structure, then V
B
W becomes a left A-module by a(v
B
w) = av
B
w for
any aA, vV andwW.
Similarly,ifC isanotherk-algebra,andiftheleftB-modulestructureonW ispartofa(B, C)-
bimodule structure, then V
B
W becomes a right C-module by (v
B
w)c = v
B
wc for any
18
cC, vV andwW.
IfV isan(A, B)-bimoduleandW isa (B, C)-bimodule,thenthesetwo structuresonV
B
W
can be combined into one (A, C)-bimodule structure on V
B
W.
(a) Let A, B, C, D be four algebras. Let V be an (A, B)-bimodule, W be a (B, C)-bimodule,
and X a (C, D)-bimodule. Prove that (V
B
W)
C
X
V
B
(W
C
X) as (A, D)-bimodules. =
The isomorphism (from left to right) is given by (v
B
w)
C
x v
B
(w
C
x) for all v V,
wW andxX.
(b) If A, B, C are three algebras, and if V is an (A, B)-bimodule and W an (A, C)-bimodule,
then the vector space Hom
A
(V, W) (the space of all left A-linear homomorphisms from V to W)
canonicallybecomesa(B, C)-bimodulebysetting(bf) (v) =f(vb)forallbB,f Hom
A
(V, W)
and vV and (fc) (v) =f(v)c for all cC, f Hom
A
(V, W) and vV.
LetA,B,C,Dbefouralgebras. LetV be a(B, A)-bimodule,W be a(C, B)-bimodule,andX a
(C, D)-bimodule. ProvethatHom
B
(V,Hom
C
(W, X))
=Hom
C
(W
B
V, X)as (A, D)-bimodules.
The isomorphism (from left to right) is given by f (w
B
vf(v)w) for all v V, w W
and f Hom
B
(V,Hom
C
(W, X)).
1.11 The tensor algebra
The notion of tensor product allows us to give more conceptual (i.e., coordinate free) denitions
of the free algebra, polynomial algebra, exterior algebra, and universal enveloping algebra of a Lie
algebra.
Namely,givenavectorspaceV,deneitstensoralgebraTV overaeldktobeTV =
n0
V
n
,
with multiplication dened by a b:=ab, aV
n
, bV
m
. Observe that a choice of a basis
x
1
, ..., x
N
in V denes an isomorphismof TV with the free algebra k < x
1
, ..., x
n
>.
Also, one can make the following denition.
Denition 1.50. (i)ThesymmetricalgebraSV ofV isthequotientofTV bytheidealgenerated
by vwwv, v, wV.
(ii) Theexterior algebraV ofV isthequotientofTV bytheidealgenerated byvv,vV.
(iii) IfV is a Lie algebra, the universal enveloping algebra U(V) of V is the quotient of TV by
the ideal generated by vwwv[v, w], v, wV.
It is easy to see that a choice of a basis x
1
, ..., x
N
in V identies SV with the polynomial
algebra k[x
1
, ..., x
N
], V with the exterior algebra
k
(x
1
, ..., x
N
), and the universal enveloping
algebra U(V) with one denedpreviously.
Also, it is easy to see that we have decompositions SV =
n0
S
n
V, V =
n0
n
V.
1.12 Hilberts third problem
Problem 1.51. It is known that if A and B are two polygons of the same area then A can be cut
by nitely many straight cuts into pieces from which one can make B. David Hilbert asked in 1900
whether it is true for polyhedra in 3 dimensions. In particular, is it true for a cube and a regular
tetrahedron of the same volume?
19
The answer is no, as was found by Dehn in 1901. The proof is very beautiful. Namely, to
any polyhedron A let us attach its Dehn invariant D(A) in V =R(R/Q) (the tensor product
of Q-vector spaces). Namely,
(a)
D(A) = l(a) ,
a
where a runs over edges of A, and l(a), (a) are the length of a and the angle at a.
(a) Show that if you cut A into B and C by a straight cut, then D(A) =D(B) +D(C).
(b) Show that =arccos(1/3)/ is not a rational number.
Hint. Assumethat= 2m/n,forintegersm, n. Deducethatrootsoftheequationx+x
1
= 2/3
are roots of unity of degree n. Conclude that x
k
+x
k
has denominator 3
k
and get a contradiction.
(c) Using (a) and (b), show that the answer to Hilberts question is negative. (Compute the
Dehn invariant of the regular tetrahedron and the cube).
1.13 Tensor products and duals of representations of Lie algebras
Denition 1.52. The tensor product of two representations V, W of a Lie algebra g is the space
V W with
VW
(x) =
V
(x)Id+Id
W
(x).
Denition 1.53. The dual representation V
to V with
V
(x) =
V
(x)
.
It is easy to check that these are indeed representations.
Problem 1.54. Let V, W, U be nite dimensional representations of a Lie algebra g. Show that
the space Hom
g
(V W, U) is isomorphic to Hom
g
(V, U W
). (Here Hom
g
:=Hom
U(g)
).
1.14 Representations of sl(2)
This subsection is devoted to the representation theory of sl(2), which is of central importance in
many areas of mathematics. It is useful to study this topic by solving the following sequence of
exercises, which every mathematician should do, in one form or another.
Problem 1.55. According to the above, a representation of sl(2) is just a vector space V with a
triple of operators E, F, H such that HE EH = 2E, HF FH = 2F, EF FE = H (the
corresponding map is given by (e) =E, (f) =F, (h) =H).
Let V be a nite dimensional representation of sl(2) (the ground eld in this problem is C).
(a) Take eigenvalues of H and pick one with the biggest real part. Call it . Let V() be the
generalized eigenspace corresponding to . Show that E|
V
()
= 0.
(b) Let W be any representation of sl(2) and w W be a nonzero vector such that Ew = 0.
For any k > 0 nd a polynomial P
k
(x) of degree k such that E
k
F
k
w = P
k
(H)w. (First compute
EF
k
w, then use induction in k).
(c) Let v V() be a generalized eigenvector of H with eigenvalue . Show that there exists
N >0 such that F
N
v= 0.
(d) Show that H is diagonalizable on V
(), and
compute E
N
F
N
v, vV
.
(g) Show that the operator C =EF +FE+H
2
/2 (the so-called Casimir operator) commutes
with E, F, H and equals
(
2
+2)
Id on V
.
Now it will be easy to prove the direct sum decomposition. Namely, assume the contrary, and
let V be a reducible representation of the smallest dimension, which is not a direct sum of smaller
representations.
(h) Show that C has only one eigenvalue on V, namely
(
2
+2)
for some nonnegative integer .
(usethatthegeneralizedeigenspacedecompositionofC mustbeadecompositionofrepresentations).
(i) Show that V has a subrepresentation W = V
v
i
). Show that V
i
are subrepresentations of V and derive a
contradiction with the fact that V cannot be decomposed.
(l)(Jacobson-Morozov Lemma)LetV beanitedimensionalcomplexvectorspaceandA:V
V a nilpotent operator. Show that there exists a unique, up to an isomorphism, representation of
sl(2) on V such that E =A. (Use the classication of the representations and the Jordan normal
form theorem)
(m) (Clebsch-Gordan decomposition) Find the decomposition into irreducibles of the represen-
tation V
of sl(2).
Hint. For a nite dimensional representation V of sl(2) it is useful to introduce the character
V
(x) = Tr(e
xH
), x C. Show that
VW
(x) =
V
(x) +
W
(x) and
VW
(x) =
V
(x)
W
(x).
ThencomputethecharacterofV
andofV
andderivethedecomposition. Thisdecomposition
is of fundamental importance in quantum mechanics.
(n) Let V =C
M
C
N
, and A=J
M
(0)Id
N
+Id
M
J
N
(0), where J
n
(0) is the Jordan block
of size n with eigenvalue zero (i.e., J
n
(0)e
i
=e
i1
, i= 2, ..., n, and J
n
(0)e
1
= 0). Find the Jordan
normal form of A using (l),(m).
1.15 Problems on Lie algebras
Problem 1.56. (Lies Theorem) The commutant K(g) of a Lie algebra g is the linear span
of elements [x, y], x, y g. This is an ideal in g (i.e., it is a subrepresentation of the adjoint
representation). A nite dimensional Lie algebra g over a eld k is said to be solvable if there
exists n such that K
n
(g) = 0. Prove the Lie theorem: if k = C and V is a nite dimensional
irreducible representation of a solvable Lie algebra g then V is 1-dimensional.
21
Hint. Prove the result by induction in dimension. By the induction assumption, K(g) has a
common eigenvector v in V, that is there is a linear function :K(g)C such that av =(a)v
for any a K(g). Show that g preserves common eigenspaces of K(g) (for this you will need to
showthat([x, a])=0forxgandaK(g). Toprovethis,considerthesmallestvectorsubspace
U containing v and invariant under x. This subspace is invariant under K(g) and any a K(g)
acts with trace dim(U)(a) in this subspace. In particular 0=Tr([x, a])=dim(U)([x, a]).).
Problem 1.57. Classify irreducible nite dimensional representations of the two dimensional Lie
algebra with basis X, Y and commutation relation [X, Y] = Y. Consider the cases of zero and
positive characteristic. Is the Lie theorem true in positive characteristic?
Problem 1.58. (hard!) For any element x of a Lie algebra g let ad(x) denote the operator g
g, y[x, y]. Consider the Lie algebra g
n
generated by two elements x, y with the dening relations
ad(x)
2
(y) =ad(y)
n+1
(x) = 0.
(a) Show that the Lie algebras g
1
,g
2
,g
3
are nite dimensional and nd their dimensions.
(b) (harder!) Show that the Lie algebra g
4
has innite dimension. Construct explicitly a basis
of this algebra.
22
2 General results of representation theory
2.1 Subrepresentations in semisimple representations
Let A be an algebra.
Denition 2.1. A semisimple (or completely reducible) representation of A is a direct sum of
irreducible representations.
Example. Let V be an irreducible representation of A of dimension n. Then Y = End(V),
withaction of Abyleft multiplication, isasemisimple representation ofA, isomorphicto nV (the
direct sum of n copies of V). Indeed, any basis v
1
, ..., v
n
of V gives rise to an isomorphism of
representations End(V)nV, given by x(xv
1
, ..., xv
n
).
Remark. Note that by Schurs lemma, any semisimple representation V of A is canonically
identiedwith
X
Hom
A
(X, V)X,whereX runsoverallirreduciblerepresentationsofA. Indeed,
wehaveanaturalmapf :
X
Hom(X, V)X V,givenbygxg(x),xX,gHom(X, V),
and it is easy to verify that this map is an isomorphism.
Well see now how Schurs lemma allows usto classify subrepresentations in nite dimensional
semisimple representations.
Proposition 2.2. Let V
i
,1 i m be irreducible nite dimensional pairwise nonisomorphic
m
representations of A, and W be a subrepresentation of V =
i=1
n
i
V
i
. Then W is isomorphic to
m
i=1
r
i
V
i
, r
i
n
i
, and the inclusion :W V is a direct sum of inclusions
i
:r
i
V
i
n
i
V
i
given
by multiplication of a row vector of elements of V
i
(of length r
i
) by a certain r
i
-by-n
i
matrix X
i
with linearly independent rows: (v
1
, ..., v
r
i
) = (v
1
, ..., v
r
i
)X
i
.
m
Proof. Theproofisbyinductioninn:=
i=1
n
i
. Thebaseofinduction(n=1)isclear. Toperform
the induction step, let us assume that W is nonzero, and x an irreducible subrepresentation
P W. SuchP exists(Problem1.20).
2
Now,bySchurslemma,P isisomorphictoV
i
forsomei,
and the inclusion |
P
:P V factors through n
i
V
i
, and upon identication of P with V
i
is given
by the formula v(vq
1
, ..., vq
n
i
), where q
l
k are not all zero.
Now note that the group G
i
= GL
n
i
(k) of invertible n
i
-by-n
i
matrices over k acts on n
i
V
i
by (v
1
, ..., v
n
i
) (v
1
, ..., v
n
i
)g
i
(and by the identity on n
j
V
j
, j = i), and therefore acts on the
set of subrepresentations of V, preserving the property we need to establish: namely, under the
action of g
i
, the matrix X
i
goes to X
i
g
i
, while X
j
, j = i dont change. Take g
i
G
i
such that
(q
1
, ..., q
n
i
)g
i
= (1,0, ...,0). Then Wg
i
contains the rst summand V
i
of n
i
V
i
(namely, it is Pg
i
),
henceWg
i
=V
i
W
,whereW
n
1
V
1
...(n
i
1)V
i
...n
m
V
m
isthekerneloftheprojection
of Wg
i
to the rst summand V
i
along the other summands. Thus the required statement follows
from the induction assumption.
Remark 2.3. InProposition 2.2, it is not important that k is algebraically closed, nor it matters
that V is nite dimensional. Ifthese assumptionsare dropped,the onlychange neededis that the
entries ofthe matrix X
i
are nolonger ink butinD
i
=End
A
(V
i
), whichis, as we know, a division
algebra. The proofof this generalized version of Proposition 2.2 is the same as before (check it!).
2
Another proof of the existence of P, which does not use the nite dimensionality of V, is by induction in n.
Namely, if W itself is not irreducible, let K be the kernel of the projection of W to the rst summand V1. Then
K is a subrepresentation of (n1 1)V1 ...nmVm, which is nonzero since W is not irreducible, so K contains an
irreducible subrepresentation by the induction assumption.
23
2.2 The density theorem
Let A be an algebra over an algebraically closed eld k.
Corollary 2.4. Let V be an irreducible nite dimensional representation of A, and v
1
, ..., v
n
V
be any linearly independent vectors. Then for any w
1
, ..., w
n
V there exists an element a A
such that av
i
=w
i
.
Proof. Assume the contrary. Then the image of the map A nV given by a (av
1
, ..., av
n
)
is a proper subrepresentation, so by Proposition 2.2 it corresponds to an r-by-n matrix X, r <
n. Thus there exist vectors u
1
, ..., u
r
V such that (u
1
, ..., u
r
)X = (v
1
, ..., v
n
). Let (q
1
, ..., q
n
)
be a nonzero vector such that X(q
1
, ..., q
n
)
T
= 0 (it exists because r < n). Then
q
i
v
i
=
(u
1
, ..., u
r
)X(q
1
, ..., q
n
)
T
= 0, i.e.
q
i
v
i
= 0 - a contradiction with the linear independence of
v
i
.
Theorem2.5. (theDensityTheorem). (i)LetV beanirreduciblenitedimensionalrepresentation
of A. Then the map :AEndV is surjective.
(ii) Let V = V
1
...V
r
, where V
i
are irreducible pairwise nonisomorphic nite dimensional
representations of A. Then the map
r
i
:A
r
End(V
i
) is surjective.
i=1 i=1
Proof. (i)LetBbetheimageofAinEnd(V). WewanttoshowthatB =End(V). LetcEnd(V),
v
1
, ..., v
n
be a basis of V, and (c
ij
) the matrix of c in this basis. Let w
j
=
v
i
c
ij
. By Corollary
2.4, there exists aA such that av
i
=w
i
. Thena maps to c, so cB, and we are done.
(ii)LetB
i
betheimageofAinEnd(V
i
),andBbetheimageofAin
r
End(V
i
). Recallthatas
i=1
arepresentationofA,
r
End(V
i
)issemisimple: itisisomorphicto
r
d
i
V
i
,whered
i
=dimV
i
.
i=1 i=1
Then by Proposition 2.2, B =
i
B
i
. On the other hand, (i) implies that B
i
=End(V
i
). Thus (ii)
follows.
2.3 Representations of direct sums of matrix algebras
In this section we consider representations of algebras A=
i
Mat
d
i
(k) for any eld k.
Theorem 2.6. Let A =
i
r
=1
Mat
d
i
(k). Then the irreducible representations of A are V
1
=
k
d
1
, . . . , V
r
= k
dr
, and any nite dimensional representation of A is a direct sum of copies of
V
1
, . . . , V
r
.
In order to prove Theorem 2.6, we shall need the notion of a dual representation.
Denition 2.7. (Dual representation) Let V be a representation of any algebra A. Then the
dual representation V
is an n-dimensional representation of A
op
. But (Mat
d
i
(k))
op
= Mat
d
i
(k) with
isomorphism (X) = X
T
, as (BC)
T
= C
T
B
T
. = A
op
Thus, A
and X
may be viewed as an
n-dimensional representation of A. Dene
:A
AX
n copies
24
by
(a
1
, . . . , a
n
) =a
1
y
1
+ +a
n
y
n
where {y
i
} is a basis of X
:X A
n
isinjective. ButA
n
=A
n
as representations ofA (check it!). Hence, Im
=X is asubrepresen-
tation of A
n
. Next, Mat
d
i
(k) =d
i
V
i
, so A=
r
i=1
d
i
V
i
, A
n
=
r
i=1
nd
i
V
i
, as a representation of A.
Hence by Proposition 2.2, X =
i
r
=1
m
i
V
i
, as desired.
Exercise. The goal of this exercise is to give an alternative proof of Theorem 2.6, not using
any of the previous results of Chapter 2.
LetA
1
,A
2
,...,A
n
benalgebraswithunits1
1
, 1
2
,..., 1
n
,respectively. LetA=A
1
A
2
...A
n
.
Clearly, 1
i
1
j
=
ij
1
i
, andthe unit of A is 1 = 1
1
+ 1
2
+...+ 1
n
.
For every representation V of A, it is easy to see that 1
i
V is a representation of A
i
for every
i {1,2, ..., n}. Conversely, if V
1
, V
2
, ..., V
n
are representations of A
1
, A
2
, ..., A
n
, respectively,
then V
1
V
2
...V
n
canonically becomes a representation of A (with (a
1
, a
2
, ..., a
n
)A acting
on V
1
V
2
...V
n
as (v
1
, v
2
, ..., v
n
)(a
1
v
1
, a
2
v
2
, ..., a
n
v
n
)).
(a) Show that a representation V of A is irreducible if and only if 1
i
V is an irreducible repre-
sentation of A
i
forexactly one i {1,2, ..., n}, while1
i
V =0 forall the other i. Thus,classify the
irreducible representations of A in terms of those of A
1
, A
2
, ..., A
n
.
(b)LetdN. Showthat theonlyirreduciblerepresentation ofMat
d
(k) isk
d
, andevery nite
dimensional representation of Mat
d
(k) is a direct sum of copies of k
d
.
Hint: Forevery(i, j) {1,2, ..., d}
2
,letE
ij
Mat
d
(k)bethematrixwith1intheithrowofthe
jthcolumnand0severywhereelse. LetV beanitedimensionalrepresentationofMat
d
(k). Show
that V =E
11
V E
22
V ...E
dd
V, andthat
i
:E
11
V E
ii
V, vE
i1
v is an isomorphismfor
every i {1,2, ..., d}. For every v E
11
V, denote S(v) =E
11
v, E
21
v, ..., E
d1
v. Prove that S(v)
is a subrepresentation of V isomorphic to k
d
(as a representation of Mat
d
(k)), and that vS(v).
Conclude that V =S(v
1
)S(v
2
)...S(v
k
), where {v
1
, v
2
, ..., v
k
} is a basis of E
11
V.
(c) Conclude Theorem 2.6.
2.4 Filtrations
Let A be an algebra. Let V be a representation of A. A (nite) ltration of V is a sequence of
subrepresentations 0=V
0
V
1
... V
n
=V.
Lemma 2.8. Any nite dimensional representation V of an algebra A admits a nite ltration
0 =V
0
V
1
... V
n
=V such that the successive quotients V
i
/V
i1
are irreducible.
Proof. The proof is by induction in dim(V). The base is clear, and only the induction step needs
to be justied. Pick an irreducible subrepresentation V
1
V, and consider the representation
U = V/V
1
. Then by the induction assumption U has a ltration 0 = U
0
U
1
... U
n1
= U
such that U
i
/U
i1
are irreducible. Dene V
i
for i 2 to be the preimages of U
i1
under the
tautological projection V V/V
1
=U. Then0=V
0
V
1
V
2
... V
n
=V is a ltration of V
with the desired property.
25
2.5 Finite dimensional algebras
Denition 2.9. TheradicalofanitedimensionalalgebraAisthesetofallelementsofAwhich
act by 0 in all irreducible representations of A. It is denoted Rad(A).
Proposition 2.10. Rad(A) is a two-sided ideal.
Proof. Easy.
Proposition 2.11. Let A be a nite dimensional algebra.
(i) Let I be a nilpotent two-sided ideal in A, i.e., I
n
= 0 for some n. Then I Rad(A).
(ii) Rad(A) is a nilpotent ideal. Thus, Rad(A) is the largest nilpotent two-sided ideal in A.
Proof. (i) Let V be an irreducible representation of A. Let vV. Then Iv V is a subrepresen-
n
EndV
i
. =
i
Proof. First, for any irreducible representation V of A, and for any nonzero v V, Av V is a
nite dimensional subrepresentation of V. (It is nite dimensional as A is nite dimensional.) As
V is irreducible and Av=0, V =Av and V is nite dimensional.
Next, suppose we have non-isomorphic irreducible representations V
1
, V
2
, . . . , V
r
. By Theorem
2.5, the homomorphism
i
:A EndV
i
i i
is surjective. So r
i
dimEndV
i
dimA. Thus, A has only nitely many non-isomorphic
irreducible representations (at most dimA).
Now, let V
1
, V
2
, . . . , V
r
be all non-isomorphic irreducible nite dimensional representations of
A. By Theorem 2.5, the homomorphism
i
:A EndV
i
i i
is surjective. The kernel of this map, by denition, is exactly Rad(A).
Corollary 2.13.
i
(dimV
i
)
2
dimA, where the V
i
s are the irreducible representations of A.
Proof. AsdimEndV
i
=(dimV
i
)
2
,Theorem2.12impliesthatdimAdimRad(A) =
dimEndV
i
=
i
(dimV
i
)
2
. As dimRad(A)0,
(dimV
i
)
2
dimA.
i i
26
Example 2.14. 1. LetA=k[x]/(x
n
). Thisalgebrahasauniqueirreduciblerepresentation,which
is a 1-dimensional space k, in which x acts by zero. So the radical Rad(A) is the ideal (x).
2. Let A be the algebra of upper triangular n by n matrices. It is easy to check that the
irreduciblerepresentationsofAareV
i
,i= 1, ..., n, whichare1-dimensional, andanymatrixxacts
by x
ii
. So the radical Rad(A) is the ideal of strictly uppertriangular matrices (as it is a nilpotent
ideal and contains the radical). A similar result holds for block-triangular matrices.
Denition 2.15. A nite dimensional algebra A is said to be semisimple if Rad(A)=0.
Proposition 2.16. For a nite dimensional algebra A, the following are equivalent:
1. A is semisimple.
2.
i
(dimV
i
)
2
=dimA, where the V
i
s are the irreducible representations of A.
3. A
Mat
d
i
(k) for some d
i
.
i
4. Any nite dimensional representation of A is completely reducible (that is, isomorphic to a
direct sum of irreducible representations).
5. A is a completely reducible representation of A.
Proof. AsdimAdimRad(A) =
(dimV
i
)
2
,clearlydimA=
(dimV
i
)
2
ifandonlyifRad(A) =
i i
0. Thus, (1)(2).
Next, by Theorem 2.12, if Rad(A) = =
i
Mat
d
i
(k) for d
i
= dimV
i
. 0, then clearly A
Thus,
(1)(3). Conversely, ifA
Mat
d
i
(k), thenbyTheorem2.6, Rad(A) = 0, soAissemisimple.
i
Thus(3)(1).
Next, (3) (4) by Theorem 2.6. Clearly (4) (5). To see that (5) (3), let A =
i
n
i
V
i
.
Consider End
A
(A) (endomorphisms of A as a representation of A). As the V
i
s are pairwise non-
isomorphic, by Schurs lemma, no copy of V
i
in A can be mapped to a distinct V
j
. Also, again by
Schurs lemma, End
A
(V
i
) =k. Thus, End
A
(A)
Mat
n
i
=A
op
by Problem = (k). But End
A
(A)
= (
(k))
op
1.22, so A
op
= Mat
n
i
(k). Thus, A
Mat
n
i
= Mat
n
i
(k), as desired.
i i i
2.6 Characters of representations
Let A be an algebra and V a nite-dimensional representation of A with action . Then the
character of V is the linear function
V
:Ak given by
V
(a)=tr
V
((a)). |
If[A, A]isthespanofcommutators [x, y]:=xyyxoverallx, yA,then[A, A]ker
V
. Thus,
we may view the character as a mapping
V
:A/[A, A]k.
Exercise. Show that if W V are nite dimensional representations of A, then
V
=
W
+
V/W
.
Theorem 2.17. (i)Characters of(distinct) irreduciblenite-dimensional representations ofAare
linearly independent.
(ii) If A is a nite-dimensional semisimple algebra, then these characters form a basis of
(A/[A, A])
.
27
Proof. (i) If V
1
, . . . , V
r
are nonisomorphic irreducible nite-dimensional representations of A, then
V
1
Vr
:AEndV
1
EndV
r
issurjectivebythedensitytheorem,so
V
1
, . . . ,
Vr
are
linearly independent. (Indeed, if
V
i
(a) = 0 for all aA, then
i
Tr(M
i
)= 0 for all M
i
End
k
V
i
. But each tr(M
i
) can range independently over k, so it must be that
1
= =
r
=0.)
(ii) First we prove that [Mat
d
(k),Mat
d
(k)] = sl
d
(k), the set of all matrices with trace 0. It is
clear that[Mat
d
(k),Mat
d
(k)]sl
d
(k). IfwedenotebyE
ij
thematrixwith1intheithrowofthe
jthcolumnand0severywhereelse,wehave[E
ij
, E
jm
] =E
im
fori= m,and[E
i,i+1
, E
i+1,i
] =E
ii
E
i+1,i+1
.Now{E
im
}{E
ii
E
i+1,i+1
}formsabasisinsl
d
(k),soindeed[Mat
d
(k),Mat
d
(k)]=sl
d
(k),
as claimed.
By semisimplicity, we can write A=Mat
d
1
(k). Then [A, A] =sl
d
1
(k) Mat
dr
(k)
sl
dr
(k), and A/[A, A]
k
r
. = By Theorem 2.6, there are exactly r irreducible representations of A
(isomorphic to k
d
1
, . . . , k
dr
, respectively), and therefore r linearly independent characters on the
r-dimensional vector space A/[A, A]. Thus,the characters form a basis.
2.7 The Jordan-Holder theorem
We will now state and prove two important theorems about representations of nite dimensional
algebras - the Jordan-H older theorem andthe Krull-Schmidt theorem.
Theorem 2.18. (Jordan-H older theorem). Let V be a nite dimensional representation of A,
and 0 = V
0
V
1
... V
n
= V, 0 = V
0
... V
m
:=V
i
/V
i
1
are irreducible for all i. Then n=m, and there
i
exists a permutation of 1, ..., n such that W
(i)
is isomorphic to W
i
.
Proof. First proof (for k of characteristic zero). The character of V obviously equals the sum
of characters of W
i
, and also the sum of characters of W
i
. In this case W
1
W
1
= 0 (as
W
1
, W
1
V. Let U =V/(W
1
W
1
), and
0 =U
0
U
1
... U
p
=U be a ltration of U with simple quotients Z
i
=U
i
/U
i1
(it exists by
Lemma 2.8). Thenwe see that:
1) V/W
1
has a ltration with successive quotients W
1
, Z
1
, ..., Z
p
, and another ltration with
successive quotients W
2
, ...., W
n
.
2) V/W
1
, ...., W
n
.
Bytheinductionassumption,thismeansthatthecollection ofirreduciblerepresentationswith
multiplicitiesW
1
, W
1
, Z
1
, ..., Z
p
coincidesononehandwithW
1
, ..., W
n
,andontheotherhand,with
W
1
, ..., W
. We are done.
m
The Jordan-H older theorem shows that the number n of terms in a ltration of V with irre-
ducible successive quotients does not depend on the choice of a ltration, and depends only on
3
This proof does not work in characteristic p because it only implies that themultiplicities of Wi andW
i
are the
samemodulop,whichisnotsucient. Infact,thecharacteroftherepresentationpV,whereV isanyrepresentation,
is zero.
28
V. This number is called the length of V. It is easy to see that n is also the maximal length of a
ltration of V in which all the inclusions are strict.
2.8 The Krull-Schmidt theorem
Theorem2.19. (Krull-Schmidttheorem)AnynitedimensionalrepresentationofAcanbeuniquely
(up to order of summands) decomposed into a direct sum of indecomposable representations.
Proof. It is clear that a decomposition of V into a direct sum of indecomposable representations
exists, so we just need to prove uniqueness. We will prove it by induction on dimV. Let V =
V
1
...V
m
=V
1
...V
n
. Leti
s
:V
s
V,i
s
:V
s
V,p
s
:V V
s
,p
s
:V V
s
bethenatural
mapsassociatedtothesedecompositions. Let
s
=p
1
i
s
p
s
i
1
:V
1
V
1
. We have
n
s=1
s
=1. Now
we need the following lemma.
Lemma 2.20. Let W be a nite dimensional indecomposable representation of A. Then
(i) Any homomorphism :W W is either an isomorphism or nilpotent;
(ii) If
s
:W W, s= 1, ..., n are nilpotent homomorphisms, then so is :=
1
+...+
n
.
Proof. (i) Generalized eigenspaces of are subrepresentations of W, and W is their direct sum.
Thus, canhaveonlyoneeigenvalue . Ifiszero, isnilpotent,otherwiseitisanisomorphism.
(ii) Theproofisbyinductioninn. Thebaseis clear. Tomake the inductionstep (n1 to n),
assumethatisnotnilpotent. Thenby(i)isanisomorphism,so
n
i=1
i
=1. Themorphisms
i
are not isomorphisms, so they are nilpotent. Thus 1
1
n
=
1
1
+...+
1
n1
is an
isomorphism, which is a contradiction with the induction assumption.
By the lemma, we nd that for some s,
s
must be an isomorphism; we may assume that
s = 1. In this case, V
1
= Im(p
1
i
1
) Ker(p
1
i
1
), so since V
1
i
1
:V
1
V
1
and g:=p
1
i
1
:V
1
V
1
are isomorphisms.
Let B =
j>1
V
j
, B
=
j>1
V
j
; then we have V = V
1
B = V
1
attached to these
decompositions. Weclaimthathisanisomorphism. Toshowthis,itsucestoshowthatKerh= 0
(as h is a map between spaces of the same dimension). Assume that vKerh B. Then vV
1
.
On the other hand, the projection of v to V
1
is zero, so gv =0. Since g is an isomorphism, we get
v=0, as desired.
Now by the induction assumption, m = n, and V
j
(j)
The theorem is proved.
Exercise. Let A be the algebra of real-valued continuous functions on R which are periodic
withperiod1. LetM betheA-moduleofcontinuousfunctionsf onRwhichareantiperiodicwith
period1, i.e., f(x+ 1)=f(x).
(i) Showthat A and M are indecomposable A-modules.
(ii) Show that A is not isomorphic to M but AA is isomorphic to M M.
Remark. Thus,weseethatingeneral,theKrull-Schmidttheoremfailsforinnitedimensional
modules. However,itstillholdsformodulesofnitelength,i.e.,modulesM suchthatanyltration
of M has length boundedabove by a certain constant l=l(M).
29
2.9 Problems
Problem 2.21. Extensions of representations. Let A be an algebra, and V, W be a pair of
representations of A. We would like to classify representations U of A such that V is a subrepre-
sentation of U, and U/V =W. Of course, there is an obvious example U =V W, but are there
any others?
Suppose we have a representation U as above. As a vector space, it can be (non-uniquely)
identied with V W, so that for any aA the corresponding operator
U
(a) has block triangular
form
V
(a) f(a)
U
(a) = ,
0
W
(a)
where f :AHom
k
(W, V) is a linear map.
(a) What is the necessary and sucient condition on f(a) under which
U
(a) is a repre-
sentation? Maps f satisfying this condition are called (1-)cocycles (of A with coecients in
Hom
k
(W, V)). They form a vector space denoted Z
1
(W, V).
(b)LetX :W V bealinearmap. ThecoboundaryofX,dX,isdenedtobethefunctionA
Hom
k
(W, V)givenbydX(a) =
V
(a)XX
W
(a). ShowthatdX isacocycle,whichvanishesifand
only if X is a homomorphism of representations. Thus coboundaries form a subspace B
1
(W, V)
Z
1
(W, V), which isisomorphic to Hom
k
(W, V)/Hom
A
(W, V). ThequotientZ
1
(W, V)/B
1
(W, V)
is
denoted Ext
1
(W, V).
(c) Show that if f, f
Z
1
(W, V) and f f
B
1
(W, V) then the corresponding extensions
U, U
W
0
then f f
B
1
(V, W). Thus, the space Ext
1
(W, V) classies extensions of W by V.
(d) Assume that W, V are nite dimensional irreducible representations of A. For any f
Ext
1
(W, V), let U
f
be the corresponding extension. Show that U
f
is isomorphic to U
f
as repre-
sentations if and only if f and f
n
+...,
30
where
i
:AEnd(V) are linear maps,
0
=, and (ab) =(a) (b).
If b(t)=1 +b
1
t+b
2
t
2
+..., where b
i
End(V), and b
1
is a formal deformation of , then b
is also a deformation of , which is said to be isomorphic to .
(a) Show that if Ext
1
(V, V) = 0, then any deformation of is trivial, i.e., isomorphic to .
(b) Is the converse to (a) true? (consider the algebra of dual numbers A=k[x]/x
2
).
Problem 2.25. The Cliord algebra. Let V be a nite dimensional complex vector space
equipped with a symmetric bilinear form (,). The Cliord algebra Cl(V) is the quotient of the
tensor algebra TV by the ideal generated by the elements vv(v, v)1, vV. More explicitly, if
x
i
,1iN isabasisofV and(x
i
, x
j
) =a
ij
thenCl(V)isgenerated byx
i
with deningrelations
x
i
x
j
+x
j
x
i
= 2a
ij
, x
i
2
=a
ii
.
Thus, if (,) = 0, Cl(V) =V.
(i) Show that if (,) is nondegenerate then Cl(V) is semisimple, and has one irreducible repre-
sentation of dimension 2
n
if dimV = 2n (so in this case Cl(V) is a matrix algebra), and two such
representations ifdim(V) = 2n+1 (i.e.,in thiscase Cl(V) isa direct sum oftwo matrix algebras).
Hint. In the even case, pick a basis a
1
, ..., a
n
, b
1
, ..., b
n
of V in which (a
i
, a
j
) = (b
i
, b
j
) = 0,
(a
i
, b
j
) =
ij
/2, and construct a representation of Cl(V) on S := (a
1
, ..., a
n
) in which b
i
acts as
dierentiation with respect to a
i
. Show that S is irreducible. In the odd case the situation is
similar, except there should be an additional basis vector c such that (c, a
i
) = (c, b
i
) = 0, (c, c) =
1, and the action of c on S may be dened either by (1)
degree
or by (1)
degree+1
, giving two
representations S
+
, S (why are they non-isomorphic?). Show that there is no other irreducible
, B
, B
, so we may assume
without loss of generality that A andB are nite dimensional.
31
Inthiscase,weclaimthatRad(AB)=Rad(A)B+ARad(B). Indeed,denotethelatter
by J. Then J is a nilpotent ideal in AB, as Rad(A) and Rad(B) are nilpotent. On the other
hand, (AB)/J = (A/Rad(A))(B/Rad(B)), which is a product of two semisimple algebras,
hence semisimple. This implies J Rad(AB). Altogether, by Proposition 2.11, we see that
J =Rad(AB), proving the claim.
Thus, we see that
(AB)/Rad(AB) =A/Rad(A)B/Rad(B).
Now, M is an irreducible representation of (A B)/Rad(A B), so it is clearly of the form
M = V W, where V is an irreducible representation of A/Rad(A) and W is an irreducible
representation of B/Rad(B), and V, W are uniquely determined by M (as all of the algebras
involved are direct sumsof matrix algebras).
32
3 Representations of nite groups: basic results
Recall that a representation of a group G over a eld k is a k-vector space V together with a
grouphomomorphism:GGL(V). Aswehave explainedabove, arepresentationofagroupG
over k is the same thing as a representation of its groupalgebra k[G].
In this section, we begin a systematic development of representation theory of nite groups.
3.1 Maschkes Theorem
Theorem3.1. (Maschke)LetGbeanitegroupandk aeldwhosecharacteristicdoesnotdivide
|G|. Then:
(i) The algebra k[G] is semisimple.
(ii) There is an isomorphism of algebras :k[G]
i
EndV
i
dened by g
i
g| , where V
i V
i
are the irreducible representations of G. In particular, this is an isomorphism of representations
of G (where G acts on both sides by left multiplication). Hence, the regular representation k[G]
decomposes into irreducibles as
i
dim(V
i
)V
i
, and one has
G =
dim(V
i
)
2
. | |
i
(the sum of squares formula).
Proof. By Proposition 2.16, (i) implies (ii), and to prove (i), it is sucient to show that if V is
a nite-dimensional representation of G and W V is any subrepresentation, then there exists a
subrepresentation W
V such that V =W W
as representations.
Chooseanycomplement W
asvectorspaces,butnotnecessarily
as representations.) Let P be the projection along W
=0. Let | |
P :=
1
(g)P(g
1
),
|G|
gG
where (g) is the action of g on V, and let
W
=kerP.
2
Now P|
W
=Id and P(V)W, so P =P, so P is a projection along W
. Thus, V =W W
as
vector spaces.
Moreover, for any hG and any yW
,
P(h)y=
1
(g)P(g
1
h)y=
1
(h)P(
1
)y=(h)Py= 0,
|G|
gG
|G|
G
so (h)y kerP = W
. Thus, W
i
r
=1
EndV
i
, where the V
i
are irreducible representations and V
1
= k is the
trivial one-dimensional representation. Then
r r
k[G] =k EndV
i
=k d
i
V
i
,
i=2 i=2
where d
i
=dimV
i
. By Schurs Lemma,
Hom
k[G]
(k, k[G])=k
Hom
k[G]
(k[G], k) =k,
fornonzerohomomorphismsofrepresentations:k[G]k and:kk[G]uniqueuptoscaling.
We can take such that (g) = 1 for all gG, and such that (1)=
gG
g. Then
(1)= g = 1 =|G|.
gG gG
If|G|=0, then has noleft inverse, as (a)(1)=0 forany ak. Thisis a contradiction.
Example 3.3. IfG=Z/pZandk hascharacteristic p, thenevery irreduciblerepresentation ofG
overk istrivial(sok[Z/pZ]indeedisnotsemisimple). Indeed,anirreduciblerepresentationofthis
groupisa1-dimensional space, onwhichthegenerator actsbyap-throotofunity, andeveryp-th
root of unity in k equals 1, as x
p
1 = (x1)
p
over k.
Problem 3.4. Let G be a group of order p
n
. Show that every irreducible representation of G over
a eld k of characteristic p is trivial.
3.2 Characters
If V is a nite-dimensional representation of a nite group G, then its character
V
: G k
is dened by the formula
V
(g) = tr
V
((g)). Obviously,
V
(g) is simply the restriction of the |
character
V
(a) of V as a representation of the algebra A=k[G] to the basis G A, so it carries
exactly thesameinformation. Thecharacter isacentral orclass function:
V
(g) dependsonlyon
the conjugacy class of g; i.e.,
V
(hgh
1
) =
V
(g).
Theorem 3.5. If the characteristic of k does not divide G , characters of irreducible representa- | |
tions of G form a basis in the space F
c
(G, k) of class functions on G.
Proof. BytheMaschketheorem,k[G]issemisimple,sobyTheorem2.17,thecharactersarelinearly
independent and are a basis of (A/[A, A])
|
which is precisely F
c
(G, k).
Corollary 3.6. The number of isomorphism classes of irreducible representations of G equals the
number of conjugacy classes of G (if G = 0 in k).
34
Exercise. Show that if G = 0 in k then the number of isomorphism classes of irreducible | |
representations of G over k is strictly less than the numberof conjugacy classes in G.
Hint. Let P =
gG
g k[G]. Then P
2
=0. So P has zero trace in every nite dimensional
representation of G over k.
Corollary 3.7. Any representation of G is determined by its character if k has characteristic 0;
namely,
V
=
W
implies V
=W.
3.3 Examples
The following are examples of representations of nite groups over C.
1. Finite abelian groups G=Z
n
1
Z
n
k
. Let G
isanabeliangroup: if
1
,
2
:GC
areirreducible
representations then so are
1
(g)
2
(g) and
1
(g)
1
. G
by (m) =e
2im/n
. Then Z
n
={
k
:k = 0, . . . , n1},
so Z
=Z
n
. In general,
n
(G
1
G
2
G
n
)
=G
1
2
G
n
,
so G
= G for any nite abelian group G. This isomorphism is, however, noncanonical:
(G
is a canonical =
isomorphism, given by :G(G
,sin 120
), (cos240
,sin 240
),
(1,0) of the coordinate plane, respectively. Thus, for example,
((12)) =
1 0
, ((123))=
cos120
sin120
.
0 1 sin120
cos 120
Toshowthatthisrepresentationisirreducible,consideranysubrepresentationV. V mustbe
the span of a subset of the eigenvectors of ((12)), which are the nonzero multiples of (1,0)
and (0,1). V must also be the span of a subset of the eigenvectors of ((123)), which are
dierent vectors. Thus, V must beeither C
2
or 0.
3. The quaternion groupQ
8
={1,i,j,k}, with deningrelations
i=jk=kj, j =ki=ik, k=ij =ji, 1 =i
2
=j
2
=k
2
.
35
The 5 conjugacy classes are {1},{1},{i},{j},{k}, so there are 5 dierent irreducible
representations, the sum of the squares of whose dimensions is 8, so their dimensions must
be 1, 1, 1, 1, and 2.
The center Z(Q
8
) is {1}, and Q
8
/Z(Q
8
)
Z
2
Z
2
. The four 1-dimensional irreducible =
representations of Z
2
Z
2
can bepulled back to Q
8
. That is, if q:Q
8
Q
8
/Z(Q
8
) isthe
quotientmap,andanyrepresentationofQ
8
/Z(Q
8
),thenq givesarepresentationofQ
8
.
The 2-dimensional representation is V =C
2
, given by (1)=Id and
0 1
1 0
0
1
These are the Pauli matrices, which arise in quantummechanics.
Exercise. Show that the 2-dimensional irreducible representation of Q
8
can be realized in
the space of functions f : Q
8
C such that f(gi) =
sentation of S
3
can be pulled back to the 2-dimensional representation of S
4
, which we will
call C
2
.
We can consider S
4
as the group of rotations of a cube acting by permuting the interior
diagonals (or, equivalently, on a regular octahedron permuting pairs of opposite faces); this
gives the 3-dimensional representation C
3
+
.
Thelast3-dimensionalrepresentation isC
3
,theproductofC
3
withthesignrepresentation.
+
C
3
andC
3
aredierent,forifgisatransposition,detg =1whiledetg = (1)
3
+
=1. | | 3
+
3
C C
NotethatanotherrealizationofC
3
isbyactionofS
4
bysymmetries(notnecessarilyrotations)
of the regular tetrahedron.
3.4 Duals and tensor products of representations
If V is a representation of a group G, then V
V
(g) = (
V
(g)
)
1
= (
V
(g)
1
)
=
V
(g
1
)
.
The character is
V
(g) =
V
(g
1
).
We have
V
(g) =
i
, where the
i
are the eigenvalues of g in V. These eigenvalues must be
roots of unity because (g)
|G|
=(g
|G|
) =(e)=Id. Thusfor complex representations
V
(g) =
V
(g
1
) =
i
=
i
=
i
=
V
(g).
Inparticular,V
=V
asrepresentations (notjustasvectorspaces)ifandonlyif
V
(g)Rforall
gG.
If V, W are representations of G, then V W is also a representation, via
VW
(g) =
V
(g)
W
(g).
36
Therefore,
VW
(g) =
V
(g)
W
(g).
Aninteresting problemdiscussedbelowis to decomposeV W (for irreducibleV, W)into the
direct sumof irreducible representations.
3.5 Orthogonality of characters
We dene a positive denite Hermitian innerproduct on F
c
(G,C) (the space of central functions)
by
1
(f
1
, f
2
) = f
1
(g)f
2
(g).
|G|
gG
ThefollowingtheoremsaysthatcharactersofirreduciblerepresentationsofGformanorthonormal
basis of F
c
(G,C) underthis innerproduct.
Theorem 3.8. For any representations V, W
(
V
,
W
)=dimHom
G
(W, V),
and
1, if V
=W,
(
V
,
W
) =
0, if V W
if V, W are irreducible.
Proof. By the denition
1
(
V
,
W
) =
V
(g)
W
(g) =
V
(g)
W
(g)
|G|
gG
|G|
gG
1
=
G
VW
(g)=Tr|
VW
(P),
| |
gG
whereP =
|G
1
|
gG
gZ(C[G]).(HereZ(C[G])denotesthecenterofC[G]). IfX isanirreducible
representation of G then
Id,if X =C,
P
X
= |
0, X = C.
Therefore,foranyrepresentationX theoperatorP
X
istheG-invariantprojectorontothesubspace |
X
G
of G-invariants in X. Thus,
Tr|
VW
(P) = dimHom
G
(C, V W
)
= dim(V W
)
G
=dimHom
G
(W, V).
Exercise. Let G be a nite group. Let V
i
be the irreducible complex representations of G.
For every i, let
i
=
dimV
i
V
i
(g) g
1
C[G].
|G|
gG
37
j
=0 for any i= j.
Hint: In (i), notice that
i
commutes with any element of k[G], and thus acts on V
j
as an
intertwining operator. Corollary 1.17 thus yields that
i
acts on V
j
as a scalar. Compute this
scalar by taking its trace in V
j
.
Here is another orthogonality formula for characters, in which summation is taken over irre-
ducible representations rather than group elements.
Theorem 3.9. Let g, hG, and let Z
g
denote the centralizer of g in G. Then
V
(g)
V
(h) =
|Z
g
| if g is conjugate to h
0, otherwise
V
where the summation is taken over all irreducible representations of G.
Proof. As noted above,
V
(h) =
V
(h), so the left hand side equals (using Maschkes theorem):
V
(g)
V
(h)=tr|
C[G]
(xgxh
1
).
V
Ifg andharenotconjugate, thistrace isclearly zero, sincethematrix oftheoperatorxgxh
1
in the basis of group elements has zero diagonal entries. On the other hand, if g and h are in the
sameconjugacyclass, thetraceisequaltothenumberofelementsxsuchthatx=gxh
1
,i.e.,the
order of the centralizer Z
g
of g. We are done.
Remark. Another proof of this result is as follows. Consider the matrix U whose rows are
labeled by irreducible representations of G and columns by conjugacy classes, with entries U
V,g
=
V
(g)/ |Z
g
|. Note that theconjugacy class ofg isG/Z
g
, thus|G|/|Z
g
| isthenumberofelements
conjugate to G. Thus, by Theorem 3.8, the rows of the matrix U are orthonormal. This means
that U is unitary and hence its columns are also orthonormal, which implies the statement.
3.6 Unitary representations. Another proof of Maschkes theorem for complex
representations
Denition 3.10. AunitarynitedimensionalrepresentationofagroupGisarepresentationofG
onacomplexnitedimensionalvectorspaceV overCequippedwithaG-invariantpositivedenite
Hermitian form
4
(,), i.e., such that
V
(g) are unitary operators: (
V
(g)v,
V
(g)w) = (v, w).
Theorem 3.11. If G is nite, then any nite dimensional representation of G has a unitary
structure. If the representation is irreducible, this structure is unique up to scaling by a positive
real number.
Proof. Take any positive denite form B on V and dene another form B as follows:
B(v, w) = B(
V
(g)v,
V
(g)w)
gG
4
We agree that Hermitian forms are linear in the rst argument and antilinear in the second one.
38
Then B is a positive denite Hermitian form on V, and
V
(g) are unitary operators. If V is
an irreducible representation and B
1
, B
2
are two positive denite Hermitian forms on V, then
B
1
(v, w) =B
2
(Av, w) for some homomorphism A: V V (since any positive denite Hermitian
form is nondegenerate). By Schurs lemma, A=Id, and clearly >0.
Theorem 3.11 implies that if V is a nite dimensional representation of a nite group G, then
thecomplexconjugaterepresentationV (i.e.,thesamespaceV withthesameadditionandthesame
action of G, but complex conjugate action of scalars) is isomorphic to the dual representation V
.
Indeed, a homomorphism of representations V V
.
Theorem 3.12. A nite dimensional unitary representation V of any group G is completely re-
ducible.
Proof. Let W be a subrepresentation of V. Let W
is a subrepresentation of W, and V = W W
.
This implies that V is completely reducible.
Theorems 3.11 and 3.12 imply Maschkes theorem for complex representations (Theorem 3.1).
Thus, we have obtained a newproof of this theorem over the eld of complex numbers.
Remark 3.13. Theorem3.12showsthatforinnitegroupsG,anitedimensionalrepresentation
may fail to admit a unitary structure (as there exist nite dimensional representations, e.g. for
G=Z, which are indecomposable but not irreducible).
3.7 Orthogonality of matrix elements
Let V be an irreducible representation of a nite group G, and v
1
, v
2
, . . . , v
n
be an orthonormal
basisofV undertheinvariantHermitianform. ThematrixelementsofV aret
V
(x) = (
V
(x)v
i
, v
j
).
ij
Proposition 3.14. (i) Matrixelementsofnonisomorphicirreduciblerepresentationsareorthog-
onal in Fun(G,C) under the form (f, g) =
|G
1
|
xG
f(x)g(x).
(ii) (t
V
i
) =
ii
jj
1
ij
, t
V
dimV
Thus,matrixelementsofirreduciblerepresentationsofGformanorthogonalbasisofFun(G,C).
Proof. LetV andW betwoirreduciblerepresentationsofG.Take{v
i
}tobeanorthonormalbasis
of V and {w
i
} to be an orthonormal basis of W under their positive denite invariant Hermitian
forms. Let w
i
(u) = (u, w
i
). Then for x G we have (xw
i
, w
j
) = (xw
i
, w
j
). Therefore, putting P =
1
|G| xG
x, we have
V W 1
(t
ij
, t
i
j
) =|G| (xv
i
, v
j
)(xw
i
, w
j
) =|G| (xv
i
, v
j
)(xw
i
, w
j
) = (P(v
i
w
i
), v
j
w
j
)
xG xG
39
If V = W, this is zero, since P projects to the trivial representation, which does not occur in
V W
. IfV =W,weneedtoconsider(P(v
i
v
i
), v
j
v
j
).WehaveaG-invariantdecomposition
= V V
CL
C = span( v
k
v
k
)
P
L = span
a:
k
a
kk
=0
( a
kl
v
k
v
l
),
k,l
andP projectstotherstsummandalongthesecondone. Theprojectionofv
i
v
i
toC CL
is thus
ii
v
k
v
k
dimV
This shows that
(P(v
i
v
i
), v
j
v
j
) =
ii
jj
dimV
which nishes the proof of (i) and (ii). The last statement follows immediately from the sum of
squares formula.
3.8 Character tables, examples
The characters of all the irreducible representations of a nite group can be arranged into a char-
acter table, with conjugacy classes of elements as the columns, and characters as the rows. More
specically, the rst row in a character table lists representatives of conjugacy classes, the second
one the numbers of elements in the conjugacy classes, and the other rows list the values of the
characters on the conjugacy classes. Due to Theorems 3.8 and 3.9 the rows and columns of a
character table are orthonormal with respect to the appropriate inner products.
Note that in any character table, the row corresponding to the trivial representation consists
of ones, and the column corresponding to the neutral element consists of the dimensions of the
representations.
S
3
#
Here is, for example, the character table of S
3
: C
+
C
C
2
Id
1
1
1
2
(12)
3
1
-1
0
(123)
2
1
1
-1
It is obtained by explicitly computing traces in the irreducible representations.
For another example consider A
4
, the group of even permutations of 4 items. There are three
one-dimensional representations (as A
4
has a normal subgroup Z
2
Z
2
, and A
4
/Z
2
Z
2
= Z
3
).
Since there are four conjugacy classes in total, there is one more irreducible representation of
dimension 3. Finally, the character table is
A
4
Id (123)
# 1 4
C 1 1
C
1
C
2 1
2
C
3
3 0
where =exp(
2
3
i
).
(132)
4
1
0
(12)(34)
3
1
1
1
1
40
The last row can be computed using the orthogonality of rows. Another way to compute the
last row is to note that C
3
is the representation of A
4
by rotations of the regular tetrahedron: in
this case (123),(132) are the rotations by 120
0
and 240
0
around a perpendicular to a face of the
tetrahedron, while (12)(34) is the rotation by 180
0
around an axis perpendicular to two opposite
edges.
Example 3.15. The following three character tables are of Q
8
, S
4
, and A
5
respectively.
1
(123) (12)(34)
+
-1 i j k Q
8
# 1 1 2 2 2
1 1 1 1 1
1
C
++
1 1 -1 -1 C
+
C 1 1 -1 1 -1
C 1 1 -1 -1 1
C
2
2 -2 0 0 0
Id (12) (1234)
#
S
4
1 6 3 8 6
1 1 1 1 1 C
+
C 1 -1 1 1 -1
C
2
2 0 2 -1 0
C
3
3 -1 -1 0 1
+
C
3
3 1 -1 0 -1
1
Id (123) (12)(34) (12345) (13245) A
5
# 1 20 15 12 12
C 1 1 1 1
C
3
3 0 -1
1+
5
+
C
3
3 0 -1
1
2
2
C
4
4 1 0 -1
C
5
5 -1 1 0 0
1
5
1+
2
5
2
-1
Indeed, the computation of the characters of the 1-dimensional representations is straightfor-
ward.
The character of the 2-dimensional representation of Q
8
is obtained from the explicit formula
(3) for this representation, or by using the orthogonality.
For S
4
, the 2-dimensional irreducible representation is obtained from the 2-dimensional irre-
ducible representation of S
3
via the surjective homomorphismS
4
S
3
, which allows to obtain its
character from the character table of S
3
.
Thecharacterofthe3-dimensionalrepresentationC
3
iscomputedfromitsgeometricrealization
+
by rotations of the cube. Namely, by rotating the cube, S
4
permutes the main diagonals. Thus
(12) is the rotation by 180
0
around an axis that is perpendicular to two opposite edges, (12)(34)
is the rotation by 180
0
around an axis that is perpendicular to two opposite faces, (123) is the
rotation arounda main diagonal by 120
0
, and(1234) is the rotation by 90
0
aroundan axis that is
perpendiculartotwooppositefaces; thisallowsustocomputethetraceseasily, usingthefactthat
the trace of a rotation by the angle in R
3
is 1 + 2 cos. Now the character of C
3
is found by
multiplying the character of C
3
by the character of the sign representation.
+
Finally,weexplainhowtoobtainthecharactertableofA
5
(evenpermutationsof5items). The
groupA
5
isthegroupofrotationsoftheregularicosahedron. Thusithasa3-dimensionalrotation
41
representation C
3
, inwhich(12)(34) istherotation by180
0
aroundanaxis perpendiculartotwo
+
opposite edges, (123) is the rotation by 120
0
around an axis perpendicular to two opposite faces,
and (12345), (13254) are the rotations by 72
0
, respectively 144
0
, around axes going through two
opposite vertices. The character of this representation is computed from this description in a
straightforward way.
Another representation ofA
5
, whichisalso 3-dimensional, isC
3
twisted bytheautomorphism
+
This representation is denoted by C
3
, except that the conjugacy classes (12345) and (13245) are interchanged.
of A
5
given by conjugation by (12) inside S
5
. . It has the
same character as C
3
+
There are two remaining irreducible representations, and by the sum of squares formula their
dimensions are 4 and 5. So we call them C
4
andC
5
.
The representation C
4
is realized on the space of functions on the set {1,2,3,4,5} with zero
sum of values, where A
5
acts by permutations (check that it is irreducible!). The character of
thisrepresentationisequaltothecharacterofthe5-dimensionalpermutationrepresentationminus
the character of the 1-dimensional trivial representation (constant functions). The former at an
element g equals to the numberof items among 1,2,3,4,5 which are xed by g.
The representation C
5
is realized on the space of functions on pairs of opposite vertices of the
icosahedron which has zero sum of values (check that it is irreducible!). The character of this
representation is computed similarly to the character of C
4
, or from the orthogonality formula.
3.9 Computing tensor product multiplicities using character tables
Character tables allow us to compute the tensor product multiplicities N
ij
k
using
V
i
V
j
=
N
ij
k
V
k
, N
ij
k
= (
i
j
,
k
)
Example 3.16. The following tables represent computed tensor product multiplicities of irre-
3
C
3
C
3
C C
+
3
C
C
2
2
C
C
3
3
C
C
3
C
3
S
3
C
+
C
C
2
C
+
C
+
C
ducible representations of S
3
, S
4
, and A
5
respectively.
C
+
C
2
C
C
+
C
C
2
C
2
C
3
+
S
4
C
+
C
C
2
C
3
+
C
+
C
+
C
C
2
C
3
+
C
C
2
3
C
+
2 3 3
C C C
+
C
3 3
C
+
3
C
+
C
3
C
3
+
C
3
C
+
C
C
3
+
C
+
C
2
C
2
C
+
C
2
C
C
3
A
5
C C
+
3
C
4
C
5
C C C
+
3
C
4
C
5
3
C
C
3
C
4
+
CC
5
C
+
3
C
4
C
5
CC
5
C
3
+
3
C
+
4 5 3 3
C C C C
+
+
C
4
+
C
3
+
C
3
C
4
C
5
C
5
CC
4
C
5
C
3
C
3
C
4
C
5
2C
5
C
4
2C
4
2C
5
3
C
+
C
3
C
5
CC
3
3.10 Problems
Problem 3.17. Let G be the group of symmetries of a regular N-gon (it has 2N elements).
42
(a) Describeall irreducible complex representations of this group (consider the cases of odd and
even N)
(b) Let V be the 2-dimensional complex representation of G obtained by complexication of the
standard representation on the real plane (the plane of the polygon). Find the decomposition of
V V in a direct sum of irreducible representations.
Problem 3.18. LetG bethe group of3by3matrices overF
p
which are upper triangular and have
ones on the diagonal, under multiplication (its order is p
3
). It is called the Heisenberg group. For
any complex number z such that z
p
= 1 we dene a representation of G on the space V of complex
functions on F
p
, by
_ _
1 1 0
(
_
0 1 0
_
f)(x) =f(x1),
0 0 1
_ _
1 0 0
(
_
0 1 1
_
f)(x) =z
x
f(x).
0 0 1
(note that z
x
makes sense since z
p
= 1).
(a) Show that such a representation exists and is unique, and compute (g) for all gG.
(b) Denote this representation by R
z
. Show that R
z
is irreducible if and only if z= 1.
(c)Classify all 1-dimensional representations of G. Show that R
1
decomposes into adirect sum
of 1-dimensional representations, where each of them occurs exactly once.
(d) Use (a)-(c) and the sum of squares formula to classify all irreducible representations of
G.
Problem 3.19. Let V be a nite dimensional complex vector space, and GL(V) be the group of
invertible linear transformations of V. Then S
n
V and
m
V (mdim(V)) are representations of
GL(V) in a natural way. Show that they are irreducible representations.
Hint: Choose a basis {e
i
} in V. Find a diagonal element H of GL(V) such that (H) has
distinct eigenvalues. (where is one of the above representations). This shows that if W is a
subrepresentation, then it is spanned by a subset S of a basis of eigenvectors of (H). Use the
invariance ofW under the operators (1+E
ij
)(where E
ij
is dened by E
ij
e
k
=
jk
e
i
)for all i=j
to show that if the subset S is nonempty, it is necessarily the entire basis.
Problem3.20. Recallthattheadjacencymatrixofagraph(withoutmultipleedges)isthematrix
in which the ij-th entry is 1 if the vertices i and j are connected with an edge, and zero otherwise.
Let be a nite graph whose automorphism group is nonabelian. Show that the adjacency matrix
of must have repeated eigenvalues.
Problem 3.21. Let I be the set of vertices of a regular icosahedron ( I =12). Let Fun(I) be the | |
space of complex functions on I. Recall that the group G = A
5
of even permutations of 5 items
acts on the icosahedron, so we have a 12-dimensional representation of G on Fun(I).
(a) Decompose this representation in a direct sum of irreducible representations (i.e., nd the
multiplicities of occurrence of all irreducible representations).
(b) Do the same for the representation of G on the space of functions on the set of faces and
the set of edges of the icosahedron.
43
Problem 3.22. Let F
q
be a nite eld with q elements, and G be the group of nonconstant inho-
mogeneous linear transformations, xax+b, over F
q
(i.e., aF
, bF
q
). Find all irreducible
q
complex representations of G, and compute their characters. Compute the tensor products of irre-
ducible representations.
Hint. Let V be the representation of G on the space of functions on F
q
with sum of all values
equal to zero. Show that V is an irreducible representation of G.
Problem 3.23. Let G = SU(2) (unitary 2 by 2 matrices with determinant 1), and V = C
2
the
standard 2-dimensional representation of SU(2). We consider V as a real representation, so it is
4-dimensional.
(a) Show that V is irreducible (as a real representation).
(b)LetHbethesubspaceofEnd
R
(V)consistingofendomorphismsofV asarealrepresentation.
Show that His4-dimensional and closed under multiplication. Show that everynonzero element in
H is invertible, i.e., H is an algebra with division.
(c) Find a basis 1, i, j, k of Hsuch that 1is the unit and i
2
=j
2
=k
2
=1, ij =ji=k, jk =
kj =i, ki=ik=j. Thus we have that Q
8
is a subgroup of the group H
of invertible elements
of H under multiplication.
The algebra H is called the quaternion algebra.
(d)Forq=a+bi+cj+dk,a, b, c, dR,letq =abicjdk,and||q||
2
=qq =a
2
+b
2
+c
2
+d
2
.
Show that q
1
q
2
=q
2
q
1
, and ||q
1
q
2
||=||q
1
||||q
2
||.
(e) Let G be the group of quaternions of norm 1. Show that this group is isomorphic to SU(2).
(Thus geometrically SU(2) is the 3-dimensional sphere).
(f ) Consider the action of G on the space V H spanned by i, j, k, by x qxq
1
, q G,
xV. Since this action preserves the norm on V, we have a homomorphism h:SU(2)SO(3),
where SO(3) is the group of rotations of the three-dimensional Euclidean space. Show that this
homomorphism is surjective and that its kernel is {1,1}.
Problem 3.24. It is known that the classication of nite subgroups of SO(3) is as follows:
1) the cyclic group Z/nZ, n1, generated by a rotation by 2/n around an axis;
2) the dihedral group D
n
of order 2n, n2 (the group of rotational symmetries in 3-space of
a plane containing a regular n-gon
5
;
3) the group of rotations of the regular tetrahedron (A
4
).
4) the group of rotations of the cube or regular octahedron (S
4
).
5) the group of rotations of a regular dodecahedron or icosahedron (A
5
).
(a) Derive this classication.
Hint. Let G be a nite subgroup of SO(3). Consider the action of G on the unit sphere. A
point of the sphere preserved by some nontrivial element of G is called a pole. Show that every
nontrivial element of G xes a unique pair of opposite poles, and that the subgroup of G xing a
particular pole P is cyclic, of some order m (called the order of P). Thus the orbit of P has n/m
elements, where n= G . Now let P
1
, ..., P
k
be the poles representing all the orbits of G on the set | |
5
A regular 2-gon is just a line segment.
44
of poles, and m
1
, ..., m
k
be their orders. By counting nontrivial elements of G, show that
1
1
2(1 ) = (1 ).
n m
i
i
Thennd allpossible m
i
and nthat can satisfy thisequation and classifythecorresponding groups.
(b)Usingthisclassication,classifynitesubgroupsofSU(2)(usethehomomorphismSU(2)
SO(3)).
Problem 3.25. Find the characters and tensor products of irreducible complex representations of
the Heisenberg group from Problem 3.18.
Problem 3.26. Let G be a nite group, and V a complex representation of G which is faithful,
i.e., the corresponding map G GL(V) is injective. Show that any irreducible representation of
G occurs inside S
n
V (and hence inside V
n
) for some n.
Hint. Show that there exists a vector u V
to the function f
u
on G given by f
u
(g) =f(gu).
Show that this map is surjective and use this to deduce the desired result.
Problem 3.27. This problem isabout anapplication ofrepresentation theory to physics (elasticity
theory). We rst describe the physical motivation and then state the mathematical problem.
Imagine a material which occupies a certain region U in the physical space V = R
3
(a space
with a positive denite inner product). Suppose the material is deformed. This means, we have
applied a dieomorphism (=change of coordinates) g : U U
,
- of real type, if V has a nondegenerate symmetric form invariant underG,
- of quaternionic type, if V has a nondegenerate skew form invariant underG.
Problem 4.1. (a) Show that End
R[G]
V is C for V of complex type, Mat
2
(R) for V of real type,
and H for V of quaternionic type, which motivates the names above.
Hint. Show that the complexication V
C
of V decomposes as V V
,C
2
areofrealtype. ForS
4
thereareveirreduciblerepresenta-
tions C
+
,C
, C
2
, C
+
3
, C
3
, whichare all of real type. Similarly, all ve irreduciblerepresentations
of A
5
C, C
+
3
, C
3
, C
4
, C
5
are of real type. As for Q
8
, its one-dimensional representations are of
V
dim(V)FS(V), i.e., the sum of dimensions of all representations of G of real type
minus the sum of dimensions of its representations of quaternionic type.
Proof. Let A:V V have eigenvalues
1
,
2
, . . . ,
n
. We have
Tr|
S
2
V
(AA) =
i
j
i j
Tr
2
V
(AA) =
i
j
|
i<j
Thus,
Tr
S
2
V
(AA)Tr
2
V
(AA) =
2
=Tr(A
2
).
i
| |
1 i n
47
V
(g
2
) =
S
2
V
(g)
2
V
(g)
Therefore,
_
1, if V isof realtype
_
|G|
1
V
( g
2
) =
S
2
V
(P)
2
V
(P)=dim(S
2
V)
G
dim(
2
V)
G
= 1, if V isof quaternionictype
gG
_
0, if V isof complextype
Finally, the numberof involutions in G equals
1
|G|
dimV
V
(
gG
g
2
)= dimV dimV.
V realV quatV
Corollary 4.5. Assume that all representations of a nite group G are dened over real numbers
(i.e., all complex representations of G are obtained by complexifying real representations). Then
the sum of dimensions of irreducible representations of G equals the number of involutions in G.
Exercise. Showthatanynontrivialnitegroupofoddorderhasanirreduciblerepresentation
which is not denedover R (i.e., not realizable by real matrices).
4.2 Frobenius determinant
EnumeratetheelementsofanitegroupGasfollows: g
1
, g
2
, . . . , g
n
.Introducenvariablesindexed
with the elements of G:
x
g
1
, x
g
2
, . . . , x
gn
.
Denition 4.6. Considerthe matrixX
G
withentries a
ij
=x
g
i
g
j
. Thedeterminant ofX
G
issome
polynomial of degree n of x
g
1
, x
g
2
, . . . , x
gn
that is called the Frobenius determinant.
The following theorem, discovered by Dedekind and proved by Frobenius, became the starting
point for creation of representation theory (see [Cu]).
Theorem 4.7.
r
detX
G
= P
j
(x)
degP
j
j=1
for some pairwise non-proportional irreducible polynomials P
j
(x), where r is the number of conju-
gacy classes of G.
We will need the following simple lemma.
Lemma 4.8. Let Y be an nn matrix with entries y
ij
. Then detY is an irreducible polynomial
of {y
ij
}.
Proof. Let detY = q
1
q
2
. . . q
k
, be the factorization of detY into irreducible polynomials (it is
dened uniquely up to scaling and permutation of factors). Since detY has degree 1 with respect
to each row and each column of Y, by uniqueness of factorization all q
i
must be homogeneous
with respect to each row and each column, of degree either 0 or 1. Now consider the factor q
1
.
It is homogeneous of degree 1 in some row. This means that it depends on all columns, so is
homogeneous of degree 1 in all columns. Thus q
1
=detY, as desired.
48
degree d, then
1
is aroot of x p(1/x). Thelast statement iseasy, since a number is algebraic
if and only if it denesa nite extension of Q.
Proposition 4.13. AQ=Z.
Proof. We will be using denition (4.9). Let z be a root of
n
p(x) =x +a
1
x
n1
+. . .+a
n1
x+a
n
,
and suppose
z =
p
q
Q, gcd(p, q) = 1.
Notice that the leading term of p(x) will have q
n
in the denominator, whereas all the other terms
will have a lower power of q there. Thus, if q = 1, then p(z) / Z, a contradiction. Thus,
z AQz Z. The reverse inclusion follows because nZis a root of xn.
Every algebraic number has a minimal polynomial p(x), which is the monic polynomial
withrationalcoecientsofthesmallestdegreesuchthatp()=0. Anyotherpolynomialq(x)with
rational coecients suchthat q()=0isdivisible byp(x). Roots ofp(x)are called thealgebraic
conjugates of; they are roots of any polynomial q withrational coecients suchthat q()=0.
50
Note that any algebraic conjugate of an algebraic integer is obviously also an algebraic inte-
ger. Therefore, by the Vieta theorem, the minimal polynomial of an algebraic integer has integer
coecients.
Below we will need the following lemma:
Lemma 4.14. If
1
, ...,
m
are algebraic numbers, then all algebraic conjugates to
1
+...+
m
are of the form
1
+...+
, where
+
2
, so we are done.
Problem 4.15. Show that if V is an irreducible complex representation of a nite group G of
dimension >1 then there exists gG such that
V
(g) = 0.
Hint. Assumethecontrary. Useorthonormality of characters to show that the arithmetic mean
of the numbers |
V
(g)|
2
for g = 1 is < 1. Deduce that their product satises 0 < < 1.
Show that all conjugates of satisfy the same inequalities (consider the Galois conjugates of the
representation V). Then derive a contradiction.
4.4 Frobenius divisibility
Theorem 4.16. Let G be a nite group, and let V be an irreducible representation of G over C.
Then
dimV divides G . | |
Proof. Let C
1
, C
2
, . . . , C
n
be the conjugacy classes of G. Set
i
=
V
(g
C
i
)
|C
i
|
,
dimV
where g
C
i
is a representative of C
i
.
Proposition 4.17. The numbers
i
are algebraic integers for all i.
Proof. LetC beaconjugacyclassinG,andP =
hC
h. ThenP isacentralelementofZ[G],soit
actsonV bysomescalar,whichisanalgebraic integer(indeed,sinceZ[G]isanitelygenerated
Z-module, any element of Z[G] is integral over Z, i.e., satises a monic polynomial equation with
integer coecients). On the other hand, taking the trace of P in V, we get C
V
(g) = dimV,
gC, so =
|C|
V
(g)
.
| |
dimV
Now, consider
V
(g
C
i
).
i
This is an algebraic integer, since:
(i)
i
are algebraic integers by Proposition 4.17,
(ii)
V
(g
C
i
) is a sumof roots ofunity (it is the sumof eigenvalues of the matrix of (g
C
i
), and
since g
C
|G
i
|
=e in G, the eigenvalues of (g
C
i
) are roots of unity), and
51
(iii) A is a ring (Proposition 4.12).
On the other hand,from the denition of
i
,
V
(g
C
i
) =
|C
i
|
V
(g
C
i
)
V
(g
C
i
)
.
dimV
C
i
i
Recalling that
V
is a class function, this is equal to
V
(g)
V
(g)
=
|G|(
V
,
V
)
.
dimV dimV
gG
Since V is an irreducible representation, (
V
,
V
) = 1, so
V
(g
C
i
) =
| |
.
dimV
C
i
Since
dim
|G|
V
Q and
C
i
V
(g
C
i
)A, by Proposition 4.13
dim
|G|
V
Z.
4.5 Burnsides Theorem
Denition 4.18. A group G is called solvable if there exists a series of nested normal subgroups
{e}=G
1
G
2
. . . G
n
=G
where G
i+1
/G
i
is abelian for all 1in1.
Remark 4.19. Such groups are called solvable because they rst arose as Galois groups of poly-
nomial equations which are solvable in radicals.
a
Theorem 4.20 (Burnside). Any group G of order p q
b
, where p and q are prime and a, b0, is
solvable.
ThisfamousresultingrouptheorywasprovedbytheBritishmathematicianWilliamBurnside
in the early 20-th century, using representation theory (see [Cu]). Here is this proof, presented in
modernlanguage.
BeforeprovingBurnsidestheoremwewillproveseveral otherresultswhichareofindependent
interest.
Theorem4.21. LetV beanirreduciblerepresentation ofanitegroupGandletC beaconjugacy
class of G with gcd(|C|,dim(V))=1. Then for any g C, either
V
(g) = 0 or g acts as a scalar
on V.
The proof will be based on the following lemma.
1
Lemma 4.22. If
1
,
2
. . .
n
are roots of unity such that (
1
+
2
+. . .+
n
) is an algebraic
n
integer, then either
1
=. . .=
n
or
1
+. . .+
n
= 0.
Proof. Let a=
1
(
1
+. . .+
n
). Ifnotall
i
areequal, then a <1. Moreover, since anyalgebraic
n
| |
conjugate of arootofunityisalso a rootofunity, |a
ofa. But
the product of all algebraic conjugates of a is an integer. Since it has absolute value <1, it must
equal zero. Therefore, a=0.
52
| |
Proof of theorem 4.21.
Let dimV = n. Let
1
,
2
, . . .
n
be the eigenvalues of
V
(g). They are roots of unity, so
V
(g) is an algebraic integer. Also, by Proposition 4.17,
n
1
|C|
V
(g) is an algebraic integer. Since
gcd(n, C )=1, this implies that | |
V
(g) 1
= (
1
+. . .+
n
).
n n
is an algebraic integer. Thus, by Lemma 4.22, we get that either
1
=. . .=
n
or
1
+. . .+
n
=
V
(g) = 0. In the rst case, since
V
(g) is diagonalizable, it must be scalar. In the second case,
V
(g)=0. Thetheorem is proved.
Theorem 4.23. Let G be a nite group, and let C be a conjugacy class in G of order p
k
where p
is prime and k >0. Then G has a proper nontrivial normal subgroup.
Proof. Choose an element gC. Since g= e, by orthogonality of columns of the character table,
dimV
V
(g) = 0. (4)
VIrrG
We can divide IrrG into three parts:
1. the trivial representation,
2. D, the set of irreducible representations whose dimension is divisible by p, and
3. N, the set of non-trivial irreducible representations whose dimensionis not divisible by p.
Lemma 4.24. There exists V N such that
V
(g) = 0 .
Proof. If V D, the number
1
dim(V)
V
(g) is an algebraic integer, so
p
1
a= dim(V)
V
(g)
p
VD
is an algebraic integer.
Now, by (4), we have
0 =
C
(g)+ dimV
V
(g)+ dimV
V
(g) = 1 +pa+ dimV
V
(g).
VD VN VN
This means that the last summandis nonzero.
Now pick V T such that
V
(g) = 0; it exists by Lemma 4.24. Theorem 4.21 implies that g
(and hence any element of C) acts by a scalar in V. Now let H be the subgroup of G generated
byelements ab
1
,a, bC. Itisnormalandacts trivially inV,soH = G,asV isnontrivial. Also
H =1, since C >1.
Proof of Burnsides theorem.
Assume Burnsides theorem is false. Then there exists a nonabelian simple group G of order
p
a
q
b
. ThenbyTheorem4.23,thisgroupcannothaveaconjugacyclassoforderp
k
orq
k
,k1. So
theorderofanyconjugacyclassinGiseither1orisdivisiblebypq. Addingtheordersofconjugacy
a
classes and equating the sum to p q
b
, we see that there has to be more than one conjugacy class
consisting just of one element. So G has a nontrivial center, which gives a contradiction.
53
n
i
V
i
, n
i
Z (i.e., n
i
are not assumed to be nonnegative).
The character of V is
V
:=
n
i
V
i
.
The following lemma is often very useful(and will be used several times below).
Lemma4.27. LetV beavirtualrepresentation withcharacter
V
. If(
V
,
V
) = 1and
V
(1)>0
then
V
is a character of an irreducible representation of G.
Proof. Let V
1
, V
2
, . . . , V
m
be the irreducible representations of G, and V =
n
i
V
i
. Then by
2 2
orthonormality of characters, (
V
,
V
) =
So
(f))(x) = g
(f)(xg) = f(xgg
) =
(gg
, xG and hH.
Remark 4.30. Notice that if we choose a representative x
)}.
H
V
Because of this,
dim(Ind
G
|G
.
H
V)=dimV
|
|H|
Problem4.31. CheckthatifK H
H
Ind
H
is isomorphic to Ind
G
GaregroupsandV arepresentationofK thenInd
G
K
V
K
V.
54
V
(xgx
1
)
|H|
xG,xgx
1
H
Proof. For a left H-coset of G, let us dene
V
={f Ind
G
|f(g) = 0 g
H
V }.
Thenone has
Ind
G
H
V =
and so
(g) =
(g),
where
.
Since g() =g is a left H-coset for any left H-coset ,
(g) = 0 if =g.
Now assume that =g. Choose x
. Then x
g =hx
where h=x
gx
1
H. Consider
the vector space homomorphism : V
). Since f V
is uniquely
determined by f(x
), is an isomorphism. We have
(gf) =g(f)(x
) =f(x
g) =f(hx
) =
V
(h)f(x
) =h(f),
and gf =
1
h(f). This means that
(g) =
V
(h). Therefore
(g) =
V
(x
gx
1
).
H\G,g=
Since it doesnot matter whichrepresentative x
G
H
V, W).
H
W) is naturally isomorphic to Hom
H
(Res
G
Proof. LetE =Hom
G
(V,Ind
G
H
W)andE
=Hom
H
(Res
G
andF
H
V, W). DeneF :E E :E
E as follows: F()v = (v)(e) for any E and (F
.
In order to check that F and F
()v Ind
G
()v)(hx) =h(F
H
W, i.e., (F ()v)(x).
Indeed, (F
()v)(x).
(c) F
() is a G-homomorphism, i.e. F
()gv =g(F
()v).
Indeed, (F
()gv)(x) =(xgv) = (F
()v)(xg) = (g(F
()v))(x).
(d) F F
=Id
E
.
This holds since F(F
())v = (F
()v)(e) =(v).
(e) F
F =Id
E
, i.e., (F
(F())v)(x) = (v)(x).
Indeed, (F
k[G]
V is a (k[H], k)-bimodule, i. e., a left k[H]-module. Prove
thatthistensorproductisisomorphictoRes
G
H
V
H
V asaleftk[H]-module. TheisomorphismRes
G
k[G]
1
k[G]
V is given by v1
k[G]
v for every vRes
G
H
V.
(b) Let W be a representation of H. Then, W is a left k[H]-module, thus a (k[H], k)-
bimodule. Then,Ind
G
=Hom
H
(k[G], W),accordingtoRemark4.29. Inotherwords,Ind
G
=
H
W
H
W
Hom
k[H]
(k[G]
1
, W). Now, use part (b) of the Exercise in Section 1.10 to conclude Theorem 4.33.
(c) Let V be a representation of G. Then, V is a left k[G]-module, thus a (k[G], k)-bimodule.
Prove that not only k[G]
1
k[G]
V, but also Hom
k[G]
(k[G]
2
, V) is isomorphic to Res
G
as a left
H
V
k[H]-module. The isomorphism Hom
k[G]
(k[G]
2
, V) Res
G
is given by f f(1) for every
H
V
f Hom
k[G]
(k[G]
2
, V).
(d) Let W be a representation of H. Then, W is a left k[H]-module, thus a (k[H], k)-
bimodule. ShowthatInd
G
, W),butalsoisomorphicto
H
W isnotonlyisomorphictoHom
k[H]
(k[G]
1
k[G]
2
k[H]
W. TheisomorphismHom
k[H]
(k[G]
1
, W)k[G]
2
k[H]
W isgivenbyf
gP
g
1
k[H]
f(g) for every f Hom
k[H]
(k[G]
1
, W), where P is a set of distinct representatives for the left
H-cosets in G. (This isomorphism is independentof the choice of representatives.)
(e) Let V be a representation of G and W a representation of H. Use (b) to prove that
Hom
G
Ind
G
W,Res
G
.
H
W, V
H
V
56
H
(V
H
V (f) Let V be a representation of H. Prove that Ind
G
= Ind
G
as representations of
G. [Hint: Write Ind
G
as k[G]
2
k[H]
V
H
(V
) as Hom
k[H]
(k[G]
1
, V
). Prove
H
V and write Ind
G
that the map Hom
k[H]
(k[G]
1
, V
Ind
G
k given by
f,
x
k[H]
v
(f(Sx)) (v) is
H
(V
)
a nondegenerate G-invariant bilinear form, where S : k[G] k[G] is the linear map dened by
Sg=g
1
for every gG.]
4.11 Examples
Here are some examples of induced representations (we use the notation for representations from
the character tables).
1. Let G = S
3
, H = Z
2
. Using the Frobenius reciprocity, we obtain:
H
C
+
= C
2
C
+
, Ind
G
Ind
G
=C
2
C
H
C
.
2. Let G=S
3
, H Thenwe obtain Ind
G
H
C
=Ind
G
. =Z
3
.
H
C
+
=C
+
C
, Ind
G
H
C
2 =C
2
3. LetG=S
4
,H =S
3
.
H
C
+
,Ind
G
=C
H
C
2
C
3
ThenInd
G
=C
+
C
3
H
C
C
3
,Ind
G
=C
2
C
+
3
.
+
Problem 4.34. Compute the decomposition into irreducibles of all the irreducible representations
of A
5
induced from
(a) Z
2
(b) Z
3
(c) Z
5
(d) A
4
(e) Z
2
Z
2
4.12 Representations of S
n
In this subsection we give a description of the representations of the symmetric group S
n
for any
n.
Denition 4.35. A partition of n is a representation of n in the form n=
1
+
2
+...+
p
,
where
i
are positive integers, and
i
i+1
.
TosuchwewillattachaYoungdiagramY
,whichistheunionofrectanglesiy i+1,
0x
i
in the coordinate plane, for i= 1, ..., p. Clearly, Y
insomeway(withoutrepetitions). Forexample,wewillconsidertheYoungtableauT
obtained
by lling in the numbersin the increasing order, left to right, top to bottom.
We can dene two subgroupsof S
n
correspondingto T
:
1. The row subgroupP
.
2. The column subgroup Q
.
Clearly, P
={1}.
57
Dene the Young projectors:
1
:= g,
|P
|
gP
:= (1)
g
g,
|Q
|
gQ
where (1)
g
denotes the sign of the permutation g. Set c
= a
. Since P
= {1}, this
element is nonzero.
The irreducible representations of S
n
are describedby the following theorem.
Theorem 4.36. The subspace V
:=C[S
n
]c
of C[S
n
] is an irreducible representation of S
n
under
left multiplication. Every irreducible representation of S
n
is isomorphic to V
for a unique .
The modules V
=S
n
,Q
={1}, soc
isthesymmetrizer,andhenceV
isthetrivial
representation.
For the partition =(1, ...,1), Q
=S
n
, P
={1}, so c
is
the sign representation.
n=3. For =(2,1), V
=C
2
.
n=4. For =(2,2), V
=C
2
; for =(3,1), V
=C
3
; for =(2,1,1), V
=C
+
3
.
Corollary4.38. AllirreduciblerepresentationsofS
n
canbegivenbymatriceswithrationalentries,
whose denominators contain no primes > n.
Remark. In fact, the irreducible representations of S
n
are realizable by matrices with integer
entries, butwe will not show this here.
Problem 4.39. Find the sum of dimensions of all irreducible representations of the symmetric
group S
n
.
Hint. Show that all irreducible representations of S
n
are real, i.e., admit a nondegenerate
invariant symmetric form. Then use the Frobenius-Schur theorem.
4.13 Proof of Theorem 4.36
Lemma 4.40. Let xC[S
n
]. Then a
xb
(x)c
, where
is a linear function.
Proof. If gP
, so a
gb
= (1)
q
c
.
Thus, to prove the required statement, we need to show that if g is a permutation which is not in
P
then a
gb
=0.
To show this, it is sucient to nda transposition t such that tP
and g
1
tgQ
; then
a
gb
=a
tgb
=a
g(g
1
tg)b
=a
gb
,
58
so a
gb
= 0. In other words, we have to nd two elements i, j standing in the same row in the
tableau T = T
,i.e., thereexistspP
,q
:=gQ
g
1
such that pT =q
gQ
).
AnytwoelementsintherstrowofT mustbeindierentcolumnsofT
,sothereexistsq
1
such that p
1
T and q
1
have
the same rst row. Now do the same procedure with the second row, nding elements p
2
, q
2
such
that p
2
p
1
T and q
2
q
1
have the same rst two rows. Continuing so, we will construct the desired
elements p, q
i
is positive.
Lemma 4.41. If > then a
C[S
n
]b
= 0.
Proof. Similarly to the previous lemma, it suces to show that for any g S
n
there exists a
transposition t P
such that g
1
tg Q
. Let T = T
and T
= gT
. Indeed,if
1
>
1
,thisis
clearbythepigeonholeprinciple(alreadyfortherstrow). Otherwise,if
1
=
1
,likeintheproof
ofthepreviouslemma,wecanndelementsp
1
P
, q
1
gQ
g
1
suchthatp
1
T andq
1
havethe
same rst row, and repeat the argument for the second row, and so on. Eventually, having done
i1suchsteps,wellhave
i
>
i
,whichmeansthatsometwoelementsofthei-throwoftherst
tableau are in the same column of the second tableau, completing the proof.
Lemma 4.42. c
=
dim
n!
V
.
Proof. Lemma 4.40 implies that c
2
is proportional to c
is 1). This
implies the statement.
Lemma 4.43. Let A be an algebra and e be an idempotent in A. Then for any left A-module M,
one has Hom
A
(Ae, M)
, V
)=Hom
Sn
(C[S
n
]c
,C[S
n
]c
) =c
C[S
n
]c
.
The latter space is zero for > by Lemma 4.41, and 1-dimensional if = by Lemmas 4.40
and4.42. Therefore,V
areirreducible,andV
isnotisomorphictoV
if=. Sincethenumber
of partitions equals the number of conjugacy classes in S
n
, the representations V
4.14 Induced representations for S
n
Denote by U
=C[S
n
]a
.
Proposition 4.44. Hom(U
, V
, V
) = 1. Thus, U
, where K
= 1.
Denition 4.45. The integers K
, V
)=Hom(C[S
n
]a
,C[S
n
]a
) =a
C[S
n
]a
,
and the result follows from Lemmas 4.40 and 4.41.
Now let us compute the character of U
. Let C
i
be the conjugacy class in S
n
having i
l
cycles
of length l for all l 1 (here i is a shorthand notation for (i
1
, ..., i
l
, ...)). Also let x
1
, ..., x
N
be
variables, and let
m
H
m
(x) = x
i
i
be the power sumpolynomials.
Theorem 4.46. Let N p(where pis the number ofparts of ). Then
U
(C
i
) is the coecient
6
of x
:=
j
j
in the polynomial
H
m
(x)
im
.
m1
Proof. The proof is obtained easily from the Mackey formula. Namely,
U
(C
i
) is the number of
elements xS
n
such that xgx
1
P
(for a representative gC
i
), divided by |P
|. Thus,
n!
(C
i
) = C
i
P
.
|C
i
|
j
j
!
| |
Now, it is easy to see that
n!
= m
im
i
m
!
m
|C
i
|
(it is the order of the centralizer Z
g
of g), so we get
im
i
m
m !
(C
i
) =
m
j
!
|C
i
P
|.
j
Now, since P
j
S
j
, we have
C
i
P
j
!
| |=
r
m1
m
r
jm
r
jm
!
,
j1
where r= (r
jm
) runsover all collections of nonnegative integers such that
mr
jm
=
j
, r
jm
=i
m
.
m j
If j > p, we dene j to be zero.
60
6
Thuswe get
i
m
!
(C
i
) =
j
r
jm
!
r m
But this is exactly the coecient of x
in
m m
(x
1
+...+x
N
)
im
m1
m
(r
jm
is the numberof times we take x
j
).
4.15 The Frobenius character formula
Let (x) =
(x
i
x
j
). Let = (N 1, N 2, ...,0) C
N
. The following theorem, due
1 i<j N
to Frobenius, gives a character formula for the Specht modules V
.
Theorem 4.47. Let N p. Then
V
(C
i
) is the coecient of x
+
:=
j
+Nj
in the polyno-
j
mial
(x)
H
m
(x)
im
.
m1
Proof. Denote
V
shortlyby
. Letusdenotetheclassfunctiondenedinthetheoremby
. It
follows from Theorem 4.46 that this function has the property
, where L
are
integers and L
)=1.
We have
(
) =
n
1
!
|C
i
|
(C
i
)
2
,
i
which is the coecient of x
+
y
+
in the series R(x, y)=(x)(y)S(x, y), where
(
m m
j,k
x
j
y
k
/m)
im
S(x, y) = .
i
m
!
i m
Summingover i and m, we get
m m
S(x, y)= exp( x
j
y
k
/m)=exp( log(1x
j
y
k
)) = (1x
j
y
k
)
1
m j,k j,k j,k
Thus,
(x
i
x
j
)(y
i
y
j
)
R(x, y) =
i<j
.
(1x
i
y
j
)
i,j
Now we need the following lemma.
Lemma 4.48.
i<j
(z
j
z
i
)(y
i
y
j
)
1
=det( ).
i,j
(z
i
y
j
) z
i
y
j
Proof. Multiplybothsidesby
(z
i
y
j
). Thentherighthandsidemustvanishonthehyperplanes
i,j
z
i
= z
j
and y
i
= y
j
(i.e., be divisible by (z)(y)), and is a homogeneous polynomial of degree
N(N 1). This implies that the right hand side and the left hand side are proportional. The
proportionality coecient (whichisequalto 1) isfoundbyinductionbymultiplying bothsidesby
z
N
y
N
and then setting z
N
=y
N
.
61
tation in S
N
, the coecient of this monomial in
Q
(1x
j
y
(j)
)
is obviously zero.
4.16 Problems
In the following problems, we do not make a distinction between Young diagrams and partitions.
Problem 4.50. For a Youngdiagram , let A() be the set ofYoung diagrams obtained by adding
a square to , and R() be the set of Young diagrams obtained by removing a square from .
(a) Show that Ind
Sn
V
=
A()
V
.
S
n1
(b) Show that Res
Sn
V
=
R()
V
.
S
n1
Problem 4.51. The content c() of a Young diagram is the sum
j
(ij). Let C =
j i=1
(ij) C[S
n
] be the sum of all transpositions. Show that C acts on the Specht module V
by
i<j
multiplication by c().
Problem 4.52. Show that the element (12)+...+(1n) acts on V
. According to the
character formula, dimV
is the coecient of x
+
in (x)(x
1
+...+x
N
)
n
. Let l
j
=
j
+N j.
Then, we get
n! n!
dimV
= (1)
s
(l
j
N +s(j))!
=
l
1
!...l
N
!
(1)
s
l
j
(l
j
1)...(l
j
N+s(j)+1)=
sS
N
:l
j
Ns(j)
j
sS
N
j
n!
l
j
!
det(l
j
(l
j
1)...(l
j
N +i+1)).
j
Using column reduction and the Vandermonde determinant formula, we see from this expression
that
dimV
=
n!
det(l
j
Ni
) =
n!
(l
i
l
j
) (5)
j
l
j
!
j
l
j
!
i<j 1 N
(where N p).
In this formula, there are many cancelations. After making some of these cancelations, we
obtainthehooklengthformula. Namely, forasquare(i, j)inaYoungdiagram(i, j 1, i
j
),
dene the hook of (i, j) to be the set of all squares (i
, j
) in with i
i, j
=j or i
=i, j
j.
Let h(i, j) bethe length of the hook of i, j, i.e., the numberof squares in it.
62
Theorem 4.53. (The hook length formula) One has
n!
dimV
= .
i
j
h(i, j)
Proof. The formula follows from formula (5). Namely, note that
l
1
!
= k.
1<jN
(l
1
l
j
)
k =l
1
l
j
1 l
1
,k
Itiseasytoseethatthefactorsinthisproductareexactlythehooklengthsh(i,1). Nowdeletethe
rst row of the diagram and proceed by induction.
4.18 Schur-Weyl duality for gl(V)
We start with a simple result which is called the Double Centralizer Theorem.
Theorem 4.54. Let A, B be two subalgebras of the algebra EndE of endomorphisms of a nite
dimensional vector space E, such that A is semisimple, and B=End
A
E. Then:
(i) A=End
B
E (i.e., the centralizer of the centralizer of A is A);
(ii) B is semisimple;
(iii) as a representation of AB, E decomposes as E =
iI
V
i
W
i
, where V
i
are all the
irreducible representations of A, and W
i
are all the irreducible representations of B. In particular,
we have a natural bijection between irreducible representations of A and B.
Proof. Since A is semisimple, we have a natural decomposition E =
iI
V
i
W
i
, where W
i
:=
Hom
A
(V
i
, E), and A=
i
EndV
i
. Therefore, by Schurs lemma, B =End
A
(E) is naturally identi-
ed with
i
End(W
i
). This implies all the statements of the theorem.
We will now apply Theorem 4.54 to the following situation: E = V
n
, where V is a nite
dimensional vector space over a eld of characteristic zero, and A is the image of C[S
n
] in EndE.
LetusnowcharacterizethealgebraB. Letgl(V)beEndV regardedasaLiealgebrawithoperation
abba.
Theorem4.55. ThealgebraB=End
A
E istheimageoftheuniversalenvelopingalgebraU(gl(V))
under its natural action on E. In other words, B is generated by elements of the form
n
(b):=b1...1 + 1b...1 +...+ 11...b,
bgl(V).
Proof. Clearly, theimage ofU(gl(V)) is contained inB,so we justneedto show that anyelement
ofB iscontainedintheimageofU(gl(V)). Bydenition,B =S
n
EndV,sotheresultfollowsfrom
part (ii) of the following lemma.
Lemma 4.56. Let k be a eld of characteristic zero.
(i) For any nite dimensional vector space U over k, the space S
n
U is spanned by elements of
the form u...u, uU.
(ii) For any algebra A over k, the algebra S
n
A is generated by elements
n
(a), aA.
63
Proof. (i)ThespaceS
n
U isanirreduciblerepresentationofGL(U)(Problem3.19). Thesubspace
spannedby u...u is a nonzero subrepresentation, so it must be everything.
(ii) By the fundamental theorem on symmetric functions, there exists a polynomial P with
rational coecients such that P(H
1
(x), ..., H
n
(x))=x
1
...x
n
(where x= (x
1
, ..., x
n
)). Then
P(
n
(a),
n
(a
2
), ...,
n
(a
n
))=a...a.
The rest follows from (i).
Now, the algebra A is semisimple by Maschkes theorem, so the double centralizer theorem
applies, and we get the following result, which goes underthe name Schur-Weyl duality.
Theorem 4.57. (i) The image A of C[S
n
] and the image B of U(gl(V)) in End(V
n
) are central-
izers of each other.
(ii) Both A and B are semisimple. In particular, V
n
is a semisimple gl(V)-module.
(iii) We have a decomposition of AB-modules V
n
=
containsb
n
. Indeed,forallvaluesoftbutnitelymany,t Id+bisinvertible,
so (t Id+b)
n
belongs to B
. This implies that this is true for all t, in particular for t=0, since
(t Id+b)
n
is a polynomial in t.
The rest follows from Lemma 4.56.
Corollary 4.59. As a representation of S
n
GL(V), V
n
decomposes as
, where
L
=Hom
Sn
(V
, V
n
) are distinct irreducible representations of GL(V) or zero.
Example 4.60. If= (n)thenL
=S
n
V,andif=(1
n
) (ncopiesof1)thenL
=
n
V. Itwas
shown in Problem 3.19 that these representations are indeed irreducible (except that
n
V is zero
if n >dimV).
4.20 Schur polynomials
Let = (
1
, ...,
p
) be a partition of n, and N p. Let
N
D
(x) =
(1)
s
j
+Nj
=det(x
j
+Nj
).
s(j) i
sS
N
j=1
64
(x)
S
(x):=
D
0
(x)
(clearly D
0
(x) is just (x)). It is easy to see that these are indeed polynomials, as D
is an-
tisymmetric and therefore must be divisible by . The polynomials S
m m
(x
1
+...+x
N
)
im
=
(C
i
)S
(x).
m :p N
Proof. The identity follows from the Frobeniuscharacter formula and the antisymmetry of
m m
(x)
(x
1
+...+x
N
)
im
.
m
Certain special values of Schur polynomials are of importance. Namely, we have
Proposition 4.62.
S
(1, z, z
2
, ..., z
N1
) =
i
i
z
j
j
z
i
z
j
1 i<j N
Therefore,
S
(1, ...,1)=
j
+ji
ji
1 i<j N
Proof. Therstidentity isobtained fromthe denitionusingthe Vandermondedeterminant. The
second identity follows from the rst one by setting z =1.
4.21 The characters of L
.
Namely, let dimV =N, g GL(V), and x
1
, ..., x
N
be the eigenvalues of g on V. To compute
thecharacter
L
(C
i
)Tr
L
(g).
ComparingthistoProposition4.61andusinglinearindependenceofcolumnsofthecharactertable
of S
n
, we obtain
Theorem 4.63. (Weyl character formula) The representation L
(x).
Therefore, the dimension of L
j
+ji
ji
1 i<j N
65
ThisshowsthatirreduciblerepresentationsofGL(V)whichoccurinV
n
forsomenarelabeled
by Young diagrams with any numberof squares but at most N =dimV rows.
Proposition 4.64. The representation L
+1
N (where 1
N
= (1,1, ...,1) Z
N
) is isomorphic to
L
N
V.
Proof. Indeed, L
N
V V
n
N
V V
n+N
, and the only component of V
n+N
that has
the same character as L
N
V is L
+1
N. Thisimplies the statement.
4.22 Polynomial representations of GL(V)
Denition 4.65. We say thata nitedimensionalrepresentation Y ofGL(V)ispolynomial (or
algebraic)ifitsmatrixelementsarepolynomialfunctionsoftheentriesofg, g
1
,gGL(V)(i.e.,
belong to k[g
ij
][1/det(g)]).
Forexample,V
n
andhenceallL
arepolynomial. AlsodeneL
r 1
N :=L
(
N
V
)
r
(this
denition makes sense by Proposition 4.64). This is also a polynomial representation. Thus we
haveattachedauniqueirreduciblepolynomialrepresentationL
ofGL(V) =GL
N
toanysequence
(
1
, ...,
N
) of integers (not necessarily positive) such that
1
...
N
. This sequence is called
the highest weight of L
.
Theorem 4.66. Every nite dimensional polynomial representation of GL(V) is completely re-
ducible, and decomposes into summands of the form L
)(
N
V
)
s
. SowemayassumethatY is
containedinaquotientofa(nite)directsumofsuchrepresentations. AsV
=
N1
V
N
V
,Y
iscontained inadirect sumofrepresentations oftheformV
n
(
N
V
)
s
,andwe aredone.
Remark 4.67. Since the Lie algebra sl(V) of traceless operators on V is a quotient of gl(V) by
scalars, theabove resultsextend inastraightforward mannerto representations ofthe Lie algebra
sl(V). Similarly, the results for GL(V) extend to the case of the group SL(V) of operators with
determinant 1. The only dierence is that in this case the representations L
and L
+1
m are
isomorphic, so theirreduciblerepresentations are parametrized byinteger sequences
1
...
N
upto a simultaneous shift by a constant.
In particular, any nite dimensional representation of sl(V) is completely reducible, and any
irreducible one is of the form L
:=C[S
n
]b
is isomorphic to V
.
Hint. Dene S
n
-homomorphisms f :V
and g:V
and
g(y) =yb
, and show that they are inverse to each other up to a nonzero scalar.
66
(b) Let : C[S
n
] C[S
n
] be the automorphism sending s to (1)
s
s for any permutation s.
Show that maps any representation V of S
n
to V C
= V
, where
x
0 x
0
x 1
Parabolic
0 x
Hyperbolic
x
0 y
0
, y=x
Elliptic
x y
, x F
q
, y
y x
F
q
, F
q
\F
2
q
(characteris-
tic polynomial over F
q
is irre-
ducible)
Number of elements in a conjugacy
class
1 (this is a central element)
q
2
1(elementsthatcommutewith
this one are of the form
0
t u
t
, t =
0)
q
2
+q (elementsthatcommutewith
thisoneareoftheform
0
t
u
0
, t, u=
0)
q
2
q (the reason will be described
below)
Numberof classes
q1(oneforeverynon-
zero x)
q1(oneforeverynon-
zero x)
1
(q1)(q2)(x, y= 0
2
and x=y)
1
q(q1)(matriceswith
2
yandyareconjugate)
More on the conjugacy class of elliptic matrices: these are the matrices whose characteristic
polynomial is irreducible over F
q
and which therefore dont have eigenvalues in F
q
. Let A be such
a matrix, and consider a quadratic extension of F
q
,
F
q
(
), F
q
\F
2
.
q
67
Over this eld, A will have eigenvalues
=
1
+
2
and
=
1
2
,
with corresponding eigenvectors
v, v (Av=v, Av=v).
Choose a basis
{e
1
=v+v, e
2
=
(vv)}.
In this basis, the matrix A will have the form
1
,
justifying the description of representative elements of this conjugacy class.
In the basis {v,v}, matrices that commute with A will have the form
0
,
0
for all
F
q
2
,
so the numberof such matrices is q
2
1.
4.24.2 1-dimensional representations
First, we describe the 1-dimensional representations of G.
Proposition 4.70. [G, G] =SL
2
(F
q
).
Proof. Clearly,
det(xyx
1
y
1
) = 1,
so
[G, G]SL
2
(F
q
).
To show the converse, it suces to show that the matrices
1 1 a 0 1 0
0 1
,
0 a
1
,
1 1
are commutators (as such matrices generate SL
2
(F
q
).) Clearly, by using transposition, it suces
to show that only the rsttwo matrices are commutators. But it is easy to see that the matrix
1 1
0 1
68
is the commutator of the matrices
1 1/2 1 0
A= , B= ,
0 1 0 1
while the matrix
a
0
0
a
1
a
0
0
1
, B =
0
1
1
0
,
This completes the proof.
Therefore,
G/[G, G]
via gdet(g). =F
q
The one-dimensional representations of G thus have the form
(g) = det(g) ,
where is a homomorphism
:F
q
C
;
so there are q1 such representations, denoted C
.
4.24.3 Principal series representations
Let
B G, B ={
0
}
(the set of uppertriangular matrices); then
|B|= (q1)
2
q,
[B, B] =U ={
1
},
0 1
and
B/[B, B]
q
F
q
=F
be a homomorphismdenedby
a b
=
1
(a)
2
(c),for some pair of homomorphisms
1
,
2
:F
q
C
.
0 c
Dene
V
1
,
2
=Ind
G
B
C
,
where C
1
,
2
) =
|G|
=q+ 1.
|B|
69
Theorem 4.71. 1.
1
=
2
V
1
,
2
is irreducible.
2.
1
=
2
=V
1
,
2
=C
, where W
= W
if and only if = ; V
1
,
2
= V
,
if and only if {
1
,
2
} =
1
,
2
} (in the
second case,
1
=
2
,
1
=
2
).
Proof. From the Mackey formula, we have
tr
V
1
,
2
(g) =
1
(aga
1
).
|B|
aG, aga
1
B
If
x 0
g= ,
0 x
the expression on the right evaluates to
(x)
|G|
=
1
(x)
2
(x)
q+ 1
.
|B|
If
x 1
g= ,
0 x
the expression evaluates to
(x) 1,
since here
aga
1
BaB.
If
x 0
g= ,
0 y
the expression evaluates to
1
(x)
2
(y) +
1
(y)
2
(x) 1,
since here
aga
1
B aB or a is an element of B multiplied by the transposition matrix.
If
x y
g= , x=y
y x
theexpressionontherightevaluatesto0becausematricesofthistypedonthaveeigenvaluesover
F
q
(and thus cannot be conjugated into B). From the denition,
i
(x)(i= 1,2) is a root of unity,
so
|G|
V
1
,
2
,
V
1
,
2
= (q+1)
2
(q1)+(q
2
1)(q1)
+2(q
2
+q)
(q1)(q2)
+ (q
2
+q)
1
(x)
2
(y)
1
(y)
2
(x).
2
x=y
70
1
=
2
=,
the last term is equal to
(q
2
+q)(q2)(q1),
and the total in this case is
(q+1)(q1)[(q+1)+(q1)2q(q2)]=(q+1)(q1)2q(q1)=2|G|,
so
1
,
2
,
V
1
,
2
= 2.
Clearly,
C
Ind
G
B
C
,
,
since
Hom
G
(C
,Ind
G
)=Hom
B
(C
,C
) =C(Theorem 4.33).
B
C
,
Therefore,Ind
G
=C
;W
isdierentfordistinct
B
C
,
isirreducible;andthecharacterofW
are distinct.
If
1
=
2
, let z =xy
1
, then the last term of the summation is
2
1
2
1
(q +q)
1
(z)
2
(z) = (q +q) (z) = (q +q)(q1) (z).
2
x=y x;z=1 z=1
Since
1
(z) = 0,
zF
q
because the sumof all roots of unity of a given order m >1 is zero, the last term becomes
1
(q
2
+q)(q1) (1) =(q
2
+q)(q1).
2
z=1
The dierence between this case andthe case of
1
=
2
is equal to
(q
2
+q)[(q2)(q1)+(q1)]=|G|,
so this is an irreducible representation by Lemma 4.27.
To prove the third assertion of the theorem, we look at the characters on hyperbolic elements
and note that the function
1
(x)
2
(y) +
1
(y)
2
(x)
determines
1
,
2
upto permutation.
71
4.24.4 Complementary series representations
LetF
q
2 F
q
beaquadraticextensionF
q
(
), F
q
\F
spaceoverF
q
;thenGisthegroupoflineartransformationsofF
2
.Weregardthisasa2-dimensionalvector
q
2 overF
q
.LetK Gbethecyclic
q
group of multiplications by elements of F
q
2
,
x y
K K
2
1.
y x
={ }, | |=q
For :K C
a homomorphism,let
Y
=Ind
G
K
.
This representation, of course, is very reducible. Let us compute its character, using the Mackey
formula. We get
x 0
=q(q1)(x);
0 x
(A) = 0 for A parabolic or hyperbolic;
q
x y x y x y
= + .
y x y x y x
The last assertion holds because if we regard the matrix as an element of F
q
2, conjugation is an
automorphismofF
q
2 overF
q
,buttheonlynontrivialautomorphismofF
q
2 overF
q
istheq
th
power
map.
We thus have
Ind
G
=Ind
G
K
V
,
,
where is the trivial character and W
V
,
V
,
Ind
G
where is the restriction of to scalars, has character
x 0
= (q1)(x);
0 x
x 1
=(x);
0 x
72
x
0
0
y
=0;
x
y
y
x
x
y
y
x
x
y
y
x
.
In all that follows, we will have
q
=.
The following two lemmas will establish that the inner product of this character with itself is
equalto1,thatitsvalueat1ispositive. AsweknowfromLemma4.27,thesetwopropertiesimply
that it is the character of an irreducible representation of G.
Lemma 4.72. Let be the character of the virtual representation dened above. Then
, = 1
and
(1)>0.
Proof.
(1)=q(q+1)(q+1)q(q1)=q1>0.
We now compute the inner product , . Since is a root of unity, this will be equal to
1
(q1) (q1)
2
1+(q1) 1 (q 1)+
q(q1)
(()+
q
())(() +
q
())
(q1)
2
q(q+1)
2
elliptic
Because is also a root of unity, the last term of the expression evaluates to
2 +
q1
() +
1q
().
elliptic
Lets evaluate the last summand.
Since F
q
2
is cyclic and
q
= ,
q1
() =
1q
() = 0.
F
q
2
F
q
2
Therefore,
(
q1
() +
1q
())= (
q1
() +
1q
())=2(q1)=
elliptic
F
q
since F
(q1) (q1)
2
1+(q1) 1 (q
2
1)+
q(q1)
(2(q
2
q)2(q1))
= 1.
(q1)
2
q(q+1)
2
V
,
V
,
Ind
G
K
73
exists and is irreducible. These characters are distinct for distinct pairs (, ) (up to switch
q
), so there are
q(q1)
such representations, each of dimensionq1.
2
We have thus found q 1 1-dimensional representations of G,
q(q1)
principal series repre-
2
2
sentations, and
q(q1)
complementary series representations, for a total of q 1 representations,
2
i.e., the number of conjugacy classes in G. This implies that we have in fact found all irreducible
representations of GL
2
(F
q
).
4.25 Artins theorem
Theorem 4.73. Let X be a conjugation-invariant system of subgroups of a nite group G. Then
two conditions are equivalent:
(i) Any element of G belongs to a subgroup H X.
(ii) The character of any irreducible representation of G belongs to the Q-span of characters of
induced representations Ind
G
is an irreducible representation of H.
H
V, where H X and V
Proof. Proof that (ii) implies (i). Assume that g G does not belong to any of the subgroups
H X. Then, since X is conjugation invariant, it cannot be conjugated into such a subgroup.
HencebytheMackey formula,
Ind
G
H
(g)=0forallH X andV. Soby(ii), foranyirreducible
(V)
representationW ofG,
W
(g)=0. Butirreduciblecharacters spanthespaceofclassfunctions,so
any class function vanishes on g, which is a contradiction.
Proof that (i) implies (ii). Let U be a virtual representation of G (i.e., a linear combination
of irreducible representations with nonzero integer coecients) such that (
U
,
Ind
G
H
)= 0 for all
V
H, V. So by Frobenius reciprocity, (
U|
H
,
V
) = 0. This means that
U
vanishes on H for any
H X. Hence by (i),
U
is identically zero. Thisimplies (ii).
Corollary 4.74. Any irreducible character of a nite group is a rational linear combination of
induced characters from its cyclic subgroups.
4.26 Representations of semidirect products
Let G, A be nite groups and : G Aut(A) be a homomorphism. For a A, denote (g)a by
g(a). The semidirect productGA is denedto be the productAG with multiplication law
(a
1
, g
1
)(a
2
, g
2
) = (a
1
g
1
(a
2
), g
1
g
2
).
Clearly, G and A are subgroupsof GA in a natural way.
We would like to study irreducible complex representations of GA. For simplicity, let us do
it whenA is abelian.
In this case, irreducible representations of A are 1-dimensional and form the character group
A
V
(a, g) =
1
x(h(a))
U
(hgh
1
).
|G
x
|
hG:hgh
1
Gx
Proof. (i) Let us decompose V =V
(O,U)
as an A-module. Thenwe get
V =
yO
V
y
,
where V
y
={vV
(O,U)
|av= (y, a)v, aA}. (Equivalently, V
y
={v V
(O,U)
|v(g)=0 unlessgy =
x}). So if W V is a subrepresentation, then W =
yO
W
y
, where W
y
V
y
. Now, V
y
is a
representation of G
y
, which goes to U under any isomorphism G
y
G
x
determined by g G
mappingxtoy. Hence,V
y
isirreducibleoverG
y
,soW
y
=0orW
y
=V
y
foreachy. Also,ifhy=z
then hW
y
=W
z
, so either W
y
=0 for all y or W
y
=V
y
for all y, as desired.
(ii) The orbit O is determined by the A-module structure of V, and the representation U by
the structure of V
x
as a G
x
-module.
(iii) We have
dimV
(
2
U,O)
= |O|
2
(dimU)
2
=
U,O U,O
2
|O| |G
x
|= |O||G/G
x
||G
x
|=|G| |O|=|G||A
|=|GA|.
O O O
(iv) The proof is essentially the same as that of the Mackey formula.
Exercise. Redo Problems 3.17(a), 3.18, 3.22 using Theorem 4.75.
Exercise. Deduce parts (i)-(iii) of Theorem 4.75 from part (iv).
75
|
5 Quiver Representations
5.1 Problems
Problem 5.1. Field embeddings. Recall that k(y
1
, ..., y
m
) denotes theeld ofrational functions
ofy
1
, ..., y
m
overaeldk. Letf :k[x
1
, ..., x
n
]k(y
1
, ..., y
m
)beaninjectivehomomorphism. Show
that m n. (Look at the growth of dimensions of the spaces W
N
of polynomials of degree N in
x
i
and their images under f as N ). Deduce that if f :k(x
1
, ..., x
n
)k(y
1
, ..., y
m
) is a eld
embedding, then mn.
Problem 5.2. Some algebraic geometry.
Let k be an algebraically closed eld, and G = GL
n
(k). Let V be a polynomial representation
of G. Show that if G has nitely many orbits on V then dim(V)n
2
. Namely:
(a) Let x
1
, ..., x
N
be linear coordinates on V. Let us say that a subset X of V is Zariski dense
if any polynomial f(x
1
, ..., x
N
) which vanishes on X is zero (coecientwise). Show that if G has
nitely many orbits on V then G has at least one Zariski dense orbit on V.
(b) Use (a) to construct a eld embedding k(x
1
, ..., x
N
)k(g
pq
), then use Problem 5.1.
(c) generalize the result of this problem to the case when G=GL
n
1
(k)...GL
nm
(k).
Problem 5.3. Dynkin diagrams.
Let be a graph, i.e., a nite set of points (vertices) connected with a certain number of edges
(we allow multiple edges). We assume that is connected (any vertex can be connected to any
other by a path of edges) and has no self-loops (edges from a vertex to itself ). Suppose the vertices
of are labeled by integers 1, ..., N. Then one can assign to an NN matrix R
= (r
ij
), where
r
ij
is the number of edges connecting vertices i and j. This matrix is obviously symmetric, and is
called the adjacency matrix. Dene the matrix A
= 2IR
is positive denite.
Dynkin diagrams appear in many areas of mathematics (singularity theory, Lie algebras, rep-
resentation theory, algebraic geometry, mathematical physics, etc.) In this problem you will get a
complete classication of Dynkin diagrams. Namely, you will prove
Theorem. is a Dynkin diagram if and only if it is one on the following graphs:
A
n
:
D
n
:
|
E
6
:
76
|
E
7
:
E
8
:
|
(a) Compute the determinant of A
where =A
N
, D
N
. (Use the row decomposition rule, and
write down a recursive equation for it). Deduce by Sylvester criterion
7
that A
N
, D
N
are Dynkin
diagrams.
8
(b) Compute the determinants of A
for E
6
, E
7
, E
8
(use row decomposition and reduce to (a)).
Show they are Dynkin diagrams.
(c) Show that if is a Dynkin diagram, it cannot have cycles. For this, show that det(A
) = 0
for a graph below
9
1 1 1
1
1
(show that the sum of rows is 0). Thus has to be a tree.
(d) Show that if is a Dynkin diagram, it cannot have vertices with 4 or more incoming edges,
andthatcanhavenomorethanonevertexwith3incomingedges. Forthis,showthatdet(A
) = 0
for a graph below:
1 1
2 2
1 1
(e) Show that det(A
3
1 2 3 4 5 6 4 2
(f ) Deduce from (a)-(e) the classication theorem for Dynkin diagrams.
(g)A(simply laced) aneDynkindiagram isaconnected graph without self-loops suchthat the
quadratic form dened by A
2
1
q(x
1
, ..., x
r
):= x
i
2
b
ij
x
i
x
j
.
iD i,jD
Hint. Itsucestochecktheresultforintegers: x
i
=n
i
. Firstassumethatn
i
0,andconsider
the space W of representations V of Qsuch that dimV
i
=n
i
. Show that the group
i
GL
n
i
(k) acts
with nitely many orbits on W k, and use Problem 5.2 to derive the inequality. Then deduce the
result in the case when n
i
are arbitrary integers.
(b) Deduce that q is a positive denite quadratic form.
Hint. Use the fact that Q is dense in R.
(c) Show that a quiver of nite type can have no self-loops. Then, using Problem 5.3, deduce
the theorem.
Problem 5.5. LetG= 1beanitesubgroup ofSU(2), and V bethe2-dimensional representation
of G coming from its embedding into SU(2). Let V
i
, iI, be all the irreducible representations of
G. Let r
ij
be the multiplicity of V
i
in V V
j
.
(a) Show that r
ij
=r
ji
.
(b) The McKay graph of G, M(G), is the graph whose vertices are labeled by i I, and i is
connected to j by r
ij
edges. Show that M(G) is connected. (Use Problem 3.26)
(c) ShowthatM(G) isananeDynkingraph(oneoftheforbiddengraphsinProblem5.3).
For this, show that the matrix a
ij
= 2
ij
r
ij
is positive semidenite but not denite, and use
Problem 5.3.
Hint. Letf =
x
i
V
i
,where
V
i
bethecharactersofV
i
. Showdirectlythat((2
V
)f, f)0.
When is it equal to 0? Next, show that M(G) has no self-loops, by using that if G is not cyclic
then G contains the central element IdSU(2).
78
(d) Which groupsfrom Problem 3.24 correspondto which diagrams?
(e) Using the McKay graph, nd the dimensions of irreducible representations of all nite
G SU(2) (namely, show that they are the numbers labeling the vertices of the ane Dynkin
diagrams on our pictures). Compare with the results on subgroups of SO(3) we obtained in
Problem 3.24.
5.2 Indecomposable representations of the quivers A
1
, A
2
, A
3
We have seen that a central question about representations of quivers is whether a certain quiver
hasonlynitelymanyindecomposablerepresentations. Intheprevioussubsectionitisshownthat
onlythosequiverswhoseunderlyingundirectedgraphisaDynkindiagrammayhavethisproperty.
Toseeiftheyactuallydohavethisproperty,werstexplicitlydecomposerepresentationsofcertain
easy quivers.
Remark 5.6. By an object of the type 1
1 isamapfromthezerospaceinto
a one-dimensional space. The object 1
0
0
W
=
ker
Im
A
0 W
1, the
third of 0
0, 1
1 and 0
1.
Example 5.9 (A
3
). The quiver A
3
consists of three vertices and two connections between them.
So we have to choose between two possible orientations.
or
1. We rst look at the orientation
.
79
Then a representation of this quiver looks like
A B
V W
.
Y
Like in Example 5.8 we rstsplit away
0 0
0 0 kerA
.
This object is a multiple of 1
0. Next, let Y
be a complement of ImB in Y.
Then we can also split away
0 0
0 0 Y
which is a multiple of the object 0
A B
V W
.
Y
Next, let X =ker(BA) and let X
be a complement of X in V. Let W
be a complement
of A(X) in W such that A(X
) W
. Thenwe get
A B A B
A B
=
, W =W
0 V W Y X Y W
The rst of these summands is a multiple of 1 1 0 . Looking at the second summand,
we now have a situation whereAis injective, B is surjective and furthermore ker(B A) = 0.
0 V W Y V X W X
A(
X)
X
=
A(
V)
Here, the rst summandis a multiple of 1
1
1. By splitting away the kernel of B,
thesecondsummandcanbedecomposedintomultiplesof 0
1
1 and 0
0.
So, on the whole, this quiver has six indecomposable representations:
1
0, 0
1, 1
1
0,
1
1
1, 0
1
1, 0
0
2. Now we look at the orientation
.
Very similarly to the other orientation, we can split away objects of the type
1
0, 0
1
which results in a situation where bothA and B are injective:
A
B
V W Y
.
80
of V Y in W, and set V
= W
V, Y
= W
Y.
Then we can decompose the representation as follows:
V W Y W V Y V Y V Y
= .
V Y
Thesecondsummandisamultipleoftheobject 1
1
1. Wegoondecomposingthe
rst summand. Again, to simplifynotation, we let
V =V
, W =W
, Y =Y
.
We can now assume that V Y =0. Next, let W
be a complement of V Y in W. Then
we get
V Y 0 0 W V V Y W
=
The second of these summands is a multiple of the indecomposable object 0 1 0 .
The rst summand can be further decomposed as follows:
V Y 0 0 V Y V V Y Y
=
These summandsare multiples of
1
0, 0
1
So - like in the other orientation - we get 6 indecomposable representations of A
3
:
1
0, 0
1, 1
1,
0
0, 1
0, 0
1
5.3 Indecomposable representations of the quiver D
4
As a last - slightly more complicated - example we consider the quiver D
4
.
Example 5.10 (D
4
). We restrict ourselves to the orientation
.
So a representation of this quiver looks like
A
1
V A
3
V
1
V
3
A
2
V
2
81
The rst thing we can do is - as usual - split away the kernels of the maps A
1
, A
2
, A
3
. More
precisely, we split away the representations
0 0 0
0 0
kerA
1
0 0 0 0 kerA
3
0
0 kerA
2
0
These representations are multiples of the indecomposable objects
0
0
0
0
0
0 1 0
So we get to a situation where all of the maps A
1
, A
2
, A
3
are injective.
A
1
V A
3
V
2
As in2, we can thenidentify thespaces V
1
, V
2
, V
3
withsubspacesof V. Sowe get to the triple of
subspaces problem of classifying a triple of subspacesof a given space V.
The next step is to split away a multiple of
1
The last summandis a multiple of the indecomposable representation
0
to reach a situation where
V
1
+V
2
+V
3
=V.
By letting Y = V
1
V
2
V
3
, choosing a complement V
of Y in V, and setting V
i
= V
V
i
,
i= 1,2,3, we can decompose this representation into
1 0 0 0 0 1
0
Y Y
Y
V
1
V
3
A
2
0 0
1 1
3
V
2
V
1
V
1
82
So - considering the rst summand and renaming the spaces to simplify notation - we are in a
situation where
V =V
1
+V
2
+V
3
, V
1
V
2
V
3
= 0.
As a next step, we let Y =V
1
V
2
and we choose a complement V
of Y in V such that V
3
V
,
and set V
1
=V
V
1
, V
2
=V
V
2
of Y. This yields the decomposition
V
V
V
3
1
V V
1
V
3
Y 0
V
2
The second summandis a multiple of the indecomposable object
1
In the resulting situation we have V
1
V
2
=0. Similarly we can split away multiples of
V
3
The rst of these summandsis a multiple of
0
By splitting these away we get to a situation where V
1
V
2
V
3
. Similarly, we can split away
objects of the type
1 1
1 0
to reach a situation in which the following conditions hold
83
0 1
to reach a situation where the spaces V
1
, V
2
, V
3
do not intersect pairwise
V
1
V
2
= V
1
V
3
= V
2
V
3
= 0.
If V
1
V
2
V
3
we let Y = V
1
(V
2
V
3
). We let V
1
be a complement of Y in V
1
. Since then
V
1
(V
2
V
3
)=0, we can select a complement V
of V
in V which contains V
2
V
3
. This gives
1
us the decomposition
V
V
1 1 0 1
and
V
2
Y
0 0 0 1
and
.
1 0
0
2
1 0
0
V
1
V
=
=
V
1
V
3
V
2
1. V
1
+V
2
+V
3
=V.
2. V
1
V
2
= 0, V
1
V
3
= 0, V
2
V
3
= 0.
3. V
1
V
2
V
3
, V
2
V
1
V
3
, V
3
V
1
V
2
.
But this implies that
V
1
V
2
=V
1
V
3
=V
2
V
3
=V.
So we get
dimV
1
=dimV
2
=dimV
3
=n
and
dimV = 2n.
Since V
3
V
1
V
2
we can write every element of V
3
in the form
xV
3
, x= (x
1
, x
2
), x
1
V
1
, x
2
V
2
.
We then can dene the projections
B
1
:V
3
V
1
, (x
1
, x
2
)x
1
,
B
2
:V
3
V
2
, (x
1
, x
2
)x
2
.
SinceV
3
V
1
= 0, V
3
V
2
=0,thesemapshavetobeinjectiveandthereforeareisomorphisms. We
then dene the isomorphism
A=B
2
B
1
:V
1
V
2
.
1
Let e
1
, . . . , e
n
be a basis for V
1
. Then we get
V
1
=Ce
1
Ce
2
Ce
n
V
2
=CAe
1
CAe
2
CAe
n
V
3
). =C(e
1
+Ae
1
)C(e
2
+Ae
2
) C(e
n
+Ae
n
So we can think of V
3
as the graph of an isomorphism A : V
1
V
2
. From this we obtain the
decomposition
V
C
2
n
V
3
C(1
,0)
C(1
,1)
=
j=1
2
C(0
,1)
These correspondto the indecomposable object
2
1 1
1
Thus the quiver D
4
with the selected orientation has 12 indecomposable objects. If one were to
explicitly decompose representations for the other possible orientations, one would also nd 12
indecomposable objects.
Itappearsas ifthe numberofindecomposablerepresentations doesnotdependontheorienta-
tion of the edges, and indeed - Gabriels theorem will generalize this observation.
84
5.4 Roots
From now on, let be a xed graph of type A
n
, D
n
, E
6
, E
7
, E
8
. We denote the adjacency matrix
of by R
.
Denition 5.11 (Cartan Matrix). We dene the Cartan matrix as
A
=2IdR
.
On the lattice Z
n
(or the space R
n
) we then dene an inner product
B(x, y) =x
T
A
y
correspondingto the graph .
Lemma 5.12. 1. B is positive denite.
2. B(x, x) takes only even values for xZ
n
.
Proof. 1. This follows by denition, since is a Dynkin diagram.
2. By the denition of the Cartan matrix we get
B(x, x) =x
T
A
x=
x
i
a
ij
x
j
= 2
x
i
2
+
x
i
a
ij
x
j
= 2
x
i
2
+ 2
a
ij
x
i
x
j
i
=(0, . . . , 1 , . . . ,0)
simple roots.
The
i
naturally form a basis of the lattice Z
n
.
Lemma 5.16. Let be a root, =
n
k
i
i
. Then either k
i
0 for all i or k
i
0 for all i.
i=1
Proof. Assume the contrary, i.e., k
i
> 0, k
j
< 0. Without loss of generality, we can also assume
that k
s
=0 forall s between i andj. We can identify the indices i, j with vertices of the graph .
i
j
85
Next,letbetheedgeconnectingiwiththenextvertextowardsj andi
bethevertexontheother
end of . We then let
1
,
2
be the graphs obtained from by removing . Since is supposed
to be a Dynkin diagram - andtherefore has no cycles or loops - both
1
and
2
will be connected
graphs, which are not connected to each other.
i
1
j
2
Thenwe have i
1
, j
2
. We dene
= k
m
m
, = k
m
m
.
m
1
m
2
With this choice we get
=+.
Since k
i
>0, k
j
<0 we know that = 0, =0 and therefore
B(, )2, B(, )2.
Furthermore,
B(, ) =k
i
k
i
,
since
1
,
2
are only connected at . But this has to be a nonnegative number, since k
i
> 0 and
k
i
0. This yields
B(, ) =B(+, +) =B(, )+2B(, ) +B(, )4.
2 0 2
But this is a contradiction, since was assumed to bea root.
Denition 5.17. We call a root =
k
i
i
a positive root if all k
i
0. A root forwhich k
i
0
i
for all i is called a negative root.
Remark 5.18. Lemma 5.16 states that every root is either positive or negative.
Example 5.19. 1. Let be of the type A
N1
. Then the lattice L=Z
N1
can be realized as
a subgroup of the lattice Z
N
by letting L Z
N
be the subgroup of all vectors (x
1
, . . . , x
N
)
such that
x
i
= 0.
i
The vectors
1
= (1,1,0, . . . ,0)
2
= (0,1,1,0, . . . ,0)
.
.
.
N1
= (0, . . . ,0,1,1)
naturally form a basis of L. Furthermore, the standard innerproduct
(x, y) = x
i
y
i
86
onZ
N
restrictstotheinnerproductB givenbyonL,sinceittakesthesamevaluesonthe
basis vectors:
(
i
,
i
) = 2
i, j adjacent
(
i
,
j
) =
1
0 otherwise
This means that vectors of the form
(0, . . . ,0,1,0, . . . ,0,1,0, . . . ,0)=
i
+
i+1
+ +
j1
and
(0, . . . ,0,1,0, . . . ,0,1,0, . . . ,0)=(
i
+
i+1
+ +
j1
)
are the roots of L. Therefore the numberof positive roots in L equals
N(N 1)
.
2
2. As a fact we also state the numberof positive roots in the other Dynkin diagrams:
D
N
N(N 1)
E
6
36 roots
E
7
63 roots
E
8
120 roots
Denition 5.20. Let Z
n
be a positive root. The reection s
(v) =vB(v, ).
We denote s
i
by s
i
and call these simple reections.
Remark 5.21. As a linear operator of R
n
, s
() =
Therefore s
Denition 5.25. Let Q be any quiver and i Q be a sink (a source). Then we let Q
i
be the
quiver obtained from Q by reversing all arrows pointing into (pointing out of) i.
We are now able to dene the reection functors (also called Coxeter functors).
Denition 5.26. Let Q be a quiver, i Q be a sink. Let V be a representation of Q. Then we
dene the reection functor
F
+
:RepQRepQ
i
i
by the rule
F
+
i
(V)
k
=V
k
ifk=i
_ _
F
+
(V)
i
=ker
_
: V
j
V
i
_
.
i
ji
Also, all maps stay the same but those now pointing out of i; these are replaced by compositions
of the inclusion of ker into V
j
with the projections V
j
V
k
.
Denition 5.27. Let Qbea quiver,iQbea source. LetV bea representation of Q. Let be
the canonical map
:V
i
V
j
.
ij
Thenwe dene the reection functor
F
i
:RepQRepQ
i
by the rule
F
i
(V)
k
=V
k
ifk=i
_ _
F
i
(V)
i
=Coker() =
_
V
j
_
/Im.
ij
Again,allmapsstaythesamebutthosenowpointingintoi;thesearereplacedbythecompositions
of the inclusions V
k
ij
V
j
with the natural map V
j
V
j
/Im.
Proposition 5.28. Let Q be a quiver, V an indecomposable representation of Q.
88
0
Since V is indecomposable, one of these summands has to be zero. If the rst summand is
zero, thenhastobesurjective. Ifthesecondsummandiszero, thentherstonehastobe
of the desired form, because else we could write it as a direct sum of several objects of the
type
1
0 0
0
which is impossible, since V was supposedto beindecomposable.
2. Follows similarly by splitting away the kernel of .
Proposition 5.29. Let Q be a quiver, V be a representation of Q.
1. If
: V
j
V
i
ji
is surjective, then
F
i
F
+
V =V.
i
2. If
:V
i
V
j
ij
is injective, then
+
F
i
V =V. F
i
Proof. In the following proof, we will always mean by i j that i points into j in the original
quiver Q. We only establish the rst statement and we also restrict ourselves to showing that the
spaces of V and F
i
F
+
V are the same. It is enough to do so for the i-th space. Let
i
: V
j
V
i
ji
89
, K gets replaced by
_ _
K
=
_
V
j
_
/(Im).
ji
But
Im=K
and therefore
_ _ _ _
K
=
_
V
j
_
/
_
ker(: V
j
V
i
)
_
=Im(: V
j
V
i
)
ji ji ji
by the homomorphismtheorem. Since was assumed to be surjective, we get
K
=V
i
.
Proposition 5.30. Let Q be a quiver, and V be an indecomposable representation of Q. Then
+
V and F
i
X =X, F F
i
Y =Y F
i i
In particular
F
i
X =0, F
i
Y =0.
Therefore
+
V =F
i
F V =F
i
XF
i
Y
i
which is a contradiction, since V was assumed to beindecomposable. So we can infer that
+
F V
i
is indecomposable.
90
V) =s
i
(d(V)).
Proof. Weonlyprovetherststatement, thesecondonefollowssimilarly. LetiQbeasinkand
let
: V
j
V
i
ji
be surjective. Let K =ker. Then
dimK = dimV
j
dimV
i
.
ji
Therefore we get
d(F
+
V)d(V) = dimV
j
2 dimV
i
=B(d(V),
i
)
i
i
ji
and
d(F
i
+
V)d(V)
j
= 0, j = i.
This implies
d(F
i
+
V)d(V) =B(d(V),
i
)
i
d(F
i
+
V) = d(V)B(d(V),
i
)
i
= s
i
(d(V)).
5.7 Coxeter elements
Denition 5.32. Let Q be a quiver and let be the underlying graph. Fix any labeling 1, . . . , r
of the vertices of . Then the Coxeter element c of Q correspondingto this labeling is denedas
c=s
1
s
2
. . . s
r
.
Lemma 5.33. Let
= k
i
i
i
with k
i
0 for all i but not all k
i
= 0. Then there is N N, such that
N
c
has at least one strictly negative coecient.
91
Proof. c belongs to a nite group W. So there is M N, such that
c
M
= 1.
We claim that
1 +c+c
2
+ +c
M1
= 0
asoperatorsonR
n
. Thisimplieswhatweneed,since hasatleastonestrictlypositivecoecient,
so one of the elements
c, c
2
, . . . , c
M1
must have at least one strictly negative one. Furthermore, it is enough to show that 1 is not an
eigenvalue for c, since
(1+c+c
2
+ +c
M1
)v=w= 0
cw=c
1 +c+c
2
+ +c
M1
v= (c+c
2
+c
3
+ +c
M1
+1)v=w.
Assume the contrary, i.e., 1 is a eigenvalue of c andlet v be a correspondingeigenvector.
cv=v s
1
. . . s
r
v=v
s
2
. . . s
r
v=s
1
v.
But since s
i
only changes the i-th coordinate of v, we get
s
1
v=v and s
2
. . . s
r
v=v.
Repeating the same procedure,we get
s
i
v=v
for all i. But this means
B(v,
i
) = 0.
for all i, and since B is nondegenerate, we get v = 0. But this is a contradiction, since v is an
eigenvector.
5.8 Proof of Gabriels theorem
LetV beanindecomposablerepresentation ofQ. We introduceaxedlabeling1, . . . nonQ,such
that i < j if one can reach j from i. This is possible, since we can assign the highest label to any
sink, remove this sink from the quiver, assign the next highest label to a sink of the remaining
quiver andso on. Thisway we create a labeling of the desired kind.
We now consider the sequence
V
(0)
=F
+
=F
+
F
+
=V, V
(1)
n
V, V
(2)
n1 n
V, . . .
This sequence is well dened because of the selected labeling: n has to be a sink of Q, n1 has
to be a sink of Q
n
(where Q
n
is obtained from Q by reversing all the arrows at the vertex r) and
so on. Furthermore, we note that V
(n)
is a representation of Q again, since every arrow has been
reversed twice (since we applied a reection functor to every vertex). This implies that we can
dene
V
(n+1)
=F
n
+
V
(n)
, . . .
and continue the sequence to innity.
92
Theorem 5.34. There is mN, such that
d V
(m)
=
p
for some p.
Proof. If V
(i)
is surjective at the appropriate vertex k, then
d V
(i+1)
=d F
k
+
V
(i)
=s
k
d V
(i)
.
This implies, that if V
(0)
, . . . , V
(i1)
are surjective at the appropriate vertices, then
d V
(i)
=. . . s
n1
s
n
d(V).
ByLemma5.33thiscannotcontinueindenitely- sinced
V
(i)
maynothaveanynegativeentries.
LetibesmallestnumbersuchthatV
(i)
isnotsurjectiveattheappropriatevertex. ByProposition
5.30 it is indecomposable. So, by Proposition 5.28, we get
d(V
(i)
) =
p
for some p.
We are now able to prove Gabriels theorem. Namely, we get the following corollaries.
Corollary 5.35. Let Q be a quiver, V be any indecomposable representation. Then d(V) is a
positive root.
Proof. By Theorem 5.34
s
i
1
. . . s
im
(d(V))=
p
.
Since the s
i
preserve B, we get
B(d(V), d(V))=B(
p
,
p
) = 2.
Corollary 5.36. Let V, V
). Then
V and V
are isomorphic.
Proof. Let i be such that
d V
(i)
=
p
.
Thenwe also get d
V
(i)
=
p
. So
V
(i)
=V
(i)
=:V
i
.
Furthermore we have
V
(i)
=F
+
. . . F
+
F
+
V
(0)
k n1 n
V
(i)
=F
k
+
. . . F
n
+
1
F
n
+
V
(0)
.
But both V
(i1)
, . . . , V
(0)
and V
(i1)
, . . . , V
(0)
have to be surjective at the appropriate vertices.
This implies
F
. . . F
V
i
=
F
n
F
n
1
. . . F
k
F
k
+
. . . F
n
+
1
F
n
+
V
(0)
=V
(0)
=V
n n1 k
F
n
F
n
1
. . . F
k
F
k
+
. . . F
n
+
1
F
n
+
V
(0)
=V
(0)
=V
93
These two corollaries show that there are only nitely many indecomposable representations
(sincethereareonlynitelymanyroots)andthatthedimensionvectorofeachofthemisapositive
root. The last statement of Gabriels theorem follows from
Corollary 5.37. For every positive root , there is an indecomposable representation V with
d(V) =.
Proof. Consider the sequence
s
n
, s
n1
s
n
, . . .
Considertherstelement ofthissequencewhichisa negative root(thishasto happenbyLemma
5.33) and look at one step before that, calling this element . So is a positive root and s
i
is a
negative root for some i. But since the s
i
only change one coordinate, we get
=
i
and
(s
q
. . . s
n1
s
n
)=
i
.
We let C
(i)
be the representation having dimensionvector
i
. Then we dene
V =F
n
F
n
1
. . . F
q
C
(i)
.
This is an indecomposable representation and
d(V) =.
Example 5.38. Letusdemonstratebyexamplehowreectionfunctorswork. Considerthequiver
D
4
with the orientation of all arrows towards the node (which is labeled by 4). Start with the
1-dimensional representation V
4
sitting at the 4-th vertex. Apply to V
4
the functor F
3
F
2
F
1
.
This yields
F
1
F
2
F
3
4
=V
1
+
2
+
3
+
4
.
Now applying F
4
we get
F
4
F
1
F
2
F
3
4
=V
1
+
2
+
3
+2
4
.
Note that this is exactly the inclusion of 3 lines into the plane, which is the most complicated
indecomposable representation of the D
4
quiver.
5.9 Problems
Problem5.39. LetQ
n
bethecyclicquiveroflengthn,i.e.,nverticesconnectedbynorientededges
forming a cycle. Obviously, the classication of indecomposable representations of Q
1
is given by
the Jordan normal form theorem. Obtain a similar classication of indecomposable representations
of Q
2
. In other words, classify pairs of linear operators A : V W and B : W V up to
isomorphism. Namely:
(a) Consider the following pairs (for n1):
1) E
n,
: V =W =C
n
, A is the Jordan block of size n with eigenvalue , B = 1 (C).
2) E
n,
: is obtained from E
n,0
by exchanging V with W and A with B.
94
3) H
n
: V = C
n
with basis v
i
, W = C
n1
with basis w
i
, Av
i
= w
i
, Bw
i
=v
i+1
for i < n, and
Av
n
= 0.
4) K
n
is obtained from H
n
by exchanging V with W and A with B.
Show that these are indecomposable and pairwise nonisomorphic.
(b) Show that if E is a representation of Q
2
such that AB is not nilpotent, then E =E
,
where E
=E
n,
for some = 0.
(c) Consider the case when AB is nilpotent, and consider the operator X on V W given
by X(v, w) = (Bw, Av). Show that X is nilpotent, and admits a basis consisting of chains (i.e.,
sequences u, Xu, X
2
u, ...X
l1
u where X
l
u= 0) which are compatible with the direct sum decompo-
sition (i.e., for every chain u V or u W). Deduce that (1)-(4) are the only indecomposable
representations of Q
2
.
(d) generalize this classication to the Kronecker quiver, which has two vertices 1 and 2 and
two edges both going from 1 to 2.
(e)(harder!) can you generalize this classication to Q
n
, n >2?
Problem 5.40. Let L
1
Z
8
be the lattice of vectors where the coordinates are either all integers
2
or all half-integers (but not integers), and the sum of all coordinates is an even integer.
(a) Let
i
=e
i
e
i+1
, i= 1, ...,6,
7
=e
6
+e
7
,
8
=1/2
8
Show that
i
are a basis
i=1
e
i
.
of L (over Z).
(b)ShowthatrootsinL(undertheusualinnerproduct)formarootsystemoftypeE
8
(compute
the inner products of
i
).
(c) Show that the E
7
and E
6
lattices can be obtained as the sets of vectors in the E
8
lattice L
where the rst two, respectively three, coordinates (in the basis e
i
) are equal.
(d) Show that E
6
, E
7
, E
8
have 72,126,240 roots, respectively (enumerate types of roots in terms
of the presentations in the basis e
i
, and count the roots of each type).
Problem 5.41. Let V
i
has a 1-dimensional space
at i and 0 everywhere else.
(a) Show that if i is a source then Ext
1
(V, V
i
) = 0 for any representation V of Q, and if i is
a sink, then Ext
1
(V
i
, V) = 0.
(b) Given an orientation of the quiver, nd a Jordan-Holder series of V
is a functor Vect
k
Vect
op
.
k
4. TheHomfunctors: IfC isalocally smallcategory thenwehave thefunctorC Setsgiven
by Y Hom(X, Y) and C
op
Sets given by Y Hom(Y, X).
5. The assignment X Fun(X,Z) is a functor SetsRings
op
.
6. LetQbeaquiver. ConsiderthecategoryC(Q)whoseobjectsaretheverticesandmorphisms
areorientedpathsbetweenthem. ThenfunctorsfromC(Q)toVect
k
arerepresentationsofQover
k.
7. Let K G be groups. Then we have the induction functor Ind
G
K
:Rep(K)Rep(G), and
Res
G
K
:Rep(G)Rep(K).
8. We have an obvious notion of the Cartesian product of categories (obtained by taking the
Cartesian productsof the classes of objects and morphismsof the factors). The functors of direct
sumandtensorproductarethenfunctorsVect
k
Vect
k
Vect
k
. AlsotheoperationsV V
n
,
V S
n
V, V
n
V are functors on Vect
k
. More generally, if is a representation of S
n
, we
havefunctorsV Hom
Sn
(, V
n
). Suchfunctors(forirreducible)arecalledtheSchurfunctors.
They are labeled by Young diagrams.
9. The reection functors F
i
: Rep(Q) Rep(Q
i
) are functors between representation cate-
gories of quivers.
97
6.3 Morphisms of functors
One of the important features of functors between categories which distinguishes them fromusual
maps or functions is that the functors between two given categories themselves form a category,
i.e., one can dene a nontrivial notion of a morphismbetween two functors.
Denition 6.6. Let C,D becategories andF, G:C D befunctorsbetween them. A morphism
a : F G (also called a natural transformation or a functorial morphism) is a collection of
morphisms a
X
: F(X) G(X) labeled by the objects X of C, which is functorial in X, i.e., for
any morphismf :X Y (for X, Y C) one has a
Y
F(f) =G(f)a
X
.
Amorphisma:F Gisanisomorphismifthereisanothermorphisma
1
:GF suchthat
aa
1
anda
1
aaretheidentities. ThesetofmorphismsfromF toGisdenotedbyHom(F, G).
Example 6.7. 1. LetFVect
k
bethecategoryofnitedimensionalvectorspacesoverk. Thenthe
functors id and on this category are isomorphic. The isomorphism is dened by the standard
maps a
V
: V V
given by a
V
(u)(f) = f(u), u V, f V
.
2. Let FVect
k
be the category of nite dimensional k-vector spaces, where the morphisms
are the isomorphisms. We have a functor F from this category to itself sending any space V to
V
)
1
. This functor satises the property that V is isomorphic to
F(V) for any V, butit is not isomorphic to the identity functor. Thisis because the isomorphism
V F(V) = V
as a representation of GL(V).
3. Let A be an algebra over a eld k, and F : Amod Vect
k
be the forgetful functor.
Thenas follows fromProblem 1.22, EndF =Hom(F, F) =A.
4. The set of endomorphismsof the identity functor on the category Amod is the center of
A (check it!).
6.4 Equivalence of categories
When two algebraic or geometric objects are isomorphic, it is usually not a good idea to say that
they are equal (i.e., literally the same). The reason is that such objects are usually equal in many
dierentways,i.e.,therearemanywaystopickanisomorphism,butbysayingthattheobjectsare
equal we are misleading the reader or listener into thinking that we are providing a certain choice
of the identication, which we actually do not do. A vivid example of this is a nite dimensional
vector space V andits dual space V
.
For this reason in category theory, one most of the time tries to avoid saying that two objects
ortwo functorsareequal. Inparticular,thisappliestothedenitionofisomorphismofcategories.
Namely, the naive notion of isomorphismof categories is denedin the obvious way: a functor
F :C D isanisomorphismifthereexistsF
1
:D C suchthatFF
1
andF
1
F areequal
to the identity functors. But this denition is not very useful. We might suspect so since we have
used the word equal for objects of a category (namely, functors) which we are not supposed to
do. Andinfacthereisanexampleoftwocategorieswhicharethesameforallpracticalpurposes
but are not isomorphic; it demonstrates the deciency of our denition.
Namely, let C
1
be the simplest possible category: Ob(C
1
) consists of one object X, with
98
Hom(X, X) = {1
X
}. Also, let C
2
have two objects X, Y and 4 morphisms: 1
X
,1
Y
, a : X Y
and b:Y X. So we must have ab= 1
Y
, ba= 1
X
.
It is easy to check that for any category D, there is a natural bijection between the collections
of isomorphism classes of functors C
1
D and C
2
D (both are identied with the collection of
isomorphismclassesofobjectsofD). ThisiswhatwemeanbysayingthatC
1
andC
2
arethesame
for all practical purposes. Nevertheless they are not isomorphic, since C
1
has one object, and C
2
has two objects (even though these two objects are isomorphic to each other).
This shows that we should adopt a more exible and less restrictive notion of isomorphism of
categories. This is accomplished by the denition of an equivalence of categories.
Denition 6.8. A functor F : C D is an equivalence of categories if there exists F
: D C
such that F F
and F
is said to be a quasi-inverse to F.
In particular, the above categories C
1
andC
2
are equivalent (check it!).
Also, the category FSet of nite sets is equivalent to the category whose objects are nonneg-
ative integers, and morphisms are given by Hom(m, n) = Maps({1, ..., m},{1, ..., n}). Are these
categories isomorphic? The answer to this question depends on whether you believe that there
is only one nite set with a given number of elements, or that there are many of those. It seems
bettertothinkthattherearemany(withoutaskinghowmany),sothatisomorphicsetsneednot
be literally equal, but this is really a matter of choice. In any case, this is not really a reasonable
question; the answer to this question is irrelevant for any practical purpose,andthinking about it
will give you nothing buta headache.
6.5 Representable functors
A fundamentalnotion in category theory is that of a representable functor. Namely, let C bea
(locally small) category, andF :C Sets be a functor. We say that F is representable if there
exists an object X C such that F is isomorphic to the functorHom(X,?).
Inasimilarway,onecantalkaboutrepresentablefunctorsfromC
op
toSets. Namely, onecalls
such a functor representable if it is of the form Hom(?, X) for some object X C.
Not every functor is representable, but if a representing object X exists, then it is unique.
Namely, we have the following lemma.
Lemma 6.9. (The Yoneda Lemma) If X exists, then it is unique up to a unique isomorphism.
I.e.,ifX, Y aretwoobjectsinC,thenforanyisomorphism offunctors:Hom(X,?)Hom(Y,?)
there is a unique isomorphism a
:X Y inducing .
Proof. (Sketch)Onesetsa
Y
1
(1
Y
),andshowsthatitisinvertible byconstructingtheinverse,
which is a
1
=
X
(1
X
). It remains to show that the composition both ways is the identity, which
wewillomithere. Thisestablishestheexistenceofa
. Itsuniquenessisveriedinastraightforward
manner.
Example 6.10. Let A be an algebra. Then the forgetful functor on the category of left A-
modules is representable, and the representing object is the free rank 1 module (=the regular
representation) M =A. But if A is innite dimensional, and we restrict attention to the category
ofnitedimensionalmodules,thentheforgetfulfunctor,ingeneral,isnotrepresentable(thisisso,
99
for example, if A is the algebra of complex functions on Z which are zero at all points butnitely
many).
6.6 Adjoint functors
Another fundamental notion in category theory is the notion of adjoint functors.
Denition6.11. FunctorsF :C DandG:D C aresaidtobeapairofadjointfunctorsiffor
anyX C,Y Dwearegivenanisomorphism
XY
:Hom
C
(F(X), Y)Hom
D
(X, G(Y))whichis
functorialinX andY;inotherwords,ifwearegiven anisomorphismoffunctorsHom(F(?),?)
Hom(?, G(?)) (C DSets). In this situation, we say that F is left adjoint to G and G is right
adjoint to F.
Remark6.12. Thisterminologyismotivatedbytheanalogybetweencategoriesandinnerproduct
spaces. Namely, the inner product on a category is the assignment X, Y Hom(X, Y) (so it
takes values in the category Sets). Yonedas lemma may be interpreted as nondegeneracy of the
innerproduct(andtherepresentabilitypropertyisanalogoustothepropertyofalinearfunctional
tobetheinnerproductwithavector,whichisnotalwaysthecaseinaninnitedimensionalspace,
even if the inner product is nondegenerate). With this analogy in mind, the above denition is
parallel to the denition of adjoint operators: F : V W and G : W V are a pair of adjoint
operators if (Fv, w) = (v, Gw) for all v, w. Note that if the inner products are not symmetric,
thentheleftandrightadjointofanoperatordontnecessarilycoincide;thesameappliestoadjoint
functors.
Not every functorhasa left or right adjoint, butifit does, it isuniqueandcan beconstructed
canonically (i.e., if we somehow found two such functors, then there is a canonical isomorphism
betweenthem). ThisfollowseasilyfromtheYonedalemma,asifF, Gareapairofadjointfunctors
then F(X) represents the functor Y Hom(X, G(Y)), and G(Y) represents the functor X
Hom(F(X), Y).
Example 6.13. 1. Let V be a nite dimensional representation of a group G. Then the left and
right adjoint to the functor V on the category of representations of G is the functor V
.
2. The functor Res
G
is left adjoint to Ind
G
This is nothing but the statement of Frobenius
K K
.
reciprocity.
3. Let Assoc
k
be the category of associative unital algebras, and Lie
k
the category of Lie
algebrasoversomeeldk. WehaveafunctorL:Assoc
k
Lie
k
,whichattachestoanassociative
algebrathesamespaceregardedasaLiealgebra,withbracket[a, b] =abba. ThenthefunctorL
hasaleftadjoint,whichisthefunctorU oftakingtheuniversalenvelopingalgebraofaLiealgebra.
4. We have the functor GL
1
:Assoc
k
Groups, given by AGL
1
(A) =A
. This functor
has a left adjoint, which is the functor Gk[G], the groupalgebra of G.
5. The left adjoint to the forgetful functor Assoc
k
Vect
k
is the functor of tensor algebra:
V TV. Also,ifwedenotebyComm
k
thecategoryofcommutativealgebras,thentheleftadjoint
to the forgetful functor Comm
k
Vect
k
is the functor of the symmetric algebra: V SV.
One can give many more examples, spanning many elds. These examples show that adjoint
functors are ubiquitousin mathematics.
100
6.7 Abelian categories
The type of categories that most often appears in representation theory is abelian categories.
The standard denition of an abelian category is rather long, so we will not give it here, referring
thereadertothetextbook[Fr];rather,wewilluseasthedenitionwhatisreallythestatement of
the Freyd-Mitchell theorem:
Denition6.14. AnabeliancategoryiscategoryequivalenttoafullsubcategoryCofthecategory
of A-mod of left modules over a ring A, which is closed under taking nite direct sums, as well as
kernels, cokernels, and images of morphisms.
Example 6.15. ThecategoryofmodulesoveranalgebraAandthecategoryofnitedimensional
modules over A are abelian categories.
Weseefromthisdenitionthatinanabeliancategory, Hom(X, Y)isanabeliangroupforeach
X, Y,compositionsaregrouphomomorphismswithrespecttoeachargument,thereisthezeroob-
ject,thenotionofaninjectivemorphism(monomorphism)andsurjectivemorphism(epimorphism),
and every morphismhas a kernel and a cokernel.
Remark 6.16. The good thing about Denition 6.14 is that it allows us to visualize objects,
morphisms, kernels, and cokernels in terms of classical algebra. But the denition also has a big
drawback, whichisthatevenifC isthewholecategory A-mod,theringAisnotdeterminedbyC.
In particular, two dierent rings can have equivalent categories of modules (such rings are called
Morita equivalent). Actually, itisworsethanthat: formanyimportantabeliancategories there
is no natural (or even manageable) ring A at all. This is why people prefer to use the standard
denition, which is free from this drawback, even though it is more abstract.
We say that an abelian category C is k-linear if the groups Hom
C
(X, Y) are equipped with
a structure of a vector space over k, and composition maps are k-linear in each argument. In
particular, the categories in Example 6.15 are k-linear.
6.8 Exact functors
Denition 6.17. A sequence of objects andmorphisms
X
0
X
1
...X
n+1
in an abelian category is said to be a complex if the composition of any two consecutive arrows
is zero. The cohomology of this complex is H
i
=Ker(d
i
)/Im(d
i1
), where d
i
:X
i
X
i+1
(thus
the cohomology is dened for 1 i n). The complex is said to be exact in the i-th term if
H
i
=0,andissaidtobean exact sequenceifitisexactinallterms. Ashort exact sequence
is an exact sequence of the form
0X Y Z 0.
Clearly, 0 X Y Z 0 is a short exact sequence if and only if X Y is injective,
Y Z is surjective, and the inducedmap Y/X Z is an isomorphism.
Denition 6.18. A functorF between two abelian categories isadditive ifit induceshomomor-
phismsonHomgroups. Also,fork-linearcategoriesonesaysthatF isk-linearifitinducesk-linear
maps between Hom spaces.
101
It is easy to show that if F is anadditive functor, then F(XY) is canonically isomorphic to
F(X)F(Y).
Example 6.19. ThefunctorsInd
G
K
,Hom
G
(V,?)inthetheoryofgrouprepresentationsover
K
,Res
G
a eld k are additive andk-linear.
Denition 6.20. An additive functor F : C D between abelian categories is left exact if for
any exact sequence
0X Y Z,
the sequence
0F(X)F(Y)F(Z)
is exact. F is right exact if for any exact sequence
X Y Z 0,
the sequence
F(X)F(Y)F(Z)0
is exact. F is exact if it is bothleft and right exact.
Denition6.21. AnabeliancategoryCissemisimpleifanyshortexactsequenceinthiscategory
splits, i.e., is isomorphic to a sequence
0X XY Y 0
(where the maps are obvious).
Example 6.22. The category of representations of a nite group G over a eld of characteristic
not dividing G (or 0) is semisimple. | |
Note that in a semisimple category, any additive functor is automatically exact on both sides.
Example 6.23. (i) The functors Ind
G
K
are exact.
K
, Res
G
(ii)ThefunctorHom(X,?)isleftexact, butnotnecessarilyrightexact. Toseethatitneednot
be right exact, it suces to consider the exact sequence
0ZZZ/2Z0,
and apply the functor Hom(Z/2Z,?).
(iii)ThefunctorX
A
forarightA-moduleX (onthecategoryofleftA-modules)isrightexact,
but not necessarily left exact. To see this, it suces to tensor multiply the above exact sequence
by Z/2Z.
Exercise. Showthatif(F, G)isapairofadjointadditivefunctorsbetweenabeliancategories,
then F is right exact and G is left exact.
Exercise. (a)LetQbeaquiverandiQa source. LetV bearepresentationofQ,andW a
representation of Q
i
(the quiver obtained from Q by reversing arrows at the vertex i). Prove that
+
F
i
. functor F
i
(b) Deduce that the functor F
+
i
is left exact, and F
i
is right exact.
102
7 Structure of nite dimensional algebras
In this section we return to studying the structure of nite dimensional algebras. Throughoutthe
section, we work over an algebraically closed eld k (of any characteristic).
7.1 Projective modules
Let A be an algebra, and P be a left A-module.
Theorem 7.1. The following properties of P are equivalent:
(i) If :M N is a surjective morphism, and :P N any morphism, then there exists a
morphism :P M such that =.
(ii)Anysurjectivemorphism:M P splits,i.e.,thereexists:P M suchthat=id.
(iii) There exists another A-module Q such that PQ is a free A-module, i.e., a direct sum of
copies of A.
(iv) The functor Hom
A
(P,?) on the category of A-modules is exact.
Proof. Toprovethat(i)implies(ii),takeN =P. Toprovethat(ii)implies(iii),takeM tobefree
(thiscanalwaysbedonesinceanymoduleisaquotientofafreemodule). Toprovethat(iii)implies
(iv),notethatthefunctorHom
A
(P,?)isexactifP isfree(asHom
A
(A, N) =N),sothestatement
follows, as if the direct sum of two complexes is exact, then each of them is exact. To prove that
(iv) implies (i), let K be the kernel of the map , and apply the exact functor Hom
A
(P,?) to the
exact sequence
0K M N 0.
Denition 7.2. A module satisfying any of the conditions (i)-(iv) of Theorem 7.1 is said to be
projective.
7.2 Lifting of idempotents
Let A be a ring, and I A a nilpotent ideal.
Proposition 7.3. Let e
0
A/I be an idempotent, i.e., e
2
=e
0
. There exists an idempotent eA
0
which isaliftofe
0
(i.e.,itprojects toe
0
underthereduction modulo I). Thisidempotent isunique
up to conjugation by an element of 1 +I.
Proof. Let us rst establish the statement in the case when I
2
= 0. Note that in this case I is a
leftandrightmoduleoverA/I. Lete
beanyliftofe
0
toA. Thene
2
e
=aI,ande
0
a=ae
0
.
We look for e in the form e=e
= e+c. For e
to be an idempotent, we
must have ec+cec = 0. This is equivalent to saying that ece = 0 and (1e)c(1e) = 0, so
c=ec(1e)+(1e)ce= [e,[e, c]]. Hence e
i
n
=1
e
i
=1.
Corollary 7.5. Lete
01
, ..., e
0m
beacompletesystemoforthogonal idempotents inA/I. Thenthere
exists a complete system of orthogonal idempotents e
1
, ..., e
m
(e
i
e
j
=
ij
e
i
,
e
i
= 1) in A which
lifts e
01
, ..., e
0m
.
Proof. The proof is by induction in m. For m = 2 this follows from Lemma 7.3. For m >
2, we lift e
01
to e
1
using Lemma 7.3, and then apply the induction assumption to the algebra
(1e
1
)A(1e
1
).
7.3 Projective covers
Obviously, every projective module over an algebra A is a direct sum of indecomposable projec-
tive modules, so to understand projective modules over A, it suces to classify indecomposable
projective modules.
Let A bea nite dimensional algebra, with simple modules M
1
, ..., M
n
.
Theorem 7.6. (i) For each i= 1, ..., n there exists a unique indecomposable projective module P
i
such that dimHom(P
i
, M
j
) =
ij
.
(ii) A=
i
n
=1
(dimM
i
)P
i
.
(iii) any indecomposable projective module over A is isomorphic to P
i
for some i.
Proof. Recall that A/Rad(A) =
i
n
=1
End(M
i
), and Rad(A) is a nilpotent ideal. Pick a basis of
M
i
, and let e
0
= E
i
the rank 1 projectors projecting to the basis vectors of this basis (j =
ij jj
,
1, ...,dimM
i
). Then e
ij
0
are orthogonal idempotents in A/Rad(A). So by Corollary 7.5 we can lift
them to orthogonal idempotents e
ij
in A. Now deneP
ij
=Ae
ij
. ThenA=
i
dimM
i
P
ij
, so P
ij
j=1
are projective. Also, we have Hom(P
ij
, M
k
) = e
ij
M
k
, so dimHom(P
ij
, M
k
) =
ik
. Finally, P
ij
is
independent of j up to an isomorphism, as e
ij
for xed i are conjugate under A
by Proposition
7.3; thus we will denote P
ij
by P
i
.
We claim that P
i
is indecomposable. Indeed, if P
i
=Q
1
Q
2
, then Hom(Q
l
, M
j
) = 0 for all j
either for l=1 or for l=2, so either Q
1
= 0 or Q
2
=0.
Also, there can be no other indecomposable projective modules, since any indecomposable
projective module has to occur in the decomposition of A. The theorem is proved.
References
[BGP] J. Bernstein, I. Gelfand, V. Ponomarev, Coxeter functors and Gabriels theorem, Russian
Math. Surveys28 (1973), no. 2, 1732.
[Cu] C.Curtis,PioneersofRepresentationTheory: Frobenius,Burnside,Schur,andBrauer,AMS,
1999.
104
[CR] C. Curtis and I. Reiner, Representation Theory of Finite Groups and Associative Algebras,
AMS, 2006.
[FH] W. Fulton andJ. Harris, Representation Theory, A rst course, Springer, New York, 1991.
[Fr] Peter J. Freyd, Abelian Categories, an Introduction to the Theory of Functors. Harper and
Row (1964).
[McL] S. MacLane, Categories for a working Mathematician: 2nd Ed., Graduate Texts in Mathe-
matics 5, Springer,1998.
105
MIT OpenCourseWare
http://ocw.mit.edu
18.712 Introduction to Representation Theory
Fall 2010
For information about citing these materials or our Terms of Use, visit: http://ocw.mit.edu/terms.