Professional Documents
Culture Documents
Femke Bekius
July 22, 2011
Bachelor Thesis
Supervisor: prof. dr. Alexander Schrijver
Data
Title: The Shannon Capacity of a Graph
Author: Femke Bekius, Femke.Bekius@student.uva.nl, 5823390
Supervisor: prof. dr. Alexander Schrijver
Second assessor: prof. dr. Monique Laurent
Enddate: July 22, 2011
Introduction 2
5 Conclusion 36
Popular summary 37
1
Introduction
2
called Perfect Graphs. In chapter 2 we use Lovász technique to determine
the Shannon capacity of C5 . For long time this was an open problem and
therefore this is a very important result.
Thereafter we proceed with the Lovász Number, which is another upper
bound for the Shannon capacity, about which we prove a couple of formulas
and properties. At the end we prove that the Lovász Number is indeed a
smaller or equal upper bound for the Shannon capacity than the one Shannon
found. In this chapter we use, among others, the techniques introduced in
chapter 1. In the last chapter we consider the three problems Lovász stated
at the end of his article for which Haemers found a counterexample.
3
Chapter 1
In this chapter we give some definitions, theorems and lemmas of the con-
cepts needed for the proofs in the rest of the thesis. If there is no reference to
an article or book we used the definition, theorem, lemma, example or corol-
lary from Lovász [5]. First we start with some definitions in graph theory
and describe some important concepts of linear algebra. In the last section
we state Shannon’s Theorem and define the perfect graphs, for which the
Shannon capacity is easy to determine. At the end we consider the cycle
graphs Cn .
4
H is an induced subgraph of G if W ⊆ V and F = {e ∈ E | e ⊆ W }.
H is a complete subgraph of G if H is an induced subgraph of G with the
additional feature that every two vertices in H are connected by an edge. A
complete subgraph is a clique, see definition 1.20.
Definition 1.5. (Strong product) The strong product G1 G2 of two graphs
G1 (V1 , E1 ) and G2 (V2 , E2 ) has vertex set V1 × V2 = {(u1 , u2 ) : u1 ∈ V1 , u2 ∈
V2 } with (u1 , u2 ) 6= (v1 , v2 ) adjacent if and only if ui = vi or ui vi ∈ Ei for
i = 1, 2.
The k-strong product of graph G, Gk = G G . . . G, with vertex set
V k = {(u1 , . . . , uk ) : ui ∈ V } with (u1 , . . . , uk ) 6= (v1 , . . . , vk ) adjacent if and
only if ui = vi or ui vi ∈ Ei ∀i [1].
It follows that α(Gk ) is the maximum number of k-letter messages which can
be sent without danger of confusion. There are at least α(G)k of such words.
Two k-letter messages are confoundable if for all 1 ≤ i ≤ k their ith letters
are adjacent.
Example 1.7. For the cycle graph C5 , also called the pentagon, α(C52 ) = 5.
In fact, if v1 , . . . , v5 are the vertices in a cyclic way, then the 2-letter messages
v1 v1 , v2 v3 , v3 v5 , v4 v2 and v5 v4 are nonconfoundable. In the figure on the next
page the red colored vertices form the maximum independent set of α(C52 ).
5
Figure 1.1: Strong product C5 C5
yj )n,m
Then x ◦ y = (xiP i=1,j=1 )n,m
, v ◦ w = (vi wjPi=1,j=1 and
(x◦y) (v◦w) = i=1 j=1 xi yj vi wj = ( ni=1 xi vi )( m
T n m T T
P P
j=1 yj wj ) = (x v)(y w).
6
Proof. Let (i1 , i2 ), (j1 , j2 ) ∈ V (G) × V (H) be nonadjacent in G H; that is,
i1 , j1 are nonadjacent in G or i2 , j2 are nonadjacent in H. Then uTi1 uj1 = 0
or v Ti2 uj2 = 0. From this we have (ui1 ◦ v i2 )T (uj1 ◦ v j2 ) = 0. On the other
hand, if (i1 , i2 ), (j1 , j2 ) ∈ V (G) × V (H) adjacent or equal in G H; that is,
i1 , j1 adjacent or equal in G and i2 , j2 adjacent or equal in H. Hence from
(1.1) follows that ui ◦ v j is an orthonormal representation of G H.
The value of an orthonormal representation (u1 , . . . , un ) is defined to be
1
min max ,
c 1≤i≤n (cT ui )2
where c ranges over all unit vectors and the minimum of all these c is called
the handle of the representation [5].
Definition 1.11. (Lovász Number) ϑ(G) is the minimum value over all the
representations mentioned above, i.e.,
1
ϑ(G) = min min max .
u1 ,...,un c 1≤i≤n (cT ui )2
7
Lemma 1.14. α(G) ≤ ϑ(G)
Proof. Let (u1 , . . . , un ) be an optimal orthonormal representation of G with
handle c. Let {1, . . . , k} be a maximum independent set in G. Then u1 , . . . , uk
are pairwise orthogonal. Extends this to an orthonormal representation
0 0 0
(u1 , . . . , un ) where ui = ui if i = 1, . . . k. Then,
n k
2 T
X
T 0 2
X k α(G)
1=c =c c= (c ui ) ≥ (cT ui )2 ≥ = .
i=1 i=1
ϑ(G) ϑ(G)
8
1.3 Shannon’s Theorem
In 1959 Shannon introduced the Shannon capacity of a graph [5].
9
Using the multiplicative version, an+m ≥ an am for all positive n, m ∈ Z and
an > 0 ∀n, we have
as Shannon defined.
Hence, Θ(G) ≥ α(G) and we conclude that α(G) is a lower bound for the
Shannon capacity of G.
A general upper bound on Θ(G) was given by Shannon. This upper bound
is denoted by α∗ (G), which is the maximum of x∈V w(x) taken over all
P
fractional vertex packings w [5].
max{cT x | x ≥ 0, Ax ≤ b} = min{y T b | y ≥ 0, y T A ≥ cT }
10
provided that ∃x ≥ 0 : Ax ≤ b and ∃y ≥ 0 : y T A ≥ cT ,[6].
Consider a matrix A whoserows are indicated by cliques, columns by vertices
1 if v ∈ C
and whose (C, v)-entry =
0 if v∈
/ C,
1 1
.. ..
and vectors c = . V (G) , b = . P (G),
1 1
here P (G) is the collection cliques in G, C ∈ P (G) and v ∈ V (G).
With this theorem α∗ (G) can be defined dually as P follows: assign nonnega-
P C of G such that x∈C q(C) ≥ 1 for each
tive weights q(C) to the cliques
point x of G and minimize C q(C).
11
Proof. If G is perfect, then G is perfect. Then ω(G) = χ(G) by definition
and it follows that α(G) = χ(G) = χ(G), because ω(G) = α(G).
In general α(G) ≤ α∗ (G) ≤ χ(G), since
12
• C5 : α(C5 ) = 2, α∗ (C5 ) = 5
2
⇒ 2 ≤ Θ(C5 ) ≤ 5
2
• C7 : α(C7 ) = 3, α∗ (C7 ) = 7
2
⇒ 3 ≤ Θ(C7 ) ≤ 7
2
• C9 : α(C9 ) = 4, α∗ (C9 ) = 9
2
⇒ 4 ≤ Θ(C9 ) ≤ 9
2
We see that if n is even, then α(Cn ) = α∗ (Cn ) = Θ(Cn ). This result follows
from the fact that if n is even Cn is a perfect graph or we can check directly
that χ(Cn ) = n2 . If n is odd we see Θ(C3 ) = 1, but for n > 3 we can only
give a lower and an upper bound for Θ(Cn ), namely n−1 2
≤ Θ(Cn ) ≤ n2 .
We should remark here that in general the exact determination of maximum
independent sets in Cnd seems to be a very hard task [2].
In the next chapter
√ we will see how Lovász proved that the Shannon capacity
of C5 equals 5, but for C7 and in general for Cn with n odd and n > 5, we
cannot give the exact value of the Shannon capacity.
13
Chapter 2
Figure 2.1: C5
√
Theorem 2.1. Θ(C5 ) = 5
In the next proof we use the umbrella technique introduced by Lovász, see
also figure 2.2 [5].
14
Proof. Consider an umbrella whose handle and five ribs have the unit length.
Open the umbrella till the point where the maximum angle between the ribs is
π
2
. Let u1 , u2 , u3 , u4 , u5 be the ribs and c the handle as vectors oriented away
from their common point. Then u1 , . . . , u5 is an orthonormal representation
T −1/4
of C5 . By the Spherical √ Cosine Theorem we can compute c ui = 5 .
To show: Θ(C5 ) ≤ 5.
Let S be a maximum independent set in C5k , so S ⊆ C5k . Then
S ⊆ {(i1 , . . . , ik ) | i1 , . . . , ik ∈ V (C5k )}. We have X := {ui1 ◦ . . . ◦ uik |
(i1 , . . . , ik ) ∈ S} is orthonormal because ui and uj are orthogonal if i, j
nonadjacent. Let c be a unit vector. Then,
1 = hc
| ◦ .{z
. . ◦ c}, c
| ◦ .{z
. . ◦ c}i
k k
X
≥ hc ◦ . . . ◦ c, xi2
x∈X
X
= hc ◦ . . . ◦ c, ui1 ◦ . . . ◦ uik i2
(i1 ,...ik )∈S
k
X Y X 1 |S|
= hc, uij i2 = √ = √
( 5)k ( 5)k
(i1 ,...ik )∈S j=1 x∈X
√
⇒| S |≤ ( 5)k .
Therefore, q√
p
k k
√
Θ(C5 ) = sup | S | ≤ sup ( 5)k = 5.
k k
15
Chapter 3
In this chapter we investigate the upper bound ϑ(G) in more detail and give
several proofs which were stated by Lovász. We first prove that ϑ(G) is an
upper bound for Θ(G). Then we prove a theorem about positive semidefinite
matrices such that we can give a relation between ϑ(G) and eigenvalues of
symmetric matrices. Next we prove a theorem about the value of ϑ(G). After
introducing an orthonormal representation of G we draw some consequences.
Then we give another way of representing the value of ϑ(G) and we prove
equality in Lemma 1.13. We give some properties of ϑ(G) if G is vertex-
or edge-transitive or if G is regular and we end with proving that ϑ(G) is a
smaller or equal upper bound for Θ(G) than the one Shannon found.
16
Theorem 3.3. Let G be a graph on vertices {1, . . . , n}. Then ϑ(G) is the
minimum of the largest eigenvalue of any symmetric matrix A = (aij )ni,j=1
such that
aij = 1 if i = j or if i and j are nonadjacent . (3.1)
Because
T T
uTi uj
ui uj T T uj ui
c− T c− T =c c−c T −c +
(c ui ) (c uj ) c uj cT u i (cT ui )(cT uj )
uT uj
= cT c − 2 + T i T
(c ui )(c uj )
uT uj
= −1 + T i T = −aij .
(c ui )(c uj )
And 2
ui 1
ϑ(G) − aii = c− T + ϑ(G) − T 2 (3.3)
(c ui ) (c ui )
Because
(cT ui ) uTi ui 1
ϑ(G) − aii = cT c − 2 T
+ T 2
+ ϑ(G) − T 2
(c ui ) (c ui ) (c ui )
1 1
= 1 − 2 + T 2 − T 2 + ϑ(G) = ϑ(G) − 1.
(c ui ) (c ui )
By equation (3.2) define D := c − cTuu1 1 , . . . , c − cTuun n and let B = DT D.
By defintion of ϑ(G) holds ∀i ϑ(G) ≥ (cT 1ui )2 . Hence ∀i we can define wi :=
ϑ(G) − (cT 1ui )2 ≥ 0, here ∆w is the matrix with on the diagonal w1 , . . . , wn
and the other entries equal 0. Then by (3.2) and (3.3),
ϑ(G)I−A = B+∆w is positive semidefinite because B is of the form DT D and
B is symmetric with non-negative eigenvalues and ∆w is clearly symmetric
17
and has positive eigenvalues. By Theorem 3.2 the largest eigenvalue of A is
at most ϑ(G).
On the other hand, let A = (aij ) be a matrix as defined above and let λ be
its largest eigenvalue. Then by Theorem 3.2 λI − A is positive semidefinite
and hence there exists vectors x1 , . . . , xn such that
(λI − A)ij = λIij − Aij = λδij − aij = xTi xj with
1 if i = j
δij = (3.4)
0 if i 6= j
1
λ= .
(cT ui )2
Therefore, ϑ(G) ≤ λ by definition of ϑ(G).
A new definition of the Lovász Number just considered is:
ϑ(G) = min λ such that λ is the largest eigenvalue of a matrix A satisfying (3.1).
Remark. The above proof shows that among the optimal orthonormal repre-
sentations there is one such that
1 1
ϑ(G) = = . . . = . (3.5)
(cT u1 )2 (cT un )2
18
The next theorem gives a good characterization of the value of ϑ(G).
Theorem 3.4. Let G be a graph on vertices {1, . . . , n} and let B = (bij )ni,j=1
range over all positive semidefinite symmetric matrices such that
bij = 0 (3.6)
for every pair (i, j) of distinct adjacent vertices and
T r(B) = 1. (3.7)
Then ϑ(G) = maxB Tr(BJ).
19
Now let C = (ϑ(G)I − A)B and (3.9) follows.
The last inequality holds because if C positive semidefinite then
∀x xT Cx ≥ 0 since C = M T M . Then xT Cx = xT M T M x =k M x k2 ≥ 0.
By the derivation above ϑ(G) − Tr(BJ) ≥ 0 and so ϑ(G) ≥ max Tr(BJ).
Second we prove ϑ(G) ≤ max Tr(BJ).
Construct a matrix B which makes that the previous inequality becomes a
equality, i.e. Tr((ϑ(G)I − A)B) = 0.
Let (i1 , j1 ), . . . , (im , jm )(ik < jk ) be the edges of G. Consider the (m + 1)-
dimensional vectors
X T
ĥ = hi1 hj1 , . . . , him hjm , ( hi )2
where h = (h1 , . . . , hn ) ∈ Rn , k h k= 1 ranges through all unit vectors and
z = (0, 0, . . . , 0, ϑ(G))T .
Claim: z is a convex hull of vectors ĥ.
By definition [6],
Xt t
X
convex hull (ĥ) = { λi fi | f1 , . . . , ft ∈ ĥ, λ1 , . . . , λt ∈ R+ , λi = 1} ⊆ Rm+1 .
i=1 i=1
Suppose the claim is not true. The vectors ĥ form a compact set, since
K = {h ∈ Rn |k h k= 1} is compact and the image of a compact space
under a continuous function is compact. Hence there exists a hyperplane
which seperate z from all the ĥ. Therefore there is a vector a and a real
number α such that aT ĥ ≤ α for all unitvectors h but aT z > α.
Set
a = (a1 , . . . , am , y)T .
Then in particular, aT ĥ ≤ α, for h = (1, 0, . . . , 0), so ĥ = (0, 0, . . . , 1) and
aT ĥ = y ≤ α. On the other hand, aT z > α implies ϑ(G)y > α.
Hence y > 0 and α > 0. To see this we have to show 0 < y ≤ α < ϑ(G)y,
i.e. (ϑ(G) − 1)y > 0.
ϑ(G) − 1 ≥ 0 because ϑ(G) ≥ 1 by definition of ϑ(G). If ϑ(G)y ≤ 0, then
y ≤ 0. Then ϑ(G)y ≤ y and (ϑ(G) − 1)y ≤ 0. Which is a contradiction, so
y > 0 and α > 0.
Up to rescaling we can suppose that y = 1 and hence α < ϑ(G).
Now define 1
a +1
2 k
if {i, j} = {ik , jk }
aij =
1 otherwise,
then aT ĥ ≤ α can be written as
Xn X n
aij hi hj ≤ α.
i=1 j=1
20
Since the P largest eigenvalue (l.e.v.) of A = (aij ) is max{xT Ax || x |= 1} and
T n Pn
x Ax = i=1 j=1 aij hi hj ≤ α it follows that the l.e.v.(aij ) is at most α.
First we show l.e.v.(A) = max{xT Ax || x |= 1}.
Let Ax = λx, we may assume | x |= 1, thus xT Ax = λxT x = λ | x |2 = λ.
Then, max{xT Ax || x |= 1} ≥ l.e.v.(A).
Since A is symmetric, there are eigenvectors e1 , . . . , en such that
eTi ej = δi,j , as in (3.4).
Agian assume | x |= 1 and x = ni=1 αi ei .
P
Then X X
1 =| x |2 = xT x = αi αj eTi ej = αi2 ,
i,j i
and
X X X
xT Ax = αi αj eTi Aej = αi αj eTi λj ej = λj αi αj eTi ej
i,j i,j i,j
X X X
= λi αi2 ≤ l.e.v.(A)αi2 = l.e.v.(A) αi2 = l.e.v.(A).
i i i
α1 + . . . + αN = 1
α1 ĥ1 + . . . + αN ĥN = z.
Set
hp = (hp,1 , . . . , hp,n )T
N
X
bij = αp hp,i hp,j
p=1
B = (bij ).
21
Also,
X N
X X N
X X X
TrBJ = bij = αp hp,i hp,j = αp ( hp,i )( hp,j )
ij p=1 ij p=1 i j
N
X
= αp (ĥp )m+1 = z m+1 = ϑ(G).
p=1
22
Indeed, ϑ(G) = (cT 1ui )2 ∀i by definition of ϑ(G) and (3.5).
Also by Lemma 3.5,
n n
X
2 2 1 X T 2
1=c d ≥ (uTi c)2 (v Ti d)2 = (v d) .
i=1
ϑ(G) i=1 i
The next theorem gives another formula for the value of the upper bound
ϑ(G). It uses the orthonormal representation of the complementary graph
G.
23
Then the vectors v i form an orthonormal representation of G by (3.10) and
(3.6). Moreover, by using the Cauchy-Schwarz inequality,
√ √
|< x, y >|≤ < x, x > < y, y >, we have
n n
! n !
X X X
(dT v i )2 = w2i (dT v i )2
i=1 i=1 i=1
n
!2 n
!2
X X
T T
≥ | wi | (d v i ) = (d wi )
i=1 i=1
n
!2 n
!2
X X
T
= d wi = wi = ϑ(G).
i=1 i=1
So, the inequality ≤ holds also and this proves the theorem.
Remark. If there is equality in the Caucht-Schwarz inequality it also follows
that
(dT v i )2 = ϑ(G)w2i = ϑ(G)bii ∀i. (3.11)
Since we derived some formulas for ϑ(G) in the previous we can now derive
some consequences, for example if G is vertex-transitive or regular.
24
Moreover, c ◦ d is a unit vector. So,
n X
m
X 2
ϑ(G H) ≥ (v i ◦ wj )T (c ◦ d)
i=1 j=1
Xn X m
= (v Ti c)2 (wTj d)2
i=1 j=1
Xn Xm
= (v Ti c)2 (wTj d)2 = ϑ(G)ϑ(H).
i=1 j=1
ϑ(G)ϑ(G) = n.
Proof. By Corollary 3.7 we have seen ϑ(G)ϑ(G) ≥ n. We will prove the op-
posite inequality.
Let Γ be the vertex-transitive automorphism group of G. Consider the el-
ements of Γ as n × n permutation matrices. Let B = (bij ) be a matrix
satisfying (3.6) and (3.7) such that TrBJ = ϑ(G). Consider
!
1 X
B = (bij ) = P −1 BP .
| Γ | P ∈Γ
25
! ! !!
1 X 1 X
TrBJ = Tr P −1 BP J = Tr P −1 BJP
|Γ| P ∈Γ
|Γ| P ∈Γ
!! !
1 X 1 X −1
= Tr P −1 BJ = P ϑ(G)
|Γ| P ∈Γ
| Γ | P ∈Γ
1
= | Γ | ϑ(G) = ϑ(G).
|Γ|
We used that P J = JP = J. This holds because P is a permutation matrix
which satisfies the condition that in each row and each column there is exactly
one entry equal to 1 and the other entries equal 0.
Beside this, B is symmetric and positive semidefinite, since B is positive
semidefinite and P −1 BP = P T BP since P is a permutation matrix.
Let {P P0 | P ∈ Γ} = Γ.
−1 −1 1 −1 1 −1 −1
P P
Then P0 BP0 = P0 |Γ| P ∈Γ P BP P 0 = |Γ| P ∈Γ P0 P BP P0 = B.
Therefore P −1 BP = B for all P ∈ Γ.
Since Γ is vertex-transitive and TrB = 1, bii = n1 for all i. Constructing
the orthonormal representation (v 1 , . . . , v n ) of G and unit vector d as in the
proof of Theorem 3.8, we have
ϑ(G)
(dT v i )2 = ϑ(G)bii =
n
by (3.11). From the definition of ϑ(G) we have,
1 n
ϑ(G) ≤ max =
1≤i≤n (v T d)2 ϑ(G)
i
and hence,
ϑ(G)ϑ(G) ≤ n.
Θ(G)Θ(G) ≤ n.
According to Theorem 3.1, where we proved that Θ(G) ≤ ϑ(G), the Corollary
follows.
Remark. Theorem 3.11 and Corollary 3.12 do not hold for all graphs be-
cause there are graphs with α(G)α(G) > n, for example a star. In this case
Theorem 3.11 an Corollary 3.12 contradict with Lemma 1.14.
26
Definition 3.13. (Regular Graph) A graph G is a regular graph if ∆(G) =
δ(G). Here ∆(G) is the maximum degree and δ(G) is the minimum degree.
The degree of a vertex v in G is the number of edges incident to v [7].
Theorem 3.14. Let G be a regular graph and let λ1 ≥ λ2 ≥ . . . ≥ λn be the
eigenvalues of the adjacency matrix A of G. Then
−nλn
ϑ(G) ≤ .
λ1 − λn
Equality holds if the automorphism group of G is transitive on the edges.
Proof. Consider the matrix J − xA where x will be chosen later. The matrix
satisfies the condition that (J −xA)ij = 1 if i = j or if i and j are nonadjacent
because aij = 0 in this case. We use Theorem 3.3 for J − xA and hence its
largest eigenvalue is at least ϑ(G). Let v i be the eigenvector of A belonging
to λi . Then since G is regular, say k−regular, we have Aj = kj. Therefore
v 1 = j is an eigenvector with eigenvalue k. Since A is symmetric we can
choose the eigenvectors all perpendicular to each other.
Let λ1 6= λ2 , Av i = λi v i and Av j = λj v j , then v Ti v j = 0. This is because
v Tj Av i = v Ti Av j , v Tj Av i = λi v Tj v i and v Ti Av j = λj v Tj v i . This is only
possible if v Ti v j = 0. Therefore j, v 2 , . . . , v n are also eigenvectors of J.
If it is true that v Ti v j = 0 if i 6= j. Then also
nv i if i = 1
Jv i =
0 if i 6= 1
−xλn = λ−nλ n
1 −λn
.
Therefore we can conclude
−nλn
ϑ(G) ≤ .
λ1 − λn
We assume the automorphism group Γ of G is transitive on the edges.
Let C = (cij ) be a symmetric matrix such that cij = 1 if i = j or if i and
27
j are nonadjacent and having largest eigenvalue ϑ(G). As in the proof of
Theorem 3.11 consider
!
1 X
C = (cij ) = P −1 CP .
| Γ | P ∈Γ
n cos( πn )
ϑ(Cn ) = .
1 + cos( πn )
l=1
2πikj
2πik −2πik
2πk
=e n e n +e n = 2 cos (fk )j .
n
2πk
Therefore, the eigenvalues
of A are {2 cos n
| k = 1, . . . n}. Hence λ1 = 2
(n−1)π
= 2 cos π − n = −2 cos( πn ). Since Cn is transitive
π
and λn = 2 cos n
π π
−n·−2 cos( n ) n cos( n )
on the edges and by Theorem 3.14 ϑ(Cn ) = π
2+2 cos( n )
= 1+cos( nπ .
)
7 cos( π7 )
Example 3.16. For the cycle graph C7 we have ϑ(C7 ) = 1+cos( π7 )
and we
know α(C7 ) = 3, thus
7
α(C7 ) = 3 ≤ Θ(C7 ) ≤ ϑ(C7 ) ≈ 3, 3 ≤ α∗ (C7 ) = .
2
In this example we see that ϑ(C7 ) a smaller upper bound is for Θ(C7 ) than
α∗ (C7 ).
28
We arrived at one of the most important theorems. This theorem says that
ϑ(G) really is a smaller or equal upper bound for the Shannon capacity than
the one Shannon found.
Theorem 3.17. ϑ(G) ≤ α∗ (G)
Proof. Use Theorem 3.4 which says ϑ(G) = maxB TrBJ if B satisfies (3.6)
and (3.7). Let (u1 , . . . , un ) be an orthonormal representation of G and c a
unit vector such that n
X
ϑ(G) = (cT ui )2 .
i=1
tion of α∗ (G)
X n
ϑ(G) = (cT ui )2 ≤ α∗ (G).
i=1
29
We have dk=1 (eTk ui )2 = 1 because V T V = 1 which says
P
Pd 2
Pd T 2
Pd T 2
j=1 (V j ) = j=1 (V ej ) = j=1 (ej V ) = 1.
In terms of the definition of ϑ(G),
1 1 √
= = d.
(ui ◦ ui )T b bT (ui ◦ ui )
We take at both sides the square and since we just take an orthonormal
representation of G and an unit vector
we get the inequality.
Therefore, ϑ(G) = minu1 ,...,un minc max1≤i≤n (cT 1ui )2 ≤ d.
30
Chapter 4
In this last chapter we discuss an article by Haemers [3] where he solved the
problems Lovász stated at the end of his article. The problems are:
1. Is ϑ(G) = Θ(G)?
2. Is Θ(G H) = Θ(G)Θ(H)?
31
Let A be a symmetric n × n matrix over a field with all diagonal entries
equal to one. Let G(A) be the graph with vertex set {1, . . . , n}, two vertices
being adjacent if and only if (A)ij 6= 0. Let A⊗k denote the kth tensor prod-
uct of A with itself. All other notation is the same as in [5]. First we need a
theorem.
Since,
32
So, B̃ 2 = 6B̃ + 10(J − I − B̃) + 15I = −4B̃ + 5I + 10J.
Write
T
0 j15 012 " #
0 aT
B̃ = j15 B1 N=
a C
012 N T B2
then
" #2 " # " # " # " #
2
0 aT aT a aT C 0 aT 1 0 1 T
j27
B̃ = = = −4 +5 +10
a C Ca aat + C 2 a C 0 I27 j27 J27
Define
" # " # " #! " #
B1 J15,12 − N −I15 0 0 J15,12 −I15 0
B= = C−
J12,15 − N T B2 0 I12 J12,15 0 0 I12
Then
33
" # " #!2 " #
2
−I15 0 0 J15,12 −I15 0
B = C−
0 I12 J12,15 0 I12 0
" # " #
−I15 0 T
−I15 0
T 2
= C − (j27 − a)a − a(j27 − a)
0 I12 0 I12
" #
−I15 0
= C 2 − C(j27 − a)aT − Ca(j27 − a)T − (j27 − a)aT C − a(j27 − a)T C
0 I12
" #!" #
12J15 0 −I15 0
+
0 15J12 0 I12
" #
−I15 0
= C 2 − Cj27 aT − Caj27
T
+ 2CaaT − j27 aT C − aj27
T
C + 2aaT C
0 I12
" #!" #
12J15 0 −I15 0
+
0 15J12 0 I12
" #
−I15 0
= − aaT − 4C + 5I27 + 10J27 − (15j27 − a)aT − (−4a + 10j27 )j27
T
0 I12
+ 2(−4a + 10j27 )aT − j27 (−4aT + 10j27 T
) − a(15j27T
− aT ) + 2a(−4aT + 10j27 T
)
" #!" #
12J15 0 −I15 0
+
0 15J12 0 I12
" # " #! " #
−I15 0 5J15 −J15,12 −I15 0
= −4C + 5I27 +
0 I12 −J12,15 5J12 0 I12
" # " # " #! " #
−I15 0 0 J15,12 J15 −J15,12 −I15 0
= −4C + 5I27 + 4 +5
0 I12 J12,15 0 −J12,15 J12 0 I12
" #" #" #
−I15 0 J15 −J15,12 −I15 0
= −4B + 5I27 + 5 .
0 I12 −J12,15 J12 0 I12
It follows that
(B − I)(B + 5I) = 5J
34
And from
" # " #! " #
−I15 0 0 J15,12 −I15 0
Bj27 = C− j27
0 I12 J12,15 0 0
I12
" # " #
−I15 0 −I 15 0
C − (j27 − a)aT − a(j27 − a)T
= j27
0 I12 0 I12
" #
−I15 0
Cj27 − 2Ca − j27 aT j27 + 2j27 aT a + aaT (j27 − 2a) − a(j27 − a)T (j27 − 2a)
=
0 I12
" #
−I15 0
= (15j27 − a − 2(−4a + 10j27 ) − 15j27 + 30j27 − 27a)
0 I12
" #
−I15 0
= (10j27 − 20a) = 10j27
0 I12
follows
BJ = 10J.
This implies that B has an all-one eigenvector with eigenvalue 10 and from
(B − I)(B + 5I) = 5J follows that the other eigenvalues are 1 and −5. Let
a, b, c be the multiplicities of the eigenvalues respectively 10, 1 and −5, then
a = 1 and 10a + 1b − 5c = TrB = 0. This implies b = 20 and c = 6.
Let A = I − B and G = G(A). The complement of G is the Schläfli graph.
The eigenvalues of A are 1 − 10 = −9, 1 − 1 = 0 and 1 − (−5) = 6 with
multiplicities 1, 20 and 6. So A has eigenvalue 0 with multiplicity 20, hence
rank(A) = 27 − 20 = 7. Thus by the previous theorem Θ(G) ≤ 7.
Define A(G) = J − I − B and let v 1 , . . . , v n be an orthogonal system of
eigenvectors. So if i ≥ 2, then j T v i = 0.
Hence (J − I − B)v i = Jv i − v i − λi v i = −v i − λi v i = (−1 − λi )v i if i ≥ 2
and (J − I − B)v 1 = 27v 1 − v 1 − λ1 v 1 = (27 − 1 − λ1 )v 1 .
Therefore A(G) = J − I − B has eigenvalues 27 − 1 − 10 = 16, −1 − 1 = −2
and −1 − (−5) = 4. Applying Theorem 3.14 to G and its complement G
gives
−27 · −5 −27 · −2
ϑ(G) ≤ = 9 and ϑ(G) ≤ = 3.
10 − (−5) 16 − (−2)
By Corollary 3.7 we have ϑ(G)ϑ(G) ≥ 27. Hence ϑ(G) = 9 and ϑ(G) = 3.
Therefore we have Θ(G) 6= ϑ(G), Θ(G)Θ(G) ≤ 21 < 27 and Θ(G G) ≥ 27
since for every graph we have α(GG) ≥| V (G) | because {(v, v) | v ∈ V (G)}
is an independent set in G G and we know α(G G) ≤ Θ(G G). Hence
also Θ(G G) ≥| V (G) |= 27.
Whence we can conclude that all three problems are answered in the negative.
35
Chapter 5
Conclusion
36
Popular summary
37
Figure 5.2: C5 C5
For example we consider the graph C52 = C5 C5 as in Figure 5.2. For this
graph we have that two vertices (u1 , u2 ) 6= (v1 , v2 ) are adjacent in C5 C5 if
and only if ui = vi or ui vi ∈ E(C5 ) for i = 1, 2. In Figure 5.2, the red ver-
tices form the maximum independent set of this graph. The 2-letter messages
which can not be confused are v1 v1 , v2 v3 , v3 v5 , v4 v2 and v5 v4 , hence α(C52 ) = 5.
38
Figure 5.3: Clique in C5
39
Concluding, it would be valuable to know the Shannon capacity for a certain
graph, but because determining the Shannon capacity is a very difficult prob-
lem, we say that this theoretical parameter is not very useful at this moment.
On the other hand, because of the connections with some central questions in
graph theory, further research in this field will maybe contribute to a better
approach and resulting in more applications of the Shannon capacity.
40
Bibliography
[1] Aigner, M. & Ziegler, G.M. (2010) Proofs from THE BOOK, 4th edition,
Springer-Verlag Berlin Heidelberg 241-250.
[2] Codenotti, B., Gerace, I. & Resta, G. (2003) Some remarks on the Shan-
non capacity of odd cycles, Ars Combinatoria, 66, 243-257.
[4] Lovász, L. (1972) Normal hypergraphs and the perfect graph conjecture,
Discrete Mathematics, 2, no 3, 253-267.
[7] West, D.B. (2001) Introduction to Graph Theory, 2th edition, Prentice
Hall.
41