Professional Documents
Culture Documents
Introduction 7 PDF
Introduction 7 PDF
Spring 2017
Tutorial 3
Problem 1
Processing the output. One is given a communication channel with transition probabilities p(y|x) and
channel capacity C = max p(x) I (X; Y). Someone processes the output by forming Y* = g(Y). He claims that
this will strictly improve the capacity.
(a) Show that he is wrong.
(b) Under what conditions does he not strictly decrease the capacity?
Answer:
a. Since: X Y Y*; then we can argue that the processing Y* = g(Y) can only reduce the
mutual information as in the lecture of the Fano’s lemma. Then we can show exactly that:
I(X;Y*) <= I(X;Y)
b. To have equality, this will only happen when the process g(Y) =Y* =Y; i.e., the process does not
do any operation on Y.
Problem 2
Additive noise channel. Find the channel capacity of the following discrete memory less channel:
Answer:
If a=0, the Z = 0 (strictly). In this case, Y = X and max {I(X;Y)} = max {H(X)} = 1 (bit or 100% per
transmission)
Information Theory COMM 1003
Spring 2017
If a! = 0, +1,-1, the Z =0 (50%) and Z = a (50%)! The output will be X = {0, 1, a, 1+a}. This means that
we can always tell if zero was sent or 1; i.e., if Y = 1, or, 0; then X was (strictly) = 1 or 0 (respectively)
or when Y = 1+a or a; then X was (strictly) = 1 or 0 (respectively)! Here C = 1 (as before).
If a =+1; then Z = 0 or 1. In this case Y \in {0, 1, 1, 2}! Here 1 mean that the signal is going to be
confused. In this case, the 1 can be marked as E and we can only detect 0 as 0 was transmitted or 2 if 1
was transmitted. This is similar to Binary Erasure channels where the capacity = 1- p(a=1) = 1 -1/2 =1/2.
If a = -1; this is similar to the previous case (a = +1).
Problem 3
Cascade of binary symmetric channels (Optical Fiber Channel): Show that a cascade of n
identical independent binary symmetric channels (see figure)
Each with raw error probability p, is equivalent to a single BSC with error probability Pe = 1/2 (1 - (1 -
2p)n) and hence
Show that limnoo I(X0;Xn) = 0 if p lies in the interval ]0; 1[:
HINT: No encoding or decoding takes place at the intermediate between the cascaded blocks
Answer:
The conditional probability distribution p(y|x) for each of the BSCs may be expressed by the
transition probability matrix A, given by
1−𝑝 𝑝
𝐓=[ ]
𝑝 1−𝑝
Since the output of the channel one undertakes the error (p) in channel two and so on; the
transition matrix for n cascaded BSC is=
1−𝑝 𝑝 𝑛
𝐓𝑛 = [ ]
𝑝 1−𝑝
To solve this, we need to compute the transmission matrix SVD (will be given in the formula
sheets)
[U,D,V] = SVD(T)
Then
1 1 1 1 0 1 1 1 1 0
𝐓 = 𝐔𝐃𝐕 𝑇 = [ ][ ] [ ] == 𝐔 [ ]𝐔
√2 1 −1 0 1 − 2𝑝 √2 1 −1 0 1 − 2𝑝
and
1 1 1 1 𝑛 1
0 1 1
𝐓 𝑛 = 𝐔𝐃𝑛 𝐔 = [ ][ ] [ ]
√2 1 −1 0 1 − 2𝑝 √2 1 −1
Information Theory COMM 1003
Spring 2017
1 1 1 1 0 1 1 1
= [ ][ (1 𝑛] [ ]
√2 1 −1 0 − 2𝑝) √2 1 −1
Then:
1 1
(1 + (1 − 2𝑝)𝑛 ) (1 − (1 − 2𝑝)𝑛 ) 𝑝𝑐 𝑝𝑒
𝑛 𝑛
𝐓 = 𝐔𝐃 𝐔 = [2 2 ] = [𝑝
1 1 𝑒 𝑝𝑐 ]
(1 − (1 − 2𝑝)𝑛 ) (1 + (1 − 2𝑝)𝑛 )
2 2
Where 𝑝𝑒 + 𝑝𝑐 = 1
Then
lim 𝐶 = 1 − ℎ(1/2) = 0
𝑛→∞
Problem 4
Total distortion channel (Erasure and Binary error). Consider a channel with binary inputs that
has both erasures and errors. Let the probability of error be 𝜖 and the probability of erasure be
𝛼 so the channel is follows
Answer:
1−𝛼−𝜖 𝛼 𝜖
𝐓=[ ] P(Y|X)
𝜖 𝛼 1−𝛼−𝜖
𝐶 = 𝐻 (𝑌) − 𝐻(𝑌|𝑋)
X / Y 0 𝜖 1
0 1−𝛼−𝜖 𝛼 𝜖
1 𝜖 𝛼 1−𝛼−𝜖
The Capacity is maximized if the input has equal probability (in this case P(x) = ½ )
1 1 1
𝑃(𝑌 = 0) = 2 (1 − 𝛼 − 𝜖) + 2 (𝜖) = 2 (1 − 𝛼) (sum of the first column* P(x) [total probability
theory])
1 1
𝑃(𝑌 = 𝜖) = 2 𝛼 + 2 𝛼 (add the second column* P(x) [total probability theory])
1 1 1
𝑃(𝑌 = 0) = 2 (1 − 𝛼 − 𝜖) + 2 (𝜖) = 2 (1 − 𝛼)
1 1−𝛼
𝐻(𝑌) = ∑ 𝑃(𝑌 = 𝑦𝑖 ) log 2 = −(1 − 𝛼) log 2 − (𝛼) log 2 (𝛼)
𝑃(𝑌 = 𝑦𝑖 ) 2
𝑖
and
𝐻(𝑌|𝑋) = ∑ 𝑃(𝑋 = 𝑥𝑗 )𝐻(𝑌|𝑋 = 𝑥𝑗 ) = 𝑃(𝑋 = 0)𝐻(𝑌|𝑋 = 0) + 𝑃(𝑋 = 1)𝐻(𝑌|𝑋 = 1)
𝑗
1
= [𝐻(𝑌|𝑋 = 1) + 𝐻(𝑌|𝑋 = 0)]
2
= −{(1 − 𝛼 − 𝜖) log 2 (1 − 𝛼 − 𝜖) + 𝜖 log 2 𝜖 + 𝛼 log 2 𝛼}