Professional Documents
Culture Documents
Channel Channel
2 3
Sent Received
A (discrete ) channel is a system consisting of an input alphabet
messages messages X and output alphabet Y and a probability transition matrix
symbols p(y|x) that expresses the probability of observing the output
symbol y given that we send the symbol x
Source Channel Channel Source
source Channel decoding decoding receiver
coding coding
Examples of channels:
4 5
1
Communication Channel Noiseless binary channel
Noiseless binary channel
Channel
Channel
0 0
input x p(y|x) output y
Transition probabilities
1 1
Memoryless:
- output only on input Transition Matrix
- input and output alphabet finite
0 1
p(y | x) = 0 1 0
1 0 1
6 7
BSC Channel
0 01
1-p
p 1-p
Error Source
0 0
e BSC Channel
1 10
xi yi = xi e
p
p 1-p
+
Output 1 1
Input p
1-p 0 0
1-p
1 1
p
BSC Channel
8 9
Symmetric Channel
(Noisy channel)
Channel
X Y
y1 y2 y3
x1 0.3 0.2 0.5
p(y | x) = x2 0.5 0.3 0.2
x3 0.2 0.5 0.3
10 11
2
Mutual Information (MI) Mutual Information (MI)
Channel Relationship between Entropy and Mutual Information
X p(y|x) Y I(X,Y)=H(Y)-H(Y|X)
Proof:
p( y | x)
MI is a measure of the amount of information that one I ( X , Y ) p ( x, y ) log 2 ( )
random variable contains about another random variable. x y p( y )
It is the reduction in the uncertainty of one random
I ( X , Y ) p ( x, y ) log 2 ( p ( y )) p ( x , y ) log 2 ( p ( y | x ))
variable due to the knowledge of the other. x y x y
H(X) H(Y) 14 15
I(X, (Y,Z) )
= I(X, Y) + I(X; Z|Y)
= H(X) - H(X|YZ)
Channel Capacity
= H(X) - H(X|Z) + H(X|Z) - H(X|YZ)
= I(X, Z) + I(X,Y|Z)
now I(X, Z|Y) = 0 (independency)
Thus: I(X, Y) I(X, Z)
16 17
3
Transmission efficiency Channel Capacity
The channel capacity C is the highest rate, in bits per channel
I need on the average H(X) bits/source output to describe the
use, at which information can be sent with arbitrarily low
source symbols X After observing Y, I need H(X|Y) bits/source
probability of error
output.
C = Max I(X , Y)
Channel p(x)
H(X) X Y H(X|Y) Where the maximum is taken over all possible input
distributions p(x)
Note that :
Reduction in description length is called the transmitted
• During data compression, we remove all redundancy in the
Information.
data to form the most compressed version possible.
Transmitted R = I(X, Y) = H(X) - H(X|Y)
• During data transmission, we add redundancy in a controlled
= H(Y) – H(Y|X) from earlier calculations
fashion to combat errors is the channel
We can maximize R by changing the input probabilities. Capacity depends on input probabilities
The maximum is called CAPACITY 18 because the transition probabilites are fixed 19
H(X) = H(p, 1-p)= H(1/2,1/2) = -1/2 log 1/2 - 1/2 log 1/2 = 1 bit = H(Y) - H(P) (Note that H(Y) ≤ 1 )