Professional Documents
Culture Documents
ITC Unit 1
ITC Unit 1
Lit P, < P, information increases with uncertainty), Further-
we with epOse & Source produces two successive and independent messages, x; and
% with joint probability Px;) = P,P; then
$0 the total information equals the sum of the individual message contributions,
Shannon's information measure log, (1/P,) is the only function that satisfies all of
the properties in Eqs, (2) and (3),161 Information Measure and Souree Encoding
Specifying the logarithmic base b determines the unit of information. The stan-
dard convention of information thery takes = 2 andthe corresponding units the
bit, @ name coined by J. W. Tukey asthe contraction for “binary digit.” Equation (1)
thus becomes J, = log; (1/P.) bits. Ths convention normalizes information measure
relative tothe most elementary soure,a source tha selets from just to equiprob-
able messages. Far if Px;) = P(x) /2, then fy = f = logy 2 = | bit In other
‘words, 1 biti the amount of information needed to choose between two eqully
likely altematives
Binary digits enter the picture simply because any two things can be represented
by the two binary digits. However, you must carefully distinguish information bits
from binary digits per se—especially since a binary digit may convey more a Iss
‘han one bit of information, depending upon its probability. To prevent possible mis-
interpretation, the abbreviation biitsis sometimes usd for binary digits as message
or code clement
When necessary, you can convert base-2 natural or common logarithms via
Tog: v = BY. one w
In ~ Tog?
IEP, = 1/10, for instance, then J, = (Iogio 10)/0.301 = 3.32 bis
Entropy and Information Rate
Now consider an information source that emits a sequence of symbols selected from
an alphabet of M different symbols, ic., an M-ary alphabet, Let X denote the entire
sot of symbols x, --..rye We can treat each symbol x) as a message that occurs
with probability P; and conveys the self-information J, The set of symbol probabili-
ties, of course, must satisfy
5)
‘We'll assume that the source is stationary so the probabilities remain constant over
time, We'll also assume that successive symbols are statistically independent and
ome from the source at an average rate of r symbols per second, These properties
define the model ofa diserete memoryless source.
(6
Yai Shannon borrowed the name, and notation H
from a similar expression in statistical mechanics. Subsequently, various physical
and philosophical arguments have been put forth relating thermodynamic entropy 10
‘communication entropy (or comentrop)’).
But we'll interpret Eq, (6) from the more pragmatic observation that when the
souree emits a sequence of n >> | symbols, the total information to be transferred is72
SHAFTER 16 © Infoemation and Detection Theory
Fae duration ofthis sequence is aber ‘yr. The information must therefore be trans.
eed atthe average rate yn) 20) bits pe second. Fooly, we define
the source information rate
rH(X) bits/sec a
0SH(X) = log M 8)
The lower bound here Fol Foods to no uncertainty oF freedom of choice, which
Sccurs when one symbol hag Probability P= 1 while P,
jource almost always emits the same Symbol. The upper bound corresponds to ma
imum uncertainty of freedom of choice, which occurs when P,
the symbols are equally likely,
__ Toillustrate the variation of H(X) between these extremes, take the special but
portant case of a binary source (Af = 2) with
Pi=p
MO) = 10) © og, 2 + (1 ~ p) toy
(9)
“Pp
in which we've introduced the “horseshoe” function Q(p). The plot of O(p) in
Fig” 16.1-1 displays a rather broad maximum centered at p = 1 ~ p = 1/2 where
FAX) = logs 2 = 1 bit/symbol; HUY) then decreases monotonically to zero ag
P>lorl=ps],
Proving the lower bound in Eq, (8) with arbitrary Mis easily done once you note
that v logs (I/v) —+0 as v > 0. The proof of te Upper bound H(X) < log, M
involves a few more steps but deserves the effort
First, we introduce another set of probabilities Q,,0,..., Qy, and replace
low: (I/F;) in Eq. (6) with log; (Q,/P). Conversion from base-2 to natural loga-
rithms gives the quantity
2
aaah
Where it's understood that all sums range from i = 1 to M, Second, we invoke the
inequality in v = v ~ 1, which becomes an equality only if v = 1, as seen in
Fig. 16.1-2. Thus, letting v = P/Q, and using Eq, (5),16.1 Information Measure and Source Encoding
2p)
»
2 Os 10
Figure 16.1-1 Binary entropy os © function of probability
ponde p85
Third, we impose the condition ®
u
DOA! (10)
80 it follows that
« Qi
2 logs pS 1- (108)
Finally, taking Q; = 1/M we have
L 1
DP los Bay = DP lows 5 ~ DSP; log M = HX) — log, Ms 0
thereby confirming that H(X) < log, M. The equality holds only in the equally
likely case P, = 1/M sow = Q,/P, = 1/Mp, = | for all
Source Entropy and Signaling Rote EXAMPLE 16.1-
Suppose 2 source emits r = 2000 symbols/se selected from an alphabet of size
‘M = 4 with symbol probabilities and self information listed in Table 16.1-1. Equa
tion (6) gives the source entropy
HX) =} 1 +EX24$X3 +45 X 3 = 175 bits/symbol
Which falls somewhat below the maximum value log, M = 2. The information rate is
R = 2000 X 1.75 = 3500 bits/see
and appropriate coding should make it possible to transmit the source information at
the binary signaling rate r, 2 3500 binits/sec.774
CHAPTER 16 @ Information and Detection Theory
AV
Figure 16.1-2 Plott» — 1 and Info).
Table 16.1-1 Symbol probabilities
cond self information
for Example 16.1-1
fa e 4
4 1/2 1
B 4 2
c. 18 3
D 8 3
ee ae
ee
FXERCISE 16.1-1 Suppose a source has M = 3 symbols with probabilities P, = p and P, = P,,
Show that H(X) = () + 1 ~ p. Then evaluate H(X)|,.. and the corresponding
value of p.
Coding for a Discrete Memoryless Channel
[hen a discrete memoryless source produces M equally likely symbols, so
Af = logs M, all symbols convey the same amount of information sod efficient
takes account ofthe variable amount of information per symbol. Here, we'll inven,
gate source encoding with a binary encoder, Equivalent results for nonbinary encod.
ing just require changing the logarithmic base.
The binary encoder in Fig. 16.13 converts incoming source symbols to code-
words consisting of binary digits produced at some fixed rate r, Viewed from itsins
16.1 Information Measure and Source Encoding
lp)"
Discrete R=AKX)
memoryless. [|
souree
Binary
encoder
Figure 161-3 Source encoding
entropy (p) and information
not generate additional infor-
that the code is uniquely
n rates, WE
‘output, the encoder looks like a binary source with
rate r,0(p) $ r; log, 2 = ry. Coding obviously does
mation, nor does it destroy information providing
decipherable, Thus, equating the encoder’s input and output informatior
conclude that R = rH(X) = rp) S ry or ro/r = HX).
The quantity = ry/r is an important parameter called the
length, Physically, corresponds to the average number of binary digits per source
symbol. Mathematically, we write the statistical average
Ne San, a
a
average code
‘where NV; represents the length of the codeword for the ith symbol. _
Shannon’s source coding theorem states that the minimum value of Nis
bounded by
HX) SN