3
SOURCE CODING
Communication systems are designed to transmit the information generated by
a source to some destination. Information sources may take a variety of
different forms. For example, in radio broadcasting, the source is generally an
audio source (voice or music). In TV broadcasting, the information source is a
video source whose output is a moving image. The outputs of these sources are
analog signals and, hence, the sources are called analog sources. In contrast,
computers and storage devices, such as magnetic or optical disks, produce
discrete outputs (usually binary or ASCII characters) and, hence, they are
called discrete sources.
Whether a source is analog or discrete, a digital communication system is
designed to transmit information in digital form. Consequently, the output of
the source must be converted to a format that can be transmitted digitally. This
conversion of the source output to a digital form is generally performed by the
source encoder, whose output may be assumed to be a sequence of binary
digits.
In this chapter, we treat source encoding based on mathematical models of
information sources and a quantitative measure of the information emitted by
a source. We consider the encoding of discrete sources first and then we discuss
the encoding of analog sources. We begin by developing mathematical models
for information sources.
3-1 MATHEMATICAL MODELS FOR INFORMATION
SOURCES
Any information source produces an output that is random, ie., the source
Output is characterized in statistical terms. Otherwise, if the source outputCHAPTER. SOURCE copInG 83
were known exactly, there would be no need to transmit it, In this section, we
consider both discrete and analog information sources, and we postulate
‘mathematical models for each type of source.
The simplest type of discrete source is one that emits a sequence of letters
selected from a finite alphabet. For example, a binary source emits a binary
sequence of the form 100101110..., where the alphabet consists of the two
letters {0, 1}. More generally, a discrete information source with an alphabet of
L possible letters, say {x,,t2,...,Xch emits a sequence of letters selected
from the alphabet.
To construct a mathematical model for a discrete source, we assume that
each letter in the alphabet {xy,.x2,...,..t;} has a given probability p, of
occurrence. That is.
Pea P(X =n), 1SK 1 and for all shifts m. In other
words, the joint probabilities for any arbitrary length sequence of source
outputs are invariant under a shift in the time origin.
An analog source has an output waveform x(t) that is a sample function of a
stochastic process X(1). We assume that X(t) is a stationary stochastic process
with autocorrelation function 4,,(t) and power spectral density ®,,(f). When
X(J) is a bandlimited stochastic process, i.c., P(f)=0 for [f|>W, the
sampling theorem may be used to represent X(t) as
oan anw(1-
xo- 3 x(5) ee iw)|
2
G11)
where {X(n/2W)} denote the samples of the process X(t) taken at the
sampling (Nyquist) rate of f, = 2W samples/s. Thus, by applying the sampling
theorem, we may convert the output of an analog source into an equivalent84° piarrat conmumicarions
discrete-time source. Then, the source output is characterized statistically by
the joint pdf p(x1,2,..., 4m) for all m>1, where X, = X(n/2W), 1