You are on page 1of 70
3 SOURCE CODING Communication systems are designed to transmit the information generated by a source to some destination. Information sources may take a variety of different forms. For example, in radio broadcasting, the source is generally an audio source (voice or music). In TV broadcasting, the information source is a video source whose output is a moving image. The outputs of these sources are analog signals and, hence, the sources are called analog sources. In contrast, computers and storage devices, such as magnetic or optical disks, produce discrete outputs (usually binary or ASCII characters) and, hence, they are called discrete sources. Whether a source is analog or discrete, a digital communication system is designed to transmit information in digital form. Consequently, the output of the source must be converted to a format that can be transmitted digitally. This conversion of the source output to a digital form is generally performed by the source encoder, whose output may be assumed to be a sequence of binary digits. In this chapter, we treat source encoding based on mathematical models of information sources and a quantitative measure of the information emitted by a source. We consider the encoding of discrete sources first and then we discuss the encoding of analog sources. We begin by developing mathematical models for information sources. 3-1 MATHEMATICAL MODELS FOR INFORMATION SOURCES Any information source produces an output that is random, ie., the source Output is characterized in statistical terms. Otherwise, if the source output CHAPTER. SOURCE copInG 83 were known exactly, there would be no need to transmit it, In this section, we consider both discrete and analog information sources, and we postulate ‘mathematical models for each type of source. The simplest type of discrete source is one that emits a sequence of letters selected from a finite alphabet. For example, a binary source emits a binary sequence of the form 100101110..., where the alphabet consists of the two letters {0, 1}. More generally, a discrete information source with an alphabet of L possible letters, say {x,,t2,...,Xch emits a sequence of letters selected from the alphabet. To construct a mathematical model for a discrete source, we assume that each letter in the alphabet {xy,.x2,...,..t;} has a given probability p, of occurrence. That is. Pea P(X =n), 1SK 1 and for all shifts m. In other words, the joint probabilities for any arbitrary length sequence of source outputs are invariant under a shift in the time origin. An analog source has an output waveform x(t) that is a sample function of a stochastic process X(1). We assume that X(t) is a stationary stochastic process with autocorrelation function 4,,(t) and power spectral density ®,,(f). When X(J) is a bandlimited stochastic process, i.c., P(f)=0 for [f|>W, the sampling theorem may be used to represent X(t) as oan anw(1- xo- 3 x(5) ee iw)| 2 G11) where {X(n/2W)} denote the samples of the process X(t) taken at the sampling (Nyquist) rate of f, = 2W samples/s. Thus, by applying the sampling theorem, we may convert the output of an analog source into an equivalent 84° piarrat conmumicarions discrete-time source. Then, the source output is characterized statistically by the joint pdf p(x1,2,..., 4m) for all m>1, where X, = X(n/2W), 1

You might also like