You are on page 1of 173

Wireless Information Transmission System Lab.

Chapter 5
Optimum Receivers for the Additive
White Gaussian Noise Channel

Institute of Communications Engineering National Sun Yat-sen University


5.1 Optimum Receiver for Signals Corrupted
by Additive White Gaussian Noise
We assume that the transmitter sends digital information by use of
M signals waveforms {sm(t)=1,2,…,M }. Each waveform is
transmitted within the symbol interval of duration T, i.e. 0≤t≤T.
The channel is assumed to corrupt the signal by the addition of
white Gaussian noise, as shown in the following figure:

r (t ) = sm (t ) + n(t ) , 0 ≤ t ≤ T
where n(t) denotes a sample function of AWGN process with
power spectral density Φnn(f )=½N0 W/Hz.

2
5.1 Optimum Receiver for Signals Corrupted
by Additive White Gaussian Noise
Our object is to design a receiver that is optimum in the sense
that it minimizes the probability of making an error.
It is convenient to subdivide the receiver into two parts—the
signal demodulator and the detector.

The function of the signal demodulator is to convert the received


waveform r(t) into an N-dimensional vector r=[r1 r2 ..…rN ] where N is the
dimension of the transmitted signal waveform.
The function of the detector is to decide which of the M possible signal
waveforms was transmitted based on the vector r.
3
5.1 Optimum Receiver for Signals Corrupted
by Additive White Gaussian Noise
Two realizations of the signal demodulator are described in the
next two sections:
One is based on the use of signal correlators.
The second is based on the use of matched filters.
The optimum detector that follows the signal demodulator is
designed to minimize the probability of error.

4
5.1.1 Correlation Demodulator
We describe a correlation demodulation that decomposes
the receiver signal and the noise into N-dimensional
vectors. In other words, the signal and the noise are
expanded into a series of linearly weighted orthonormal
basis functions {fn(t)}.
It is assumed that the N basis function {f n(t)}span the
signal space, so every one of the possible transmitted
signals of the set {sm(t)=1≤m≤M } can be represented as a
linear combination of {f n(t)}.
In case of the noise, the function {f n(t)} do not span the
noise space. However we show below that the noise terms
that fall outside the signal space are irrelevant to the
detection of the signal.
5
5.1.1 Correlation Demodulator
Suppose the receiver signal r(t) is passed through a parallel bank
of N basis functions {f n(t)}, as shown in the following figure:

[ sm (t ) + n (t ) ] f k (t ) dt
T T
∫0
r (t ) f k (t ) dt = ∫
0

⇒ rk = smk + nk , k = 1, 2,..... N
T
smk = ∫ sm (t ) f k (t )dt , k = 1, 2,......N
0
T
nk = ∫ n(t ) f k (t )dt , k = 1, 2,.......N
0

The signal is now represented by the vector sm with


components smk, k=1,2,…N. Their values depend on which of
the M signals was transmitted.
6
5.1.1 Correlation Demodulator
In fact,we can express the receiver signal r(t) in the interval 0 ≤ t
≤ T as: N N
r (t ) = ∑ smk f k (t ) + ∑ nk f k (t ) + n′(t )
k =1 k =1
N
= ∑ rk f k (t ) + n′(t )
k =1
The term n'(t), defined as
N
n' (t ) = n(t ) − ∑ nk f k (t )
k =1

is a zero-mean Gaussian noise process that represents the


difference between original noise process n(t) and the part
corresponding to the projection of n(t) onto the basis functions
{fk(t)}.
7
5.1.1 Correlation Demodulator
We shall show below that n'(t) is irrelevant to the decision as to
which signal was transmitted. Consequently, the decision may
be based entirely on the correlator output signal and noise
components rk=smk+nk, k=1,2,…,N.
The noise components {nk} are Gaussian and mean values are:
E ( nk ) = ∫ E [ n (t ) ] f k (t ) dt = 0 for all n.
T

0
Power spectral density
is Φnn(f)=½N0 W/Hz
and their covariances are:
Autocorrelation
∫ E [ n(t )n(τ )] f
T T
E (nk nm ) = ∫ k (t ) f m (τ )dtdτ Conclusion: The N noise
0 0
Components {nk} are
1 T T
= N 0 ∫ ∫ δ (t − τ ) f k (t ) f m (τ )dtdτ zero-mean uncorrelated
2 0 0 Gaussian random
1 T 1 variables with a common
= N 0 ∫ f k (t ) f m (t )dt = N 0δ mk varianceσn2=½N0.
2 0 2
8
5.1.1 Correlation Demodulator
From the above development, it follows that the correlator
output {rk} conditioned on the mth signal being transmitted are
Gaussian random variables with mean

E (rk ) = E ( smk + nk ) = smk

and equal variance 1


σ = σ = N0
2
r
2
n
2
Since the noise components {nk} are uncorrelated Gaussian
random variables, they are also statistically independent. As a
consequence, the correlator outputs {rk} conditioned on the mth
signal being transmitted are statistically independent Gaussian
variables.

9
5.1.1 Correlation Demodulator
The conditional probability density functions of the random
variables r=[r1 r2 … rN] are:
N
p (r | s m ) = ∏ p (rk | smk ) , m = 1,2,...., M ----(A)
k =1

1 ⎡ (rk − smk ) 2 ⎤
p (rk | smk ) = exp ⎢− ⎥, k = 1,2,....., N ---(B)
πN 0 ⎣ N0 ⎦

By substituting Equation (A) into Equation (B), we obtain the joint


conditional PDFs

1 ⎡ N (rk − smk ) 2 ⎤
p (r|s m ) = exp ⎢ −∑ ⎥ , m = 1, 2,......, M
(π N 0 ) N 2
⎣ k =1 N0 ⎦

10
5.1.1 Correlation Demodulator
The correlator outputs (r1 r2 … rN) are sufficient statistics for
reaching a decision on which of the M signals was transmitted,
i.e., no additional relevant information can be extracted from the
remaining noise process n'(t).
Indeed, n'(t) is uncorrelated with the N correlator outputs {rk}:
⎧⎪ ⎡ N ⎤ ⎫⎪
E [ n′(t )rk ] = E [ n′(t ) ] smk + E [ n′(t )nk ] = E [ n′(t )nk ] = E ⎨ ⎢ n(t ) − ∑ n j f j (t ) ⎥ nk ⎬
⎩⎪ ⎣ j =1 ⎦ ⎭⎪
N
= ∫ E [ n(t )n(τ ) ] f k (τ )dτ − ∑ E (n j nk ) f j (t )
T
nk = ∫ n (t ) f k (t ) dt
T

0 0
j =1
N
1 1
N 0δ ( t − τ ) f k (τ )dτ − ∑ N 0δ jk f j (t )
T
=∫
0 2 j =1 2

1 1
= N 0 f k (t ) − N 0 f k (t ) = 0 Q.E.D.
2 2
11
5.1.1 Correlation Demodulator
Since n' (t) and {rk} are Gaussian and uncorrelated, they are also
statistically independent.
Consequently, n'(t) does not contain any information that is
relevant to the decision as to which signal waveform was
transmitted.
All the relevant information is contained in the correlator outputs
{rk} and, hence, n'(t) can be ignored.

12
5.1.1 Correlation Demodulator

Example 5.1-1.
Consider an M-ary baseband PAM signal set in which the basic
pulse shape g(t) is rectangular as shown in following figure.
The additive noise is a zero-mean white Gaussian noise process.
Let us determine the basis function f(t) and the output of the
correlation-type demodulator.
The energy in the rectangular pulse is

ε =∫
T T
g (t )dt = ∫0 a dt = a T
2 2 2
g 0

13
5.1.1 Correlation Demodulator
Example 5.1-1.(cont.)
Since the PAM signal set has dimension N=1, there is only
one basis function f(t) given as:
1 ⎧1 T (0 ≤ t ≤ T )
f (t ) = g (t ) = ⎨
a 2T ⎩0 (otherwise)
The output of the correlation-type demodulator is:
T 1 T
r = ∫ r (t ) f (t )dt = ∫ r (t )dt
0 0
T
The correlator becomes a simple integrator when f(t) is
rectangular: if we substitute for r(t), we obtain:
r=
1
T ∫{
0
T
[ s m (t ) + n (t ) ] } dt =
1 ⎡ T
T ⎢⎣ ∫0
s m ( t ) dt +
T
∫0 n ( t ) dt ⎤
⎥⎦
= sm + n
14
5.1.1 Correlation Demodulator
Example 5.1-1.(cont.)
The noise term E(n)=0 and:
⎡1 T T ⎤
σ = E ⎡⎣ n ( t ) n ( t ) ⎤⎦ = E ⎢ ∫ ∫ n(t )n(τ )dtdτ ⎥
2
n
⎣T 0 0 ⎦
1 T T N0 T T
= ∫ ∫ E [ n(t )n(τ ) ] dtdτ = ∫ ∫ δ (t − τ )dtdτ
T 0 0 2T 0 0

N0 T 1
=
2T ∫0
1 ⋅ dτ = N 0
2
The probability density function for the sampled output is:
1 ⎡ (r − sm ) 2 ⎤
p(r | sm ) = exp ⎢ − ⎥
πN 0 ⎣ N 0 ⎦

15
5.1.2 Matched-Filter Demodulator
Instead of using a bank of N correlators to generate the variables
{rk}, we may use a bank of N linear filters. To be specific, let us
suppose that the impulse responses of the N filters are:
hk (t ) = f k (T − t ) , 0≤t ≤T

where {fk(t)} are the N basis functions and hk(t)=0 outside of the
interval 0 ≤ t ≤ T.
The outputs of these filters are :
yk (t ) = r ( t ) ∗ hk ( t )
t
= ∫ r (τ )hk (t − τ )dτ
0
t
= ∫ r (τ ) f k (T − t + τ )dτ , k = 1, 2,......, N
0

16
5.1.2 Matched-Filter Demodulator
If we sample the outputs of the filters at t = T, we obtain
T
yk (T ) = ∫ r (τ ) f k (τ )dτ = rk , k = 1, 2,....., N
0
A filter whose impulse response h(t) = s(T–t), where s(t) is
assumed to be confined to the time interval 0 ≤ t ≤ T, is called
matched filter to the signal s(t).
An example of a signal and its matched filter are shown in the
following figure.

17
5.1.2 Matched-Filter Demodulator
The response of h(t) = s(T–t) to the signal s(t) is:
y (t ) = s ( t ) ∗ h ( t ) = ∫ s (τ ) h ( t − τ ) dτ = ∫ s (τ ) s (T − t + τ ) dτ
t t

0 0

which is the time-autocorrelation function of the signal s(t).


Note that the autocorrelation function y(t) is an even function of
t , which attains a peak at t=T.

18
5.1.2 Matched-Filter Demodulator
Matched filter demodulator that generates the observed
variables {rk}

19
5.1.2 Matched-Filter Demodulator
Properties of the matched filter.
If a signal s(t) is corrupted by AWGN, the filter with an
impulse response matched to s(t) maximizes the output
signal-to-noise ratio (SNR).
Proof: let us assume the receiver signal r(t) consists of the
signal s(t) and AWGN n(t) which has zero-mean and
1
Φ nm ( f ) = N 0 W/Hz.
2
Suppose the signal r(t) is passed through a filter with
impulse response h(t), 0≤t≤T, and its output is sampled
at time t=T. The output signal of the filter is:
t
y (t ) = s (t ) ∗ h(t ) = ∫ r (τ )h(t − τ )dτ
0
t t
= ∫ s (τ )h(t − τ )dτ + ∫ n(τ )h(t − τ )dτ
0 0

20
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
At the sampling instant t=T:
T T
y (T ) = ∫ s (τ ) h (T − τ ) dτ + ∫ n (τ ) h (T − τ ) dτ
0 0

= y s (T ) + yn (T )
This problem is to select the filter impulse response that
maximizes the output SNR0 defined as:
y s2 (T )
SNR 0 =
[ ]
E y n2 (T )
E ⎡⎣ yn (T ) ⎤⎦ = ∫ ∫ E [ n(τ )n(t ) ] h(T − τ )h(T − t )dtdτ
T T
2
0 0

1 T T 1 T
= N 0 ∫ ∫ δ (t − τ )h(T − τ )h(T − t )dtdτ = N 0 ∫ h 2 (T − t )dt
2 0 0 2 0

21
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
By substituting for ys (T) and E ⎡⎣ yn2 (T ) ⎤⎦ into SNR0.
τ ' = T −τ
2 2
⎡ s (τ ) h(T − τ ) dτ ⎤
T
⎡ h (τ ') s (T − τ ') dτ '⎤
T

⎢⎣ ∫0 ⎥⎦ ⎢⎣ ∫0 ⎥⎦
SNR 0 = =
1 T 1 T
N 0 ∫ h (T − t ) dt
2
N 0 ∫ h 2 (T − t ) dt
2 0 2 0

Denominator of the SNR depends on the energy in h(t).


The maximum output SNR over h(t) is obtained by
maximizing the numerator subject to the constraint that the
denominator is held constant.

22
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
Cauchy-Schwarz inequality: if g1(t) and g2(t) are finite-
energy signals, then
2
⎡ g (t ) g (t ) dt ⎤ ≤ ∞ g 2 (t ) dt ∞ g 2 (t ) dt

⎢⎣ ∫− ∞ 1 2 ⎥⎦ ∫−∞ 1 ∫−∞ 2
with equality when g1(t)=Cg2(t) for any arbitrary constant
C.
If we set g1(t)=h1(t) and g2(t)=s(T−t), it is clear that the
SNR is maximized when h(t)=Cs(T–t).

23
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
The output (maximum) SNR obtained with the matched
filter is:
2 2
⎡ s (τ ) h(T − τ ) dτ ⎤
T
⎡ s (τ )Cs (T − (T − τ ) ) dτ ⎤
T

⎢ ∫0
⎣ ⎥
⎦ 2 ⎣⎢ ∫0 ⎦⎥
SNR 0 = =
N 0 ∫ C 2 s 2 (T − (T − t ) ) dt
1 T N0 1 T
N 0 ∫ h 2 (T − t ) dt
2 0 2 0

2 T 2 2ε
=
N0 ∫0
s (t ) dt =
N0

Note that the output SNR from the matched filter depends
on the energy of the waveform s(t) but not on the detailed
characteristics of s(t).
24
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
Since h(t)=s(T–t), the Fourier transform of this relationship
is: T
H ( f ) = ∫ s (T − t )e − j 2π ft dt let τ = T - t
0

= ∫ s (τ )e j 2π f τ dτ ⎤ e − j 2π fT = S * ( f )e − j 2π fT
⎡ T

⎢⎣ 0 ⎥⎦
The matched filter has a frequency response that is the
complex conjugate of the transmitted signal spectrum
multiplied by the phase factor e-j2πfT (sampling delay of T).
In other worlds, |H( f )|=|S( f )|, so that the magnitude
response of the matched filter is identical to the transmitted
signal spectrum.
On the other hand, the phase of H( f ) is the negative of the
phase of S( f ).
25
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
If the signal s(t) with spectrum S(f ) is passed through the
matched filter, the filter output has a spectrum
2
Y( f )=|S( f )| e -j2πfT. S ( f ) ⋅ S *
( f ) e − j 2π fT

The output waveform is:


∞ ∞
ys (t ) = ∫ Y ( f )e j 2π ft
df = ∫ S ( f ) e − j 2π fT e j 2π ft df
2
−∞ −∞

By sampling the output of the matched filter at t = T, we obtain


(from Parseval’s relation):
ys (T ) = ∫ S ( f ) df = ∫ s 2 (t )dt = ε
∞ 2 T

−∞ 0
The noise at the output of the matched filter has a power
spectral density (from equation 2.2-27) Φ ( f ) = 1 H ( f ) 2 N
0 0
2
26
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
The total noise power at the output of the matched filter is

Pn = ∫ Φ 0 ( f )df
−∞

= N 0 ∫ H ( f ) df = N 0 ∫ S ( f ) df = ε N 0
1 ∞ 2 1 ∞ 2 1
2 −∞ 2 −∞ 2

The output SNR is simply the ratio of the signal power Ps ,


given by Ps = y s2 (T ) , to the noise power Pn.

SNR 0 =
Ps
=
ε 2
=

Pn 1 ε N N0
0
2
27
5.1.2 Matched-Filter Demodulator
Example 5.1-2:
M=4 biorthogonal signals are constructed from the two
orthogonal signals shown in the following figure for
transmitting information over an AWGN channel. The noise is
assumed to have a zero-mean and power spectral density ½ N0.

28
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
The M=4 biorthogonal signals have dimensions N=2 (shown
in figure a): ⎧⎪ 1
f1 (t ) = ⎨ 2 T (0 ≤ t ≤ T )
2
⎪⎩0 (otherwise)
⎧⎪ 1
f 2 (t ) = ⎨ 2 T ( T ≤ t ≤ T)
2
⎪⎩0 (otherwise)
The impulse responses of the two matched filters are (figure b):
⎧⎪ 1
h1 (t ) = f1 (T − t ) = ⎨ 2 T ( T ≤ t ≤ T)
2
⎪⎩0 (otherwise)
⎧⎪ 1
h2 (t ) = f 2 (T − t ) = ⎨ 2 T (0 ≤ t ≤ T )
2
⎪⎩0 (otherwise)
29
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
If s1(t) is transmitted, the responses of the two matched filters
are as shown in figure c, where the signal amplitude is A.
Since y1(t) and y2(t) are sampled at t=T, we observe that
y1S (T)= 1 A2T , y2S(T)=0.
2
From equation 5.1-27, we have ½A2T=ε and the received
vector is:
r = [r1 r2 ] = ⎢ ε + n1 n2 ⎥
⎡ ⎤
⎣ ⎦
where n1(t)=y1n(T) and n2=y2n(T) are the noise components at
the outputs of the matched filters, given by
T
ykn (T ) = ∫ n(t ) f k (t )dt , k = 1,2
0

30
5.1.2 Matched-Filter Demodulator
Example5.1-2(cont.)
Clearly, E(nk)=E[ykn(T)] and their variance is

σ = E [y (T )] = ∫ ∫ E[n(t )n(τ )] f
T T
2
n
2
kn k (t ) f k (τ )dtdτ
0 0

1 T T
= N 0 ∫ ∫ δ (t − τ ) f k (τ ) f k (t )dtdτ
2 0 0

1 T 1
= N 0 ∫ f k (t )dt = N 0
2

2 0 2
Observe that the SNR0 for the first matched filter is

SNR 0 =
( ε) 2

=

1 N0
N0
2
31
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
This result agrees with our previous result.
We can also note that the four possible outputs of the
two matched filters, corresponding to the four possible
transmitted signals are:

( r1 , r2 ) = ( ε + n1 , n2 ) , ( n1 , ε + n2 ) , ( − ε + n1 , n2 ) and ( n1 , − ε + n2 )

32
5.1.3 The Optimum Detector
Our goal is to design a signal detector that makes a decision
on the transmitted signal in each signal interval based on
the observation of the vector r in each interval such that the
probability of a correct decision is maximized.
We assume that there is no memory in signals transmitted
in successive signal intervals.
We consider a decision rule based on the computation of
the posterior probabilities defined as
P(sm|r)≣P(signal sm was transmitted|r), m=1,2,…,M.
The decision criterion is based on selecting the signal
corresponding to the maximum of the set of posterior
probabilities { P(sm|r)}. This decision criterion is called the
maximum a posterior probability (MAP) criterion.
33
5.1.3 The Optimum Detector
Using Bayes’ rule, the posterior probabilities may be expressed
as p (r | s m ) P (s m )
P (s m | r ) = ---(A)
p (r )
where P(sm) is the a priori probability of the mth signal being
transmitted.
The denominator of (A), which is independent of which signal is
transmitted, may be expressed as
M
p (r) = ∑ p (r | s m ) P (s m )
m =1
Some simplification occurs in the MAP criterion when the M
signal are equally probable a priori, i.e., P(sm)=1/M.
The decision rule based on finding the signal that maximizes
P(sm|r) is equivalent to finding the signal that maximizes P(r|sm).
34
5.1.3 The Optimum Detector
The conditional PDF P(r|sm) or any monotonic function of it is
usually called the likelihood function.
The decision criterion based on the maximum of P(r|sm) over
the M signals is called maximum-likelihood (ML) criterion.
We observe that a detector based on the MAP criterion and one
that is based on the ML criterion make the same decisions as
long as a priori probabilities P(sm) are all equal.
In the case of an AWGN channel, the likelihood function p(r|sm)
is given by:
1 N
⎡ ( rk − smk ) ⎤
2
p (r | s m ) = exp ⎢ − ∑ ⎥ , m = 1, 2,......, M
(π N 0 ) N 2
⎣ k =1 N0 ⎦ constant
1 1 N
ln p (r | s m ) = − N ln(πN 0 ) −
2
∑ k mk
N 0 k =1
( r − s ) 2

35
5.1.3 The Optimum Detector
The maximum of ln p(r|sm) over sm is equivalent to finding the
signal sm that minimizes the Euclidean distance:
N
D (r, s m ) = ∑ (rk − smk ) 2
k =1

We called D(r,sm), m=1,2,…,M, the distance metrics.


Hence, for the AWGN channel, the decision rule based on the
ML criterion reduces to finding the signal sm that is closest in
distance to the receiver signal vector r. We shall refer to this
decision rule as minimum distance detection.

36
5.1.3 The Optimum Detector
Expanding the distance metrics:
N N N
D(r, s m ) = ∑ r − 2∑ rn smn + ∑ smn
n
2 2

n =1 n =1 n =1

= r − 2r ⋅ s m + s m , m = 1, 2,......., M
2 2

The term || r||2 is common to all distance metrics, and, hence, it


may be ignored in the computations of the metrics.
The result is a set of modified distance metrics.

D (r, s m ) = −2r ⋅ s m + s m
2

Note that selecting the signal sm that minimizes D'(r, sm ) is


equivalent to selecting the signal that maximizes the metrics
C(r, sm)= –D'(r, sm ),
C (r , s m ) = 2r ⋅ s m − s m
2

37
5.1.3 The Optimum Detector
The term r⋅sm represents the projection of the signal vector onto
each of the M possible transmitted signal vectors.
The value of each of these projection is a measure of the
correlation between the receiver vector and the mth signal. For
this reason, we call C(r, sm), m=1,2,…,M, the correlation
metrics for deciding which of the M signals was transmitted.
Finally, the terms || sm||2 =εm, m=1,2,…,M, may be viewed as
bias terms that serve as compensation for signal sets that have
unequal energies.
If all signals have the same energy, || sm||2 may also be ignored.
Correlation metrics can be expressed as:
C (r, s m ) = 2 ∫ r (t ) sm (t ) dt − ε
T
, m = 0,1,......., M
0 m

38
5.1.3 The Optimum Detector
These metrics can be generated by a demodulator that cross-
correlates the received signal r(t) with each of the M possible
transmitted signals and adjusts each correlator output for the
bias in the case of unequal signal energies.

39
5.1.3 The Optimum Detector
We have demonstrated that the optimum ML detector computes
a set of M distances D(r,sm) or D'(r,sm) and selects the signal
corresponding to the smallest (distance) metric.
Equivalently, the optimum ML detector computes a set of M
correlation metrics C(r, sm) and selects the signal corresponding
to the largest correlation metric.
The above development for the optimum detector treated the
important case in which all signals are equal probable. In this
case, the MAP criterion is equivalent to the ML criterion.
When the signals are not equally probable, the optimum MAP
detector bases its decision on the probabilities given by:
p (r | s m ) P (s m )
P (s m | r ) = or PM ( r , s m ) = p ( r | s m ) P ( s m )
p (r )
40
5.1.3 The Optimum Detector
Example 5.1-3:
Consider the case of binary PAM signals in which the two
possible signal points are s1 = −s2 = ε b ,, where εb is the
energy per bit. The priori probabilities are P(s1)=p and
P(s2)=1-p. Let us determine the metrics for the optimum
MAP detector when the transmitted signal is corrupted with
AWGN.
The receiver signal vector for binary PAM is:
r = ± ε b + y n (T )
where yn(T) is a zero mean Gaussian random variable with
1
variance σ n2 = N 0 .
2

41
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
The conditional PDF P(r|sm) for two signals are
1 ⎡ (r − ε b )2 ⎤
p (r | s1 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦
1 ⎡ (r + ε n ) 2 ⎤
p (r | s2 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦
Then the metrics PM(r,s1) and PM(r,s2) are
p ⎡ (r − ε b ) 2 ⎤
PM (r, s1 ) = p ⋅ p (r | s1 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦

PM (r, s 2 ) = (1 − p ) ⋅ p (r | s2 ) =
1− p ⎡ (r +
exp ⎢ −
ε ) 2


b

2πσ n ⎢⎣ 2σ n
2
⎥⎦
42
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
If PM(r,s1) > PM(r,s2), we select s1 as the transmitted signal:
otherwise, we select s2. This decision rule may be expressed
as: PM (r, s1 ) s1
1
PM (r, s 2 ) s2
PM (r, s1 ) p ⎡ (r + ε b ) 2 − (r − ε b ) 2 ⎤
= exp ⎢ ⎥
PM (r, s 2 ) 1 − p ⎢⎣ 2σ n2
⎥⎦
(r + ε b ) 2 − (r − ε b ) 2 s1 1 − p
ln
2σ n
2
s2
p
1 2 1− p 1 1− p
s1

ε br
2
σ n ln
p
= N 0 ln
4 p
s2

43
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
1 1− p
The threshold is 4
N 0 ln
p , denoted by τh, divides the real
line into two regions, say R1 and R2, where R1 consists of the
set of points that are greater than τh and R2 consists of the set
of points that are less than τh.
If r ε b > τ h , the decision is made that s1 was transmitted.
If r ε b < τ h , the decision is made that s2 was transmitted.

44
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
The threshold τh depends on N0 and p. If p=1/2, τh=0.
If p>1/2, the signal point s1 is more probable and, hence,
τh<0. In this case, the region R1 is larger than R2 , so that s1 is
more likely to be selected than s2.
The average probability of error is minimized
It is interesting to note that in the case of unequal priori
probabilities, it is necessary to know not only the values of
the priori probabilities but also the value of the power
spectral density N0 , or equivalently, the noise-to-signal ratio,
in order to compute the threshold.
When p=1/2, the threshold is zero, and knowledge of N0 is not
required by the detector.
45
5.1.3 The Optimum Detector
Proof of “the decision rule based on the maximum-likelihood
criterion minimizes the probability of error when the M signals
are equally probable a priori”.
Let us denote by Rm the region in the N-dimensional space for
which we decide that signal sm(t) was transmitted when the
vector r=[r1r2….. rN] is received.
The probability of a correct decision given that sm(t) was
P ( c | s m ) ⋅ p ( s m ) = ∫ p ( r | s m ) ⋅ p ( s m ) dr
transmitted is:
Rm 1
The average probability of a correct decision is: M
M M
1 1
P (c ) = ∑ P (c | s m ) = ∑ ∫ p (r | s m )dr
m =1 M m =1 M Rm
Note that P(c) is maximized by selecting the signal sm if p(r|sm)
is larger than p(r|sk) for all m≠k. Q.E.D.
46
5.1.3 The Optimum Detector
Similarly for the MAP criterion, when the M signals are not
equally probable, the average probability of a correct decision
is M
P(c) = ∑ ∫ P(s m | r ) p (r )dr
m =1 Rm
same for all sm.
In order for P (c) to be as large as possible, the points that are
to be included in each particular region Rm are those for which
P(sm|r) exceeds all the other posterior probabilities. Q.E.D.

We conclude that MAP criterion maximize the probability of


correct detection.

47
5.1.4 The Maximum-Likelihood
Sequence Detector
When the signal has no memory, the symbol-by-symbol
detector described in the preceding section is optimum in the
sense of minimizing the probability of a symbol error.
When the transmitted signal has memory, i.e., the signals
transmitted in successive symbol intervals are interdependent,
the optimum detector is a detector that bases its decisions on
observation of a sequence of received signals over successive
signal intervals.
In this section, we describe a maximum-likelihood sequence
detection algorithm that searches for minimum Euclidean
distance path through the trellis that characterizes the memory
in the transmitted signal.

48
5.1.4 The Maximum-Likelihood
Sequence Detector
To develop the maximum-likelihood sequence detection
algorithm, let us consider, as an example, the NRZI signal
described in Section 4.3-2. Its memory is characterized by the
trellis shown in the following figure:

The signal transmitted in each signal interval is binary PAM.


Hence, there are two possible transmitted signals corresponding
to the signal points s1 = −s2 = ε b where εb is the energy per bit.
49
5.1.4 The Maximum-Likelihood
Sequence Detector
The output of the matched-filter or correlation demodulator for
binary PAM in the kth signal interval may be expressed as
rk = ± ε b + nk
where nk is a zero-mean Gaussian random variable with
variance σ n2 = N 0 .
2
The conditional PDFs for the two possible transmitted signals
are
p ( rk | s1 ) =
1 (
⎡ r − ε 2⎤
exp ⎢ −
k b ) ⎥
2πσ n ⎢ 2σ n2 ⎥
⎢⎣ ⎥⎦

(
⎡ r + ε
) ⎤
2

p ( rk | s2 ) =
1
exp −⎢ k b ⎥
2πσ n ⎢ 2σ n2 ⎥
⎢⎣ ⎥⎦
50
5.1.4 The Maximum-Likelihood
Sequence Detector
Since the channel noise is assumed to be white and Gaussian
and f (t= iT), f (t =jT) for i≠j are orthogonal, it follows that
E(nknj)=0, k≠j.
Hence, the noise sequence n1, n2, …, nk is also white.
Consequently, for any given transmitted sequence s(m), the joint
PDF of r1, r2, …, rk may be expressed as a product of K
marginal PDFs, K
p (r1 , r2 ,..., rk | s ( m ) ) = ∏ p (rk | sk( m ) )
k =1
K
1 ⎡ (rk − sk( m ) ) 2 ⎤
=∏ exp ⎢− ⎥
k =1 2π σ n ⎣ 2σ 2
n ⎦
K
⎛ 1 ⎞ ⎡ K (rk − sk( m ) ) 2 ⎤
= ⎜⎜ ⎟ exp ⎢− ∑ ⎥ (A)
⎟ σ
⎝ 2π σ n ⎠
2
⎣ k =1 2 n ⎦
51
5.1.4 The Maximum-Likelihood
Sequence Detector
Then, given the received sequence r1,r2,…,rk at the output of
the matched filter or correlation demodulator, the detector
determines the sequence s(m)={s1(m),s2(m),..,sK(m)} that
maximizes the conditional PDF p(r1,r2,…,rk|s(m)). Such a
detector is called the maximum-likelihood (ML) sequence-
detector.
By taking the logarithm of Equation (A) and neglecting the
terms that are independent of (r1,r2,…,rk), we find that an
equivalent ML sequence detector selects the sequence s(m) that
minimizes the Euclidean distance metric
K
D(r, s ( m ) ) = ∑ (rk − sk( m ) ) 2
k =1

52
5.1.4 The Maximum-Likelihood
Sequence Detector
In searching through the trellis for the sequence that minimizes
the Euclidean distance, it may appear that we must compute the
distance for every possible sequence.
For the NRZI example, which employs binary modulation, the
total number of sequence is 2K , where K is the number of
outputs obtained form the demodulator.
However, this is not the case. We may reduce the number of
the sequences in the trellis search by using the Viterbi
algorithm to eliminate sequences as new data is received from
the demodulator.
The Viterbi algorithm is a sequence trellis search algorithm for
performing ML sequence detection. We describe it below in
the context of the NRZI signal. We assume that the search
process begins initially at state S0.

53
Viterbi Decoding Algorithm
Basic concept
Generate the code trellis at the decoder
The decoder penetrates through the code trellis level by level in
search for the transmitted code sequence
At each level of the trellis, the decoder computes and
compares the metrics of all the partial paths entering a node
The decoder stores the partial path with the larger metric and
eliminates all the other partial paths. The stored partial path is
called the survivor.

54
5.1.4 The Maximum-Likelihood
Sequence Detector
The corresponding trellis is shown in the following figure :

At time t=T, we receive r1=s1(m)+n1 from the demodulator, and at


t=2T, we receive r2=s2(m)+n2. Since the signal memory is one bit,
which we denote by L=1, we observe that the trellis reaches its
regular (steady state) form after two transitions. Differential encoding.
55
5.1.4 The Maximum-Likelihood
Sequence Detector
Thus, upon receipt of r2 at t=2T, we observe that there are two
signal paths entering each of the nodes and two signal paths
leaving each node. The two paths entering node S0 at t=2T
correspond to the information bits (0,0) and (1,1) or, equivalently,
to the signal points ( − ε b , − ε b ) and ( ε b , − ε b ) , respectively.
The two paths entering node S1 at t=2T correspond to the
information bits (0,1) and (1,0) or, equivalently, to the signal
points ( − ε b , ε b ) and ( ε b , ε b ) , respectively.
For the two paths entering node S0, we compute the two
Euclidean distance metrics.
D0 (0, 0) = (r1 + ε b ) 2 + (r2 + ε ) 2
b
D0 (1,1) = (r1 − ε b ) 2
+ (r2 + εb ) 2

by using the outputs r1 and r2 from demodulator.


56
5.1.4 The Maximum-Likelihood
Sequence Detector
The Viterbi algorithm compares these two metrics and discards
the path having the larger metrics. The other path with the
lower metric is saved and is called survivor at t=2T.
Similarly, for two paths entering node S1 at t=2T, we compute
the two Euclidean distance metrics
D1 (0,1) = (r1 + ε b ) 2
+ (r2 − εb ) 2

D1 (1,0) = (r1 − ε b ) 2
+ (r2 − εb ) 2

by using the output r1 and r2 from the demodulator. The two


metrics are compared and the signal path with the larger metric
is eliminated.
Thus, at t=2T, we are left with two survivor paths, one at node
S0 and the other at node S1 , and their corresponding metrics.
57
5.1.4 The Maximum-Likelihood
Sequence Detector
Upon receipt of r3 at t=3T, we compute the metrics of the two
paths entering state S0. Suppose the survivors at t=2T are the
paths (0,0) at S0 and (0,1) at S1. Then, the two metrics for the
paths entering S0 at t=3T are
D0 (0, 0, 0) = D0 (0, 0) + (r3 + ε ) 2
b
D0 (0,1,1) = D1 (0,1) + (r3 + εb ) 2

These two metrics are compared and the path with the larger
(greater-distance) metric is eliminated.
Similarly, the metrics for the two paths entering S1 at t=3T are
D1 (0,0,1) = D0 (0,0) + (r3 − ε b ) 2
D1 (0,1,0) = D1 (0,1) + (r3 − ε b ) 2
These two metrics are compared and the path with the larger
metrics is eliminated.
58
5.1.4 The Maximum-Likelihood
Sequence Detector
It is relatively easy to generalize the trellis search performed by
the Viterbi algorithm for M-ary modulation.
Delay modulation with M=4 signals: such signal is characterized
by the four-state trellis shown in the following figure

59
5.1.4 The Maximum-Likelihood
Sequence Detector
Delay modulation with M=4 signals: (cont.)
We observe that each state has two signal paths entering and
two signals paths leaving each node.
The memory of the signal is L=1.
The Viterbi algorithm will have four survivors at each stage
and their corresponding metrics.
Two metrics corresponding to the two entering paths are
computed at each node, and one of the two signal paths
entering the node is eliminated at each state of the trellis.
The Viterbi algorithm minimizes the number of trellis paths
searched in performing ML sequence detection.
From the description of the Viterbi algorithm given above, it is
unclear as to how decisions are made on the individual detected
information symbols given the surviving sequences.
60
5.1.4 The Maximum-Likelihood
Sequence Detector
If we have advanced to some stage, say K, where K>>L in the
trellis, and we compare the surviving sequences, we shall find
that with probability approaching one all surviving sequences
will be identical in bit (or symbol) positions K-5L and less.
In a practical implementation of the Viterbi algorithm, decisions
on each information bit (or symbol) are forced after a delay of
5L bits, and, hence, the surviving sequences are truncated to the
5L most recent bits (or symbol). Thus, a variable delay in bit or
symbol detection is avoided.
The loss in performance resulting from the sub-optimum
detection procedure is negligible if the delay is at least 5L.

61
5.2.1 Probability of Error for Binary
Modulation
Let us consider binary PAM signals where the two signal
waveforms are s1(t)=g(t) and s1(t)= −g(t), and g(t) is an arbitrary
pulse that is nonzero in the interval 0≤ t ≤Tb and zero elsewhere.
Since s1(t)= −s2(t), these signals are said to be antipodal.
The energy in the pulse g(t) is εb.
As indicated in section 4.3.1, PAM signals are one-dimensional,
and, their geometric representation is simply the one-
dimensional vector s1 = ε b , s2 = − ε b .

Figure (A)
62
5.2.1 Probability of Error for Binary
Modulation
Let us assume that the two signals are equally likely and that
signal s1(t) was transmitted. Then, the received signal from the
(matched filter or correlation) demodulator is

r = s1 + n = ε b + n

where n represents the additive Gaussian noise component,


1
which has zero mean and variance σ n2 = N 0 .
2

In this case, the decision rule based on the correlation metric


given by Equation 5.1-44 compares r with the threshold zero. If
r >0, the decision is made in favor of s1(t), and if r<0, the
decision is made that s2(t) was transmitted.

63
5.2.1 Probability of Error for Binary
Modulation

The two conditional PDFs of r are:


−( r − )
2
1
p ( r | s1 ) =
εb / N0
e
π N0
−( r + )
2
1
p ( r | s2 ) =
εb / N0
e
π N0

64
5.2.1 Probability of Error for Binary
Modulation
Given that s1(t) was transmitted, the probability of error is simply
the probability that r<0.
P ( e | s1 ) = ∫ p ( r | s1 )dr =
0 1 0
⎡ r− ε 2⎤
⎢ ⎥ dr ( )
π N 0 ∫−∞
b
exp −
−∞ ⎢ N0 ⎥
⎢⎣ ⎥⎦

1 − 2ε b / N 0 ( r− εb )

− x2 / 2
= e dx x=
2π −∞ N0
2
1 ∞
= ∫ e − x2 / 2
dx ( x = −x)
2π 2ε b / N 0


⎛ 2ε b ⎞ Q (t ) =
1

− x2 / 2
e dx t ≥ 0
= Q⎜ ⎟⎟ 2π
⎜ N t
⎝ 0 ⎠
65
5.2.1 Probability of Error for Binary
Modulation
If we assume that s2(t) was transmitted, r = − ε b + n and the
( )
probability that r>0 is also P ( e | s2 ) = Q 2ε b / N 0 . Since the
signal s1(t) and s2(t) are equally likely to be transmitted, the
average probability of error is
1 1 ⎛ 2ε b

Pb = P ( e | s1 ) + P ( e | s2 ) = Q ⎜⎜ ⎟⎟ --(A)
2 2 ⎝ N0

Two important characteristics of this performance measure:
First, we note that the probability of error depends only on
the ratio εb /N0.
Secondly, we note that. 2εb /N0 is also the output SNR0 from
the matched-filter (and correlation) demodulator.
The ratio εb /N0 is usually called the signal-to-noise ratio per bit.
66
5.2.1 Probability of Error for Binary
Modulation
We also observe that the probability of error may be expressed
in terms of the distance between that the two signals s1and s2 .
From figure (A), w observe that the two signals are separated by
1
the distance d12=2 ε b . By substituting ε b = d122 into Equation
4
(A),we obtain
⎛ d2 ⎞
Pb = Q⎜ 12 ⎟
⎜ 2N 0 ⎟
⎝ ⎠

This expression illustrates the dependence of the error


probability on the distance between the two signals points.

67
5.2.1 Probability of Error for Binary
Modulation

Error probability for binary orthogonal signals


The signal vectors s1and s2 are two-dimensional.

s1 = [ ε b 0]
s2 = [0 εb ]

where εb denote the energy for each of the waveforms. Note that
the distance between these signal points is d12 = 2ε b .
68
5.2.1 Probability of Error for Binary
Modulation
Error probability for binary orthogonal signals
To evaluate the probability of error, let us assume that s1 was
transmitted. Then, the received vector at the output of the
demodulator is r = [ ε b + n1 n2 ].
We can now substitute for r into the correlation metrics
given by C (r, s m ) = 2r ⋅ s m − s m to obtain C(r, s1) and C(r,
2

s2).
The probability of error is the probability that C(r, s2) > C(r,
s1).
n2 > n1 + ε b
( ) (
⇒ 2 n2 ε b − ε b > 2ε b + 2 n1 ε b − ε b )
P ( e | s1 ) = P[C ( r , s2 ) > C ( r1 , s2 )] = P[n2 − n1 > ε b ]

69
5.2.1 Probability of Error for Binary
Modulation
Error probability for binary orthogonal signals
Since n1 and n2 are zero-mean statistically independent
Gaussian random variables each with variance ½ N0, the
random variable x= n2 – n1 is zero-mean Gaussian with
variance N0. Hence,
( ) 1 ∞
P n2 − n1 > ε b = ∫
− x2 2 N0
e dx
2π N 0 εb

1 ∞ ⎛ εb ⎞

− x2 2
= e dx = Q ⎜⎜ ⎟⎟
2π εb N0
⎝ N0 ⎠
The same error probability is obtained when we assume that
s2 is transmitted:
⎛ εb ⎞
Pb = Q ⎜
⎜ N ⎟⎟ ( )
= Q γ b where γ b is the SNR per bit.
⎝ 0 ⎠

70
5.2.1 Probability of Error for Binary
Modulation
If we compare the probability of error for binary antipodal
signals with that for binary orthogonal signals, we find that
orthogonal signals required a factor of 2 increase in energy to
achieve the same error probability as antipodal signals.
Since 10 log10 2=3 dB, we say that orthogonal signals are 3dB
poorer than antipodal signals. The difference of 3dB is simply
due to the distance between the two signal points, which is d122 = 2ε b
for orthogonal signals, whereas d122 = 4ε b for antipodal signals.
The error probability versus 10 log10 εb/N0 for these two types of
signals is shown in the following figure (B). As observed from
this figure, at any given error probability, the εb/N0 required for
orthogonal signals is 3dB more than that for antipodal signals.

71
5.2.1 Probability of Error for Binary
Modulation
Probability of error for binary signals

Figure (B)
72
5.2.2 Probability of Error for M-ary
Orthogonal Signals
For equal-energy orthogonal signals, the optimum detector
selects the signal resulting in the largest cross correlation
between the received vector r and each of the M possible
transmitted signals vectors {sm}, i.e.,
M
C (r , sm ) = r ⋅ sm = ∑ rk smk , m = 1,2, … , M
k =1

To evaluate the probability of error, let us suppose that the signal


s1 is transmitted. Then the received signal vector is
r = ⎡⎣ ε b + n1 n2 n3 nM ⎤⎦
where εs denotes the symbol energy and n1,n2,…,nM are zero-
mean, mutually statistically independent Gaussian random
variable with equal variance σ n2 = 1 N 0 .
2
73
5.2.2 Probability of Error for M-ary
Orthogonal Signals
In this case, the outputs from the bank of M correlations are
C (r , s1 ) = ε s ( ε s + n1 )
C (r , s2 ) = ε s n2

C (r , sM ) = ε s nM
Note that the scale factor εs may be eliminated from the
correlator outputs dividing each output by ε s .
With this normalization, the PDF of the first correlator output is

pr1 ( x1 ) =
1 ⎡ x − ε
exp ⎢− 1
(s ) ⎤⎥
2

πN 0 ⎢ N0 ⎥
⎣ ⎦
74
5.2.2 Probability of Error for M-ary
Orthogonal Signals
And the PDFs of the other M−1 correlator outputs are

prm ( xm ) =
1 − xm2 N 0
e , m = 2,3, … , M
πN 0
It is mathematically convenient to first derive the probability that
the detector makes a correct decision. This is the probability that
r1 is large than each of the other M−1 correlator outputs n2,n3,…,
nM. This probability may be expressed as

Pc = ∫ P (n2 < r1 , n3 < r1 , … , nM < r1 | r1 ) p (r1 )dr1
−∞

where P ( n2 < r1 , n3 < r1 ,… , nM < r1 | r1 ) denotes the joint


probability that n2,n3,…, nM are all less than r1, conditioned on
any given r1. Then this joint probability is averaged over all r1.

75
5.2.2 Probability of Error for M-ary
Orthogonal Signals
Since the {rm} are statistically independent, the joint probability
factors into a product of M−1 marginal probabilities of the form:

P(nm < r1 | r1 ) = ∫ prm ( xm )dxm


r1
m = 2,3, … , M --(B)
−∞

1 r1 2 N 0

− x2 2
= e dx x = 2 N 0 xm
2π −∞

This probabilities are identical for m=2,3,…,M, and, the joint


probability under consideration is simply the result in Equation
(B) raised to the (M−1)th power. Thus, the probability of a
correct decision is
M −1
∞ ⎛ 1 r1 2 N0 − x2 2 ⎞
Pc = ∫ ⎜ ∫ e dx ⎟ p ( r1 ) dr1
−∞
⎝ 2π −∞

76
5.2.2 Probability of Error for M-ary
Orthogonal Signals
The probability of a (k-bit) symbol error is
2
y = r1
N0
PM = 1 − Pc

⎧ M −1 ⎡ ⎛ ⎞
2
⎤ ⎫
⎪ 1 ∞ ⎛ 1 y ⎞ 1 2ε ⎪
∫−∞ ⎜⎝ 2π ∫−∞ ⎢ ⎥
−x 2
PM = 1 − ⎨ − −
2

⎜⎜ ⎟⎟ ⎬
s
e dx ⎟ exp y dy
⎪⎩ 2π ⎠ ⎢ 2⎝

N0 ⎠ ⎥ ⎪
⎦ ⎭
--- (C)
The same expression for the probability of error is obtained
when any one of the other M−1 signals is transmitted. Since all
the M signals are equally likely, the expression for PM given
above is the average probability of a symbol error.
This expression can be evaluated numerically.
77
5.2.2 Probability of Error for M-ary
Orthogonal Signals
In comparing the performance of various digital modulation
methods, it is desirable to have the probability of error
expressed in terms of the SNR per bit, εb/N0, instead of the SNR
per symbol, εs/N0.
With M=2k, each symbol conveys k bits of information, and
hence εs= k εb. Thus, Equation (C) may be expressed in terms of
εb/N0 by substituting for εs.
It is also desirable to convert the probability of a symbol error
into an equivalent probability of a binary digit error. For
equiprobable orthogonal signals, all symbol errors are
equiprobable and occur with probability
PM P
= kM
M −1 2 −1
78
5.2.2 Probability of Error for M-ary
Orthogonal Signals
⎛k⎞
Furthermore, there are ⎜ n ⎟ ways in which n bits out of k may be
⎝ ⎠
in error. Hence, the average number of bit errors per k-bit
symbol is k
⎛ k ⎞ PM 2k −1
∑ n⎜ ⎟ k
n =1 ⎝ n ⎠ 2 − 1
=k k
2 −1
PM --(D)

and the average bit error probability is just the result in Equation
(D) divided by k, the number of bits per symbol. Thus,
2k −1 P
Pb = k PM ≈ M k >> 1
2 −1 2
The graphs of the probability of a binary digit error as a function
of the SNR per bit, εb/N0, are shown in Figure (C) for
M=2,4,8,16,32, and 64. This figure illustrates that, by increasing
the number M of waveforms, one can reduce the SNR per bit
required to achieve a given probability of a bit error.
79
5.2.2 Probability of Error for M-ary
Orthogonal Signals
For example, to achieve a Pb=10 –5, the required SNR per bit is a
little more than 12dB for M=2, but if M is increased to 64 signal
waveforms, the required SNR per bit is approximately 6dB. Thus,
a savings of over 6dB is realized in transmitter power required to
achieve a Pb=10 –5 by increasing M from M=2 to M=64.

Figure (C)

80
5.2.2 Probability of Error for M-ary
Orthogonal Signals
What is the minimum required εb/N0 to achieve an arbitrarily
small probability of error as M→∞?
A union bound on the probability of error.
Let us investigate the effect of increasing M on the
probability of error for orthogonal signals. To simplify the
mathematical development, we first derive an upper bound
on the probability of a symbol error that is much simpler than
the exact form given in the following Equation (5.2-21)

1 ∞⎡ ⎛ 1 y ⎞
M −1
⎤ ⎡ 1⎛ 2ε s ⎞
2

⎜ ⎟
PM = ∫ 1− ⎜ ∫ e−x ⎢ − − ⎥ dy
2

2
dx ⎟ ⎥ exp ⎜ y ⎟
2π −∞ ⎢⎣ ⎝ 2π −∞
⎠ ⎥⎦ ⎢ 2⎝ N0 ⎠ ⎥
⎣ ⎦

81
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error (cont.)
Recall that the probability of error for binary orthogonal
signals is given by 5.2-11: ⎛ εb ⎞
Pb = Q⎜⎜ ( )
⎟ = Q γb

⎝ N0 ⎠
Now, if we view the detector for M orthogonal signals as one
that makes M −1 binary decisions between the correlator
outputs C(r,s1) that contains the signal and the other M−1
correlator outputs C(r,sm), m=2,3,…,M, the probability of
error is upper-bounded by union bound of the M −1 events.
That is, if Ei represents the eventM that C(r,si)> C(r,s1) for i≠1,
( )
then we have PM = P ∪ i = 2 Ei ≤ ∑ P ( Ei ) . Hence,
M

i =2

PM ≤ ( M − 1) Pb = ( M − 1) Q ( )
ε S / N 0 < MQ ( ε S / N0 )
82
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error(cont.)
This bound can be simplified further by upper-bounding
(
Q εS / N0 )
We have
( )
Q ε s N 0 < e −ε s 2 N 0 --(E)
thus,
PM < Me −ε s 2 N0
= 2k e − kε b 2 N0
--(F)
− k ( ε b N 0 − 2ln 2 ) 2
PM < e
As k→∞ or equivalently, as M→∞, the probability of error
approaches zero exponentially, provided that εb/N0 is greater
εb
than 2ln 2,
> 2 ln 2 = 1.39 (1.42dB)
N0
83
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error(cont.)
The simple upper bound on the probability of error given by
Equation (F) implies that, as long as SNR>1.42 dB, we can
achieve an arbitrarily low PM..
However, this union bound is not a very tight upper bound as
a sufficiently low SNR due to the fact that upper bound for
the Q function in Equation (E) is loose.
In fact, by more elaborate bounding techniques, it is shown in
Chapter 7 that the upper bound in Equation (F) is sufficiently
tight for εb/N0>4 ln2.
For εb/N0<4 ln2, a tighter upper bound on PM is
−k ( ε 0 N 0 − ln 2 )2
PM < 2e
84
5.2.2 Probability of Error for M-ary
Orthogonal Signals

A union bound on the probability of error(cont.)


Consequently, PM →0 as k→∞, provided that
εb
> ln 2 = 0.693 (− 1.6dB)
N0
Hence, −1.6 dB is the minimum required SNR per bit to
achieve an arbitrarily small probability of error in the limit as
k→∞(M→∞). This minimum SNR per bit is called the
Shannon limit for an additive Gaussian noise channel.

85
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
As indicated in Section 4.3, a set of M=2k biorthogonal signals
are constructed from ½ M orthogonal signals by including the
negatives of the orthogonal signals. Thus, we achieve a
reduction in the complexity of the demodulator for the
biorthogonal signals relative to that for orthogonal signals, since
the former is implemented with ½ M cross correlation or
matched filters, whereas the latter required M matched filters or
cross correlators.
Let us assume that the signal s1(t) corresponding to the vector
s1=[ ε s 0 0…0] was transmitted. The received signal vector is
r = [ ε s + n1 n2 nM 2 ]
where the {nm} are zero-mean, mutually statistically independent
and identically distributed Gaussian random variables with
1
variance σ n2 = N 0 .
2
86
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
The optimum detector decides in favor of the signal
corresponding to the largest in magnitude of the cross
correlators M 2
1
C ( r,s m ) = r ⋅ s m = ∑ rk smk , m = 1, 2,..., M
k =1 2
while the sign of this largest term is used to decide whether sm(t)
or −sm(t) was transmitted.
According to this decision rule, the probability of a correct
decision is equal to the probability that r1 = ε s + n1 > 0 and r1
exceeds |rm|=|nm| for m=2,3,… ½ M. But
P ( nm < r1 | r1 > 0 ) =
1 r1 1 r1 N0 2 − y 2 2
∫ ∫
− x2 N0
e dx = e dy
π N0 − r1
2π − r1 N 0 2
N0
y= x
2
87
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
Then, the probability of a correct decision is
M 2 −1
⎛ 1∞
−x 2 ⎞
p ( r1 ) dr1
r1 N0 2
Pc = ∫ ⎜ ∫
2
e ⎟
0
⎝ 2π − r1 N0 2 ⎠
from which, upon substitution for p(r1), we obtain
M 2 −1
1 ∞ ⎛ 1 v + 2ε s N 0 ⎞
∫− 2ε s N0 ⎜⎝ 2π ∫−(v+ 2ε s N0 ) e dx ⎟⎠ e dv --(G)
−x 2 −v2 2
Pc =
2

p r1 ( x1 ) =
1 ⎡

exp −
(
x1 − ε s ⎤
2


) v = r1 − (
ε s
N0
)
πN 0 ⎢ N0 ⎥ 2
⎣ ⎦
where we have used the PDF of r1 given in Equation 5.2-15.
The probability of a symbol error PM=1−Pc

88
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
Pc, and PM may be evaluated numerically for different values of
M. The graph shown in the following figure (D) illustrates PM
as a function of εb/N0, where εs=k εb, for M=2,4,8,16 and 32.

89
5.2.3 Probability of Error for M-ary
Biorthogonal Signals

In this case, the probability of error for M=4 is greater than that
for M=2. This is due to the fact that we have plotted the symbol
error probability PM in Figure(D).

If we plotted the equivalent bit error probability, we should find


that the graphs for M=2 and M=4 coincide.
As in the case of orthogonal signals, as M→∞ (k→∞), the
minimum required εb/N0 to achieve an arbitrarily small
probability of error is −1.6 dB, the Shannon limit.

90
5.2.4 Probability of Error for
Simplex Signals
Next we consider the probability of error for M simplex signals.
Recall from Section 4.3 that simplex signals are a set of M equally
correlated with mutual cross-correlation coefficient ρmn= −1/(M −1).
These signals have the same minimum separation of 2ε s between
adjacent signal points in M-dimensional space as orthogonal signals.
They achieve this mutual separation with a transmitted energy of
εs(M −1)/M, which is less than that required for orthogonal signals
by a factor of (M −1)/M.
Consequently, the probability of error for simplex signals is
identical to the probability of error for orthogonal signals, but this
performance is achieved with saving of
10 log(1 − ρ ) = 10 log
M
dB
M −1
For M=2, the saving is 3dB. As M is increased, the saving in SNR
approaches 0 dB.
91
5.2.5 Probability of Error for M-ary
Binary-Coded Signals
In Section 4.3, we have shown that binary-coded signal
waveform are represented by signal vectors (4.3-38) :
s m = [s m1 s m 2 s mN ] , m = 1,2, … , M
where s mj = ± ε N for all m and j. N is the block length of
the code and is also the dimension of the M signal waveform
(e)
If d min is the minimum Euclidean distance, then the probability
of a symbol error is upper-bounded as
⎛ (d (e ) )2 ⎞
PM < (M − 1)Pb = (M − 1)Q⎜⎜ min ⎟
⎜ 2 N 0 ⎟⎟
⎝ ⎠

< 2 exp ⎢−
k (d min )
(e ) 2 ⎤

⎢⎣ 4 N 0 ⎥⎦
92
5.2.6 Probability of Error for M-ary
PAM
Recall (4.3-6) that M-ary PAM signals are represent
geometrically as M one-dimensional signal points with value :
sm =
1
2
ε g Am m = 1, 2, ,M

where Am = (2 m − 1 − M )d , m = 1,2, , M
The Euclidean distance between adjacent signal points is d 2ε g .
Assuming equally probable signals, the average energy is :
1 M
1 M
2
d 2
ε
ε av = M ∑ ε m = M ∑ sm = 2M ∑ ( 2m − 1 − M )
g
M
2

m =1 m =1 m =1

d 2ε g ⎡ 1 M ( M + 1)
⎤ 1
ε
M

=
2M ⎣ 3⎢ (
M M 2
−1) =
⎥ 6

M 2
(
− 1 d 2
) ∑
g m =1
m=
2
M M ( M + 1)( 2 M + 1)
∑m =
2

m =1 6

93
5.2.6 Probability of Error for M-ary
PAM
Equivalently, we may characterize these signals in terms of their
average power, where is :
ε d εg
( )
2
1
Pav = = M −1
av2
(5.2-40)
T 6 T
The average probability of error for M-ary PAM :
The detector compares the demodulator output r with a set of M-1
thresholds, which are placed at the midpoints of successive amplitude
level and decision is made in favor of the amplitude level that is close to r.
1
2d εg
2

We note that if the mth amplitude level is transmitted, the demodulator


output is
r = sm + n =
1
2
ε g Am + n
1
where the noise variable n has zero-mean and variance σ n =
2
N0
2
94
5.2.6 Probability of Error for M-ary
PAM
Assuming all amplitude levels are equally likely a priori, the
average probability of a symbol error is the probability that the
noise variable n exceeds in magnitude one-half of the distance
between levels.
However, when either one of the two outside levels ±(M-1) is
transmitted, an error can occur in one direction only.
As a result, we have the error probability:
M −1 ⎛ ⎞ M −1
PM = P ⎜⎜ r − sm > d
1
ε ⎟⎟ =
2

ε
e − x 2 N0
dx
2π N 0 2
g
M ⎝ 2 ⎠ M d g 2

M −1 2 ∞

− y2 2
= e dy where y = x N0 2
M 2π d2 ε g N0
1 ∞
Q (t ) = ∫ e− x / 2 dx t ≥ 0
2

2 ( M − 1) ⎛ d 2ε g ⎞ 2π t

= Q⎜ ⎟ 1
i( M − 2 ) +
1 1
i i2
M ⎜ N0 ⎟ M M 2
⎝ ⎠
95
5.2.6 Probability of Error for M-ary
PAM
From (5.2-40), we note that d ε
2
g =
6
M −1
2
PavT

By substituting for d
2
ε , we obtain the average probability
g
of a symbol error for PAM in terms of the average power :
2 ( M − 1) ⎛ 6 PavT ⎞ 2 ( M − 1) ⎛ 6ε av ⎞
PM = Q⎜ ⎟= Q ⎜ ⎟
M ⎜
⎝ ( )
M 2 − 1 N0 ⎟

M ⎜
⎝ ( )
M 2 − 1 N0 ⎟

It is customary for us to use the SNR per bit as the basic
parameter, and since T=kTb and k =log2 M:

PM =
2(M − 1) ⎛⎜ (6 log 2 M )ε b av ⎞⎟
M
Q⎜
⎝ (M 2
)
− 1 N 0 ⎟⎠
where ε bav = PavTb is the average bit energy and ε bav N o is the average
SNR per bit.

96
5.2.6 Probability of Error for M-ary
PAM
The case M=2
corresponds to the error
probability for binary
antipodal signals.
The SNR per bit
increase by over 4 dB for
every factor-of-2 increase
in M.
For large M, the
additional SNR per bit
required to increase M by
a factor of 2 approaches 6
Probability of a symbol error for PAM dB.
97
5.2.7 Probability of Error for M-ary
PSK
Recall from section 4.3 that digital phase-modulated signal
waveforms may be expressed as (4.3-11):
⎡ 2π
sm (t ) = g (t ) cos ⎢ 2πf c t + (m − 1)⎤⎥, 1 ≤ m ≤ M , 0 ≤ t ≤ T
⎣ M ⎦
Energy in each waveform.
and have the vector representation: (from 4.3-12)
2π 2π
s m (t ) = ⎢ ε s cos (m − 1) ε s sin (m − 1)⎥ , ε s = ε g
⎡ ⎤ 1
⎣ M M ⎦ 2
Since the signal waveforms have equal energy, the optimum
detector for the AWGN channel given by Equation 5.1-44
computes the correlation metrics

C (r, s m ) = r ⋅ s m , m = 1,2, ,M

98
5.2.7 Probability of Error for M-ary
PSK
In other word, the received signal vector r=[r1 r2] is projected
onto each of the M possible signal vectors and a decision is
made in favor of the signal with the largest projection.
This correlation detector is equivalent to a phase detector that
computes the phase of the received signal from r. We selects
the signal vector sm whose phase is closer to r.
−1 r2
The phase of r is Θ r = tan
r1

We will determine the PDF of Θ r , and compute the probability


of error from it.
Consider the case in which the transmitted signal phase is
Θ r = 0 , corresponding to the signal s1(t).
99
5.2.7 Probability of Error for M-ary
PSK
The transmitted signal vector is s1 = ⎡⎣ ε s 0 ⎤⎦ , and the received
signal vector has components:
r1 = ε s + n1 r2 = n2
Because n1 and n2 are jointly Gaussian random variable, it
follow that r1 and r2 are jointly Gaussian random variable
variables with E (r1 ) = ε s , E (r2 ) = 0, and σ r2 = σ r2 = 1 N 0 = σ r2 .
1 2
2

p r (r1 , r2 ) =
1
exp ⎢
(
− 1 s )
⎡ r − ε + r2 ⎤2
2

2πσ r
2
⎢ 2σ r 2

⎣ ⎦
The PDF of the phaseΘr is obtained by a change in variables
from (r1,r2) to : r
V = r12 + r22 Θ r = tan −1 2
r1
100
5.2.7 Probability of Error for M-ary
PSK
The joint PDF of V and Θr : dr1dr2 = VdVd Θ

⎛ V 2 + ε s − 2 ε s V cos Θ r ⎞
pV ,Θ r (V , Θ r ) =
V
⎜− ⎟
exp
2πσ r2 ⎜ 2σ 2 ⎟
⎝ r ⎠
Integration of pV , Θ r (V , Θ r ) over the range of V yields pΘr ( Θ r )

pΘr ( Θ r ) = ∫ pV ,Θr (V , Θ r ) dV
0
2
⎛ V ⎞
∞ −⎜ − 2ε s N 0 cos Θ r ⎟ 2σ r2
1 V
e −ε s ∫
N 0 sin Θ r
=
2
⎝ N0 2 ⎠
e dV
2πσ 2
r
0 N0 2
( )
2
1 ∞ − V ′ − 2γ s cos Θ r / 2σ r2 V
∫0 dV ′ V ′ =
− γ s sin Θr
= V ′e
2
e
2πσ r2 N0 2
where we define the symbol SNR as γ s = ε s N0 .
101
5.2.7 Probability of Error for M-ary
PSK

f Θ r (Θ r ) becomes
narrower and more
peaked about Θ r = 0 as
the SNR γ s increases.

Probability density function pΘr (Θr ) for γ S = 1, 2, 4, and 10.


102
5.2.7 Probability of Error for M-ary
PSK
When s1(t) is transmitted, a decision error is made if the noise
causes the phase to fall outside the range − π M ≤ Θ r ≤ π M .
Hence, the probability of a symbol error is
π M
PM = 1 − ∫ p Θ r (Θ r ) dΘ r
−π M

In general, the integral of p Θ r (Θ r ) doesn’t reduced to a simple


form and must be evaluated numerically, except for M =2 and
M =4.
For binary phase modulation, the two signals s1(t) and s2(t) are
antipodal. Hence, the error probability is
⎛ 2ε b ⎞
P2 = Q⎜⎜ ⎟

⎝ N 0 ⎠
103
5.2.7 Probability of Error for M-ary
PSK
When M = 4, we have in effect two binary phase-modulation
signals in phase quadrature. Since there is no crosstalk or
interference between the signals on the two quadrature carriers,
the bit error probability is identical to that of M = 2. (5.2-57)
Then the symbol error probability for M=4 is determined by
⎛ 2ε b ⎞
2
noting that ⎡ ⎤
Pc = (1 − P2 ) = ⎢1 − Q ⎜⎜ ⎟⎥
2

⎢⎣ ⎝ N 0 ⎠⎥⎦
where Pc is the probability of a correct decision for the 2-bit
symbol.
There, the symbol error probability for M = 4 is
⎛ 2ε b ⎞ ⎡ 1 ⎛ 2ε b ⎞⎤
P4 = 1 − Pc = 2Q ⎜ ⎟⎟ ⎢1 − Q ⎜⎜ ⎟⎟ ⎥
⎜ N
⎝ 0 ⎠ ⎢⎣ 2 ⎝ N 0 ⎠ ⎥⎦
104
5.2.7 Probability of Error for M-ary
PSK
For M > 4, the symbol error probability PM is obtained by
π M
numerically integrating Equation PM = 1 − ∫ p Θ r (Θ r ) dΘ .r
−π M

For large values of


M, doubling the
number of phases
requires an
additional 6 dB/bit
to achieve the same
performance

105
5.2.7 Probability of Error for M-ary
PSK
An approximation to the error probability for large M and for
large SNR may be obtained by first approximating p Θ r (Θ ) .
For ε s N 0 >> 1 and Θ r ≤ 0.5π , pΘ r (Θ ) is well approximated
as : γs
PΘ r (Θ r ) ≈ cos Θ r e −γ s sin 2 Θ r

π
Performing the change in variable from Θ r to u = ν s sin Θ r ,
π M γs
PM ≈ 1 − ∫−π cos Θ r e −γ s sin 2 Θ r
dΘ r
M
π k = log 2 M
2 ∞ ν s = kν b
∫ 2γ
−u 2
≈ sin (π M )
e du
π s

⎛ π ⎞ ⎛ π ⎞
= 2Q⎜ 2γ s sin ⎟ = 2Q⎜ 2kγ b sin ⎟
⎝ M⎠ ⎝ M⎠
106
5.2.7 Probability of Error for M-ary
PSK
When a Gray Code is used in the mapping of k-bits symbols
into the corresponding signal phases, two k-bit symbols
corresponding to adjacent signal phases differ in only a signal
bit.
The most probable error result in the erroneous selection of an
adjacent phase to the true one.
Most k-bit symbol error contain only a single-bit error. The
equivalent bit error probability for M-ary PSK is well
1
approximated as : Pb ≈ PM
k
In practice, the carrier phase is extracted from the received
signal by performing some nonlinear operation that introduces
a phase ambiguity.
107
5.2.7 Probability of Error for M-ary
PSK
For binary PSK, the signal is often squared in order to remove
the modulation, and the double-frequency component that is
generated is filtered and divided by 2 in frequency in order to
extract an estimate of the carrier frequency and phase φ.
This operation result in a phase ambiguity of 180°in the carrier phase.
For four-phase PSK, the received signal is raised to the fourth
power to remove the digital modulation, and the resulting fourth
harmonic of the carrier frequency is filtered and divided by 4 in
order to extract the carrier component.
These operations yield a carrier frequency component containing φ, but
there are phase ambiguities of ±90° and 180° in the phase estimate.
Consequently, we do not have an absolute estimate of the carrier
phase for demodulation.

108
5.2.7 Probability of Error for M-ary
PSK

The phase ambiguity problem can be overcome by encoding the


information in phase differences between successive signal
transmissions as opposed to absolute phase encoding.
For example, in binary PSK, the information bit 1 may be
transmitted by shifting the phase of carrier by 180° relative to the
previous carrier phase. Bit 0 is transmitted by a zero phase shift
relative to the phase in the previous signaling interval.
In four-phase PSK, the relative phase shifts between successive
intervals are 0, 90°, 180°, and -90°,corresponding to the
information bits 00, 01, 11, and 10, respectively.
The PSK signals resulting from the encoding process are said to
be differentially encoded.

109
5.2.7 Probability of Error for M-ary
PSK
The detector is a relatively simple phase comparator that
compares the phase of the demodulated signal over two
consecutive interval to extract the information.
Coherent demodulation of differentially encoded PSK results in
a higher probability of error than that derived for absolute phase
encoding.
With differentially encoded PSK, an error in the demodulated
phase of the signal in any given interval will usually result in
decoding errors over two consecutive signaling intervals.
The probability of error in differentially encoded M-ary PSK is
approximately twice the probability of error for M-ary PSK with
absolute phase encoding.

110
5.2.8 Differential PSK (DPSK) and Its
Performance
The received signal of a differentially encoded phase-modulated
signal in any given signaling interval is compared to the phase of
the received signal from the preceding signaling interval.
We demodulate the differentially encoded signal by multiplying
r(t) by cos2πfct and sin2πfct integrating the two products over
the interval T.
At the kth signaling interval, the demodulator output :
rk = ⎡⎣ ε s cos (θ k − φ ) + nk1 ε s sin (θ k − φ ) + nk2 ⎤⎦

or equivalently,
rk = ε s e j (θ −φ ) + nk
k

where θk is the phase angle of the transmitted signal at the kth


signaling interval, φ is the carrier phase, and nk=nk1+jnk2 is the
noise vector.
111
5.2.8 Differential PSK (DPSK) and Its
Performance
Similarly, the received signal vector at the output of the
demodulator in the preceding signaling interval is :
rk −1 = ε s e j (θ k −1 −φ ) + nk −1
The decision variable for the phase detector is the phase
difference between these two complex numbers. Equivalently,
we can project rk onto rk-1 and use the phase of the resulting
complex number :
rk rk*−1 = ε s e j (θ k −θ k −1 ) + ε s e j (θ k −φ ) *
n k −1 + ε s e − j (θ k −1 − φ )
nk + nk nk* −1

which, in the absence of noise, yields the phase differenceθk-


θk-1.
Differentially encoded PSK signaling that is demodulated and
detected as described above is called differential PSK (DPSK).
112
5.2.8 Differential PSK (DPSK) and Its
Performance
If the pulse g(t) is rectangular, the matched filter
may be replaced by integrate-and-dump filter

Block diagram of DPSK demodulator

113
5.2.8 Differential PSK (DPSK) and Its
Performance
The error probability performance of a DPSK demodulator and
detector
The derivation of the exact value of the probability of error for
M-ary DPSK is extremely difficult, except for M = 2.
Without loss of generality, suppose the phase difference
θk-θk-1=0. Furthermore, the exponential factor e-j(θk-1-φ)
and ej(θk-φ) can be absorbed into Gaussian noise components
nk-1 and nk, without changing their statistical properties.
rk rk*−1 = ε s + ε s (nk + nk*−1 ) + nk nk*−1
The complication in determining the PDF of the phase is the
term nkn*k-1.

114
5.2.8 Differential PSK (DPSK) and Its
Performance
However, at SNRs of practical interest, the term nkn*k-1 is
small relative to the dominant noise term ε s (nk+n*k-1).
We neglect the term nkn*k-1 and normalize rkr*k-1 by dividing
through by ε s , the new set of decision metrics becomes :
(
x = ε s + Re nk + nk* −1 )
(
y = Im nk + nk* −1 )
The variables x and y are uncorrelated Gaussian random
variable with identical variances σn2=N0. The phase is
−1 y
Θ r = tan
x
The noise variance is now twice as large as in the case of
PSK. Thus we can conclude that the performance of DPSK is
3 dB poorer than that for PSK.
115
5.2.8 Differential PSK (DPSK) and Its
Performance
This result is relatively good for M ≥4 , but it is pessimistic for M
= 2 that the loss in binary DPSK relative to binary PSK is less
than 3 dB at large SNR.
In binary DPSK, the two possible transmitted phase
differences are 0 and π rad. Consequently, only the real part
of rkr*k-1 is need for recovering the information.
( )
*
(
1
)
Re rk rk −1 = rk rk*−1 + rk*rk −1
2
Because the phase difference the two successive signaling
intervals is zero, an error is made if Re(rkr*k-1)<0.
The probability that rkr*k-1+r*krk-1<0 is a special case of a
derivation, given in Appendix B.

116
5.2.8 Differential PSK (DPSK) and Its
Performance

Appendix B concerned with the probability that a general


quadratic form in complex-valued Gaussian random Variable
is less than zero. According to Equation B-21, we find it
depend entirely on the first and second moments of the
complex-valued Gaussian random variables rk and rk-1.
We obtain the probability of error for binary DPSK in the form
1 −ε b N 0
Pb = e
2
where εb/N0 is the SNR per bit.

117
5.2.8 Differential PSK (DPSK) and Its
Performance
The probability of a binary digit error for four-phase DPSK with
Gray coding can be express in terms of well-known functions,
but it's derivation is quite involved.
According to Appendix C, it is expressed in the form :

(
⎡ 1 2
)
2 ⎤
Pb = Q1 (a , b ) − I 0 (ab ) exp ⎢ − a + b ⎥
1
2 ⎣ 2 ⎦
where Q1(a,b) is the Marcum Q function (2.1-122),
I0(x) is the modified Bessel function of order zero(2.1-123),
and the parameters a and b are defined as
⎛ 1⎞ ⎛ 1⎞
a = 2γ b ⎜⎜1 − ⎟ , and b = 2γ b ⎜1 +
⎟ ⎜

⎝ 2⎠ ⎝ 2 ⎟⎠

118
5.2.8 Differential PSK (DPSK) and Its
Performance

Because binary DPSK is only


slightly inferior to binary PSK
at large SNR, and DPSK does
not require an elaborate method
for estimate the carrier phase, it
is often used in digital
communication system.

Probability of bit error for binary and four-phase


PSK and DPSK
119
5.2.9 Probability of Error for QAM

Recall from Section 4.3 that QAM signal waveforms may be


expressed as (4.3-19)
s m (t ) = Amc g (t ) cos 2πf c t − Ams (t ) sin 2πf c t

where Amc and Ams are the information-bearing signal amplitudes


of the quadrature carriers and g(t) is the signal pulse.
The vector representation of these waveform is
⎡ ⎤
s m = ⎢ Amc
1
εg Ams
1
εg ⎥
⎣ 2 2 ⎦

To determine the probability of error for QAM, we must specify


the signal point constellation.
120
5.2.9 Probability of Error for QAM
QAM signal sets that have M = 4 points.

Figure (a) is a four-phase modulated signal and Figure (b) is


with two amplitude levels, labeled A1 and A2, and four phases.

Because the probability of error is dominated by the minimum


distance between pairs of signal points, let us impose the condition
(e)
that d min = 2 A and we evaluate the average transmitter power, based
on the premise that all signal points are equally probable.

121
5.2.9 Probability of Error for QAM
For the four-phase signal, we have
Pav = (4 )2 A 2 = 2 A 2
1
4
For the two-amplitude, four-phase QAM, we place the points
(e)
on circles of radii A and 3 A. Thus, d min = 2 A , and
[ ]
Pav = 2(3 A 2 ) + 2 A 2 = 2 A 2
1
2
which is the same average power as the M = 4-phase signal
constellation.
Hence, for all practical purposes, the error rate performance
of the two signal sets is the same.
There is no advantage of the two-amplitude QAM signal set over
M = 4-phase modulation.

122
5.2.9 Probability of Error for QAM
QAM signal sets that have M = 8 points.
We consider the four signal constellations :
Assuming that the signal
points are equally probable,
the average transmitted
signal power is :

Pav =
1 M

M m =1
( A 2
mc + Ams )
2

∑ (a )
A2 M
= 2
mc + a ms
2

M m =1

The coordinates (Amc,Ams) for each


signal point are normalized by A
123
5.2.9 Probability of Error for QAM

The two sets (a) and (c) contain signal points that fall on a
rectangular grid and have Pav = 6A2.
The signal set (b) requires an average transmitted power
Pav= 6.83A2, and (d) requires Pav = 4.73A2.
The fourth signal set (d) requires approximately 1 dB less
power than the first two and 1.6 dB less power than the third
to achieve the same probability of error.
The fourth signal constellation is known to be the best eight-
point QAM constellation because it requires the least power for
a given minimum distance between signal points.

124
5.2.9 Probability of Error for QAM
QAM signal sets for M ≥ 16
For 16-QAM, the signal points at a given
amplitude level are phase-rotated by relative
to the signal points at adjacent amplitude
levels.
However, the circular 16-QAM constellation
is not the best 16-point QAM signal
constellation for the AWGN channel.
Rectangular M-ary QAM signal are most frequently used in
practice. The reasons are :
Rectangular QAM signal constellations have the distinct
advantage of being easily generated as two PAM signals
impressed on phase-quadrature carriers.
125
5.2.9 Probability of Error for QAM
The average transmitted power required to achieve a given
minimum distance is only slightly greater than that of the best
M-ary QAM signal constellation.
For rectangular signal constellations in which M = 2k, where k is
even, the QAM signal constellation is equivalent to two PAM
signals on quadrature carriers, each having M = 2 k 2 signal
points.
The probability of error for QAM is easily determined from the
probability of error for PAM.
Specifically, the probability of a correct decision for the M-ary
QAM system is (
P = 1− P
c M
)
2

where p M
is the probability of error of an M − ary PAM.

126
5.2.9 Probability of Error for QAM

By appropriately modifying the probability of error for M-ary


PAM, we obtain
⎛ 1 ⎞ ⎛⎜ 3 ε av ⎞⎟
P M = 2⎜ 1 − ⎟Q ⎜
⎝ M ⎠ ⎝ M − 1 N 0 ⎟⎠
where εav/No is the average SNR per symbol.
Therefore, the probability of a symbol error for M-ary QAM is
(
PM = 1 − 1 − P M
)
2

Note that this result is exact for M = 2k when k is even.


When k is odd, there is no equivalent M -ary PAM system.
There is still ok, because it is rather easy to determine the error
rate for a rectangular signal set.

127
5.2.9 Probability of Error for QAM
We employ the optimum detector that
bases its decisions on the optimum
distance metrics (5.1-43), it is
relatively straightforward to show that
the symbol error probability is tightly
upper-bounded as
3ε av
2
⎡ ⎛ ⎞⎤

PM ≤ 1 − ⎢ 2Q ⎜ ⎟⎥
⎢⎣ ⎝ (M − 1)N 0

⎠ ⎥⎦
⎛ 3kε bav ⎞
≤ 4Q ⎜⎜ ⎟
⎝ ( M − 1)N 0


for any k ≥ 1, whereεbav/N0 is the
average SNR per bit.
128
5.2.9 Probability of Error for QAM

For nonrectangular QAM signal constellation, we may upper-


bound the error probability by use of a union bound :


[ ]
PM < (M − 1)Q ⎜ d min 2 N 0 ⎞⎟
⎛ (e ) 2

(e )
where d min is the minimum Euclidean distance between signal
points.
This bound may be loose when M is large.
We approximate PM by replacing M-1 by Mn, where Mn is the
(e )
largest number of neighboring points that are at distance d min
from any constellation point.
It is interesting to compare the performance of QAM with that
of PSK for any given signal size M.

129
5.2.9 Probability of Error for QAM

For M-ary PSK, the probability of a symbol error is approximate


as ⎛ π ⎞
PM ≈ 2Q ⎜ 2γ s sin ⎟
⎝ M⎠
where γs is the SNR per symbol.
For M-ary QAM, the error probability is :
⎛ 1 ⎞ ⎛⎜ 3 ε av ⎞⎟
P M = 2⎜ 1 − ⎟Q ⎜
⎝ M ⎠ ⎝ M − 1 N 0 ⎟⎠
We simply compare the arguments of Q function for the two
signal formats. The ratio of these two arguments is
3 (M − 1)
RM =
2 sin 2 (π M )

130
5.2.9 Probability of Error for QAM
We can find out that when M = 4, we have RM = 1.
4-PSK and 4-QAM yield comparable performance for the same
SNR per symbol.
For M>4, M-ary QAM yields better performance than M-ary
PSK.

131
5.2.10 Comparison of Digital
Modulation Methods
One can compare the digital modulation methods on the basis of
the SNR required to achieve a specified probability of error.
However, such a comparison would not be very meaningful,
unless it were made on the basis of some constraint, such as a
fixed data rate of transmission or, on the basis of a fixed
bandwidth.
For multiphase signals, the channel bandwidth required is
simply the bandwidth of the equivalent low-pass signal pulse g(t)
with duration T and bandwidth W, which is approximately equal
to the reciprocal of T.
R
Since T=k/R=(log2M)/R, it follows that W =
log 2 M

132
5.2.10 Comparison of Digital
Modulation Methods
As M is increased, the channel bandwidth required, when the bit
rate R is fixed, decreases. The bandwidth efficiency is measured
by the bit rate to bandwidth ratio, which is
R
= log 2 M
W
The bandwidth-efficient method for transmitting PAM is single-
sideband. The channel bandwidth required to transmit the signal
is approximately equal to 1/2T and,
R
= 2 log 2 M
W
this is a factor of 2 better than PSK.
For QAM, we have two orthogonal carriers, with each carrier
having a PAM signal.

133
5.2.10 Comparison of Digital
Modulation Methods
Thus, we double the rate relative to PAM. However, the QAM
signal must be transmitted via double-sideband. Consequently,
QAM and PAM have the same bandwidth efficiency when the
bandwidth is referenced to the band-pass signal.
As for orthogonal signals, if the M = 2k orthogonal signals are
constructed by means of orthogonal carriers with minimum
frequency separation of 1/2T, the bandwidth required for
transmission of k = log2M information bits is
M M M
W = = = R
2T 2(k R ) 2 log 2 M
In the case, the bandwidth increases as M increases.
In the case of biorthogonal signals, the required bandwidth is
one-half of that for orthogonal signals.
134
5.2.10 Comparison of Digital
Modulation Methods

A compact and meaningful


comparison of modulation
methods is one based on the
normalized data rate R/W (bits per
second per hertz of bandwidth)
versus the SNR per bit (εb/N0 )
required to achieve a given error
probability.
In the case of PAM, QAM, and
PSK, increasing M results in a
higher bit-to-bandwidth ratio R/W.

135
5.2.10 Comparison of Digital
Modulation Methods
However, the cost of achieving the higher data rate is an
increase in the SNR per bit.
Consequently, these modulation methods are appropriate for
communication channels that are bandwidth limited, where we
desire a R/W >1 and where there is sufficiently high SNR to
support increases in M.
Telephone channels and digital microwave ratio channels are
examples of such band-limited channels.
In contrast, M-ary orthogonal signals yield a R/W ≤ 1. As M
increases, R/W decreases due to an increase in the required
channel bandwidth.
The SNR per bit required to achieve a given error probability
decreases as M increases.
136
5.2.10 Comparison of Digital
Modulation Methods
Consequently, M-ary orthogonal signals are appropriate for
power-limited channels that have sufficiently large bandwidth to
accommodate a large number of signals.
As M→∞, the error probability can be made as small as desired,
provided that SNR>0.693 (-1.6dB). This is the minimum SNR
per bit required to achieve reliable transmission in the limit as
the channel bandwidth W→∞ and the corresponding R/W→0.
The figure above also shown the normalized capacity of the
band-limited AWGN channel, which is due to Shannon (1948).
The ratio C/W, where C (=R) is the capacity in bits/s, represents
the highest achievable bit rate-to-bandwidth ratio on this channel.
Hence, it serves the upper bound on the bandwidth efficiency of
any type of modulation.

137
5.4 Optimum Receiver For Signals with
Random Phase In AWGN Channel
In this section, we consider the design of the optimum receiver for
carrier modulated signals when the carrier phase is unknown and
no attempt is made to estimate its value.
Uncertainty in the carrier phase of the receiver signal may be due
to one or more of the following reasons:
The oscillators that are used at the transmitter and the receiver to generate
the carrier signals are generally not phase synchronous.
The time delay in the propagation of the signal from the transmitter to the
receiver is not generally known precisely.
Assuming a transmitted signal of the form
[ ]
s (t ) = Re g (t )e j 2πf ct
that propagates through a channel with delay t0 will be received as:
s ( t − t0 ) = Re ⎡ g ( t − t0 ) e c ( 0 ) ⎤ = Re ⎡⎣ g ( t − t0 ) e − j 2π fct0 e j 2π fc ⎤⎦
j 2π f t − t
⎣ ⎦
138
5.4 Optimum Receiver For Signals with
Random Phase In AWGN Channel
The carrier phase shift due to the propagation delay t0 is
φ = −2πf ct0
Note that large changes in the carrier phase can occur due to
relatively small changes in the propagation delay.
For example, if the carrier frequency fc=1 MHz, an uncertainty
or a change in the propagation delay of 0.5μs will cause a
phase uncertainty of π rad.
In some channels the time delay in the propagation of the signal
from the transmitter to the receiver may change rapidly and in
an apparently random fashion.
In the absence of the knowledge of the carrier phase, we may
treat this signal parameter as a random variable and determine
the form of the optimum receiver for recovering the transmitted
information from the received signal.
139
5.4.1 Optimum Receiver for Binary Signals
We consider a binary communication system that uses the two
carrier modulated signal s1(t) and s2(t) to transmit the information,
where:
sm ( t ) = Re ⎡⎣ slm ( t ) e j 2π fct ⎤⎦ , m = 1, 2, 0 ≤ t ≤ T
and slm(t), m=1,2 are the equivalent low-pass signals.
The two signals are assumed to have equal energy

4.1-22 4.1-24
2
ε = ∫ s (t )dt = ∫ sml (t ) dt
T
2 1 T
m
0 2 0
The two signals are characterized by the complex-valued
correlation coefficient
ρ12 ≡ ρ = ∫ sl1 (t )sl 2 (t )dt
1 T *
2ε 0
140
5.4.1 Optimum Receiver for Binary Signals

The received signal is assumed to be a phase-shifted version of


the transmitted signal and corrupted by the additive noise

{ } [
n (t ) = Re [nc (t ) + jns (t )]e j 2πf ct = Re z (t )e j 2πf ct ]
The received signal may be expressed as

{[ ]
r (t ) = Re slm (t )e jφ + z (t ) e j 2πf ct }
where rl ( t ) = slm ( t ) e jφ + z ( t ) , 0 ≤ t ≤ T
rl(t) is the equivalent low-pass received signal.
This received signal is now passed through a demodulator whose
sampled output at t =T is passed to the detector.

141
5.4.1 Optimum Receiver for Binary Signals
The optimum demodulator
In section 5.1.1, we demonstrated that if the received signals
was correlated with a set of orthogonal functions {fn(t)} that
spanned the signal space, the outputs from the bank of
correlators provide a set of sufficient statistics for the detector
to make a decision that minimizes the probability error.
We also demonstrated that a bank of matched filters could be
substituted for the bank of correlations.
A similar orthogonal decomposition can be employed for a
received signal with an unknown carrier phase.
It is mathematically convenient to deal with the equivalent
low-pass signal and to specify the signal correlators or
matched filters in terms of the equivalent low-pass signal
waveforms.
142
5.4.1 Optimum Receiver for Binary Signals

The optimum demodulator


The impulse response hl(t) of a filter that is matched to the
complex-valued equivalent low-pass signal sl(t), 0≤t≤T is given
as (from Problem 5.6) hl(t)=sl*(T-t) and the output of such
2

∫0 sl (t ) dt =where
2ε ε is the
T
filter at t=T is simply (4.1-24)
signal energy.
A similar result is obtained if the signal sl(t) is correlated with
sl*(t) and the correlator is sampled signal sl(t) at t=T.
The optimum demodulator for the equivalent low-pass
received signal sl(t) may be realized by two matcher filters in
parallel, one matched to sl1(t) and the other to sl2(t).

143
5.4.1 Optimum Receiver for Binary Signals

The optimum demodulator


Optimum receiver for binary signals

The output of the matched filters or correlators at the sampling


instant are the two complex numbers
rm = rmc + jrms , m = 1,2
144
5.4.1 Optimum Receiver for Binary Signals

The optimum demodulator


Suppose that the transmitted signal is s1(t). It can be shown
that (Problem 5.41)
r1 = 2ε cos φ + n1c + j (2ε sin φ + n1s ).
r2 = 2ε ρ cos(φ − α 0 ) + n2 c + j [2ε ρ sin (φ − α 0 ) + n2 s ]
where ρ is the complex-valued correlation coefficient of two
signals sl1(t) and sl2(t), which may be expressed as
ρ=|ρ|exp(jα0).
The random noise variables n1c, n1s, n2c, and n2s are jointly
Gaussian, with zero-mean and equal variance.

145
5.4.1 Optimum Receiver for Binary Signals

The optimum detector


The optimum detector observes the random variables [r1c r1s
r2c r2s]=r, where r1=r1c+jr1s, and r2=r2c+jr2s, and bases its
decision on the posterior probabilities P(sm|r), m=1,2.
These probabilities may be expressed as
p (r | sm )P (sm )
P (sm | r ) = , m = 1,2
p (r ) Likelihood ratio.

The optimum decision rule may be expressed as


s1
> P ( s 2 | r ) or p ( r | s1 ) >s1
P (s2 )
P ( s1 | r ) < <
s2
p ( r | s 2 ) s2 P ( s1 )

146
5.4.1 Optimum Receiver for Binary Signals

The optimum detector


The ratio of PDFs on the left-hand side is the likelihood ratio,
which we denote as
p (r | s1 )
Λ (r ) =
p (r | s 2 )
The right-hand side is the ratio of the two prior probabilities,
which takes the value of unity when the two signals are
equally probable.
The probability density functions p(r|s1) and p(r|s2) can be
obtained by averaging the PDFs p(r|sm,φ) over the PDF of the
random carrier phase, i.e.,

p (r | s m ) = ∫ p (r | s m , φ ) p (φ )dφ
0

147
5.4.1 Optimum Receiver for Binary Signals

The optimum detector


For the special case in which the two signals are orthogonal,
i.e., ρ=0, the outputs of the demodulator are (from 5.4-10):
r1 = r1c + jr1s
= 2ε cos φ + n1c + j (2ε sin φ + n1s )
r2 = r2 c + jr2 s
= n2 c + jn2 s
where (n1c, n1s, n2c, n2s) are mutually uncorrelated and,
statistically independent, zero-mean Gaussian random variable.

148
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
The joint PDF of r=[r1c r1s r2c r2s] may be expressed as a
product of the marginal PDFs:
1 ⎡ (r1c − 2ε cos φ ) 2 + (r1s − 2ε sin φ ) 2 ⎤
p (r1c , r1s | s1 , φ ) = exp ⎢ − ⎥
2πσ 2
⎣ 2 σ 2

1 ⎛ r22c + r22s ⎞
p (r2 c , r2 s | s1 , φ ) = exp ⎜ − ⎟
2πσ 2
⎝ 2 σ 2

where σ2=2εN0.
The uniform PDF for the carrier phase φ (p(φ)=1/2π)
represents the most ignorance that can be exhibited by the
detector.
This is called the least favorable PDF for φ.
149
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
With p(φ)=1/2π, 0≤φ≤2π, substituted into the integral in
p(r|sm), we obtain:
1 2π
∫ p ( r1c , r1s | s1 , φ )d φ
2π 0

1 ⎛ r12c + r12c + 4ε 2 ⎞ 1 2π ⎡ 2ε ( r1c cos φ + r1s sin φ ) ⎤


=
2πσ 2
exp ⎜ −
⎝ 2σ 2 ⎟
⎠ 2π

0
exp ⎢
⎣ σ 2 ⎥d φ

1 2π ⎡ 2ε ( r1c cos φ + r1s sin φ ) ⎤ ⎞ ⎛ 2ε r 2 + r 2


∫0 φ⎟= ⎜ 1c 1s
exp ⎢ ⎥d I
2π ⎣ σ 2
⎦ ⎟ 0
⎜ σ 2
⎠ ⎝
where I0(x) is the modified Bessel function of zeroth order.

150
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
By performing a similar integration under the assumption
that the signal s2(t) was transmitted, we obtain the result
1 ⎛ r22c + r22s + 4ε 2 ⎞ ⎛ 2ε r22c + r22s ⎞
p ( r2 c , r2 s | s 2 ) = exp ⎜ − ⎟ I 0 ⎜⎜ ⎟
2πσ 2
⎝ 2σ 2
⎠ ⎝ σ 2


When we substitute these results into the likelihood ratio
given by equation Λ(r), we obtain the result
(
Λ (r ) =
)
I 0 2ε r12c + r12s σ 2 s1 P ( s )
>
( ) <
I 0 2ε r2 c + r2 s σ s2 P ( s1 )
2 2 2
2

?
The optimum detector computes the two envelopes r12c + r12s and r22c + r22s
(
and the corresponding values of the Bessel function I 2ε r12c + r12s / σ 2 )
( )
and I 2ε r22c + r22s / σ 2 to form the likelihood ratio.
0
0

151
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
we observe that this computation requires knowledge of the
noise variance σ2.
The likelihood ratio is then compared with the threshold
P(s2)/P(s1) to determine which signal was transmitted.
A significant simplification in the
implementation of the optimum detector
occurs when the two signals are equally
probable. In such a case the threshold
becomes unity, and, due to the
monotonicity of Bessel function shown
in figure, the optimums detection rule
1
simplifies to r12c + r12s <> r22c + r22s
s2

152
5.4.1 Optimum Receiver for Binary Signals

The optimum detector


The optimum detector bases its decision on the two envelopes
and r12c + r12s , and r22c + r22s , it is called an envelope detector.
We observe that the computation of the envelopes of the
received signal samples at the output of the demodulator
renders the carrier phase irrelevant in the decision as to which
signal was transmitted.
Equivalently, the decision may be based on the computation
of the squared envelope r1c2+r1s2 and r2c2+r2s2, in which case
the detector is call a square-law detector.

153
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
Recall that in binary FSK we employ two different
frequencies, say f1 and f2=f1+Δf, to transmit a binary
information sequence.
The chose of minimum frequency separation Δf = f2 - f1 is
considered below:
s1 ( t ) = 2ε b Tb cos 2π f1t , s2 ( t ) = 2ε b Tb cos 2π f 2t , 0 ≤ t ≤ Tb
The equivalent low-pass counterparts are:
sl1 ( t ) = 2ε b Tb , sl 2 ( t ) = 2ε b Tb e j 2πΔft , 0 ≤ t ≤ Tb

The received signal may be expressed as:


2ε b
r (t ) = cos ( 2π f mt + φm ) + n ( t ) , m = 0,1
Tb
154
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
Demodulation and square-law detection:

2
f kc ( t ) = cos ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ , k = 0,1
Tb
2
f ks ( t ) = sin ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ , k = 0,1
Tb

155
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
If the mth signal is transmitted, the four samples at the detector
may be expressed as:
⎧ 2ε b
T ⎪ ⎪⎫ ⎪⎧ 2 ⎪⎫
rkc = ∫ r ( t )i f kc ( t ) dt = ∫ ⎨ cos ⎡⎣ 2π ( f1 + mΔf ) t + φm ⎤⎦ + n ( t ) ⎬i⎨ cos ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ ⎬ dt
T b b

⎩⎪ Tb ⎭⎪ ⎩⎪ Tb ⎭⎪
0 0

2 εb
∫ {cos ⎡⎣2π ( f }{ }
+ mΔf ) t ⎤⎦ cos φm − sin ⎡⎣ 2π ( f1 + mΔf ) t ⎤⎦ sin φm i cos ⎡⎣ 2π ( f1 + k Δf ) t ⎤⎦ dt + nkc
Tb
= 1
Tb 0

εb
{cos ⎡⎣2π ( m − k ) Δft ⎤⎦ + cos ⎡⎣2π ( 2 f + ( m + k ) Δf ) t ⎤⎦} cos φ −
Tb
=
Tb ∫ 0 1 m

{sin ⎡⎣2π ( m − k ) Δft ⎤⎦ − sin ⎣⎡2π ( 2 f + ( m + k ) Δf ) t ⎦⎤} sin φ dt + n


1 m kc

ε b ⎧⎪ sin ⎡⎣ 2π ( m − k ) Δft ⎤⎦ cos ⎡⎣ 2π ( m − k ) Δft ⎤⎦


Tb
⎫⎪
= ⎨ cos φm + sin φm ⎬ + nkc
Tb ⎪⎩ 2π ( m − k ) Δf 2π ( m − k ) Δf ⎪⎭0
⎪⎧ sin ⎡⎣ 2π ( m − k ) ΔfTb ⎤⎦ cos ⎡⎣ 2π ( m − k ) ΔfTb ⎤⎦ − 1 ⎪⎫
= εb ⎨ cos φm + sin φm ⎬ + nkc k , m = 0,1
⎩⎪ 2 π ( m − k ) b
Δ fT 2 π ( m − k ) b
Δ fT ⎭⎪
156
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
If the mth signal is transmitted, the four samples at the detector
may be expressed as (cont.):
⎧ 2ε b
T ⎪ ⎪⎫ ⎪⎧ 2 ⎪⎫
rks = ∫ r ( t )i f ks ( t ) dt = ∫ ⎨ cos ⎡⎣ 2π ( f1 + mΔf ) t + φm ⎤⎦ + n ( t ) ⎬i⎨ sin ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ ⎬ dt
T b b

0 0
⎪⎩ Tb ⎭⎪ ⎩⎪ Tb ⎭⎪
2 εb
∫ {cos ⎡⎣2π ( f }{ }
+ mΔf ) t ⎤⎦ cos φm − sin ⎡⎣ 2π ( f1 + mΔf ) t ⎤⎦ sin φm i sin ⎡⎣ 2π ( f1 + k Δf ) t ⎤⎦ dt + nks
Tb
= 1
Tb 0

εb
{− sin ⎡⎣2π ( m − k ) Δft ⎤⎦ + sin ⎡⎣2π ( 2 f + ( m + k ) Δf ) t ⎤⎦} cos φ −
Tb
=
Tb ∫ 0 1 m

{cos ⎡⎣2π ( m − k ) Δft ⎤⎦ − cos ⎡⎣2π ( 2 f + ( m + k ) Δf ) t ⎤⎦} sin φ dt + n


1 m ks

ε b ⎧⎪ cos ⎡⎣ 2π ( m − k ) Δft ⎤⎦ sin ⎡⎣ 2π ( m − k ) Δft ⎤⎦


Tb
⎫⎪
= ⎨ cos φm − sin φm ⎬ + nks
Tb ⎪⎩ 2π ( m − k ) Δf 2π ( m − k ) Δf ⎪⎭0
⎧⎪ cos ⎡⎣ 2π ( m − k ) ΔfTb ⎤⎦ − 1 sin ⎡⎣ 2π ( m − k ) ΔfTb ⎤⎦ ⎫⎪
= εb ⎨ cos φm − sin φm ⎬ + nks k , m = 0,1
⎩⎪ 2 π ( m − k ) b
Δ fT 2 π ( m − k ) b
Δ fT ⎭⎪
157
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
We observe that when k=m, the sampled values to the
detector are
rmc = ε b cos φm + nmc
k=m
rms = − ε b sin φm + nms
We observe that when k≠m, the signal components in the
samples rkc and rks will vanish, independently of the values
of the phase shifts φk, provided that the frequency
separation between successive frequency is Δf= 1/T.
In such case, the other two correlator outputs consist of
noise only, i.e.,
rkc=nkc, rks=nks, k≠m.

158
5.4.2 Optimum Receiver for M-ary
Orthogonal Signals
If the equal energy and equally probable signal waveforms are
represented as
sm ( t ) = Re ⎡⎣slm ( t ) e j 2π fct ⎤⎦ , m = 1, 2,..., M , 0 ≤ t ≤ T
where slm(t) are the equivalent low-pass signals.
The optimum correlation-type or matched-filter-type
demodulator produces the M complex-valued random variables
rm = rmc + jrms = ∫ rl ( t ) hlm (T − t ) dt
T
*
0

= ∫ rl ( t ) s (T − (T − t ) ) dt = ∫ rl ( t ) slm ( t ) dt , m = 1, 2,..., M
T T
* *
0 lm 0
where rl(t) is the equivalent low-pass signals.
The optimum detector, based on a random, uniformly distributed
carrier phase, computes the M envelopes
rm = rmc2 + rms2 , m = 1, 2,..., M
or, the squared envelopes |rm|2, and selects the signal with the
largest envelope.
159
5.4.2 Optimum Receiver for M-ary
Orthogonal Signals
Optimum receiver for
M-ary orthogonal
FSK signals.
There are 2M
correlators: two for
each possible
transmitted frequency.
The minimum
frequency separation
between adjacent
frequencies to
maintain
orthogonality is Δf=
1/T. Demodulation of M-ary signals for
noncoherent detection
160
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
We assume that the M signals are equally probable a priori and
that the signal s1(t) is transmitted in the signal interval 0≦t≦T.
The M decision metrics at the detector are the M envelopes

rm = rmc2 + rms2 , m = 1, 2,..., M

where r1c = ε s cos φ1 + n1c


r1s = ε s sin φ1 + n1s
and rmc = nmc , rms = nms , m = 2,3,..., M
The additive noise components {nmc} and {nms} are mutually
statistically independent zero-mean Gaussian variables with equal
variance σ2=N0/2.

161
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The PDFs of the random variables at the input to the detector are
⎛ ε (r2 + r2 ) ⎞
1 ⎛ r1c + r1s + ε s ⎞ ⎜ s 1c 1s ⎟
2 2
pr1 ( r1c , r1s ) = exp ⎜ − ⎟ I0 ⎜
2πσ 2
⎝ 2σ 2
⎠ ⎜ σ 2 ⎟⎟
⎝ ⎠
1 ⎛ rmc2 + rms2 ⎞
prm ( rmc , rms ) = exp ⎜ − ⎟, m = 2,3,..., M
2πσ 2
⎝ 2σ ⎠
2

We define the normalized variables


rmc2 + rms2
Rm =
σ
rms
−1
Θ m = tan
rmc

162
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
Clearly, rmc=σRmcosΘm and rms =σRmsinΘm. The Jacobian of
this transformation is
σ cos Θ m σ sin Θ m
J = = σ 2 Rm
−σ Rm sin Θ m σ Rm cos Θ m
Consequently,
R1 ⎡ 1⎛ 2 ε s ⎞ ⎤ ⎛ 2ε s ⎞
p( R1 , Θ1 ) = exp ⎢ − ⎜ R1 + 2 ⎟ ⎥ I 0 ⎜⎜ R1 ⎟⎟
2π ⎣ 2⎝ N0 ⎠⎦ ⎝ N0 ⎠
Rm ⎛ 1 2⎞
p ( Rm , Θ m ) = exp ⎜ − Rm ⎟ , m = 2,3,..., M
2π ⎝ 2 ⎠
Finally, by averaging p(Rm,Θm) over Θm, the factor of 2π is
eliminated.
Thus, we find that R1 has a Rice probability distribution and Rm,
m=2,3,…,M, are each Rayleigh-distribued.
163
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The probability of a correct decision is simply the probability that
R1>R2, and R1>R3,…,and R1>Rm. Hence,
Pc = P ( R2 < R1 , R3 < R1 ,… , RM < R1 )

= ∫ P ( R2 < R1 , R3 < R1 ,… , RM < R1 | R1 = x ) pR1 ( x ) dx
0

Because the random variables Rm, m=2,3,…,M, are statistically


independent and identically distributed, the joint probability
conditioned on R1 factors into a product of M-1 identical terms.
∞ M −1
where Pc = ∫ ⎡⎣ P ( R2 < R1 | R1 = x ) ⎤⎦ pR1 ( x ) dx (5.4-42)
0

P ( R2 < R1 | R1 = x ) = ∫ pR2 ( r2 ) dr2 = 1 − e


x
− x2 / 2
(From Eq 2.1-129)
0

164
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The (M-1)th power may be expressed as
⎛ M − 1⎞ − nx2 2
( )
M −1 M −1
1− e = ∑ ( −1) ⎜
− x2 / 2 n
⎟e
n =0 ⎝ n ⎠
Substitution of this result into Equation 5.4-42 and integration
over x yield the probability of a correct decision as
M −1
n ⎛ M − 1⎞ 1 ⎡ nε s ⎤
Pc = ∑ ( −1) ⎜ ⎟ exp ⎢ ⎥
n =0 ⎝ n ⎠ n +1 ⎣ ( n + 1) N 0 ⎦
where εs/N0 is the SNR per symbol.
The probability of a symbol error, which is Pm=1-Pc, becomes
M −1
⎛ M − 1⎞ 1 ⎡ nkε b ⎤
PM = ∑ ( −1) ⎜
n +1
⎟ exp ⎢ − ⎥
n =1 ⎝ n ⎠ n +1 ⎣ ( n + 1) N 0 ⎦
where εb/N0 is the SNR per bit.
165
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
For binary orthogonal signals (M=2),
Equation reduces to the simple form
P2 = 12 e − ε b 2 N 0
For M>2, we may compute the
probability of a bit error by making
use of the relationship
2k −1
Pb = k PM
2 −1
which was established in Section 5.2.
Figure shows the bit-error
probability as a function of SNR per
bit γb for M=2,4,8,16, and 32.

166
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
Just as in the case of coherent detection of M-ary orthogonal
signals, we observe that for any given bit-error probability, the
SNR per bit decreases as M increase.
It will be show in Chapter 7 that, in the limit as M→∞, the
probability of bit error Pb can be made arbitrarily small provided
that the SNR per bit is greater than the Shannon limit of -1.6dB.
The cost for increasing M is the bandwidth required to transmit
the signals.
For M-ary FSK, the frequency separation between adjacent
frequencies is Δf=1/T for signal orthogonality.
The bandwidth required for the M signals is W=M Δf=M/T.

167
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The bit rate is R=k/T, where k=log2M.
Therefore, the bit rate-to-bandwidth ratio is

R log 2 M
=
W M

168
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
In this section, we consider the performance of the envelope
detector for binary, equal-energy correlated signals.
When the two signals are correlated, the input to the detector is
the complex-valued random variables given by Equation 5.4-10.
We assume that the detector bases its decision on the envelopes
|r1| and |r2|, which are correlated (statistically dependent).
The marginal PDFs of R1=|r1| and R2=|r2| are Ricean distributed
and may be expressed as
⎧ Rm ⎛ Rm2 + β m2 ⎞ ⎛ β m Rm ⎞
⎪ exp ⎜ − ⎟ I0 ⎜ ⎟ ( Rm > 0 )
p ( Rm ) = ⎨ 2ε s N 0 ⎝ 4ε N 0 ⎠ ⎝ 2ε N 0 ⎠
⎪0
⎩ ( Rm < 0 )
m=1,2, where β1=2ε and β2=2ε|ρ|, based on the
assumption that signal s1(t) was transmitted.
169
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
Since R1 and R2 are statistically dependent as a consequence of
the nonorthogonality of the signals, the probability of error may
be obtained by evaluating the double integral
∞ ∞
Pb = P ( R2 > R1 ) = ∫ ∫ p ( x , x )dx dx
1 2 1 2
0 x1

where p(x1,x2) is the joint PDF of the envelopes R1 and R2.


This approach was first used by Helstrom (1995), who
determined the joint PDF of R1 and R2 and evaluated the double
integral.
An alternative approach is based on the observation that the
probability of error may also be expressed as
Pb = P ( R2 > R1 ) = P ( R22 > R12 ) = P ( R22 − R12 > 0 )
170
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
But R22-R12 is a special case of a general quadratic form in
complex-valued Gaussian random variable, treated later in
Appendix B.
For the special case under consideration, the derivation yields
the error probability in the form

Pb = Q1 ( a, b ) − e
1
2
− ( a 2 +b2 ) 2
I 0 ( ab )

where
a=
εb
2N (1−
0
1− ρ
2
) b=
εb
2N (1+
0
1− ρ
2
)
Q1(a, b) is the Marcum Q function defined in 2.1-123 and I0(x)
is the modified Bessel function of order zero.
171
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
The error probability Pb is illustrated in Figure for several
values of |ρ|.

Probability of
error for
noncoherent
detection of
binary FSK

Pb is minimized when ρ=0; that is, when the signals are


orthogonal.
172
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
For this case, a = 0, b = ε b N 0, and Equation reduces to
⎛ ε b ⎞ 1 −εb
Pb = Q1 ⎜⎜ 0, ⎟⎟ − 2 e
2 N0

⎝ N0 ⎠
From the definition of Q1(a, b) in Equation 2.1-123, it follows
that
⎛ ε b ⎞ −εb 2 N0
Q1 ⎜⎜ 0, ⎟⎟ = e
⎝ N0 ⎠

Substitution of these relations yields the desired result given


previously in Equation 5.4-47. On the other hand, when |ρ|=1,
the error probability becomes Pb=1/2, as expected.

173

You might also like