Professional Documents
Culture Documents
Chapter 5
Optimum Receivers for the Additive
White Gaussian Noise Channel
r (t ) = sm (t ) + n(t ) , 0 ≤ t ≤ T
where n(t) denotes a sample function of AWGN process with
power spectral density Φnn(f )=½N0 W/Hz.
2
5.1 Optimum Receiver for Signals Corrupted
by Additive White Gaussian Noise
Our object is to design a receiver that is optimum in the sense
that it minimizes the probability of making an error.
It is convenient to subdivide the receiver into two parts—the
signal demodulator and the detector.
4
5.1.1 Correlation Demodulator
We describe a correlation demodulation that decomposes
the receiver signal and the noise into N-dimensional
vectors. In other words, the signal and the noise are
expanded into a series of linearly weighted orthonormal
basis functions {fn(t)}.
It is assumed that the N basis function {f n(t)}span the
signal space, so every one of the possible transmitted
signals of the set {sm(t)=1≤m≤M } can be represented as a
linear combination of {f n(t)}.
In case of the noise, the function {f n(t)} do not span the
noise space. However we show below that the noise terms
that fall outside the signal space are irrelevant to the
detection of the signal.
5
5.1.1 Correlation Demodulator
Suppose the receiver signal r(t) is passed through a parallel bank
of N basis functions {f n(t)}, as shown in the following figure:
[ sm (t ) + n (t ) ] f k (t ) dt
T T
∫0
r (t ) f k (t ) dt = ∫
0
⇒ rk = smk + nk , k = 1, 2,..... N
T
smk = ∫ sm (t ) f k (t )dt , k = 1, 2,......N
0
T
nk = ∫ n(t ) f k (t )dt , k = 1, 2,.......N
0
0
Power spectral density
is Φnn(f)=½N0 W/Hz
and their covariances are:
Autocorrelation
∫ E [ n(t )n(τ )] f
T T
E (nk nm ) = ∫ k (t ) f m (τ )dtdτ Conclusion: The N noise
0 0
Components {nk} are
1 T T
= N 0 ∫ ∫ δ (t − τ ) f k (t ) f m (τ )dtdτ zero-mean uncorrelated
2 0 0 Gaussian random
1 T 1 variables with a common
= N 0 ∫ f k (t ) f m (t )dt = N 0δ mk varianceσn2=½N0.
2 0 2
8
5.1.1 Correlation Demodulator
From the above development, it follows that the correlator
output {rk} conditioned on the mth signal being transmitted are
Gaussian random variables with mean
9
5.1.1 Correlation Demodulator
The conditional probability density functions of the random
variables r=[r1 r2 … rN] are:
N
p (r | s m ) = ∏ p (rk | smk ) , m = 1,2,...., M ----(A)
k =1
1 ⎡ (rk − smk ) 2 ⎤
p (rk | smk ) = exp ⎢− ⎥, k = 1,2,....., N ---(B)
πN 0 ⎣ N0 ⎦
1 ⎡ N (rk − smk ) 2 ⎤
p (r|s m ) = exp ⎢ −∑ ⎥ , m = 1, 2,......, M
(π N 0 ) N 2
⎣ k =1 N0 ⎦
10
5.1.1 Correlation Demodulator
The correlator outputs (r1 r2 … rN) are sufficient statistics for
reaching a decision on which of the M signals was transmitted,
i.e., no additional relevant information can be extracted from the
remaining noise process n'(t).
Indeed, n'(t) is uncorrelated with the N correlator outputs {rk}:
⎧⎪ ⎡ N ⎤ ⎫⎪
E [ n′(t )rk ] = E [ n′(t ) ] smk + E [ n′(t )nk ] = E [ n′(t )nk ] = E ⎨ ⎢ n(t ) − ∑ n j f j (t ) ⎥ nk ⎬
⎩⎪ ⎣ j =1 ⎦ ⎭⎪
N
= ∫ E [ n(t )n(τ ) ] f k (τ )dτ − ∑ E (n j nk ) f j (t )
T
nk = ∫ n (t ) f k (t ) dt
T
0 0
j =1
N
1 1
N 0δ ( t − τ ) f k (τ )dτ − ∑ N 0δ jk f j (t )
T
=∫
0 2 j =1 2
1 1
= N 0 f k (t ) − N 0 f k (t ) = 0 Q.E.D.
2 2
11
5.1.1 Correlation Demodulator
Since n' (t) and {rk} are Gaussian and uncorrelated, they are also
statistically independent.
Consequently, n'(t) does not contain any information that is
relevant to the decision as to which signal waveform was
transmitted.
All the relevant information is contained in the correlator outputs
{rk} and, hence, n'(t) can be ignored.
12
5.1.1 Correlation Demodulator
Example 5.1-1.
Consider an M-ary baseband PAM signal set in which the basic
pulse shape g(t) is rectangular as shown in following figure.
The additive noise is a zero-mean white Gaussian noise process.
Let us determine the basis function f(t) and the output of the
correlation-type demodulator.
The energy in the rectangular pulse is
ε =∫
T T
g (t )dt = ∫0 a dt = a T
2 2 2
g 0
13
5.1.1 Correlation Demodulator
Example 5.1-1.(cont.)
Since the PAM signal set has dimension N=1, there is only
one basis function f(t) given as:
1 ⎧1 T (0 ≤ t ≤ T )
f (t ) = g (t ) = ⎨
a 2T ⎩0 (otherwise)
The output of the correlation-type demodulator is:
T 1 T
r = ∫ r (t ) f (t )dt = ∫ r (t )dt
0 0
T
The correlator becomes a simple integrator when f(t) is
rectangular: if we substitute for r(t), we obtain:
r=
1
T ∫{
0
T
[ s m (t ) + n (t ) ] } dt =
1 ⎡ T
T ⎢⎣ ∫0
s m ( t ) dt +
T
∫0 n ( t ) dt ⎤
⎥⎦
= sm + n
14
5.1.1 Correlation Demodulator
Example 5.1-1.(cont.)
The noise term E(n)=0 and:
⎡1 T T ⎤
σ = E ⎡⎣ n ( t ) n ( t ) ⎤⎦ = E ⎢ ∫ ∫ n(t )n(τ )dtdτ ⎥
2
n
⎣T 0 0 ⎦
1 T T N0 T T
= ∫ ∫ E [ n(t )n(τ ) ] dtdτ = ∫ ∫ δ (t − τ )dtdτ
T 0 0 2T 0 0
N0 T 1
=
2T ∫0
1 ⋅ dτ = N 0
2
The probability density function for the sampled output is:
1 ⎡ (r − sm ) 2 ⎤
p(r | sm ) = exp ⎢ − ⎥
πN 0 ⎣ N 0 ⎦
15
5.1.2 Matched-Filter Demodulator
Instead of using a bank of N correlators to generate the variables
{rk}, we may use a bank of N linear filters. To be specific, let us
suppose that the impulse responses of the N filters are:
hk (t ) = f k (T − t ) , 0≤t ≤T
where {fk(t)} are the N basis functions and hk(t)=0 outside of the
interval 0 ≤ t ≤ T.
The outputs of these filters are :
yk (t ) = r ( t ) ∗ hk ( t )
t
= ∫ r (τ )hk (t − τ )dτ
0
t
= ∫ r (τ ) f k (T − t + τ )dτ , k = 1, 2,......, N
0
16
5.1.2 Matched-Filter Demodulator
If we sample the outputs of the filters at t = T, we obtain
T
yk (T ) = ∫ r (τ ) f k (τ )dτ = rk , k = 1, 2,....., N
0
A filter whose impulse response h(t) = s(T–t), where s(t) is
assumed to be confined to the time interval 0 ≤ t ≤ T, is called
matched filter to the signal s(t).
An example of a signal and its matched filter are shown in the
following figure.
17
5.1.2 Matched-Filter Demodulator
The response of h(t) = s(T–t) to the signal s(t) is:
y (t ) = s ( t ) ∗ h ( t ) = ∫ s (τ ) h ( t − τ ) dτ = ∫ s (τ ) s (T − t + τ ) dτ
t t
0 0
18
5.1.2 Matched-Filter Demodulator
Matched filter demodulator that generates the observed
variables {rk}
19
5.1.2 Matched-Filter Demodulator
Properties of the matched filter.
If a signal s(t) is corrupted by AWGN, the filter with an
impulse response matched to s(t) maximizes the output
signal-to-noise ratio (SNR).
Proof: let us assume the receiver signal r(t) consists of the
signal s(t) and AWGN n(t) which has zero-mean and
1
Φ nm ( f ) = N 0 W/Hz.
2
Suppose the signal r(t) is passed through a filter with
impulse response h(t), 0≤t≤T, and its output is sampled
at time t=T. The output signal of the filter is:
t
y (t ) = s (t ) ∗ h(t ) = ∫ r (τ )h(t − τ )dτ
0
t t
= ∫ s (τ )h(t − τ )dτ + ∫ n(τ )h(t − τ )dτ
0 0
20
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
At the sampling instant t=T:
T T
y (T ) = ∫ s (τ ) h (T − τ ) dτ + ∫ n (τ ) h (T − τ ) dτ
0 0
= y s (T ) + yn (T )
This problem is to select the filter impulse response that
maximizes the output SNR0 defined as:
y s2 (T )
SNR 0 =
[ ]
E y n2 (T )
E ⎡⎣ yn (T ) ⎤⎦ = ∫ ∫ E [ n(τ )n(t ) ] h(T − τ )h(T − t )dtdτ
T T
2
0 0
1 T T 1 T
= N 0 ∫ ∫ δ (t − τ )h(T − τ )h(T − t )dtdτ = N 0 ∫ h 2 (T − t )dt
2 0 0 2 0
21
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
By substituting for ys (T) and E ⎡⎣ yn2 (T ) ⎤⎦ into SNR0.
τ ' = T −τ
2 2
⎡ s (τ ) h(T − τ ) dτ ⎤
T
⎡ h (τ ') s (T − τ ') dτ '⎤
T
⎢⎣ ∫0 ⎥⎦ ⎢⎣ ∫0 ⎥⎦
SNR 0 = =
1 T 1 T
N 0 ∫ h (T − t ) dt
2
N 0 ∫ h 2 (T − t ) dt
2 0 2 0
22
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
Cauchy-Schwarz inequality: if g1(t) and g2(t) are finite-
energy signals, then
2
⎡ g (t ) g (t ) dt ⎤ ≤ ∞ g 2 (t ) dt ∞ g 2 (t ) dt
∞
⎢⎣ ∫− ∞ 1 2 ⎥⎦ ∫−∞ 1 ∫−∞ 2
with equality when g1(t)=Cg2(t) for any arbitrary constant
C.
If we set g1(t)=h1(t) and g2(t)=s(T−t), it is clear that the
SNR is maximized when h(t)=Cs(T–t).
23
5.1.2 Matched-Filter Demodulator
Proof: (cont.)
The output (maximum) SNR obtained with the matched
filter is:
2 2
⎡ s (τ ) h(T − τ ) dτ ⎤
T
⎡ s (τ )Cs (T − (T − τ ) ) dτ ⎤
T
⎢ ∫0
⎣ ⎥
⎦ 2 ⎣⎢ ∫0 ⎦⎥
SNR 0 = =
N 0 ∫ C 2 s 2 (T − (T − t ) ) dt
1 T N0 1 T
N 0 ∫ h 2 (T − t ) dt
2 0 2 0
2 T 2 2ε
=
N0 ∫0
s (t ) dt =
N0
Note that the output SNR from the matched filter depends
on the energy of the waveform s(t) but not on the detailed
characteristics of s(t).
24
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
Since h(t)=s(T–t), the Fourier transform of this relationship
is: T
H ( f ) = ∫ s (T − t )e − j 2π ft dt let τ = T - t
0
= ∫ s (τ )e j 2π f τ dτ ⎤ e − j 2π fT = S * ( f )e − j 2π fT
⎡ T
⎢⎣ 0 ⎥⎦
The matched filter has a frequency response that is the
complex conjugate of the transmitted signal spectrum
multiplied by the phase factor e-j2πfT (sampling delay of T).
In other worlds, |H( f )|=|S( f )|, so that the magnitude
response of the matched filter is identical to the transmitted
signal spectrum.
On the other hand, the phase of H( f ) is the negative of the
phase of S( f ).
25
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
If the signal s(t) with spectrum S(f ) is passed through the
matched filter, the filter output has a spectrum
2
Y( f )=|S( f )| e -j2πfT. S ( f ) ⋅ S *
( f ) e − j 2π fT
−∞ 0
The noise at the output of the matched filter has a power
spectral density (from equation 2.2-27) Φ ( f ) = 1 H ( f ) 2 N
0 0
2
26
5.1.2 Matched-Filter Demodulator
Frequency-domain interpretation of the matched filter
The total noise power at the output of the matched filter is
∞
Pn = ∫ Φ 0 ( f )df
−∞
= N 0 ∫ H ( f ) df = N 0 ∫ S ( f ) df = ε N 0
1 ∞ 2 1 ∞ 2 1
2 −∞ 2 −∞ 2
SNR 0 =
Ps
=
ε 2
=
2ε
Pn 1 ε N N0
0
2
27
5.1.2 Matched-Filter Demodulator
Example 5.1-2:
M=4 biorthogonal signals are constructed from the two
orthogonal signals shown in the following figure for
transmitting information over an AWGN channel. The noise is
assumed to have a zero-mean and power spectral density ½ N0.
28
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
The M=4 biorthogonal signals have dimensions N=2 (shown
in figure a): ⎧⎪ 1
f1 (t ) = ⎨ 2 T (0 ≤ t ≤ T )
2
⎪⎩0 (otherwise)
⎧⎪ 1
f 2 (t ) = ⎨ 2 T ( T ≤ t ≤ T)
2
⎪⎩0 (otherwise)
The impulse responses of the two matched filters are (figure b):
⎧⎪ 1
h1 (t ) = f1 (T − t ) = ⎨ 2 T ( T ≤ t ≤ T)
2
⎪⎩0 (otherwise)
⎧⎪ 1
h2 (t ) = f 2 (T − t ) = ⎨ 2 T (0 ≤ t ≤ T )
2
⎪⎩0 (otherwise)
29
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
If s1(t) is transmitted, the responses of the two matched filters
are as shown in figure c, where the signal amplitude is A.
Since y1(t) and y2(t) are sampled at t=T, we observe that
y1S (T)= 1 A2T , y2S(T)=0.
2
From equation 5.1-27, we have ½A2T=ε and the received
vector is:
r = [r1 r2 ] = ⎢ ε + n1 n2 ⎥
⎡ ⎤
⎣ ⎦
where n1(t)=y1n(T) and n2=y2n(T) are the noise components at
the outputs of the matched filters, given by
T
ykn (T ) = ∫ n(t ) f k (t )dt , k = 1,2
0
30
5.1.2 Matched-Filter Demodulator
Example5.1-2(cont.)
Clearly, E(nk)=E[ykn(T)] and their variance is
σ = E [y (T )] = ∫ ∫ E[n(t )n(τ )] f
T T
2
n
2
kn k (t ) f k (τ )dtdτ
0 0
1 T T
= N 0 ∫ ∫ δ (t − τ ) f k (τ ) f k (t )dtdτ
2 0 0
1 T 1
= N 0 ∫ f k (t )dt = N 0
2
2 0 2
Observe that the SNR0 for the first matched filter is
SNR 0 =
( ε) 2
=
2ε
1 N0
N0
2
31
5.1.2 Matched-Filter Demodulator
Example 5.1-2: (cont.)
This result agrees with our previous result.
We can also note that the four possible outputs of the
two matched filters, corresponding to the four possible
transmitted signals are:
( r1 , r2 ) = ( ε + n1 , n2 ) , ( n1 , ε + n2 ) , ( − ε + n1 , n2 ) and ( n1 , − ε + n2 )
32
5.1.3 The Optimum Detector
Our goal is to design a signal detector that makes a decision
on the transmitted signal in each signal interval based on
the observation of the vector r in each interval such that the
probability of a correct decision is maximized.
We assume that there is no memory in signals transmitted
in successive signal intervals.
We consider a decision rule based on the computation of
the posterior probabilities defined as
P(sm|r)≣P(signal sm was transmitted|r), m=1,2,…,M.
The decision criterion is based on selecting the signal
corresponding to the maximum of the set of posterior
probabilities { P(sm|r)}. This decision criterion is called the
maximum a posterior probability (MAP) criterion.
33
5.1.3 The Optimum Detector
Using Bayes’ rule, the posterior probabilities may be expressed
as p (r | s m ) P (s m )
P (s m | r ) = ---(A)
p (r )
where P(sm) is the a priori probability of the mth signal being
transmitted.
The denominator of (A), which is independent of which signal is
transmitted, may be expressed as
M
p (r) = ∑ p (r | s m ) P (s m )
m =1
Some simplification occurs in the MAP criterion when the M
signal are equally probable a priori, i.e., P(sm)=1/M.
The decision rule based on finding the signal that maximizes
P(sm|r) is equivalent to finding the signal that maximizes P(r|sm).
34
5.1.3 The Optimum Detector
The conditional PDF P(r|sm) or any monotonic function of it is
usually called the likelihood function.
The decision criterion based on the maximum of P(r|sm) over
the M signals is called maximum-likelihood (ML) criterion.
We observe that a detector based on the MAP criterion and one
that is based on the ML criterion make the same decisions as
long as a priori probabilities P(sm) are all equal.
In the case of an AWGN channel, the likelihood function p(r|sm)
is given by:
1 N
⎡ ( rk − smk ) ⎤
2
p (r | s m ) = exp ⎢ − ∑ ⎥ , m = 1, 2,......, M
(π N 0 ) N 2
⎣ k =1 N0 ⎦ constant
1 1 N
ln p (r | s m ) = − N ln(πN 0 ) −
2
∑ k mk
N 0 k =1
( r − s ) 2
35
5.1.3 The Optimum Detector
The maximum of ln p(r|sm) over sm is equivalent to finding the
signal sm that minimizes the Euclidean distance:
N
D (r, s m ) = ∑ (rk − smk ) 2
k =1
36
5.1.3 The Optimum Detector
Expanding the distance metrics:
N N N
D(r, s m ) = ∑ r − 2∑ rn smn + ∑ smn
n
2 2
n =1 n =1 n =1
= r − 2r ⋅ s m + s m , m = 1, 2,......., M
2 2
37
5.1.3 The Optimum Detector
The term r⋅sm represents the projection of the signal vector onto
each of the M possible transmitted signal vectors.
The value of each of these projection is a measure of the
correlation between the receiver vector and the mth signal. For
this reason, we call C(r, sm), m=1,2,…,M, the correlation
metrics for deciding which of the M signals was transmitted.
Finally, the terms || sm||2 =εm, m=1,2,…,M, may be viewed as
bias terms that serve as compensation for signal sets that have
unequal energies.
If all signals have the same energy, || sm||2 may also be ignored.
Correlation metrics can be expressed as:
C (r, s m ) = 2 ∫ r (t ) sm (t ) dt − ε
T
, m = 0,1,......., M
0 m
38
5.1.3 The Optimum Detector
These metrics can be generated by a demodulator that cross-
correlates the received signal r(t) with each of the M possible
transmitted signals and adjusts each correlator output for the
bias in the case of unequal signal energies.
39
5.1.3 The Optimum Detector
We have demonstrated that the optimum ML detector computes
a set of M distances D(r,sm) or D'(r,sm) and selects the signal
corresponding to the smallest (distance) metric.
Equivalently, the optimum ML detector computes a set of M
correlation metrics C(r, sm) and selects the signal corresponding
to the largest correlation metric.
The above development for the optimum detector treated the
important case in which all signals are equal probable. In this
case, the MAP criterion is equivalent to the ML criterion.
When the signals are not equally probable, the optimum MAP
detector bases its decision on the probabilities given by:
p (r | s m ) P (s m )
P (s m | r ) = or PM ( r , s m ) = p ( r | s m ) P ( s m )
p (r )
40
5.1.3 The Optimum Detector
Example 5.1-3:
Consider the case of binary PAM signals in which the two
possible signal points are s1 = −s2 = ε b ,, where εb is the
energy per bit. The priori probabilities are P(s1)=p and
P(s2)=1-p. Let us determine the metrics for the optimum
MAP detector when the transmitted signal is corrupted with
AWGN.
The receiver signal vector for binary PAM is:
r = ± ε b + y n (T )
where yn(T) is a zero mean Gaussian random variable with
1
variance σ n2 = N 0 .
2
41
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
The conditional PDF P(r|sm) for two signals are
1 ⎡ (r − ε b )2 ⎤
p (r | s1 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦
1 ⎡ (r + ε n ) 2 ⎤
p (r | s2 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦
Then the metrics PM(r,s1) and PM(r,s2) are
p ⎡ (r − ε b ) 2 ⎤
PM (r, s1 ) = p ⋅ p (r | s1 ) = exp ⎢ − ⎥
2πσ n ⎢⎣ 2σ n
2
⎥⎦
PM (r, s 2 ) = (1 − p ) ⋅ p (r | s2 ) =
1− p ⎡ (r +
exp ⎢ −
ε ) 2
⎤
⎥
b
2πσ n ⎢⎣ 2σ n
2
⎥⎦
42
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
If PM(r,s1) > PM(r,s2), we select s1 as the transmitted signal:
otherwise, we select s2. This decision rule may be expressed
as: PM (r, s1 ) s1
1
PM (r, s 2 ) s2
PM (r, s1 ) p ⎡ (r + ε b ) 2 − (r − ε b ) 2 ⎤
= exp ⎢ ⎥
PM (r, s 2 ) 1 − p ⎢⎣ 2σ n2
⎥⎦
(r + ε b ) 2 − (r − ε b ) 2 s1 1 − p
ln
2σ n
2
s2
p
1 2 1− p 1 1− p
s1
ε br
2
σ n ln
p
= N 0 ln
4 p
s2
43
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
1 1− p
The threshold is 4
N 0 ln
p , denoted by τh, divides the real
line into two regions, say R1 and R2, where R1 consists of the
set of points that are greater than τh and R2 consists of the set
of points that are less than τh.
If r ε b > τ h , the decision is made that s1 was transmitted.
If r ε b < τ h , the decision is made that s2 was transmitted.
44
5.1.3 The Optimum Detector
Example 5.1-3: (cont.)
The threshold τh depends on N0 and p. If p=1/2, τh=0.
If p>1/2, the signal point s1 is more probable and, hence,
τh<0. In this case, the region R1 is larger than R2 , so that s1 is
more likely to be selected than s2.
The average probability of error is minimized
It is interesting to note that in the case of unequal priori
probabilities, it is necessary to know not only the values of
the priori probabilities but also the value of the power
spectral density N0 , or equivalently, the noise-to-signal ratio,
in order to compute the threshold.
When p=1/2, the threshold is zero, and knowledge of N0 is not
required by the detector.
45
5.1.3 The Optimum Detector
Proof of “the decision rule based on the maximum-likelihood
criterion minimizes the probability of error when the M signals
are equally probable a priori”.
Let us denote by Rm the region in the N-dimensional space for
which we decide that signal sm(t) was transmitted when the
vector r=[r1r2….. rN] is received.
The probability of a correct decision given that sm(t) was
P ( c | s m ) ⋅ p ( s m ) = ∫ p ( r | s m ) ⋅ p ( s m ) dr
transmitted is:
Rm 1
The average probability of a correct decision is: M
M M
1 1
P (c ) = ∑ P (c | s m ) = ∑ ∫ p (r | s m )dr
m =1 M m =1 M Rm
Note that P(c) is maximized by selecting the signal sm if p(r|sm)
is larger than p(r|sk) for all m≠k. Q.E.D.
46
5.1.3 The Optimum Detector
Similarly for the MAP criterion, when the M signals are not
equally probable, the average probability of a correct decision
is M
P(c) = ∑ ∫ P(s m | r ) p (r )dr
m =1 Rm
same for all sm.
In order for P (c) to be as large as possible, the points that are
to be included in each particular region Rm are those for which
P(sm|r) exceeds all the other posterior probabilities. Q.E.D.
47
5.1.4 The Maximum-Likelihood
Sequence Detector
When the signal has no memory, the symbol-by-symbol
detector described in the preceding section is optimum in the
sense of minimizing the probability of a symbol error.
When the transmitted signal has memory, i.e., the signals
transmitted in successive symbol intervals are interdependent,
the optimum detector is a detector that bases its decisions on
observation of a sequence of received signals over successive
signal intervals.
In this section, we describe a maximum-likelihood sequence
detection algorithm that searches for minimum Euclidean
distance path through the trellis that characterizes the memory
in the transmitted signal.
48
5.1.4 The Maximum-Likelihood
Sequence Detector
To develop the maximum-likelihood sequence detection
algorithm, let us consider, as an example, the NRZI signal
described in Section 4.3-2. Its memory is characterized by the
trellis shown in the following figure:
(
⎡ r + ε
) ⎤
2
p ( rk | s2 ) =
1
exp −⎢ k b ⎥
2πσ n ⎢ 2σ n2 ⎥
⎢⎣ ⎥⎦
50
5.1.4 The Maximum-Likelihood
Sequence Detector
Since the channel noise is assumed to be white and Gaussian
and f (t= iT), f (t =jT) for i≠j are orthogonal, it follows that
E(nknj)=0, k≠j.
Hence, the noise sequence n1, n2, …, nk is also white.
Consequently, for any given transmitted sequence s(m), the joint
PDF of r1, r2, …, rk may be expressed as a product of K
marginal PDFs, K
p (r1 , r2 ,..., rk | s ( m ) ) = ∏ p (rk | sk( m ) )
k =1
K
1 ⎡ (rk − sk( m ) ) 2 ⎤
=∏ exp ⎢− ⎥
k =1 2π σ n ⎣ 2σ 2
n ⎦
K
⎛ 1 ⎞ ⎡ K (rk − sk( m ) ) 2 ⎤
= ⎜⎜ ⎟ exp ⎢− ∑ ⎥ (A)
⎟ σ
⎝ 2π σ n ⎠
2
⎣ k =1 2 n ⎦
51
5.1.4 The Maximum-Likelihood
Sequence Detector
Then, given the received sequence r1,r2,…,rk at the output of
the matched filter or correlation demodulator, the detector
determines the sequence s(m)={s1(m),s2(m),..,sK(m)} that
maximizes the conditional PDF p(r1,r2,…,rk|s(m)). Such a
detector is called the maximum-likelihood (ML) sequence-
detector.
By taking the logarithm of Equation (A) and neglecting the
terms that are independent of (r1,r2,…,rk), we find that an
equivalent ML sequence detector selects the sequence s(m) that
minimizes the Euclidean distance metric
K
D(r, s ( m ) ) = ∑ (rk − sk( m ) ) 2
k =1
52
5.1.4 The Maximum-Likelihood
Sequence Detector
In searching through the trellis for the sequence that minimizes
the Euclidean distance, it may appear that we must compute the
distance for every possible sequence.
For the NRZI example, which employs binary modulation, the
total number of sequence is 2K , where K is the number of
outputs obtained form the demodulator.
However, this is not the case. We may reduce the number of
the sequences in the trellis search by using the Viterbi
algorithm to eliminate sequences as new data is received from
the demodulator.
The Viterbi algorithm is a sequence trellis search algorithm for
performing ML sequence detection. We describe it below in
the context of the NRZI signal. We assume that the search
process begins initially at state S0.
53
Viterbi Decoding Algorithm
Basic concept
Generate the code trellis at the decoder
The decoder penetrates through the code trellis level by level in
search for the transmitted code sequence
At each level of the trellis, the decoder computes and
compares the metrics of all the partial paths entering a node
The decoder stores the partial path with the larger metric and
eliminates all the other partial paths. The stored partial path is
called the survivor.
54
5.1.4 The Maximum-Likelihood
Sequence Detector
The corresponding trellis is shown in the following figure :
D1 (1,0) = (r1 − ε b ) 2
+ (r2 − εb ) 2
These two metrics are compared and the path with the larger
(greater-distance) metric is eliminated.
Similarly, the metrics for the two paths entering S1 at t=3T are
D1 (0,0,1) = D0 (0,0) + (r3 − ε b ) 2
D1 (0,1,0) = D1 (0,1) + (r3 − ε b ) 2
These two metrics are compared and the path with the larger
metrics is eliminated.
58
5.1.4 The Maximum-Likelihood
Sequence Detector
It is relatively easy to generalize the trellis search performed by
the Viterbi algorithm for M-ary modulation.
Delay modulation with M=4 signals: such signal is characterized
by the four-state trellis shown in the following figure
59
5.1.4 The Maximum-Likelihood
Sequence Detector
Delay modulation with M=4 signals: (cont.)
We observe that each state has two signal paths entering and
two signals paths leaving each node.
The memory of the signal is L=1.
The Viterbi algorithm will have four survivors at each stage
and their corresponding metrics.
Two metrics corresponding to the two entering paths are
computed at each node, and one of the two signal paths
entering the node is eliminated at each state of the trellis.
The Viterbi algorithm minimizes the number of trellis paths
searched in performing ML sequence detection.
From the description of the Viterbi algorithm given above, it is
unclear as to how decisions are made on the individual detected
information symbols given the surviving sequences.
60
5.1.4 The Maximum-Likelihood
Sequence Detector
If we have advanced to some stage, say K, where K>>L in the
trellis, and we compare the surviving sequences, we shall find
that with probability approaching one all surviving sequences
will be identical in bit (or symbol) positions K-5L and less.
In a practical implementation of the Viterbi algorithm, decisions
on each information bit (or symbol) are forced after a delay of
5L bits, and, hence, the surviving sequences are truncated to the
5L most recent bits (or symbol). Thus, a variable delay in bit or
symbol detection is avoided.
The loss in performance resulting from the sub-optimum
detection procedure is negligible if the delay is at least 5L.
61
5.2.1 Probability of Error for Binary
Modulation
Let us consider binary PAM signals where the two signal
waveforms are s1(t)=g(t) and s1(t)= −g(t), and g(t) is an arbitrary
pulse that is nonzero in the interval 0≤ t ≤Tb and zero elsewhere.
Since s1(t)= −s2(t), these signals are said to be antipodal.
The energy in the pulse g(t) is εb.
As indicated in section 4.3.1, PAM signals are one-dimensional,
and, their geometric representation is simply the one-
dimensional vector s1 = ε b , s2 = − ε b .
Figure (A)
62
5.2.1 Probability of Error for Binary
Modulation
Let us assume that the two signals are equally likely and that
signal s1(t) was transmitted. Then, the received signal from the
(matched filter or correlation) demodulator is
r = s1 + n = ε b + n
63
5.2.1 Probability of Error for Binary
Modulation
64
5.2.1 Probability of Error for Binary
Modulation
Given that s1(t) was transmitted, the probability of error is simply
the probability that r<0.
P ( e | s1 ) = ∫ p ( r | s1 )dr =
0 1 0
⎡ r− ε 2⎤
⎢ ⎥ dr ( )
π N 0 ∫−∞
b
exp −
−∞ ⎢ N0 ⎥
⎢⎣ ⎥⎦
1 − 2ε b / N 0 ( r− εb )
∫
− x2 / 2
= e dx x=
2π −∞ N0
2
1 ∞
= ∫ e − x2 / 2
dx ( x = −x)
2π 2ε b / N 0
∞
⎛ 2ε b ⎞ Q (t ) =
1
∫
− x2 / 2
e dx t ≥ 0
= Q⎜ ⎟⎟ 2π
⎜ N t
⎝ 0 ⎠
65
5.2.1 Probability of Error for Binary
Modulation
If we assume that s2(t) was transmitted, r = − ε b + n and the
( )
probability that r>0 is also P ( e | s2 ) = Q 2ε b / N 0 . Since the
signal s1(t) and s2(t) are equally likely to be transmitted, the
average probability of error is
1 1 ⎛ 2ε b
⎞
Pb = P ( e | s1 ) + P ( e | s2 ) = Q ⎜⎜ ⎟⎟ --(A)
2 2 ⎝ N0
⎠
Two important characteristics of this performance measure:
First, we note that the probability of error depends only on
the ratio εb /N0.
Secondly, we note that. 2εb /N0 is also the output SNR0 from
the matched-filter (and correlation) demodulator.
The ratio εb /N0 is usually called the signal-to-noise ratio per bit.
66
5.2.1 Probability of Error for Binary
Modulation
We also observe that the probability of error may be expressed
in terms of the distance between that the two signals s1and s2 .
From figure (A), w observe that the two signals are separated by
1
the distance d12=2 ε b . By substituting ε b = d122 into Equation
4
(A),we obtain
⎛ d2 ⎞
Pb = Q⎜ 12 ⎟
⎜ 2N 0 ⎟
⎝ ⎠
67
5.2.1 Probability of Error for Binary
Modulation
s1 = [ ε b 0]
s2 = [0 εb ]
where εb denote the energy for each of the waveforms. Note that
the distance between these signal points is d12 = 2ε b .
68
5.2.1 Probability of Error for Binary
Modulation
Error probability for binary orthogonal signals
To evaluate the probability of error, let us assume that s1 was
transmitted. Then, the received vector at the output of the
demodulator is r = [ ε b + n1 n2 ].
We can now substitute for r into the correlation metrics
given by C (r, s m ) = 2r ⋅ s m − s m to obtain C(r, s1) and C(r,
2
s2).
The probability of error is the probability that C(r, s2) > C(r,
s1).
n2 > n1 + ε b
( ) (
⇒ 2 n2 ε b − ε b > 2ε b + 2 n1 ε b − ε b )
P ( e | s1 ) = P[C ( r , s2 ) > C ( r1 , s2 )] = P[n2 − n1 > ε b ]
69
5.2.1 Probability of Error for Binary
Modulation
Error probability for binary orthogonal signals
Since n1 and n2 are zero-mean statistically independent
Gaussian random variables each with variance ½ N0, the
random variable x= n2 – n1 is zero-mean Gaussian with
variance N0. Hence,
( ) 1 ∞
P n2 − n1 > ε b = ∫
− x2 2 N0
e dx
2π N 0 εb
1 ∞ ⎛ εb ⎞
∫
− x2 2
= e dx = Q ⎜⎜ ⎟⎟
2π εb N0
⎝ N0 ⎠
The same error probability is obtained when we assume that
s2 is transmitted:
⎛ εb ⎞
Pb = Q ⎜
⎜ N ⎟⎟ ( )
= Q γ b where γ b is the SNR per bit.
⎝ 0 ⎠
70
5.2.1 Probability of Error for Binary
Modulation
If we compare the probability of error for binary antipodal
signals with that for binary orthogonal signals, we find that
orthogonal signals required a factor of 2 increase in energy to
achieve the same error probability as antipodal signals.
Since 10 log10 2=3 dB, we say that orthogonal signals are 3dB
poorer than antipodal signals. The difference of 3dB is simply
due to the distance between the two signal points, which is d122 = 2ε b
for orthogonal signals, whereas d122 = 4ε b for antipodal signals.
The error probability versus 10 log10 εb/N0 for these two types of
signals is shown in the following figure (B). As observed from
this figure, at any given error probability, the εb/N0 required for
orthogonal signals is 3dB more than that for antipodal signals.
71
5.2.1 Probability of Error for Binary
Modulation
Probability of error for binary signals
Figure (B)
72
5.2.2 Probability of Error for M-ary
Orthogonal Signals
For equal-energy orthogonal signals, the optimum detector
selects the signal resulting in the largest cross correlation
between the received vector r and each of the M possible
transmitted signals vectors {sm}, i.e.,
M
C (r , sm ) = r ⋅ sm = ∑ rk smk , m = 1,2, … , M
k =1
C (r , sM ) = ε s nM
Note that the scale factor εs may be eliminated from the
correlator outputs dividing each output by ε s .
With this normalization, the PDF of the first correlator output is
pr1 ( x1 ) =
1 ⎡ x − ε
exp ⎢− 1
(s ) ⎤⎥
2
πN 0 ⎢ N0 ⎥
⎣ ⎦
74
5.2.2 Probability of Error for M-ary
Orthogonal Signals
And the PDFs of the other M−1 correlator outputs are
prm ( xm ) =
1 − xm2 N 0
e , m = 2,3, … , M
πN 0
It is mathematically convenient to first derive the probability that
the detector makes a correct decision. This is the probability that
r1 is large than each of the other M−1 correlator outputs n2,n3,…,
nM. This probability may be expressed as
∞
Pc = ∫ P (n2 < r1 , n3 < r1 , … , nM < r1 | r1 ) p (r1 )dr1
−∞
75
5.2.2 Probability of Error for M-ary
Orthogonal Signals
Since the {rm} are statistically independent, the joint probability
factors into a product of M−1 marginal probabilities of the form:
1 r1 2 N 0
∫
− x2 2
= e dx x = 2 N 0 xm
2π −∞
⎧ M −1 ⎡ ⎛ ⎞
2
⎤ ⎫
⎪ 1 ∞ ⎛ 1 y ⎞ 1 2ε ⎪
∫−∞ ⎜⎝ 2π ∫−∞ ⎢ ⎥
−x 2
PM = 1 − ⎨ − −
2
⎜⎜ ⎟⎟ ⎬
s
e dx ⎟ exp y dy
⎪⎩ 2π ⎠ ⎢ 2⎝
⎣
N0 ⎠ ⎥ ⎪
⎦ ⎭
--- (C)
The same expression for the probability of error is obtained
when any one of the other M−1 signals is transmitted. Since all
the M signals are equally likely, the expression for PM given
above is the average probability of a symbol error.
This expression can be evaluated numerically.
77
5.2.2 Probability of Error for M-ary
Orthogonal Signals
In comparing the performance of various digital modulation
methods, it is desirable to have the probability of error
expressed in terms of the SNR per bit, εb/N0, instead of the SNR
per symbol, εs/N0.
With M=2k, each symbol conveys k bits of information, and
hence εs= k εb. Thus, Equation (C) may be expressed in terms of
εb/N0 by substituting for εs.
It is also desirable to convert the probability of a symbol error
into an equivalent probability of a binary digit error. For
equiprobable orthogonal signals, all symbol errors are
equiprobable and occur with probability
PM P
= kM
M −1 2 −1
78
5.2.2 Probability of Error for M-ary
Orthogonal Signals
⎛k⎞
Furthermore, there are ⎜ n ⎟ ways in which n bits out of k may be
⎝ ⎠
in error. Hence, the average number of bit errors per k-bit
symbol is k
⎛ k ⎞ PM 2k −1
∑ n⎜ ⎟ k
n =1 ⎝ n ⎠ 2 − 1
=k k
2 −1
PM --(D)
and the average bit error probability is just the result in Equation
(D) divided by k, the number of bits per symbol. Thus,
2k −1 P
Pb = k PM ≈ M k >> 1
2 −1 2
The graphs of the probability of a binary digit error as a function
of the SNR per bit, εb/N0, are shown in Figure (C) for
M=2,4,8,16,32, and 64. This figure illustrates that, by increasing
the number M of waveforms, one can reduce the SNR per bit
required to achieve a given probability of a bit error.
79
5.2.2 Probability of Error for M-ary
Orthogonal Signals
For example, to achieve a Pb=10 –5, the required SNR per bit is a
little more than 12dB for M=2, but if M is increased to 64 signal
waveforms, the required SNR per bit is approximately 6dB. Thus,
a savings of over 6dB is realized in transmitter power required to
achieve a Pb=10 –5 by increasing M from M=2 to M=64.
Figure (C)
80
5.2.2 Probability of Error for M-ary
Orthogonal Signals
What is the minimum required εb/N0 to achieve an arbitrarily
small probability of error as M→∞?
A union bound on the probability of error.
Let us investigate the effect of increasing M on the
probability of error for orthogonal signals. To simplify the
mathematical development, we first derive an upper bound
on the probability of a symbol error that is much simpler than
the exact form given in the following Equation (5.2-21)
1 ∞⎡ ⎛ 1 y ⎞
M −1
⎤ ⎡ 1⎛ 2ε s ⎞
2
⎤
⎜ ⎟
PM = ∫ 1− ⎜ ∫ e−x ⎢ − − ⎥ dy
2
⎢
2
dx ⎟ ⎥ exp ⎜ y ⎟
2π −∞ ⎢⎣ ⎝ 2π −∞
⎠ ⎥⎦ ⎢ 2⎝ N0 ⎠ ⎥
⎣ ⎦
81
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error (cont.)
Recall that the probability of error for binary orthogonal
signals is given by 5.2-11: ⎛ εb ⎞
Pb = Q⎜⎜ ( )
⎟ = Q γb
⎟
⎝ N0 ⎠
Now, if we view the detector for M orthogonal signals as one
that makes M −1 binary decisions between the correlator
outputs C(r,s1) that contains the signal and the other M−1
correlator outputs C(r,sm), m=2,3,…,M, the probability of
error is upper-bounded by union bound of the M −1 events.
That is, if Ei represents the eventM that C(r,si)> C(r,s1) for i≠1,
( )
then we have PM = P ∪ i = 2 Ei ≤ ∑ P ( Ei ) . Hence,
M
i =2
PM ≤ ( M − 1) Pb = ( M − 1) Q ( )
ε S / N 0 < MQ ( ε S / N0 )
82
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error(cont.)
This bound can be simplified further by upper-bounding
(
Q εS / N0 )
We have
( )
Q ε s N 0 < e −ε s 2 N 0 --(E)
thus,
PM < Me −ε s 2 N0
= 2k e − kε b 2 N0
--(F)
− k ( ε b N 0 − 2ln 2 ) 2
PM < e
As k→∞ or equivalently, as M→∞, the probability of error
approaches zero exponentially, provided that εb/N0 is greater
εb
than 2ln 2,
> 2 ln 2 = 1.39 (1.42dB)
N0
83
5.2.2 Probability of Error for M-ary
Orthogonal Signals
A union bound on the probability of error(cont.)
The simple upper bound on the probability of error given by
Equation (F) implies that, as long as SNR>1.42 dB, we can
achieve an arbitrarily low PM..
However, this union bound is not a very tight upper bound as
a sufficiently low SNR due to the fact that upper bound for
the Q function in Equation (E) is loose.
In fact, by more elaborate bounding techniques, it is shown in
Chapter 7 that the upper bound in Equation (F) is sufficiently
tight for εb/N0>4 ln2.
For εb/N0<4 ln2, a tighter upper bound on PM is
−k ( ε 0 N 0 − ln 2 )2
PM < 2e
84
5.2.2 Probability of Error for M-ary
Orthogonal Signals
85
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
As indicated in Section 4.3, a set of M=2k biorthogonal signals
are constructed from ½ M orthogonal signals by including the
negatives of the orthogonal signals. Thus, we achieve a
reduction in the complexity of the demodulator for the
biorthogonal signals relative to that for orthogonal signals, since
the former is implemented with ½ M cross correlation or
matched filters, whereas the latter required M matched filters or
cross correlators.
Let us assume that the signal s1(t) corresponding to the vector
s1=[ ε s 0 0…0] was transmitted. The received signal vector is
r = [ ε s + n1 n2 nM 2 ]
where the {nm} are zero-mean, mutually statistically independent
and identically distributed Gaussian random variables with
1
variance σ n2 = N 0 .
2
86
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
The optimum detector decides in favor of the signal
corresponding to the largest in magnitude of the cross
correlators M 2
1
C ( r,s m ) = r ⋅ s m = ∑ rk smk , m = 1, 2,..., M
k =1 2
while the sign of this largest term is used to decide whether sm(t)
or −sm(t) was transmitted.
According to this decision rule, the probability of a correct
decision is equal to the probability that r1 = ε s + n1 > 0 and r1
exceeds |rm|=|nm| for m=2,3,… ½ M. But
P ( nm < r1 | r1 > 0 ) =
1 r1 1 r1 N0 2 − y 2 2
∫ ∫
− x2 N0
e dx = e dy
π N0 − r1
2π − r1 N 0 2
N0
y= x
2
87
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
Then, the probability of a correct decision is
M 2 −1
⎛ 1∞
−x 2 ⎞
p ( r1 ) dr1
r1 N0 2
Pc = ∫ ⎜ ∫
2
e ⎟
0
⎝ 2π − r1 N0 2 ⎠
from which, upon substitution for p(r1), we obtain
M 2 −1
1 ∞ ⎛ 1 v + 2ε s N 0 ⎞
∫− 2ε s N0 ⎜⎝ 2π ∫−(v+ 2ε s N0 ) e dx ⎟⎠ e dv --(G)
−x 2 −v2 2
Pc =
2
2π
p r1 ( x1 ) =
1 ⎡
⎢
exp −
(
x1 − ε s ⎤
2
⎥
) v = r1 − (
ε s
N0
)
πN 0 ⎢ N0 ⎥ 2
⎣ ⎦
where we have used the PDF of r1 given in Equation 5.2-15.
The probability of a symbol error PM=1−Pc
88
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
Pc, and PM may be evaluated numerically for different values of
M. The graph shown in the following figure (D) illustrates PM
as a function of εb/N0, where εs=k εb, for M=2,4,8,16 and 32.
89
5.2.3 Probability of Error for M-ary
Biorthogonal Signals
In this case, the probability of error for M=4 is greater than that
for M=2. This is due to the fact that we have plotted the symbol
error probability PM in Figure(D).
90
5.2.4 Probability of Error for
Simplex Signals
Next we consider the probability of error for M simplex signals.
Recall from Section 4.3 that simplex signals are a set of M equally
correlated with mutual cross-correlation coefficient ρmn= −1/(M −1).
These signals have the same minimum separation of 2ε s between
adjacent signal points in M-dimensional space as orthogonal signals.
They achieve this mutual separation with a transmitted energy of
εs(M −1)/M, which is less than that required for orthogonal signals
by a factor of (M −1)/M.
Consequently, the probability of error for simplex signals is
identical to the probability of error for orthogonal signals, but this
performance is achieved with saving of
10 log(1 − ρ ) = 10 log
M
dB
M −1
For M=2, the saving is 3dB. As M is increased, the saving in SNR
approaches 0 dB.
91
5.2.5 Probability of Error for M-ary
Binary-Coded Signals
In Section 4.3, we have shown that binary-coded signal
waveform are represented by signal vectors (4.3-38) :
s m = [s m1 s m 2 s mN ] , m = 1,2, … , M
where s mj = ± ε N for all m and j. N is the block length of
the code and is also the dimension of the M signal waveform
(e)
If d min is the minimum Euclidean distance, then the probability
of a symbol error is upper-bounded as
⎛ (d (e ) )2 ⎞
PM < (M − 1)Pb = (M − 1)Q⎜⎜ min ⎟
⎜ 2 N 0 ⎟⎟
⎝ ⎠
⎡
< 2 exp ⎢−
k (d min )
(e ) 2 ⎤
⎥
⎢⎣ 4 N 0 ⎥⎦
92
5.2.6 Probability of Error for M-ary
PAM
Recall (4.3-6) that M-ary PAM signals are represent
geometrically as M one-dimensional signal points with value :
sm =
1
2
ε g Am m = 1, 2, ,M
where Am = (2 m − 1 − M )d , m = 1,2, , M
The Euclidean distance between adjacent signal points is d 2ε g .
Assuming equally probable signals, the average energy is :
1 M
1 M
2
d 2
ε
ε av = M ∑ ε m = M ∑ sm = 2M ∑ ( 2m − 1 − M )
g
M
2
m =1 m =1 m =1
d 2ε g ⎡ 1 M ( M + 1)
⎤ 1
ε
M
=
2M ⎣ 3⎢ (
M M 2
−1) =
⎥ 6
⎦
M 2
(
− 1 d 2
) ∑
g m =1
m=
2
M M ( M + 1)( 2 M + 1)
∑m =
2
m =1 6
93
5.2.6 Probability of Error for M-ary
PAM
Equivalently, we may characterize these signals in terms of their
average power, where is :
ε d εg
( )
2
1
Pav = = M −1
av2
(5.2-40)
T 6 T
The average probability of error for M-ary PAM :
The detector compares the demodulator output r with a set of M-1
thresholds, which are placed at the midpoints of successive amplitude
level and decision is made in favor of the amplitude level that is close to r.
1
2d εg
2
ε
e − x 2 N0
dx
2π N 0 2
g
M ⎝ 2 ⎠ M d g 2
M −1 2 ∞
∫
− y2 2
= e dy where y = x N0 2
M 2π d2 ε g N0
1 ∞
Q (t ) = ∫ e− x / 2 dx t ≥ 0
2
2 ( M − 1) ⎛ d 2ε g ⎞ 2π t
= Q⎜ ⎟ 1
i( M − 2 ) +
1 1
i i2
M ⎜ N0 ⎟ M M 2
⎝ ⎠
95
5.2.6 Probability of Error for M-ary
PAM
From (5.2-40), we note that d ε
2
g =
6
M −1
2
PavT
By substituting for d
2
ε , we obtain the average probability
g
of a symbol error for PAM in terms of the average power :
2 ( M − 1) ⎛ 6 PavT ⎞ 2 ( M − 1) ⎛ 6ε av ⎞
PM = Q⎜ ⎟= Q ⎜ ⎟
M ⎜
⎝ ( )
M 2 − 1 N0 ⎟
⎠
M ⎜
⎝ ( )
M 2 − 1 N0 ⎟
⎠
It is customary for us to use the SNR per bit as the basic
parameter, and since T=kTb and k =log2 M:
PM =
2(M − 1) ⎛⎜ (6 log 2 M )ε b av ⎞⎟
M
Q⎜
⎝ (M 2
)
− 1 N 0 ⎟⎠
where ε bav = PavTb is the average bit energy and ε bav N o is the average
SNR per bit.
96
5.2.6 Probability of Error for M-ary
PAM
The case M=2
corresponds to the error
probability for binary
antipodal signals.
The SNR per bit
increase by over 4 dB for
every factor-of-2 increase
in M.
For large M, the
additional SNR per bit
required to increase M by
a factor of 2 approaches 6
Probability of a symbol error for PAM dB.
97
5.2.7 Probability of Error for M-ary
PSK
Recall from section 4.3 that digital phase-modulated signal
waveforms may be expressed as (4.3-11):
⎡ 2π
sm (t ) = g (t ) cos ⎢ 2πf c t + (m − 1)⎤⎥, 1 ≤ m ≤ M , 0 ≤ t ≤ T
⎣ M ⎦
Energy in each waveform.
and have the vector representation: (from 4.3-12)
2π 2π
s m (t ) = ⎢ ε s cos (m − 1) ε s sin (m − 1)⎥ , ε s = ε g
⎡ ⎤ 1
⎣ M M ⎦ 2
Since the signal waveforms have equal energy, the optimum
detector for the AWGN channel given by Equation 5.1-44
computes the correlation metrics
C (r, s m ) = r ⋅ s m , m = 1,2, ,M
98
5.2.7 Probability of Error for M-ary
PSK
In other word, the received signal vector r=[r1 r2] is projected
onto each of the M possible signal vectors and a decision is
made in favor of the signal with the largest projection.
This correlation detector is equivalent to a phase detector that
computes the phase of the received signal from r. We selects
the signal vector sm whose phase is closer to r.
−1 r2
The phase of r is Θ r = tan
r1
p r (r1 , r2 ) =
1
exp ⎢
(
− 1 s )
⎡ r − ε + r2 ⎤2
2
⎥
2πσ r
2
⎢ 2σ r 2
⎥
⎣ ⎦
The PDF of the phaseΘr is obtained by a change in variables
from (r1,r2) to : r
V = r12 + r22 Θ r = tan −1 2
r1
100
5.2.7 Probability of Error for M-ary
PSK
The joint PDF of V and Θr : dr1dr2 = VdVd Θ
⎛ V 2 + ε s − 2 ε s V cos Θ r ⎞
pV ,Θ r (V , Θ r ) =
V
⎜− ⎟
exp
2πσ r2 ⎜ 2σ 2 ⎟
⎝ r ⎠
Integration of pV , Θ r (V , Θ r ) over the range of V yields pΘr ( Θ r )
∞
pΘr ( Θ r ) = ∫ pV ,Θr (V , Θ r ) dV
0
2
⎛ V ⎞
∞ −⎜ − 2ε s N 0 cos Θ r ⎟ 2σ r2
1 V
e −ε s ∫
N 0 sin Θ r
=
2
⎝ N0 2 ⎠
e dV
2πσ 2
r
0 N0 2
( )
2
1 ∞ − V ′ − 2γ s cos Θ r / 2σ r2 V
∫0 dV ′ V ′ =
− γ s sin Θr
= V ′e
2
e
2πσ r2 N0 2
where we define the symbol SNR as γ s = ε s N0 .
101
5.2.7 Probability of Error for M-ary
PSK
f Θ r (Θ r ) becomes
narrower and more
peaked about Θ r = 0 as
the SNR γ s increases.
105
5.2.7 Probability of Error for M-ary
PSK
An approximation to the error probability for large M and for
large SNR may be obtained by first approximating p Θ r (Θ ) .
For ε s N 0 >> 1 and Θ r ≤ 0.5π , pΘ r (Θ ) is well approximated
as : γs
PΘ r (Θ r ) ≈ cos Θ r e −γ s sin 2 Θ r
π
Performing the change in variable from Θ r to u = ν s sin Θ r ,
π M γs
PM ≈ 1 − ∫−π cos Θ r e −γ s sin 2 Θ r
dΘ r
M
π k = log 2 M
2 ∞ ν s = kν b
∫ 2γ
−u 2
≈ sin (π M )
e du
π s
⎛ π ⎞ ⎛ π ⎞
= 2Q⎜ 2γ s sin ⎟ = 2Q⎜ 2kγ b sin ⎟
⎝ M⎠ ⎝ M⎠
106
5.2.7 Probability of Error for M-ary
PSK
When a Gray Code is used in the mapping of k-bits symbols
into the corresponding signal phases, two k-bit symbols
corresponding to adjacent signal phases differ in only a signal
bit.
The most probable error result in the erroneous selection of an
adjacent phase to the true one.
Most k-bit symbol error contain only a single-bit error. The
equivalent bit error probability for M-ary PSK is well
1
approximated as : Pb ≈ PM
k
In practice, the carrier phase is extracted from the received
signal by performing some nonlinear operation that introduces
a phase ambiguity.
107
5.2.7 Probability of Error for M-ary
PSK
For binary PSK, the signal is often squared in order to remove
the modulation, and the double-frequency component that is
generated is filtered and divided by 2 in frequency in order to
extract an estimate of the carrier frequency and phase φ.
This operation result in a phase ambiguity of 180°in the carrier phase.
For four-phase PSK, the received signal is raised to the fourth
power to remove the digital modulation, and the resulting fourth
harmonic of the carrier frequency is filtered and divided by 4 in
order to extract the carrier component.
These operations yield a carrier frequency component containing φ, but
there are phase ambiguities of ±90° and 180° in the phase estimate.
Consequently, we do not have an absolute estimate of the carrier
phase for demodulation.
108
5.2.7 Probability of Error for M-ary
PSK
109
5.2.7 Probability of Error for M-ary
PSK
The detector is a relatively simple phase comparator that
compares the phase of the demodulated signal over two
consecutive interval to extract the information.
Coherent demodulation of differentially encoded PSK results in
a higher probability of error than that derived for absolute phase
encoding.
With differentially encoded PSK, an error in the demodulated
phase of the signal in any given interval will usually result in
decoding errors over two consecutive signaling intervals.
The probability of error in differentially encoded M-ary PSK is
approximately twice the probability of error for M-ary PSK with
absolute phase encoding.
110
5.2.8 Differential PSK (DPSK) and Its
Performance
The received signal of a differentially encoded phase-modulated
signal in any given signaling interval is compared to the phase of
the received signal from the preceding signaling interval.
We demodulate the differentially encoded signal by multiplying
r(t) by cos2πfct and sin2πfct integrating the two products over
the interval T.
At the kth signaling interval, the demodulator output :
rk = ⎡⎣ ε s cos (θ k − φ ) + nk1 ε s sin (θ k − φ ) + nk2 ⎤⎦
or equivalently,
rk = ε s e j (θ −φ ) + nk
k
113
5.2.8 Differential PSK (DPSK) and Its
Performance
The error probability performance of a DPSK demodulator and
detector
The derivation of the exact value of the probability of error for
M-ary DPSK is extremely difficult, except for M = 2.
Without loss of generality, suppose the phase difference
θk-θk-1=0. Furthermore, the exponential factor e-j(θk-1-φ)
and ej(θk-φ) can be absorbed into Gaussian noise components
nk-1 and nk, without changing their statistical properties.
rk rk*−1 = ε s + ε s (nk + nk*−1 ) + nk nk*−1
The complication in determining the PDF of the phase is the
term nkn*k-1.
114
5.2.8 Differential PSK (DPSK) and Its
Performance
However, at SNRs of practical interest, the term nkn*k-1 is
small relative to the dominant noise term ε s (nk+n*k-1).
We neglect the term nkn*k-1 and normalize rkr*k-1 by dividing
through by ε s , the new set of decision metrics becomes :
(
x = ε s + Re nk + nk* −1 )
(
y = Im nk + nk* −1 )
The variables x and y are uncorrelated Gaussian random
variable with identical variances σn2=N0. The phase is
−1 y
Θ r = tan
x
The noise variance is now twice as large as in the case of
PSK. Thus we can conclude that the performance of DPSK is
3 dB poorer than that for PSK.
115
5.2.8 Differential PSK (DPSK) and Its
Performance
This result is relatively good for M ≥4 , but it is pessimistic for M
= 2 that the loss in binary DPSK relative to binary PSK is less
than 3 dB at large SNR.
In binary DPSK, the two possible transmitted phase
differences are 0 and π rad. Consequently, only the real part
of rkr*k-1 is need for recovering the information.
( )
*
(
1
)
Re rk rk −1 = rk rk*−1 + rk*rk −1
2
Because the phase difference the two successive signaling
intervals is zero, an error is made if Re(rkr*k-1)<0.
The probability that rkr*k-1+r*krk-1<0 is a special case of a
derivation, given in Appendix B.
116
5.2.8 Differential PSK (DPSK) and Its
Performance
117
5.2.8 Differential PSK (DPSK) and Its
Performance
The probability of a binary digit error for four-phase DPSK with
Gray coding can be express in terms of well-known functions,
but it's derivation is quite involved.
According to Appendix C, it is expressed in the form :
(
⎡ 1 2
)
2 ⎤
Pb = Q1 (a , b ) − I 0 (ab ) exp ⎢ − a + b ⎥
1
2 ⎣ 2 ⎦
where Q1(a,b) is the Marcum Q function (2.1-122),
I0(x) is the modified Bessel function of order zero(2.1-123),
and the parameters a and b are defined as
⎛ 1⎞ ⎛ 1⎞
a = 2γ b ⎜⎜1 − ⎟ , and b = 2γ b ⎜1 +
⎟ ⎜
⎟
⎝ 2⎠ ⎝ 2 ⎟⎠
118
5.2.8 Differential PSK (DPSK) and Its
Performance
121
5.2.9 Probability of Error for QAM
For the four-phase signal, we have
Pav = (4 )2 A 2 = 2 A 2
1
4
For the two-amplitude, four-phase QAM, we place the points
(e)
on circles of radii A and 3 A. Thus, d min = 2 A , and
[ ]
Pav = 2(3 A 2 ) + 2 A 2 = 2 A 2
1
2
which is the same average power as the M = 4-phase signal
constellation.
Hence, for all practical purposes, the error rate performance
of the two signal sets is the same.
There is no advantage of the two-amplitude QAM signal set over
M = 4-phase modulation.
122
5.2.9 Probability of Error for QAM
QAM signal sets that have M = 8 points.
We consider the four signal constellations :
Assuming that the signal
points are equally probable,
the average transmitted
signal power is :
Pav =
1 M
∑
M m =1
( A 2
mc + Ams )
2
∑ (a )
A2 M
= 2
mc + a ms
2
M m =1
The two sets (a) and (c) contain signal points that fall on a
rectangular grid and have Pav = 6A2.
The signal set (b) requires an average transmitted power
Pav= 6.83A2, and (d) requires Pav = 4.73A2.
The fourth signal set (d) requires approximately 1 dB less
power than the first two and 1.6 dB less power than the third
to achieve the same probability of error.
The fourth signal constellation is known to be the best eight-
point QAM constellation because it requires the least power for
a given minimum distance between signal points.
124
5.2.9 Probability of Error for QAM
QAM signal sets for M ≥ 16
For 16-QAM, the signal points at a given
amplitude level are phase-rotated by relative
to the signal points at adjacent amplitude
levels.
However, the circular 16-QAM constellation
is not the best 16-point QAM signal
constellation for the AWGN channel.
Rectangular M-ary QAM signal are most frequently used in
practice. The reasons are :
Rectangular QAM signal constellations have the distinct
advantage of being easily generated as two PAM signals
impressed on phase-quadrature carriers.
125
5.2.9 Probability of Error for QAM
The average transmitted power required to achieve a given
minimum distance is only slightly greater than that of the best
M-ary QAM signal constellation.
For rectangular signal constellations in which M = 2k, where k is
even, the QAM signal constellation is equivalent to two PAM
signals on quadrature carriers, each having M = 2 k 2 signal
points.
The probability of error for QAM is easily determined from the
probability of error for PAM.
Specifically, the probability of a correct decision for the M-ary
QAM system is (
P = 1− P
c M
)
2
where p M
is the probability of error of an M − ary PAM.
126
5.2.9 Probability of Error for QAM
127
5.2.9 Probability of Error for QAM
We employ the optimum detector that
bases its decisions on the optimum
distance metrics (5.1-43), it is
relatively straightforward to show that
the symbol error probability is tightly
upper-bounded as
3ε av
2
⎡ ⎛ ⎞⎤
⎜
PM ≤ 1 − ⎢ 2Q ⎜ ⎟⎥
⎢⎣ ⎝ (M − 1)N 0
⎟
⎠ ⎥⎦
⎛ 3kε bav ⎞
≤ 4Q ⎜⎜ ⎟
⎝ ( M − 1)N 0
⎟
⎠
for any k ≥ 1, whereεbav/N0 is the
average SNR per bit.
128
5.2.9 Probability of Error for QAM
⎝
[ ]
PM < (M − 1)Q ⎜ d min 2 N 0 ⎞⎟
⎛ (e ) 2
⎠
(e )
where d min is the minimum Euclidean distance between signal
points.
This bound may be loose when M is large.
We approximate PM by replacing M-1 by Mn, where Mn is the
(e )
largest number of neighboring points that are at distance d min
from any constellation point.
It is interesting to compare the performance of QAM with that
of PSK for any given signal size M.
129
5.2.9 Probability of Error for QAM
130
5.2.9 Probability of Error for QAM
We can find out that when M = 4, we have RM = 1.
4-PSK and 4-QAM yield comparable performance for the same
SNR per symbol.
For M>4, M-ary QAM yields better performance than M-ary
PSK.
131
5.2.10 Comparison of Digital
Modulation Methods
One can compare the digital modulation methods on the basis of
the SNR required to achieve a specified probability of error.
However, such a comparison would not be very meaningful,
unless it were made on the basis of some constraint, such as a
fixed data rate of transmission or, on the basis of a fixed
bandwidth.
For multiphase signals, the channel bandwidth required is
simply the bandwidth of the equivalent low-pass signal pulse g(t)
with duration T and bandwidth W, which is approximately equal
to the reciprocal of T.
R
Since T=k/R=(log2M)/R, it follows that W =
log 2 M
132
5.2.10 Comparison of Digital
Modulation Methods
As M is increased, the channel bandwidth required, when the bit
rate R is fixed, decreases. The bandwidth efficiency is measured
by the bit rate to bandwidth ratio, which is
R
= log 2 M
W
The bandwidth-efficient method for transmitting PAM is single-
sideband. The channel bandwidth required to transmit the signal
is approximately equal to 1/2T and,
R
= 2 log 2 M
W
this is a factor of 2 better than PSK.
For QAM, we have two orthogonal carriers, with each carrier
having a PAM signal.
133
5.2.10 Comparison of Digital
Modulation Methods
Thus, we double the rate relative to PAM. However, the QAM
signal must be transmitted via double-sideband. Consequently,
QAM and PAM have the same bandwidth efficiency when the
bandwidth is referenced to the band-pass signal.
As for orthogonal signals, if the M = 2k orthogonal signals are
constructed by means of orthogonal carriers with minimum
frequency separation of 1/2T, the bandwidth required for
transmission of k = log2M information bits is
M M M
W = = = R
2T 2(k R ) 2 log 2 M
In the case, the bandwidth increases as M increases.
In the case of biorthogonal signals, the required bandwidth is
one-half of that for orthogonal signals.
134
5.2.10 Comparison of Digital
Modulation Methods
135
5.2.10 Comparison of Digital
Modulation Methods
However, the cost of achieving the higher data rate is an
increase in the SNR per bit.
Consequently, these modulation methods are appropriate for
communication channels that are bandwidth limited, where we
desire a R/W >1 and where there is sufficiently high SNR to
support increases in M.
Telephone channels and digital microwave ratio channels are
examples of such band-limited channels.
In contrast, M-ary orthogonal signals yield a R/W ≤ 1. As M
increases, R/W decreases due to an increase in the required
channel bandwidth.
The SNR per bit required to achieve a given error probability
decreases as M increases.
136
5.2.10 Comparison of Digital
Modulation Methods
Consequently, M-ary orthogonal signals are appropriate for
power-limited channels that have sufficiently large bandwidth to
accommodate a large number of signals.
As M→∞, the error probability can be made as small as desired,
provided that SNR>0.693 (-1.6dB). This is the minimum SNR
per bit required to achieve reliable transmission in the limit as
the channel bandwidth W→∞ and the corresponding R/W→0.
The figure above also shown the normalized capacity of the
band-limited AWGN channel, which is due to Shannon (1948).
The ratio C/W, where C (=R) is the capacity in bits/s, represents
the highest achievable bit rate-to-bandwidth ratio on this channel.
Hence, it serves the upper bound on the bandwidth efficiency of
any type of modulation.
137
5.4 Optimum Receiver For Signals with
Random Phase In AWGN Channel
In this section, we consider the design of the optimum receiver for
carrier modulated signals when the carrier phase is unknown and
no attempt is made to estimate its value.
Uncertainty in the carrier phase of the receiver signal may be due
to one or more of the following reasons:
The oscillators that are used at the transmitter and the receiver to generate
the carrier signals are generally not phase synchronous.
The time delay in the propagation of the signal from the transmitter to the
receiver is not generally known precisely.
Assuming a transmitted signal of the form
[ ]
s (t ) = Re g (t )e j 2πf ct
that propagates through a channel with delay t0 will be received as:
s ( t − t0 ) = Re ⎡ g ( t − t0 ) e c ( 0 ) ⎤ = Re ⎡⎣ g ( t − t0 ) e − j 2π fct0 e j 2π fc ⎤⎦
j 2π f t − t
⎣ ⎦
138
5.4 Optimum Receiver For Signals with
Random Phase In AWGN Channel
The carrier phase shift due to the propagation delay t0 is
φ = −2πf ct0
Note that large changes in the carrier phase can occur due to
relatively small changes in the propagation delay.
For example, if the carrier frequency fc=1 MHz, an uncertainty
or a change in the propagation delay of 0.5μs will cause a
phase uncertainty of π rad.
In some channels the time delay in the propagation of the signal
from the transmitter to the receiver may change rapidly and in
an apparently random fashion.
In the absence of the knowledge of the carrier phase, we may
treat this signal parameter as a random variable and determine
the form of the optimum receiver for recovering the transmitted
information from the received signal.
139
5.4.1 Optimum Receiver for Binary Signals
We consider a binary communication system that uses the two
carrier modulated signal s1(t) and s2(t) to transmit the information,
where:
sm ( t ) = Re ⎡⎣ slm ( t ) e j 2π fct ⎤⎦ , m = 1, 2, 0 ≤ t ≤ T
and slm(t), m=1,2 are the equivalent low-pass signals.
The two signals are assumed to have equal energy
4.1-22 4.1-24
2
ε = ∫ s (t )dt = ∫ sml (t ) dt
T
2 1 T
m
0 2 0
The two signals are characterized by the complex-valued
correlation coefficient
ρ12 ≡ ρ = ∫ sl1 (t )sl 2 (t )dt
1 T *
2ε 0
140
5.4.1 Optimum Receiver for Binary Signals
{ } [
n (t ) = Re [nc (t ) + jns (t )]e j 2πf ct = Re z (t )e j 2πf ct ]
The received signal may be expressed as
{[ ]
r (t ) = Re slm (t )e jφ + z (t ) e j 2πf ct }
where rl ( t ) = slm ( t ) e jφ + z ( t ) , 0 ≤ t ≤ T
rl(t) is the equivalent low-pass received signal.
This received signal is now passed through a demodulator whose
sampled output at t =T is passed to the detector.
141
5.4.1 Optimum Receiver for Binary Signals
The optimum demodulator
In section 5.1.1, we demonstrated that if the received signals
was correlated with a set of orthogonal functions {fn(t)} that
spanned the signal space, the outputs from the bank of
correlators provide a set of sufficient statistics for the detector
to make a decision that minimizes the probability error.
We also demonstrated that a bank of matched filters could be
substituted for the bank of correlations.
A similar orthogonal decomposition can be employed for a
received signal with an unknown carrier phase.
It is mathematically convenient to deal with the equivalent
low-pass signal and to specify the signal correlators or
matched filters in terms of the equivalent low-pass signal
waveforms.
142
5.4.1 Optimum Receiver for Binary Signals
∫0 sl (t ) dt =where
2ε ε is the
T
filter at t=T is simply (4.1-24)
signal energy.
A similar result is obtained if the signal sl(t) is correlated with
sl*(t) and the correlator is sampled signal sl(t) at t=T.
The optimum demodulator for the equivalent low-pass
received signal sl(t) may be realized by two matcher filters in
parallel, one matched to sl1(t) and the other to sl2(t).
143
5.4.1 Optimum Receiver for Binary Signals
145
5.4.1 Optimum Receiver for Binary Signals
146
5.4.1 Optimum Receiver for Binary Signals
147
5.4.1 Optimum Receiver for Binary Signals
148
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
The joint PDF of r=[r1c r1s r2c r2s] may be expressed as a
product of the marginal PDFs:
1 ⎡ (r1c − 2ε cos φ ) 2 + (r1s − 2ε sin φ ) 2 ⎤
p (r1c , r1s | s1 , φ ) = exp ⎢ − ⎥
2πσ 2
⎣ 2 σ 2
⎦
1 ⎛ r22c + r22s ⎞
p (r2 c , r2 s | s1 , φ ) = exp ⎜ − ⎟
2πσ 2
⎝ 2 σ 2
⎠
where σ2=2εN0.
The uniform PDF for the carrier phase φ (p(φ)=1/2π)
represents the most ignorance that can be exhibited by the
detector.
This is called the least favorable PDF for φ.
149
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
With p(φ)=1/2π, 0≤φ≤2π, substituted into the integral in
p(r|sm), we obtain:
1 2π
∫ p ( r1c , r1s | s1 , φ )d φ
2π 0
150
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
By performing a similar integration under the assumption
that the signal s2(t) was transmitted, we obtain the result
1 ⎛ r22c + r22s + 4ε 2 ⎞ ⎛ 2ε r22c + r22s ⎞
p ( r2 c , r2 s | s 2 ) = exp ⎜ − ⎟ I 0 ⎜⎜ ⎟
2πσ 2
⎝ 2σ 2
⎠ ⎝ σ 2
⎟
⎠
When we substitute these results into the likelihood ratio
given by equation Λ(r), we obtain the result
(
Λ (r ) =
)
I 0 2ε r12c + r12s σ 2 s1 P ( s )
>
( ) <
I 0 2ε r2 c + r2 s σ s2 P ( s1 )
2 2 2
2
?
The optimum detector computes the two envelopes r12c + r12s and r22c + r22s
(
and the corresponding values of the Bessel function I 2ε r12c + r12s / σ 2 )
( )
and I 2ε r22c + r22s / σ 2 to form the likelihood ratio.
0
0
151
5.4.1 Optimum Receiver for Binary Signals
The optimum detector
we observe that this computation requires knowledge of the
noise variance σ2.
The likelihood ratio is then compared with the threshold
P(s2)/P(s1) to determine which signal was transmitted.
A significant simplification in the
implementation of the optimum detector
occurs when the two signals are equally
probable. In such a case the threshold
becomes unity, and, due to the
monotonicity of Bessel function shown
in figure, the optimums detection rule
1
simplifies to r12c + r12s <> r22c + r22s
s2
152
5.4.1 Optimum Receiver for Binary Signals
153
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
Recall that in binary FSK we employ two different
frequencies, say f1 and f2=f1+Δf, to transmit a binary
information sequence.
The chose of minimum frequency separation Δf = f2 - f1 is
considered below:
s1 ( t ) = 2ε b Tb cos 2π f1t , s2 ( t ) = 2ε b Tb cos 2π f 2t , 0 ≤ t ≤ Tb
The equivalent low-pass counterparts are:
sl1 ( t ) = 2ε b Tb , sl 2 ( t ) = 2ε b Tb e j 2πΔft , 0 ≤ t ≤ Tb
2
f kc ( t ) = cos ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ , k = 0,1
Tb
2
f ks ( t ) = sin ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ , k = 0,1
Tb
155
5.4.1 Optimum Receiver for Binary Signals
Detection of binary FSK signal
If the mth signal is transmitted, the four samples at the detector
may be expressed as:
⎧ 2ε b
T ⎪ ⎪⎫ ⎪⎧ 2 ⎪⎫
rkc = ∫ r ( t )i f kc ( t ) dt = ∫ ⎨ cos ⎡⎣ 2π ( f1 + mΔf ) t + φm ⎤⎦ + n ( t ) ⎬i⎨ cos ⎡⎣( 2π f1 + 2π k Δf ) t ⎤⎦ ⎬ dt
T b b
⎩⎪ Tb ⎭⎪ ⎩⎪ Tb ⎭⎪
0 0
2 εb
∫ {cos ⎡⎣2π ( f }{ }
+ mΔf ) t ⎤⎦ cos φm − sin ⎡⎣ 2π ( f1 + mΔf ) t ⎤⎦ sin φm i cos ⎡⎣ 2π ( f1 + k Δf ) t ⎤⎦ dt + nkc
Tb
= 1
Tb 0
εb
{cos ⎡⎣2π ( m − k ) Δft ⎤⎦ + cos ⎡⎣2π ( 2 f + ( m + k ) Δf ) t ⎤⎦} cos φ −
Tb
=
Tb ∫ 0 1 m
0 0
⎪⎩ Tb ⎭⎪ ⎩⎪ Tb ⎭⎪
2 εb
∫ {cos ⎡⎣2π ( f }{ }
+ mΔf ) t ⎤⎦ cos φm − sin ⎡⎣ 2π ( f1 + mΔf ) t ⎤⎦ sin φm i sin ⎡⎣ 2π ( f1 + k Δf ) t ⎤⎦ dt + nks
Tb
= 1
Tb 0
εb
{− sin ⎡⎣2π ( m − k ) Δft ⎤⎦ + sin ⎡⎣2π ( 2 f + ( m + k ) Δf ) t ⎤⎦} cos φ −
Tb
=
Tb ∫ 0 1 m
158
5.4.2 Optimum Receiver for M-ary
Orthogonal Signals
If the equal energy and equally probable signal waveforms are
represented as
sm ( t ) = Re ⎡⎣slm ( t ) e j 2π fct ⎤⎦ , m = 1, 2,..., M , 0 ≤ t ≤ T
where slm(t) are the equivalent low-pass signals.
The optimum correlation-type or matched-filter-type
demodulator produces the M complex-valued random variables
rm = rmc + jrms = ∫ rl ( t ) hlm (T − t ) dt
T
*
0
= ∫ rl ( t ) s (T − (T − t ) ) dt = ∫ rl ( t ) slm ( t ) dt , m = 1, 2,..., M
T T
* *
0 lm 0
where rl(t) is the equivalent low-pass signals.
The optimum detector, based on a random, uniformly distributed
carrier phase, computes the M envelopes
rm = rmc2 + rms2 , m = 1, 2,..., M
or, the squared envelopes |rm|2, and selects the signal with the
largest envelope.
159
5.4.2 Optimum Receiver for M-ary
Orthogonal Signals
Optimum receiver for
M-ary orthogonal
FSK signals.
There are 2M
correlators: two for
each possible
transmitted frequency.
The minimum
frequency separation
between adjacent
frequencies to
maintain
orthogonality is Δf=
1/T. Demodulation of M-ary signals for
noncoherent detection
160
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
We assume that the M signals are equally probable a priori and
that the signal s1(t) is transmitted in the signal interval 0≦t≦T.
The M decision metrics at the detector are the M envelopes
161
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The PDFs of the random variables at the input to the detector are
⎛ ε (r2 + r2 ) ⎞
1 ⎛ r1c + r1s + ε s ⎞ ⎜ s 1c 1s ⎟
2 2
pr1 ( r1c , r1s ) = exp ⎜ − ⎟ I0 ⎜
2πσ 2
⎝ 2σ 2
⎠ ⎜ σ 2 ⎟⎟
⎝ ⎠
1 ⎛ rmc2 + rms2 ⎞
prm ( rmc , rms ) = exp ⎜ − ⎟, m = 2,3,..., M
2πσ 2
⎝ 2σ ⎠
2
162
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
Clearly, rmc=σRmcosΘm and rms =σRmsinΘm. The Jacobian of
this transformation is
σ cos Θ m σ sin Θ m
J = = σ 2 Rm
−σ Rm sin Θ m σ Rm cos Θ m
Consequently,
R1 ⎡ 1⎛ 2 ε s ⎞ ⎤ ⎛ 2ε s ⎞
p( R1 , Θ1 ) = exp ⎢ − ⎜ R1 + 2 ⎟ ⎥ I 0 ⎜⎜ R1 ⎟⎟
2π ⎣ 2⎝ N0 ⎠⎦ ⎝ N0 ⎠
Rm ⎛ 1 2⎞
p ( Rm , Θ m ) = exp ⎜ − Rm ⎟ , m = 2,3,..., M
2π ⎝ 2 ⎠
Finally, by averaging p(Rm,Θm) over Θm, the factor of 2π is
eliminated.
Thus, we find that R1 has a Rice probability distribution and Rm,
m=2,3,…,M, are each Rayleigh-distribued.
163
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The probability of a correct decision is simply the probability that
R1>R2, and R1>R3,…,and R1>Rm. Hence,
Pc = P ( R2 < R1 , R3 < R1 ,… , RM < R1 )
∞
= ∫ P ( R2 < R1 , R3 < R1 ,… , RM < R1 | R1 = x ) pR1 ( x ) dx
0
164
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The (M-1)th power may be expressed as
⎛ M − 1⎞ − nx2 2
( )
M −1 M −1
1− e = ∑ ( −1) ⎜
− x2 / 2 n
⎟e
n =0 ⎝ n ⎠
Substitution of this result into Equation 5.4-42 and integration
over x yield the probability of a correct decision as
M −1
n ⎛ M − 1⎞ 1 ⎡ nε s ⎤
Pc = ∑ ( −1) ⎜ ⎟ exp ⎢ ⎥
n =0 ⎝ n ⎠ n +1 ⎣ ( n + 1) N 0 ⎦
where εs/N0 is the SNR per symbol.
The probability of a symbol error, which is Pm=1-Pc, becomes
M −1
⎛ M − 1⎞ 1 ⎡ nkε b ⎤
PM = ∑ ( −1) ⎜
n +1
⎟ exp ⎢ − ⎥
n =1 ⎝ n ⎠ n +1 ⎣ ( n + 1) N 0 ⎦
where εb/N0 is the SNR per bit.
165
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
For binary orthogonal signals (M=2),
Equation reduces to the simple form
P2 = 12 e − ε b 2 N 0
For M>2, we may compute the
probability of a bit error by making
use of the relationship
2k −1
Pb = k PM
2 −1
which was established in Section 5.2.
Figure shows the bit-error
probability as a function of SNR per
bit γb for M=2,4,8,16, and 32.
166
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
Just as in the case of coherent detection of M-ary orthogonal
signals, we observe that for any given bit-error probability, the
SNR per bit decreases as M increase.
It will be show in Chapter 7 that, in the limit as M→∞, the
probability of bit error Pb can be made arbitrarily small provided
that the SNR per bit is greater than the Shannon limit of -1.6dB.
The cost for increasing M is the bandwidth required to transmit
the signals.
For M-ary FSK, the frequency separation between adjacent
frequencies is Δf=1/T for signal orthogonality.
The bandwidth required for the M signals is W=M Δf=M/T.
167
5.4.3 Probability of Error for Envelope
Detection of M-ary Orthogonal Signals
The bit rate is R=k/T, where k=log2M.
Therefore, the bit rate-to-bandwidth ratio is
R log 2 M
=
W M
168
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
In this section, we consider the performance of the envelope
detector for binary, equal-energy correlated signals.
When the two signals are correlated, the input to the detector is
the complex-valued random variables given by Equation 5.4-10.
We assume that the detector bases its decision on the envelopes
|r1| and |r2|, which are correlated (statistically dependent).
The marginal PDFs of R1=|r1| and R2=|r2| are Ricean distributed
and may be expressed as
⎧ Rm ⎛ Rm2 + β m2 ⎞ ⎛ β m Rm ⎞
⎪ exp ⎜ − ⎟ I0 ⎜ ⎟ ( Rm > 0 )
p ( Rm ) = ⎨ 2ε s N 0 ⎝ 4ε N 0 ⎠ ⎝ 2ε N 0 ⎠
⎪0
⎩ ( Rm < 0 )
m=1,2, where β1=2ε and β2=2ε|ρ|, based on the
assumption that signal s1(t) was transmitted.
169
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
Since R1 and R2 are statistically dependent as a consequence of
the nonorthogonality of the signals, the probability of error may
be obtained by evaluating the double integral
∞ ∞
Pb = P ( R2 > R1 ) = ∫ ∫ p ( x , x )dx dx
1 2 1 2
0 x1
Pb = Q1 ( a, b ) − e
1
2
− ( a 2 +b2 ) 2
I 0 ( ab )
where
a=
εb
2N (1−
0
1− ρ
2
) b=
εb
2N (1+
0
1− ρ
2
)
Q1(a, b) is the Marcum Q function defined in 2.1-123 and I0(x)
is the modified Bessel function of order zero.
171
5.4.4 Probability of Error for Envelope
Detection of Correlated Binary Signals
The error probability Pb is illustrated in Figure for several
values of |ρ|.
Probability of
error for
noncoherent
detection of
binary FSK
⎝ N0 ⎠
From the definition of Q1(a, b) in Equation 2.1-123, it follows
that
⎛ ε b ⎞ −εb 2 N0
Q1 ⎜⎜ 0, ⎟⎟ = e
⎝ N0 ⎠
173