Professional Documents
Culture Documents
0.9 0.9
0.8 0.8
0.7 0.7
0.6 0.6
I2
I2
0.6 0.6
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
I1 I1
(a) (b)
1 1
0.9 0.9
0.8 0.8
0.7 0.7
0.6 0.6
I2
I2
0.6 0.6
0.4 0.4
0.3 0.3
0.2 0.2
0.1 0.1
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
I1 I1
(c) (d)
Figure 10.33 (a) Extrinsic information transfer characteristic of decoder 1; (b) Extrinsic information
transfer characteristic of decoder 2; (c) Input-output extrinsic transfer characteristic of the two
constituent decoders working together; (d) EXIT chart, including the staircase (shown dashed)
embracing the extrinsic information transfer characteristics of both constituent decoders.
0.9
0.8
0.7
0.6
Ie
0.6
0.4
0.3
0.2
Figure 10.34
0.1
EXIT chart demonstrating nonconvergent
0
behavior of the turbo decoder when the EbN0 is 0 0.2 0.4 0.6 0.8 1
Ia
reduced compared to that in Figure 10.33d.
Haykin_ch10_pp3.fm Page 661 Friday, January 4, 2013 5:03 PM
a
2
L a m j = ------ m j + n a (10.129)
2
where na is the sample value of a zero-mean Gaussian random variable with variance 2a .
The rationale for the approximate Gaussian model just described is motivated by the
following two points (Lin and Costello, 2003):
a. For an AWGN channel with soft (i.e., unquantized) output, the log-likelihood ratio,
L-value, denoted by La(m j |rj (0)) of a transmitted message bit mj given the receiver
signal rj (0), may be modeled as follows (see Problem 10.36):
0 0
La m j rj = Lc rj + La mj (10.130)
where Lc = 4(Es N0 ) is the channel reliability factor defined in (10.91) and La(mj) is
the a priori L-value of message bit mj. The point to note here is that the product
terms Lcrj (0) for varying j are independent Gaussian random variables with variance
2Lc and mean Lc.
b. Extensive Monte Carlo simulations of the a posteriori extrinsic L-values, Le(mj), for
a constituent decoder with large block length appear to support the Gaussian-model
assumption of (10.129); see Wiberg et al. (1999).
Accordingly, using the Gaussian approximation of (10.129), we may express the
conditional probability density function of the a priori L-value as follows:
2
1 – mj a 2 2
f L m j = ------------------ exp – -----------------------------------
- (10.131)
a
2 a 2 2 a
where is a dummy variable, representing a sample value of La(mj). Note also that is
continuous whereas, of course, mj is discrete. It follows that in formulating the mutual
information between the message bit mj = +1 and a priori L-value La(mj) we have a binary
Haykin_ch10_pp3.fm Page 662 Friday, January 4, 2013 5:03 PM
input AWGN channel to deal with; such a channel was discussed previously in Example 5
of Chapter 5 on information theory. Building on the results of that example, we may
express the first desired mutual information, denoted by I1(mj; La), as follows:
1 2f L m j
I 1 m j ; L a = --- – fL mj log2 -----------------------------------------------------------------------------
- d
a
(10.132)
2 a f L m j = –1 + f L m j = +1
m j = –1, +1 a a
where the summation accounts for the binary nature of the information bit mj and the
integral accounts for the continuous nature of La. Using (10.131) and (10.132) and
manipulating the results, we get (ten Brink, 2001):
2
– m j 2a 2
exp – ----------------------------------- -
2 2a
I 1 m j ; L a = 1–
–
------------------------------------------------------ log 2 1 + exp – d
2 a
(10.133)
2
which, as expected, depends solely on the variance a . To emphasize this fact, let the new
function
𝒥 a := I1 mj ; La (10.134)
and with it, the corresponding Gaussian random variable La(mj) defined in (10.129) is
obtained.
Referring back to (10.128), we note that for us to construct the EXIT chart we also need
to know the second mutual information between the message bit mj and the a posteriori
extrinsic L-value Lp(mj). To this end, we may build on the formula of (10.132) to write
1 2f L m j
I 2 m j ; L p = --- – fL mj log2 -----------------------------------------------------------------------------
- d
p
(10.137)
2 p f L m j = –1 + f L m j = +1
m j = –1, +1 p p
and the corresponding L-values produced at the output of the corresponding BCJR decoder
in the receiver. Due to the inherently nonlinear input–output characteristic of the BCJR
decoder, the underlying distribution of the L-values is not only unknown, but also highly
likely to be non-Gaussian as well, thereby complicating the measurement. To get around this
difficulty, we may invoke the ergodic theorem, which was discussed previously in Chapter 4
on stochastic processes. As explained therein, under certain conditions, it is feasible to
replace the operation of ensemble averaging (i.e., expectation) with time averaging. Thus, in
proceeding along this ergodic path, we have a new nonlinear transformation, where the time
average of a large set of L-value samples, available at the output of the BCJR decoder,
provides an estimate of the desired mutual information; moreover, it does so without
requiring knowledge of the original data (i.e., the mj). It is for this reason that the second
method of computing EXIT charts is called the averaging method.15
Just as the use of a single constituent decoder suffices for computing an EXIT chart in
the histogram method, the same decoding scenario applies equally well to the averaging
method. It is with this point in mind that the underlying scheme for the averaging method
is as depicted in Figure 10.35. Most importantly, this scheme is designed in such a way
that the following requirements are satisfied:
1. Implementations of channel estimation, carrier receiver, modulation, and
demodulation are all perfect.
2. The turbo decoder is perfectly synchronized with the turbo encoder.
3. The BCJR algorithm or exact equivalent (i.e., the log-MAP algorithm) is used to
optimize the turbo decoder.
Moreover, the following analytic correspondences between the constituent encoder 1 at
the top of Figure 10.35 and the turbo decoder 2 at the bottom of the figure are carefully
noted: the code vectors c(0), c(1), and termination vector t(1) in the encoder map onto the a
posteriori L-values Lp(r(0)), Lp(r(1)), and Lp(z(1)) in the decoder, respectively.
It can therefore be justifiably argued that in light of these rigorous requirements, the
underlying algorithm for the averaging method is well designed and therefore trustworthy
in the following sense: in the course of computing the EXIT chart, the algorithm trusts
what the computed L-values actually say; that is, they do not under or over represent their
confidence in the message bits. This important characteristic of the averaging method is to
be contrasted against the histogram method. Indeed, it is for this reason that the histogram
method compares the L-values against the true values of the message bits, hence requiring
knowledge of them.
In summary, we may say that trustworthy L-values are those L-values that satisfy the
consistency condition. A simple way of testing this condition is to do the following: use
the averaging and histogram methods to compute two respective sets of L-values. If, then,
both methods yield the same value for the mutual information, then the consistency
condition is satisfied (Maunder, 2012).