Professional Documents
Culture Documents
fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
1
Abstract—In this article, we develop an end-to-end wireless assumed channel models [1]. Recently, deep learning has
communication system using deep neural networks (DNNs), been applied to refine the traditional block-structure commu-
where DNNs are employed to perform several key functions, nication systems, including the multiple-input and multiple-
including encoding, decoding, modulation, and demodulation.
However, an accurate estimation of instantaneous channel trans- output (MIMO) detection [2], [3], channel decoding [4]–[9],
fer function, i.e., channel state information (CSI), is needed in and channel estimation [10], [11]. In addition, deep learning
order for the transmitter DNN to learn to optimize the receiver based methods have also shown impressive improvement by
gain in decoding. This is very much a challenge since CSI jointly optimizing the conventional communication blocks,
varies with time and location in wireless communications and including joint channel estimation and detection [12], joint
is hard to obtain when designing transceivers. We propose to
use a conditional generative adversarial net (GAN) to represent channel encoding and source encoding [13].
channel effects and to bridge the transmitter DNN and the Besides enhancing the traditional communication blocks,
receiver DNN so that the gradient of the transmitter DNN deep learning provides a new paradigm for future commu-
can be back-propagated from the receiver DNN. In particular, nication systems. As a pure data-driven method, the features
a conditional GAN is employed to model the channel effects and the parameters of a deep learning model can be learned
in a data-driven way, where the received signal corresponding
to the pilot symbols is added as a part of the conditioning directly from the data, without handcraft or ad-hoc designs,
information of the GAN. To address the curse of dimensionality by optimizing an end-to-end loss function. Inspired by this
when the transmit symbol sequence is long, convolutional layers methodology, end-to-end learning based communication sys-
are utilized. From the simulation results, the proposed method tems have been investigated in several prior works [14]–[18],
is effective on additive white Gaussian noise (AWGN) chan- where both the transmitter and the receiver are represented
nels, Rayleigh fading channels, and frequency-selective channels,
which opens a new door for building data-driven DNNs for end- by deep neural networks (DNNs), as shown in Fig. 1(b).
to-end communication systems. This framework can be interpreted as an auto-encoder system,
which is widely used in learning representations of the data
Index Terms—Channel GAN, CNN, end-to-end communication
system, channel coding. [19]–[21]. The transmitter and the receiver correspond to the
auto-encoder and auto-decoder, respectively. From Fig. 1(b),
I. I NTRODUCTION the transmitter learns to encode the transmitted symbols into
encoded data, x, which is then sent to the channel while the
In a traditional wireless communication system shown in
receiver learns to recover the transmitted symbols based on
Fig. 1(a), the data transmission entails multiple signal pro-
the received signal, y, from the channel. As a result, the
cessing blocks in the transmitter and the receiver. While
traditional communication modules at the transmitter, such as
the technologies in this system are quite mature, individual
the encoding and modulation, are replaced by a DNN while
blocks therein are separately designed and optimized, often
the modules at the receiver, such as the decoding and the
with different assumptions and objectives, making it difficult,
demodulation, are replaced by another DNN. The transmitter
if not impossible, to ascertain global optimality of the sys-
and the receiver DNNs are trained offline using measurement
tem. In addition, the channel propagation is expressed as an
or simulated data set in a supervised learning manner opti-
assumed mathematical model embedded in the design. The
mizing a loss function that reflects recovery accuracy. From
assumed model may not correctly or accurately reflect the
Fig. 1(b), the loss function depends on the weight/parameter
actual transmission scenario, thereby compromising the system
set of the transmit DNN, θT , the weight/parameter set of
performance.
the receiver DNN, θR , and channel realization, h, and thus
On the contrary, the learning based data-driven methods
can be expressed as L(θT , θR , h). However, in the offline
provide a new way for handling the imperfection of the
training phase, the channel realization, h, is unknown. We
This work was supported in part by a research gift from Intel Corporation only know that it is from a sample space/set, H, with certain
and the National Science Foundation under Grants 1815637 and 1731017. features or characteristics, such as indoor channels or typical
H. Ye, G. Y. Li, and B-H. F. Juang are with the School of Elec-
trical and Computer Engineering, Georgia Institute of Technology, At- urban channels. Therefore, the parameter sets of the transmitter
lanta, GA 30332 USA (email: yehao@gatech.edu; liye@ece.gatech.edu; and receiver DNNs are essentially optimized to minimize
juang@ece.gatech.edu) E(h∈H) {L(θT , θR , h)}.
L. Liang is with Intel Labs, Hillsboro, OR 97124 USA. This work was done
when he was with the School of Electrical and Computer Engineering, Georgia From the above discussion, several critical challenges in
Institute of Technology, Atlanta, GA 30332 USA (email: lliang@gatech.edu) the learning based end-to-end communication system need
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
2
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
3
decoders. In this section, previous works in the related topics with a small block length. For example, blocks of eight
are briefly reviewed. information bits are used in [17]. How to extend the end-
to-end framework to a large block size and how to model the
A. GANs and Conditional GANs unknown channel using a data-driven approach are still open
problems, which are addressed in our proposed approach.
GANs have been proposed in [25] as a generative frame-
work, where a generator and a discriminator are competing
with each other in the training stage. By the feedback of the C. Learning based Decoders
discriminator, the generator improves its ability to generate Our proposed method is also closely related to learning
samples that are similar to the real samples. GAN and its based encoding and decoding. Learning based approach has
variants [26], [27] are most widely used in computer vision. been utilized in improving decoding performance for a long
Much of the recent GAN research is focusing on improving time. It dates back to the 1990s when several attempts had been
the quality of the generated images [28]. made to decode the codes with the recurrent neural network
In order to generate samples with a specific property, a (RNN) [31].
conditional GAN is proposed based on the GAN framework, With the widely used deep learning approaches, DNN has
where the context information is added to the generator and been utilized in decoding with a wide range of applications. In
the discriminator. Originally, the condition added is the label [7], a fully connected neural network is trained for decoding
information so that the generator can generate corresponding short Polar codes. The performance of decoding is similar to
samples given a particular category. Nowadays, conditional maximum likelihood decoding. But they find it difficult to
GANs are widely used in changing the style and the content break through the curse of dimensionality. In order to train
of the input [29], [33]. For instance, GANs have been utilized long codewords, a partition method has been employed in [7].
to generate high-resolution images from low-resolution images Moreover, there have been several trials to incorporate some
[29]. prior information in the decoding process. RNN is used in [5]
Apart from applications in computer vision, recently GANs for decoding the convolutional and turbo codes. In [4], the
have been exploited to model the channel effects of additive traditional belief-propagation decoding algorithm is extended
white Gaussian noise (AWGN) channels [24], similar to our as deep learning layers to decode linear codes. Recently, the
work. However, our approach can be applied to more realistic curse of dimensionality is mitigated when the convolutional
fading channels by using conditional GAN, which employs and circular convolutional layers are utilized to learn the
the received pilot information as a part of the condition encoding and decoding modules simultaneously rather than
information when generating the channel outputs. just learning to decode human-designed codewords [6].
B. DNN based End-to-End Communications III. M ODELING C HANNEL WITH C ONDITIONAL GAN
The objective of the paper is to build an end-to-end com- An end-to-end communication system learns to optimize
munication system for wireless channels, where the channel DNNs for the transmitter and the receiver. However, the back-
effects are modeled with conditional GAN. The idea of end- propagation algorithm, which is used to train the weights of
to-end learning communication systems has been proposed in DNNs, is blocked by the unknown CSI, preventing the overall
[14] and has been shown to have a similar performance as the learning of the end-to-end system. To address the issue, we use
traditional approaches with block structures under the AWGN a conditional GAN to learn the channel effects and to act as
condition. In [15], the end-to-end method has been extended to a bridge for the gradients to pass through. By the conditional
handle various hardware imperfection. In [16], an end-to-end GAN, the output distribution of the channel can be learned in
learning method is adopted within the orthogonal frequency- a data-driven manner and therefore many complicated effects
division multiplexing (OFDM) system. In [30], CNNs are of the channel can be addressed. In this section, we introduce
employed for modulation and demodulation, where improved the conditional GAN and discuss how to use it to model the
results have been shown for very high order modulation. In channel effects.
addition, source coding can also be considered as a part of
the end-to-end communication system for transmitting text or
image [13]. A. Conditional GAN
Training the end-to-end communication system without GAN [25] is a new class of generative methods for distri-
channel models has been investigated recently. A reinforce- bution learning, where the objective is to learn a model that
ment learning based framework has been employed in [17] can produce samples close to some target distribution, pdata .
to optimize the end-to-end communication system without In our system, a GAN is applied to model the distribution of
requiring the channel transfer function or CSI, where the the channel output and the learned model is then used as a
channel and the receiver are considered as the environment surrogate of the real channel when training the transmitter so
when training the transmitter. The recovery performance at that the gradients can pass through to the transmitter.
the receiver is considered as the reward, which guides the As shown in Fig. 2, the GAN system consists of two
training of the transmitter. In [18], a model-free end-to-end DNNs, i.e., the generator, G, and the discriminator, D. The
learning method has been developed based on simultaneous input to the generator, G, is a noise vector z sampled from
perturbation methods. However, both works are conducted a prior distribution pz , e.g., uniform distribution. The input
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
4
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
5
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
6
transmitter can learn the constellation of the embedding, x, so TABLE I: Model Parameters of FCN
that the received signal can be easily detected at the receiver. Parameters Values
Transmitter: Neurons in each hidden layers 32, 32
Learning rate 0.001
D. Training Channel GAN Receiver: Neurons in each hidden layers 32, 32
Learning rate 0.001
The training procedure of channel GAN is illustrated in Generator: Neurons in each hidden layers 128, 128, 128
Algorithm 1 in detail. In each iteration, the generator and Discriminator: Neurons in each hidden layers 32, 32, 32
the discriminator are trained iteratively. The parameters of Learning rate 0.0001
one model will be fixed while training the other. With the
learned transmitter, the real data can be obtained with the
encoded signal from the transmitter going through the real TABLE II: Model Parameters of CNN
channel while the fake data is obtained from the encoded data
Type of layer Kernel size/Annotation Output size
going through the channel generator. The parameters of the Transmitter
generator and the discriminator are updated according to the Input Input layer K×1
loss function of equations (3) and (4), respectively. Conv+Relu 5 K × 256
Conv+Relu 3 K × 128
Conv+Relu 3 K × 64
Algorithm 1 Channel GAN Training Algorithm Conv 3 K×2
1: for number of training iterations do Normalization Power normalization K×2
Receiver
2: % Updating the Generator Conv+Relu 5 K × 256
3: Sample minibatch of transmit data {s} and channel {h} Conv+Relu 5 K × 128
4: Get the minibatch of condition information {m} from Conv+Relu 5 K × 128
the output of the transmitter DNN and the received pilot Conv+Relu 5 K × 128
Conv+Relu 5 K × 64
signal from the channel {h}. Conv+Relu 5 K × 64
5: Sample minibatch of samples {z}. Conv+Relu 5 K × 64
6: Update the generator by ascending the stochastic gra- Conv+Sigmoid 3 K×1
Generator
dient of the loss function (3). Conv+Relu 5 K × 256
7: % Updating the Discriminator Conv+Relu 3 K × 128
8: Sample minibatch of transmit data {s} and channel Conv+Relu 3 K × 64
{h}. Conv 3 K×2
Discrimator
9: Get the minibatch of condition information {m} from Conv+Relu 5 K × 256
the output of the transmitter DNN and the received pilot Conv+Relu 3 K × 128
signal from the channel {h}. Conv+Relu 3 K × 64
Conv+Relu 3 K × 16
10: Sample minibatch of examples real data by collecting
FC +Relu 100 100
the output of channel. FC+Sigmoid 1 1
11: Sample minibatch of samples {z}
12: Update the discriminator by descending the stochastic
gradient of the loss function (4).
13: end for
and CNN are shown in Table I and Table II, respectively. The
weights of both models are updated by Adam [32] and the
batch size for training is 320.
V. E XPERIMENTS
In this section, the implementation details of the end-to- 2) Channel Types: Three types of channels are considered
end learning based approach are provided and the simulation in our experiments, i.e., AWGN channels, Rayleigh channels,
results are presented. For several types of most commonly used and frequency-selective multipath channels. In an AWGN
channels, the channel GAN has shown the ability to model channel, the output of the channel, y, is the summation of the
the channel effects in a data-driven way. In addition, the end- input signal, x, and Gaussian noise, w, that is, y = x + w.
to-end communication system, which is built on the channel Rayleigh fading is a reasonable model for narrow-band wire-
GAN, can achieve similar or better results even the channel less channels when many objects in the environment scatter
information is unknown during training and optimizing the the radio signal before arriving at the receiver. In a Rayleigh
transmitter and the receiver. channel, the channel output is determined by y = hn · x + w,
where hn ∼ CN (0, 1). The channel coefficient hn is time-
varying and is unknown when design transceivers. Therefore,
A. Experimental Settings channel estimation is required to get the instantaneous CSI,
1) Implementation Details: Two types of DNN models are which is then used by the receiver to detect the transmit data.
designed in our experiments. One is fully connected networks With frequency-selective channels, radio signal propagates via
(FCN) and the other is the CNN. The FCN is used for a small multiple paths, which differ in amplitudes, phases, and delay
block size and the CNN is used in the a large block size to times, and cause undesired frequency-selective fading and
avoid the curse of dimensionality. The parameters of the FCN time dispersion of the received signal. The baseband complex
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
7
channel impulse response can be expressed as with different channel gains and phase rotations according to
Kp
conditioning information.
X
h(t) = bk ejθk p(t − τk ),
k=0
can work with various pilot signals other than the delta signal
3
as indicated in our prior work [12]. In addition, for AWGN
and Rayleigh fading channels, the proposed conditional GAN 2
Gaussian noise directly to the hidden layer [14]. For the Iteration
Rayleigh fading channel, an additional equalization layer is Fig. 5: KL divergence of the generated channel distribution
employed [17]. The input of the auto-decoder is the equalized and real channel distribution.
signal ŷ = yĥ , where ĥ is obtained from the pilot information.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
8
BER
10−3
100
Hamming(7,4) wi h MLD
E2E-GAN
E2E-AWGN 10−4
10−1
10−5
10−2 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
SNR (dB)
BER
10−3
100
RSC
E2E-GAN
10−4 E2E-AWGN
10−1
10−5
0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
SNR (dB)
BLER
10−2
100
Hamming(7,4) wi h MLD
E2E-GAN 10−3
E2E-AWGN
10−1
10−4
0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
SNR (dB)
BLER
10−2
Fig. 7: BER and BLER with a block size of 128 bits under
an AWGN channel.
10−3
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
9
100 100
QAM+RSC OFDM+QAM+RSC
E2E-GAN E2E
E2E-Rayleigh E2E witho t pilots
10−1 10−1
10−2 10−2
BER
BER
10−3 10−3
10−4 10−4
10−5 10−5
0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
SNR (dB) SNR (dB)
100 100
QAM+RSC
E2E-GAN
E2E-Rayleigh
10 1 10−1
BLER
BLER
10 2 10−2
10 3 10−3
OFDM+QAM+RSC
E2E
E2E wi hou pilo s
10 4 10−4
0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0 0.0 2.5 5.0 7.5 10.0 12.5 15.0 17.5 20.0
SNR (dB) SNR (dB)
Fig. 8: BER and BLER under a Rayleigh channel. Fig. 9: BER and BLER under a frequency-selective multipath
channel.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
10
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.
This article has been accepted for publication in a future issue of this journal, but has not been fully edited. Content may change prior to final publication. Citation information: DOI 10.1109/TWC.2020.2970707, IEEE
Transactions on Wireless Communications
11
[34] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard, W. [36] S. Kullback and R. A. Leibler, “On information and sufficiency,” Ann.
Hubbard, and L. D. Jackel, “Backpropagation applied to handwritten zip Math. Statist., vol. 22, no. 1, pp. 79–86, Mar. 1951.
code recognition,” Neural Computation, vol. 1, no. 4, pp. 541–551, Dec. [37] Q. Wang, S. R. Kulkarni, and S. Verdú, “Divergence estimation for mul-
1989. tidimensional densities via k-nearest-neighbor distances,” IEEE Transac-
[35] A. Viterbi, “Error bounds for convolutional codes and an asymptotically tions on Information Theory, vol. 55 no. 5, pp. 2392-2405, May 2009.
optimum decoding algorithm,” IEEE Trans. Inf. Theory, vol. 13, no. 2, [38] P. Kyosti, “IST-4-027756 WINNER II D1.1.2 v.1.1: WINNER II Chan-
pp. 260-269, Apr. 1967. nel Models,” 2007, [online] Available: http://www.ist-winner.org.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/.