Professional Documents
Culture Documents
Communications
Moh. Khalid Hasan*, Md. Shahjalal, Md. Mainul Islam, Md. Morshed Alam, Md. Faisal Ahmed, and Yeong Min Jang**
Dept. of Electronics Engineering, Kookmin University, Seoul, South Korea
Email: *khalidrahman45@ieee.org, **yjang@kookmin.ac.kr
Abstract— In the coming future, it is obvious that the wireless In NOMA, a multiuser signal is multiplexed using
networks will be congested with massive amounts of data traffic superposition coding in the transmitter and then sent to the users
with the increasing number of users. Current multiple access using different power levels in a non-orthogonal basis. Usually,
techniques will certainly not have the capability to efficiently serve the users with bad channel conditions are allocated higher
in the massively congested scenarios. In recent times, non-
power than the users with good channel conditions. Therefore,
orthogonal multiple access (NOMA) has been recognized as an
immensely potential technique for 5G and beyond
having proper channel state information (CSI) in the transmitter
communications that can increase spectral efficiency to a greater has a profound significance in NOMA. When the signal is
extent serving a vast number of users. However, several collected in the receiver, the user with the stronger channel gain
circumscriptions are observed in NOMA such as the requirement retrieves its signal immediately. On the other hand, the user
of a perfect channel state information in the transmitter and high with the weaker channel recognizes the other signals as
computational complexity in the receiver. The use of deep learning interference and performs successive interference cancellation
(DL) techniques is a great resolution to deal with the challenges. (SIC) technique to retrieve its original signal. In the
This paper discusses the applications of the DL methods in NOMA conventional orthogonal multiple access systems, the
for 5G and beyond communications. Firstly, the deep neural
orthogonality between channels is achieved using guard periods
networks that are employed in NOMA are listed up and discussed.
After that, their functions are studied specifically focusing on how to avoid interference, which particularly decreases the spectral
to improve the NOMA performance. Finally, the possible future efficiency to a great extent. This problem is completely
challenges and research issues are identified at the end of the remediated using NOMA, where no guard period is used and so
paper. signals overlap with each other. NOMA provides fantastic
throughput with maintaining an excellent level of user fairness
Index terms— Non-orthogonal multiple access (NOMA), deep
[9].
learning (DL), and 5G.
In terms of theoretical perspective, SIC can be considered
I. INTRODUCTION as an optimal scheme for both uplink and downlink
communication [10]. With the perfect SIC, the performance of
Next-generation wireless networks will require an
NOMA is superior compared to the existing OFDMA scheme
extensive capacity to support an immense amount of advanced
[11]. Perfect CSI is very crucial for performing SIC in the
multimedia applications [1, 2]. The fifth-generation (5G)
receiver. It is pretty difficult to design an efficient power
technologies face significant challenges with the explosive
allocation scheme without having the exact CSI at the
growth of internet-of-thing-based large scale heterogeneous
transmitter. Acquiring a perfect or near-perfect CSI is a
networks. The uplink/downlink transmission of future massive
challenging task. Fortunately, deep learning (DL) techniques
incremental data to a huge number of users will face numerous
are already being proved to be a useful resolution to this
hindrances using the current multiple-access techniques due to
problem. It can be incorporated in NOMA to detect a
their incapability to support efficient spectrum usage [3, 4]. The
completely unknown and sharply changing channel condition.
necessity of a new method for 5G and beyond communication
In this paper, we enlist the functions of DL to improve the
in handling a large quantity of data is realized by researchers
performance of NOMA systems. We first provide a short
over the last few years [5].
overview of the DL methods that are useful in NOMA and other
Non-orthogonal multiple access (NOMA) is a promising wireless networking applications. Afterward, we specifically
technique that has drawn significant research interest in recent discuss the applications of DL in NOMA. Finally, we provide
years due to its potential applications in 5G and beyond a short overview of the challenges that can be observed using
networks [5, 6]. It brings numerous advantages to the current DL methods in NOMA.
wireless networks such as an increase in spectral efficiency,
reliability, and connectivity reducing overall latency. NOMA II. DEEP LEARNING IN NOMA
exploits the existing resources in a more efficient basis than the
The current improvements in DL algorithms have brought
conventional multiple access techniques [7, 8].
significant benefits to the area of wireless communications.
x xˆ ( H , P, yi )
Data in wireless systems, such as CSI, are increasingly real-time and the input signals can be trained with the help of
heterogeneous and display intricate correlations [12]. Efficient that data. This type of learning is called online learning. A block
training of these data can exhibit better performance. In NOMA, diagram of the NOMA system with the integration of both
the employment of DL is extensively researched over the last online and offline training for the auto-detection of CSI is
few years. Different kinds of learning algorithms are utilized in depicted in Fig. 2. DL can also be used for estimating the
different aspects and demonstrate better performance compared channel considering adverse weather. For example, in [14], a
to traditional schemes. In general, three types of learning DL technique, called long-short term memory (LSTM) is used
methods are developed in the literature, such as supervised, for detecting dynamically fluctuating channel. Herein, the
unsupervised, and reinforcement. Supervised learning is hidden layers can act as a memory for the network state, which
extensively utilized to estimate the CSI. In addition, it can also particularly permits the DL to save, recall, and process the
be used in other applications such as spectrum sensing, preceding complex data.
localization, and throughput prediction. Unsupervised learning
DL can also be employed in NOMA for other purposes as
is mostly used in user clustering and congestion control. Deep
follows.
reinforcement learning is a different type of DL method that is
considered to have immense significance in wireless a. Resource allocation
communications. Until now, the most recognized application of
Scheduling can be very important in the cases where
reinforcement learning is resource management. It uses a
NOMA subchannels cannot provide sufficient bandwidth to the
special type of learning algorithm, called Q learning, and
users. A proper resource allocation algorithm can maximize the
produces the output without having any preliminary knowledge
overall sum rate. Therefore, the performance of the NOMA
about the learning environment.
system greatly depends on the resource allocation strategies
In NOMA, DL is employed in different applications. DL [11]. DL is a useful method that can be applied to overcome the
can be extremely useful for complex data processing and to resource allocation problem in the case of multiple users [15,
acquire perfect CSI. Two types of training can be identified in 16]. In addition, dynamic resource allocation is a challenge for
NOMA, such as online and offline training. A DL-based a fast-changing channel condition. Optimizing energy
NOMA model [13] is depicted in Fig. 1. The hidden layer is consumption with efficient resource allocation is a significant
employed for training and recognition [13]. Hidden layers are research issue using NOMA. Intelligent learning methods are
particularly equipped with multiple neurons. In Fig. 1, one of extensively developed to deal with this challenge. In [17], a DL
the layers is named as noise layer, where the processed signal method, called deep belief network, is proposed for resource
can be corrupted using additive-white Gaussian noise. The CSI allocation by minimizing the total transmit power.
can be acquired automatically by extensive training of the input Reinforcement learning can also be useful in solving the
data using existing channel models. Arbitrary sequences of the dynamic resource allocation problem [15]. In reinforcement
input signals are extensively trained using CSI of different learning, a deep deterministic policy gradient is used for
environments collected from simulations. This type of training deriving the resource allocation decision. In addition, the
is called offline training. In addition, the output signals can be subcarrier assignment decision is evaluated with the help of Q
fed into the training data to accurately predict the channel learning.
environment. Pilot signals can be utilized to carry the CSI in
304
Transmitter Channel Receiver
P1
Signal 1 Signal 2
Signal 1 detection detection
Signal 1 Power
Fading
Modulation allocation ∑ AWGN
channel
Signal 2 deployment P1>P2
Signal 1 Subtraction of
Signal 2 detection signal 1
P2
Offline Real-time
Simulated data CSI attainment Online training
training channel data
Fig. 2. Block diagram of the NOMA system employing online and offline DL mechanisms.
305
Table 1: Applications of DL-methods in NOMA
Reference Type of learning Deep learning model Application
[11] Supervised Recurrent neural network Rapid and optimized resource allocation
[14] [13] Supervised LSTM Channel estimation
[15] Reinforcement Deep Q-network Subcarrier assignment and power allocation
[16] Reinforcement Attention-based neural network Joint power allocation and channel assignment
[17] Supervised and Deep belief network Power optimization
unsupervised
[18] Supervised Deep neural network Automatic realization of the CSI
[21] Unsupervised Deep neural network Signal constellation design
[25] Supervised Deep neural network Reliability improvement of grant-free access
[26] Supervised learning Deep neural network Optimization for energy-efficient scheduling
[27] Reinforcement Deep neural network [27] based on Power allocation optimization
Q-learning [28]
306
[16] C. He, Y. Hu, Y. Chen, and B. Zeng, “Joint power allocation and
channel assignment for NOMA with deep reinforcement
learning,” IEEE Journal on Selected Areas in Communications,
vol. 37, no. 10, pp. 2200-2210, Oct. 2019.
[17] J. Luo, J. Tang, D. K. C. So, G. Chen, K. Cumanan, and J. A.
Chambers, “A deep learning-based approach to power
minimization in multi-carrier NOMA with SWIPT,” IEEE
Access, vol. 7, pp. 17450-17460, 2019.
[18] C. Lin, Q. Chang, and X. Li, “A Deep Learning Approach for
MIMO-NOMA Downlink Signal Detection,” Sensors, vol. 19,
2019.
[19] J. Kang, I. Kim, and C. Chun, “Deep learning-based MIMO-
NOMA with imperfect SIC decoding,” IEEE Systems Journal,
pp. 1-4.
[20] D. A. Awan, R. L.G. Cavalcante, M. Yukawa, and S. Stanczak,
“Detection for 5G-NOMA: an online adaptive machine learning
approach," arXiv:1711.00355.
[21] L. Jiang, X. Li, N. Ye, and A. Wang, “Deep learning-aided
constellation design for downlink NOMA,” in Proc.
International Wireless Communications & Mobile Computing
Conference (IWCMC), Tangier, Morocco, 2019.
[22] F. Alberge, “Constellation design with deep learning for
downlink non-orthogonal multiple access,” in Proc. IEEE
Annual International Symposium on Personal, Indoor and
Mobile Radio Communications (PIMRC), Bologna, 2018.
[23] Y. Liu, H. Zhang, K. Long, A. Nallanathan, and V. C. M. Leung,
“Energy-efficient subchannel matching and power allocation in
NOMA autonomous driving vehicular networks,” IEEE
Wireless Communications, vol. 26, no. 4, pp. 88-93, Aug. 2019.
[24] P. Yang, L. Li, W. Liang, H. Zhang, and Z. Ding, “Latency
optimization for multi-user NOMA-MEC offloading using
reinforcement learning,” in Proc. Wireless and Optical
Communications Conference (WOCC), Beijing, China, 2019.
[25] N. Ye, X. Li, H. Yu, A. Wang, W. Liu, and X. Hou, “Deep
learning aided grant-free NOMA toward reliable low-latency
access in tactile internet of things,” IEEE Transactions on
Industrial Informatics, vol. 15, no. 5, pp. 2995-3005, May 2019.
[26] L. Lei et al., “Learning-assisted optimization for energy-
efficient scheduling in deadline-aware NOMA systems,” IEEE
Transactions on Green Communications and Networking, vol.
3, no. 3, pp. 615-627, Sep. 2019.
[27] K. N. Doan, M. Vaezi, W. Shin, H. V. Poor, H. Shin, and Tony
Q. S. Quek, “Power allocation in cache-aided NOMA systems:
optimization and deep reinforcement learning approaches,”
arXiv:1909.11074.
[28] L. Xiao, Y. Li, C. Dai, H. Dai, and H. V. Poor, “Reinforcement
learning-based NOMA power allocation in the presence of smart
jamming,” IEEE Transactions on Vehicular Technology, vol.
67, no. 4, pp. 3377-3389, Apr. 2018.
307