Professional Documents
Culture Documents
This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
Date of publication xxxx 00, 0000, date of current version xxxx 00, 0000.
Digital Object Identifier 10.1109/ACCESS.2022.Doi Number
INDEX TERMS Deep learning, Gamma-rays, Gamma-ray telescopes, LSTM, Signal classification
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
Because gamma rays are absorbed in the atmosphere, atmospheric Cherenkov devices comprises multidimensional
studying the universe using gamma rays from Earth is data. In this article, a classification model based on deep
extremely difficult [4]. Prior to the successful adoption of learning have been employed to sort gamma and hadron
ground-based detection of gamma rays, satellites were the signals from MAGIC gamma telescope data.
only means of exploring the high energy universe [5]. The II. Related Works
most significant constraint on the lower sensitivity of satellite- Changfan Zhang et al. [17] made an effective method for
based investigations is their confined effective area. As long extracting parameters from gamma-ray emissions of special
as there was no significant improvement in the effective area, nuclear materials (SNM) and identifying SNM classes using a
there was very little hope of efficiently probing the VHE backpropagation neural network (BPNN) and template
universe. This goal was achieved through the successful matching approach.
operation of ground-based telescopes detecting and For a ground-based atmospheric Cherenkov telescope,
identifying atmospheric Cherenkov radiation [6]. Ground- Sharma Mradul et al. [18] looked into gamma-hadron
based telescopes have an effective area of ∼105 m2 as separation in great detail. They employed Monte Carlo event
compared to ∼1 m2 for satellite-based telescopes. Despite the simulation to evaluate and compare various supervised ML
fact that the gamma ray to hadron event ratio is ∼10-3, gamma- techniques. These included the Random Forest (RF) method,
hadron classification is a major issue. Furthermore, adopting a Artificial Neural Networks (ANN), Linear Discriminant
stereoscopic system with two or more telescopes enhances the analysis, Naive Bayes (NB) Classifiers, Support Vector
rejection of hadronic events by a factor of 100 [7]. As high- Machines (SVM), and the conventional dynamic supercut
energy cosmic rays pass through the atmosphere, they produce technique. For gamma-hadron segregation, the Random
secondary particles, resulting in widespread air showers. Forest approach has proved to be the most effective ML
These particles travel at relativistic speeds, resulting in method.
Cherenkov radiation in the atmosphere. This radiation is A case study that compares multivariate classification
detected using ground-based detectors. The IACT-based approaches was presented by R.K. Bock et al. [19]. The input
telescopes gather the Cherenkov photons that fall on these for the imaging gamma-ray Cherenkov telescope consists of
telescopes. These telescopes have a camera that collects Monte Carlo data, which has been generated and pre-
Cherenkov photons that are reflected from the mirror. The processed. Both incoming gamma rays and hadronic showers
camera is made up of photomultiplier tubes that are linked by contribute to this information, which is then separated into its
rapid electronics that digitize and observe the Cherenkov respective categories. The data is well-suited for testing
photon pulse [8]. classification algorithms due to the low contrast between the
Deep learning methods [9-12] provide better solutions for signal (gamma) and background (hadrons).
multivariate problems. Deep Learning is effective due to its Hirashima H et al. [20] utilized a ML approach along with
superior accuracy when handling large amounts of data. They novel features, including 3D dosiomics characteristics coupled
are very good at processing visual, speech, and text data [13– with plan and dosiomics features, to estimate and categorize
16]. the gamma passing rate value for volumetric modulated arc
RNN (Recurrent Neural Network) is well suited for therapy (VMAT) plans. Using the high-energy stereoscopic
sequential data analysis tasks, such as time series or language system (HESS), S. Ohm et al. [21] demonstrated the stability
data. In the context of the MAGIC gamma telescope dataset and ability to minimize background noise of their tree
classification, RNN has been used to take into account the classification algorithm compared to the HESS standard
temporal relationship between the features. The sequence of analysis.
features extracted from the telescope camera for each event Very Energetic Radiation Imaging Telescope Array System
can be considered as a sequence of input vectors. By using (VERITAS) data was classified using "boosted decision trees"
RNN, the information from the previous time steps can be by Krause, M., et al. [22], and the results indicated improved
retained and used to help classify the current input. This is sensitivity compared to the usual VERITAS analysis.
particularly useful in scenarios where the sequence of features To do this, Brill et al. [23] presented a combination of CNNs
is important for making accurate predictions, and where and RNNs. Using CTLearn, a freely available Python
traditional machine learning algorithms may not be able to programme that use deep learning to analyse data from IACTs,
capture this information. Therefore, by using RNNs, the model the team created a CNN-RNN network and discovered
has been taken into account the sequence of features and inadequate evidence that ordering telescope pictures by
improve the accuracy of classification. cumulative magnitude improves background rejection
The main objective of this research study is to develop a performance. Nieto et al. [24] employed CTLearn, a python
deep learning-based classification model by adopting the package that uses deep learning in order to analyse data from
MAGIC gamma telescope dataset from the UCI machine IACT arrays.
learning (ML) repository. Multidimensional datasets are Using the water Cherenkov detector with a smaller water
extremely challenging to manage using traditional methods. volume and four PMTs and analysing the PMT signal's spatial
The standard for signal characterization in ground-based
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
and time patterns with a ML method enables optimum muon III. Methodology
tagging, as stated by Conceiço et al. [25]. Figure 1 illustrates the block diagram of the proposed
Algorithms are the backbone of ML, which examines data, classification model. It is composed of several stages,
learns from it, and then makes well-informed judgments based including pre-processing, exploratory data analysis, square
on the information it has acquired. One of the key advantages root transformation, dataset segregation, deep learning
of deep learning compared to traditional ML methods is its classifier, and experimental validation.
ability to perform feature engineering automatically. In order
to learn quickly, a deep learning system may analyse the data A. Data
for related elements without being explicitly instructed. We The MAGIC gamma telescope dataset, sourced from the
used the MAGIC gamma telescope dataset, which has 19020 UCI ML repository [26], is designed to simulate the
observations with ten characteristics (excluding the target). As registration of high-energy gamma particles using the IACT.
a result, we proposed a deep learning-based strategy for The Cherenkov gamma telescope studies high-energy gamma
identifying gamma rays. rays by detecting the Cherenkov radiation emitted by charged
The major contributions of this article are: particles generated in electromagnetic showers established by
• We propose an efficient classification model to gamma rays. This dataset comprises 19020 instances, each
identify high-energy gamma particles in Cherenkov with 10 attributes, and its description is provided in Table 1.
Gamma Telescope data using deep learning The target variable 'class' has two values: 'g' (gamma) for
techniques. The proposed classification model signal and 'h' (hadron) for background, occurring with
consists of several stages: pre-processing, exploratory frequencies of 12332 and 6688, respectively.
data analysis, square root transformation, dataset The dataset contains information about the pulses left by
segregation, deep learning classifier, and experimental incoming Cherenkov photons on a plane of photomultiplier
validation. tubes, which is commonly referred to as the camera. This data
• We present an exploratory data analysis that involves allows for discriminating between signal and background by
analyzing the relationships between the target reconstructing the shower image, which represents the
attribute 'class' and various features. Scatter plots and resulting pattern from the Cherenkov photons. The attribute
heat maps are used to visualize correlations and information includes 11 continuous variables, such as the
patterns. We further normalize the skewed major and minor axis of the ellipse, size, concentration,
distributions of attributes using square root asymmetry, and angle of the major axis with the vector to
transformation. origin.
• We train and evaluate three deep learning models: However, it is worth noting that the dataset has a limitation,
Long Short-Term Memory (LSTM), Gated Recurrent as it does not provide information about the distribution of
Unit (GRU), and Bidirectional LSTM (Bi-LSTM). energy and zenith for the 19020 events.
Different activation functions (ReLU and Swish) are
used in hidden layers to improve model performance.
The models are evaluated using various metrics:
accuracy, precision, recall, F1 score, and Area Under
the ROC Curve (AUC).
• Our proposed model, LSTM+ReLU, achieves the
highest accuracy of 88.71% with all attributes and
88.76% when the least correlated attribute is removed.
The AUC value for the LSTM+ReLU model is
0.9377, indicating strong performance in
distinguishing between classes.
• The proposed model outperforms existing methods,
FIGURE 1. The Block diagram of Proposed Classification Model
including logistic regression, linear discriminant
analysis, and CART. This research study
TABLE 1
demonstrates that the proposed LSTM+ReLU model DESCRIPTION OF THE DATASET
is effective in identifying gamma particles, with an Feature Description
accuracy of up to 88.76%. The model's AUC-ROC fLength The length of the major axis of an ellipse fitted to the
value further validates its efficacy in classifying particle image in millimetres.
fWidth The length of the minor axis of an ellipse fitted to the
gamma and hadron particles. particle image in millimetres.
The research contributes to the field of gamma-ray fSize The logarithm of the sum of the content of all pixels in the
astronomy by providing insights into the classification of high- particle image.
fConc The ratio of the sum of the two highest pixel values to fSize.
energy gamma particles using deep learning techniques. fConc1 The ratio of the highest pixel value to fSize.
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
(a)
(i)
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
(b)
Figure 4. (a) Original dataset distribution plot; (b) Square Root
Transformation dataset distribution plot
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
transformation can be used to normalise it [28]. Figure 4 (a) sigmoid (x). Their findings suggest that when it comes to
shows the distribution plot of the original dataset with 18905 deeper models, Swish is superior to ReLU. So, it is decided to
unique instances. It is observed that most of the features are validate the deep learning approaches with Swish and ReLU
skewed, and their values can be found in table 2. Except for activation functions.
‘fAsym’, 'fM3Long', and ‘fM3Tran’ attributes, the other 7 The ReLU activation function [34] can be defined as,
attributes have been performed with square root 𝑥 𝑖𝑓 𝑥 > 0
transformation. The minimised skew and kurtosis values can 𝑓(𝑥) = { } (1)
0 𝑖𝑓 𝑥 < 0
be found in table 2. The same can be observed in figure 4 (b).
It shows the distribution plot after square root transformation. where x is the input to a neuron.
TABLE 2. The swish activation function [35] can be expressed as,
MAGIC GAMMA TELESCOPE DATASET – SKEW AND KURTOSIS VALUES
𝑥
Attributes Original Dataset After Square Root Transform 𝑠𝑤𝑖𝑠ℎ (𝑥) = 𝑥. 𝑠𝑖𝑔𝑚𝑜𝑖𝑑 (𝛽𝑥) = (2)
1+𝑒 −𝛽𝑥
skew kurtosis skew kurtosis
fLength 2.022 5.031 1.137 1.062 where β is either constant or a trainable parameter
fWidth 3.395 17.013 1.597 4.809 depending on the model.
fSize 0.873 0.723 0.657 0.195
fConc 0.489 -0.517 -0.023 -0.572 2) LSTM
fConc1 0.687 0.031 0.088 -0.428 Figure 5 depicts the proposed LSTM model. It is comprised
fAsym -1.038 8.231 -1.038 8.231
fM3Long -1.130 4.717 -1.130 4.717
of three hidden layers, the first of which is an LSTM base
fM3Tran 0.124 8.676 0.124 8.676 layer, followed by a standard feedforward output layer. The
fAlpha 0.857 -0.521 0.272 -1.125 activation functions "ReLU" and "swish" were used in hidden
fDis 0.229 -0.112 -0.391 0.171
layers. Because the dataset is binary in nature, the 'sigmoid'
activation function was used in the dense layer. The
E. Deep Learning "sparse_categorical_crossentropy" as loss function and
Long short-term memory (LSTM) and gated recurrent unit "Adam" as optimizer have been employed in the model.
(GRU) belong to the category of recurrent neural networks The equations (3) to (8) show in concise form the forward
(RNNs). Each layer in these networks takes its input, pass of an LSTM cell with a forget gate [36]. The lowercase
multiplies it by a linear layer, and then adds the result to the variables represent vectors.
hidden layer weights. This output is passed on to the next
iteration of the network, creating a feedback loop 𝑓𝑡 = 𝜎𝑔 (𝑊𝑓 𝑥𝑡 + 𝑈𝑓 ℎ𝑡−1 + 𝑏𝑓 ) (3)
characteristic of recurrent neural networks. In contrast to
𝑖𝑡 = 𝜎𝑔 (𝑊𝑖 𝑥𝑡 + 𝑈𝑖 ℎ𝑡−1 + 𝑏𝑖 ) (4)
traditional feedforward neural networks, LSTMs include
feedback connections [29]. GRUs require fewer training 𝑂𝑡 = 𝜎𝑔 (𝑊𝑜 𝑥𝑡 + 𝑈𝑜 ℎ𝑡−1 + 𝑏𝑜 ) (5)
parameters, leading to reduced memory usage and faster
execution compared to LSTMs. However, on larger datasets, 𝐶̃𝑡 = 𝜎𝑐 (𝑊𝑐 𝑥𝑡 + 𝑈𝑐 ℎ𝑡−1 + 𝑏𝑐 ) (6)
LSTM tends to be more accurate [30]. Both LSTM and GRU 𝐶𝑡 = 𝑓𝑡 ⊙ 𝑐𝑡−1 + 𝑖𝑡 ⊙ 𝐶̃𝑡 (7)
models have an advantage over ordinary RNNs as they
mitigate the problem of the vanishing gradient. Specifically, ℎ𝑡 = 𝑂𝑡 ⊙ 𝜎ℎ (𝐶𝑡 ) (8)
LSTM and GRU show better performance in terms of
validation and prediction accuracy [31].
Additionally, a Bidirectional LSTM (Bi-LSTM) is a
sequence processing model that comprises two LSTMs—one
processing inputs in a forward direction and the other
processing inputs in a backward direction [32]. In this study,
LSTM, GRU, and Bi-LSTM networks were utilized for
classification purposes.
1) ACTIVATION FUNCTION
With different NN models and datasets, activation functions
behave differently [33]. The ReLU function (rectified linear
activation function) is a piecewise linear function that returns
the input value unmodified if the value is positive and returns
zero otherwise. Two famous types of nonlinear activation
functions are the sigmoid and the hyperbolic tangent. With
both the sigmoid and tanh functions, saturation is a frequent
problem. Therefore, large tanh and sigmoid values snap to 1.0,
and small ones to -1.0 or 0.0. The Google Brain Team released
a activation function called Swish, and it's as easy as f(x) = x Figure 5: Proposed LSTM+ReLU & LSTM+swish model
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
3) GRU
The GRU's operation is analogous to that of an LSTM
equipped with a forget gate, however it has fewer parameters
due to the absence of an output gate [37]. Figure 6 depicts the
proposed GRU+ReLU model. It is comprised of three hidden
layers, the first of which is a base layer, followed by a standard
feedforward output layer. Because the dataset is binary in
nature, the 'sigmoid' activation function was used in the dense
layer. The "sparse_categorical_crossentropy" as loss function
and "Adam" as optimizer have been employed in the model.
Equations (9) and (10), respectively, express the update gate
vector and reset gate vector of GRU.
𝑧𝑡 = 𝜎𝑔 (𝑈𝑧 ℎ𝑡−1 + 𝑏𝑧 ) (9)
𝑟𝑡 = 𝜎𝑔 (𝑈𝑟 ℎ𝑡−1 + 𝑏𝑟 ) (10)
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
(e)
(a)
(f)
Figure 8. Loss Function Plot (a) LSTM+ReLU (b) GRU+ReLU (c) Bi-LSTM
+ ReLU (d) LSTM+Swish (e) LSTM+Swish with 9 attributes (f) LSTM+ReLU
with 9 attributes
(b)
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
(a) (b)
(b)
(c) (d)
(e) (f)
Figure 9. Confusion matrix (a) LSTM+ReLU (b) GRU+ReLU (c) Bi-LSTM +
ReLU (d) LSTM+Swish (e) LSTM+Swish with 9 attributes (f) LSTM+ReLU
with 9 attributes
(c)
(a)
(d)
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
[24] Nieto, D., Brill, A., Feng, Q., Humensky, T.B., Kim, B., Miener, T.,
Mukherjee, R., & Sevilla, J. (2019). CTLearn: Deep Learning for Dr. K. KARTHICK is working as an Associate
Gamma-ray Astronomy. Proceedings of 36th International Cosmic Professor in the Department of Electrical and
Ray Conference — PoS(ICRC2019). Electronics Engineering at GMR Institute of
[25] Conceição, R., González, B.S., Guillén, A. et al. Muon identification Technology in Razam, India. He holds a B.E.
in a compact single-layered water Cherenkov detector and degree in Electrical and Electronics Engineering
gamma/hadron discrimination using machine learning techniques. from Periyar University, Salem, India, as well as
Eur. Phys. J. C 81, 542 (2021). https://doi.org/10.1140/epjc/s10052- an M.E. degree in Power Electronics and Drives
021-09312-4 from Anna University, Chennai, India. He
[26] Dua, D. and Graff, C. (2019). UCI Machine Learning Repository completed his Doctorate in Electrical
[http://archive.ics.uci.edu/ml]. Irvine, CA: University of California, Engineering from Anna University, Chennai.
School of Information and Computer Science. With over 16 years of teaching experience and
[27] Bono R, Arnau J, Alarcón R, Blanca MJ. Bias, Precision, and membership in ISTE, his research interests include data analytics, machine
Accuracy of Skewness and Kurtosis Estimators for Frequently Used learning, text detection and recognition, image processing, and Electrical
Continuous Distributions. Symmetry. 2020; 12(1):19. Drives.
https://doi.org/10.3390/sym12010019 Dr. S. AKILA AGNES is an assistant professor
[28] McDonald, J.H. 2014. Handbook of Biological Statistics (3rd ed.). at GMR Institute of Technology, Rajam, Andhra
Sparky House Publishing, Baltimore, Maryland. Pradesh. She holds a Ph.D. in Computer Science
[29] Zhang Z, Cheng H, Yang T (2020) A recurrent neural network and Engineering from Karunya Institute of
framework for flexible and adaptive decision making based on Technology and Sciences, India (2022), and
sequence learning. PLOS Computational Biology 16(11): e1008342. previously obtained her bachelor's and master's
https://doi.org/10.1371/journal.pcbi.1008342 degrees in Information Technology from Anna
[30] Wang C, Du W, Zhu Z, Yue Z. The real-time big data processing University, India. Her research focuses on
method based on LSTM or GRU for the smart job shop production computer vision, machine learning, deep learning,
process. Journal of Algorithms & Computational Technology. January and medical image processing.
2020. doi:10.1177/1748302620962390
[31] Noh S-H. Analysis of Gradient Vanishing of RNNs and Performance
Comparison. Information. 2021; 12(11):442. Dr.S.SENDIL KUMAR received a BE degree in
https://doi.org/10.3390/info12110442 Electrical and Electronics Engineering from
[32] Rusul L. Abduljabbar, Hussein Dia, Pei-Wei Tsai, "Unidirectional and Adhiyamaan College of Engineering, Hosur, in
Bidirectional LSTM Models for Short-Term Traffic Prediction", 2000. He earned an M.Tech in Energy System
Journal of Advanced Transportation, vol. 2021, Article ID 5589075, Engineering from VIT University, Vellore, in
16 pages, 2021. https://doi.org/10.1155/2021/5589075 2004. He completed his PhD degree in the area of
[33] Israt Jahan, Md. Faisal Ahmed, Md. Osman Ali, Yeong Min Jang, Power System Engineering from Anna University,
Self-gated rectified linear unit for performance improvement of deep Chennai, in 2012. His area of research interest
neural networks, ICT Express, 2022, , ISSN 2405-9595, includes Signal Processing Techniques applied to
https://doi.org/10.1016/j.icte.2021.12.012. Power System Engineering. Currently, he is a Professor at S.A. Engineering
[34] Chaity Banerjee, Tathagata Mukherjee, and Eduardo Pasiliao. 2020. College, Chennai.
The Multi-phase ReLU Activation Function. In Proceedings of the
2020 ACM Southeast Conference (ACM SE '20). Association for Dr.SULTAN ALFARHOOD received the Ph.D.
Computing Machinery, New York, NY, USA, 239–242. degree in computer science from the University of
https://doi.org/10.1145/3374135.3385313 Arkansas. He is an Assistant Professor with the
[35] G. C. Tripathi, M. Rawat and K. Rawat, "Swish Activation Based Department of Computer Science, King Saud
Deep Neural Network Predistorter for RF-PA," TENCON 2019 - 2019 University (KSU). Since joining KSU in 2007, he
IEEE Region 10 Conference (TENCON), 2019, pp. 1239-1242, doi: has made several contributions to the field of
10.1109/TENCON.2019.8929500. computer science through his research and
[36] H. Mayer, F. Gomez, D. Wierstra, I. Nagy, A. Knoll and J. publications. Sultan’s research spans a variety of
Schmidhuber, "A System for Robotic Heart Surgery that Learns to Tie domains including machine learning, recommender systems, linked open
Knots Using Recurrent Neural Networks," 2006 IEEE/RSJ data, text mining and ML-based IoT systems. His work includes proposing
International Conference on Intelligent Robots and Systems, 2006, pp. innovative approaches and techniques to enhance the accuracy and
543-548, doi: 10.1109/IROS.2006.282190. effectiveness of these systems. His recent publications have focused on
[37] Bi L, Hu G, Raza MM, Kandel Y, Leandro L, Mueller D. A Gated using deep learning and machine learning techniques to address challenges
Recurrent Units (GRU)-Based Model for Early Detection of Soybean in these domains. Sultan’s research continues to make significant
Sudden Death Syndrome through Time-Series Satellite Imagery. contributions to the field of computer science and machine learning. His
Remote Sensing. 2020; 12(21):3621. work has been published in several high-impact journals and conferences.
https://doi.org/10.3390/rs12213621
[38] Z. Xueqing, Z. Zhansong and Z. Chaomo, "Bi-LSTM Deep Neural Dr. MEJDL SAFRAN is a passionate researcher
Network Reservoir Classification Model Based on the Innovative and educator in the field of artificial intelligence,
Input of Logging Curve Response Sequences," in IEEE Access, vol. with a focus on deep learning and its applications
9, pp. 19902-19915, 2021, doi: 10.1109/ACCESS.2021.3053289. in various domains. He is currently an Assistant
[39] M. Fatourechi, R. K. Ward, S. G. Mason, J. Huggins, A. Schlögl and Professor of Computer Science at King Saud
G. E. Birch, "Comparison of Evaluation Metrics in Classification University, where he has been a faculty member
Applications with Imbalanced Datasets," 2008 Seventh International since 2008. He obtained his bachelor’s degree in
Conference on Machine Learning and Applications, 2008, pp. 777- computer science from King Saud University in
782, doi: 10.1109/ICMLA.2008.34. 2007, his master’s degree in computer science
[40] Hajian-Tilaki K. Receiver Operating Characteristic (ROC) Curve from Southern Illinois University Carbondale in
Analysis for Medical Diagnostic Test Evaluation. Caspian J Intern 2013, and his doctoral degree in computer science from the same university
Med. 2013;4(2):627-635. in 2018. His doctoral dissertation was on developing efficient learning-
[41] Emmanuel Dadzie, Kelvin Kwakye. Developing a Machine Learning based recommendation algorithms for top-N tasks and top-N workers in
Algorithm-Based Classification Models for the Detection of High- large-scale crowdsourcing systems. He has published more than 20 articles
Energy Gamma Particles. 2021. ffhal-03425661 in peer-reviewed journals and conference proceedings. He has been leading
grant projects in the fields of AI in medical imaging and AI in smart farming.
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4
This article has been accepted for publication in IEEE Access. This is the author's version which has not been fully edited and
content may change prior to final publication. Citation information: DOI 10.1109/ACCESS.2024.3359533
This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 License. For more information, see https://creativecommons.org/licenses/by-nc-nd/4