Professional Documents
Culture Documents
APPLICATION OF NEURAL
NETWORKS IN PREDICTION OF
LEPROSY DISEASE USING
DEMOGRAPHIC AND CLINICAL
DATA
Abstract
classify, analyze and predict an outcome for complex data. This network model has
Now-a-days ANN is broadly used in the medical field for early diagnosis of diseases.
network and Radial Basis Function were used for diagnosis of leprosy disease. This
study facilitates to compare the performance of two neural networks based on its
accuracy on both training and testing data set. For constructing the network, the
whole data set was divided into inputs and outputs. The demographic, physical and
clinical symptoms were considered as inputs and the outputs are identified as cases
with ‘1’ and non-cases with ‘0’. The original data were divided into training and
testing sample. The training sample was used to develop a model to diagnose the
disease.
Basis Function
91
5.1 INTRODUCTION
method and its architecture imitates the original functioning of biological nervous
system of the human brain. Early diagnosis or prediction of disease is one of the
most important requirements in the medical field. The Neural network models are
broadly used for this purpose throughout the world. It is used not only in the medical
field, but also in engineering, agriculture, education, business and manufacturing. For
the past three decades, the neural network model is depicted as a better predictive
ANN models are limited in the field of leprosy, especially in the phases of diagnosis.
However, the main aim to employ such models is to achieve the accuracy in its
prediction and for being easily applicable in the diagnosis of health problems.
The structure of ANN has logical features termed as neurons or Node s which
involve three major layers with feed forward architecture, the input layer (neurons),
hidden layer and output layer. Neurons in each layer describe the connectivity of
inputs and outputs. The input layer is a set of input units, which receives all the
information from various sources. The input units (neurons) are completely linked
to the hidden layer. Similarly, the hidden layer with hidden units (neurons) is
connected to output layer with output units (neurons). The connection between the
neurons (or Node) is made by weights (Fig. 5.1). Each neuron converts the received
weighted inputs. The output layer sends the information to activation pattern applied
92
Thus
∑ ( ) (5.1)
where O denotes the output, W is a weight element and X is an input element andf is
an activation function.
The activation pattern of the input neurons is to transmit input signal to all the
neurons in the consecutive layer (output layer or hidden layer). Fig. 5.2 represents
the actual functioning of neural networks. Input neurons in input layers are identified
based on the number of independent variables; likewise each output neuron in the
output layer corresponds to one dependent variable. Hidden layers are lying between
input and output layer. The output is obtained by computing input vector with
weights which can be determined through the methods like regression coefficients in
93
linear regression. The number of neurons and hidden layers play an important role in
HIDDEN
INPUT
OUTPUT
layer feed forward, feed backward and lateral connectivity. This study considers two
network designs; Multilayer Perceptron (MLP) and Radial Basis Function (RBF).
The main difference between these two types is that the parameters of MLP are
nonlinear and that of the RBF is linear [Jayawardena A.W et al. (1997)]. The purpose
of utilizing both networks (MLP and RBF) would be same, but their structures of
manipulation are different. However, MLP network design with Back Propagation
(BP) algorithm is commonly used for involving more than one hidden layer in the
prediction of outcome. This study attempts to use both the approaches to determine
the parameter estimates and to compare the performances of these models in the
prediction of disease.
94
5.1.3 Multilayer Perceptron (MLP) design
The significant aspect of MLP design is to fix the number of hidden layers
and the number of units in the layers, since determining the appropriate number of
hidden layers is essential to achieve the accuracy in prediction. MLP design with
error back propagation algorithm choose best parameter estimates out of a set of
parameter estimates obtained from each hidden layer. This method performs the
method.
above figure
( ) (5.2)
Wherei is the input vector of the neuron, w is the weights between the
neurons, bis the bias, a is the output signal of the neuron and f is the activation
95
Fig. 5.3 MLP networkwith inputs
The logarithmic and hyperbolic tangent functions are the most important
nonlinear activation functions for MLP and helps to know nonlinear relationships
A Radial Basis Function (RBF) network is a type of two layers feed forward
network. In the input layer, the input neuron is transformed by a basic function to the
hidden layer and a linear combination of the responses of hidden layer neurons is
added to form the output unit at the output layer. The training methods, supervised
and unsupervised method are adopted for the training of RBF networks. The hidden
layers are trained by an unsupervised learning method and determine the Radial
Basis Function for each Node and the output layer is trained in supervised learning
96
Fig. 5.4 Radial basis network with i inputs
(| | ) (5.3)
Eq. 5.3 represents the numeric expression of the Fig. 5.4, where i is the input
vector of the input layer and w is the weight of activation function, ||dist|| is a vector
of the distance between the weight and the input vector. The hidden layer consists of
Radial Basis Function which is obtained from a distance between its weight (w)
and the input vector (i) and is multiplied by the bias b[Rejane B. Santos et al.
(2013)].
develop the technique to predict the chances of occurrence of the disease after a
certain period of time. The technique discussed below has been applied in anti-
leprosy vaccination trial; nevertheless it may also be applicable in a trial for any
97
other diseases. Identification and prediction of disease are also essential in the field
The diagnosis of leprosy has become quite complex, since symptoms of the
disease appeared 3--5 years after contracting the infection. Due to this long
incubation period, the diagnosis of leprosy has become quite complex for physicians
to determine when a person is infected. Even though many leprosy patients can be
clinically cured, people often suffer with the consequences such as nerve and skin
the diagnosis of leprosy disease based on physical and clinical examination may lead
to over diagnosis and misdiagnosis of the disease. To overcome this problem, this
methods when data are multivariate with multi-co linearity between factors. ANN is
a non parametric method which does not make any prior assumption about the
distribution and also it provides highly precise results compared to other regression
models. ANN is more flexible to study large, complex and non-linear relationship
between the factors [Irfan Y. Khan et al. (2013)]. Hence, ANN model enables to
assess the relationship between the factors, hidden manipulation among factors and
prediction of outcome for complex problems. Ali A. El-Solh et al. (1999) developed
a predictive model known as the General regression neural network (GRNN) and
compared the predictive accuracy of the neural network with Clinician assessment in
98
the study “Predicting Active Pulmonary Tuberculosis Using an Artificial Neural
connection weights from the input layers to hidden layer and hidden layer to output
layer by Mei-Sheng Duh et al. (1998). In another study, Yang Benfu et al. (2009)
artificial neural network and confirmed its efficiency in the study “Study on the
Qeethara Kadhim Al-Shayea and Itedal S. H. Bahia (2010) analyzed the data
obtained from UCI Machine Learning Repository in order to diagnose diseases in the
this article they showed that the neural network provided significant result in the
algorithm, Radial Basis Function (RBF) and Learning vector quantization (LVQ) in
the article “Artificial Neural Networks for Diagnosis of Kidney Stones Disease”. In
this article, they compare the performance of all three neural networks on the basis of
accuracy, time taken and size of training data. Mohd Khalid Awang and Fadzilah
Siraj (2013) aimed to evaluate the application of Neural network in predicting the
presence of heart disease, in particular the angina in patients in the paper “Utilization
paper, they revealed that the artificial Neural Network would give a better diagnosis
99
incorporated in the current information system. Irfan Y. Khan et al. (2013)
investigated two acute cases, nephritis disease and heart disease and analyzed
dataset by the feed forward, back propagation neural network with supervised
Network in Medical Diagnosis disease like an acute nephritis disease and heart
disease”. They also suggested that ANN is a very flexible and powerful tool in
medical diagnosis.
5.3 METHODS
Leprosy vaccination trial was done. Two consecutive resurveys were conducted from
1993 to 1998 with one and half years to two years of gap. Background of vaccination
trial and availability of individuals during vaccination intake and all successive two
determined the predictors such as age, sex, region and other physically and clinically
proved symptoms which were the influencing factors for various diseases in
and Multi-Layer Perceptron feed forward back propagation network for second
resurvey in this study. In second resurvey (table 5.1), out of 1,21,904; 85,322 and
85,311 individuals were randomly assigned in the training set; 36,578 and 36,586
were randomly assigned in the validation set to construct the MLP and RBF neural
100
of BCG vaccine, contact of leprosy at home and vaccination status (vaccinated or
unvaccinated) were assigned as the input to the both networks (figure 5.5 and figure
5.6). The input layer (Figure 5.5) is defined by 13 input units, one hidden layer is
for first resurvey. Similarly, The input layer (Figure 5.7) is defined by 13 input units,
one hidden layer is formed by 9 numbers of neurons and the output layer is formed
diseased (incidence=0) for second resurvey. All independent variables (inputs) were
normalized by scaling the value over a range between zero and one.
(1) To construct the MLP and RBF Neural networks to predict the
input data.
101
5.5 THE METHODOLOGY OF NETWORK MODELS IN DIAGNOSIS
OF LEPROSY
models based on its accuracy on both training and testing data set. For constructing
the network, the variables of whole data set was divided into inputs and outputs. The
demographic, physical and clinical symptoms were considered as inputs and the
outputs are identified as cases with ‘1’ and non-cases with ‘0’. The original data were
divided into training and testing sample. The training sample was used to develop a
model and the testing data could to assess the performance of models in prediction of
the disease.
current resurveys were used for constructing MLP and RBF neural network
separately. The Sensitivity, Specificity and ROC curve were used to assess the
resulting from the both Multi layer perceptron and Radial Basis Function was
evaluated based on validation set. The sample of validation set (30%) is randomly
allotted to the entire data set. A receiver operating characteristic curve (ROC) was
generated from both the suggested models and represented by a diagram obtained
from Sensitivity against Specificity for various thresholds that are used to estimate
the diagnostic accuracy of models. Those measures, Sensitivity and Specificity were
102
and every individual, whether infected or not. Here the dependent variable has two
categories and each curve represents at subject as the disease versus other category.
5.7 RESULTS
Analysis was carried out for the second resurvey. The whole dataset of
second resurvey (121904) are classified into the training set (around 70%) and the
testing set (around 30%) randomly for MLP and RBF models as mentioned in the
table 5.1. Out of 1,21,904 dataset, 85,467 are allocated into training set, 36,434 are
allocated into testing set and 3 are excluded due to the missing value for constructing
MLP network. Similarly, 85,248 and 36,649 are allocated into training and testing set
respectively ; and 7 are excluded due to the missing value for constructing RBF
network.
Table 5.1 Selection of dataset for MLP and RBF Networks (second Resurvey)
Radial Basis
Multi LayerPerceptron
Case summary of data Function
selection
N Percent N Percent
Excluded 3 7
103
Table 5.2 Network information of Multi layer Perceptron Layer
Number of units in
14
Input Layer
Input Layer
Rescaling Method Standardized
for Covariates
Number of Units in
5
Hidden Layer
Hidden Layer
Activation Function Hyperbolic Tangent
Dependent
1
Variables
Incidence Rate
Number of Units 1
Rescaling Method
Output Layer for Scale Standardized
Dependents
Table 5.2 depicts the network information of Multi Layer Perceptron model
.The 14 independent variables (including Bias) are given in the input layer. The
hidden layer consist 5 number of units. The output layer consist only one unit,
namely, the incidence rate. Table 5.3 represents the model summary of Multi Layer
Perceptron model. It shows that in training set, the sum of square error is 42,768.124
and the relative error is 1. In the testing set, the sum of square is 17,945.82 and
104
Table 5.3 Model summary of Multi Layer perceptron model
Table 5.4 depicts the network information of Radial Basis Function model.
The input layer consists 13 independent variables. The hidden layer consists 10
number of units. The output layer consist only one unit, namely, the incidence rate.
The Table 5.5 represents the model summary of Radial Basis Function model. It
shows that in the training set, the sum of square error is 264.999 and the Percent of
Incorrect predictions is only 0.3%. In the testing set, the sum of square is 102.622
The ANN models are constructed ( figure 5.5 and figure 5.6) using the
characteristics such sex, age, symptoms of the previous resurvey (first resurvey),
duration after vaccination (till second resurvey) and contact at leprosy are
represented in the input layer and the models were designed to generate the output
105
Table 5.4 Network information of Radial Basis Function
Number of units
13
in Input Layer
Input Layer
Rescaling Method
Standardized
for Covariates
Number of Units
10
in Hidden Layer
Hidden Layer
Activation
Hyperbolic Tangent
Function
Dependent
1
Variables Incidence Rate
Number of Units 1
Rescaling Method
Output Layer for Scale Standardized
Dependents
Activation
Identify
Function
Error Function Sum of Squares
Percent Incorrect
0.30%
Training Predictions
106
Table 5.6 Classification table of observed and predicted outcome of Multi Layer
Predicted
Table 5.7 Classification table of observed and predicted outcome of Radial Basis
Predicted
107
Table 5.8 Area under the ROC Curve (second resurvey without Symptoms)
MLP RBF
Incidence Rate Others 0.67 0.611
Incidence 0.67 0.611
symptoms of both MLP and RBF neural networks are represented in the table 5.6 and
table 5.7. In which, the suggested models are not able to predict the disease category
in second resurvey. At the same time, we can assess the performance of models
Multi Layer Perceptron and Radial Basis Function models through the values 0.67
and 0.61 respectively (table 5.8) of area under the curves achieved by the ROC
108
Figure 5.5. The Network diagram obtained using MLP Network for second
Foot note: PP1- Presence of Patches in First resurvey, NP1- Number of Patches in
First resurvey, PLS1-Patchy loss of Sensation at First resurvey, IS1- Infilteration of
Skin in First resurvey, OD1- Overall Deformity in First resurvey, CC0 – Contact
with Leprosy patients at intake, CC1- Contact with Leprosy patients in first
resurvey, CC2- Contact with Leprosy patients in second resurvey, Dur-2, Duration
at resurvey2
109
Figure 5.6 Receivers Operating Characteristic Curve obtained for Predicting
110
Figure 5.7 The Network model obtained using Radial Basis Function of second
Foot note: PP1- Presence of Patches in First resurvey, NP1- Number of Patches in
First resurvey, PLS1-Patchy loss of Sensation at First resurvey, IS1- Infilteration of
Skin in First resurvey, OD1- Overall Deformity in First resurvey, CC0 – Contact
with Leprosy patients at intake, CC1- Contact with Leprosy patients in first resurvey,
CC2- Contact with Leprosy patients in second resurvey, Dur2- Duration at
resurvey2
111
Figure 5.8 Receivers Operating Characteristic Curve obtained for
112
5.8 DISCUSSION
utilize the neural network for the diagnosis of leprosy. The review of various
literatures revealed that this predictive model provides a high level of Sensitivity and
(1999)] a general regression neural network (GRNN) was used to develop the
the patients more accurately than clinical assessment. The neural network model had
achieved the Sensitivity of 100% (95% CI, 72--100) and Specificity of 69% (95% CI,
61--78) with the predictor variables, upper zone disease on chest roentgenogram,
fever, weight loss, etc.,. In a similar study, the neural network model was used in
[Yang Benfu (2009)] . In which, they had defined personal and clinical information,
medical histories, chest X-rays and clinical laboratory results as input variables to
In the same year, two different studies evaluate the performance of neural
network model in prediction of heart disease. In the first study, they identified eight
cholesterols, hypertension, diabetes mellitus and hyper- chilseteroemia that were the
major risk factors of anginha . In this study, the model achieved the prediction
distinguish between infected and non-infected person in acute nephritis disease and
113
heart disease. Here also the symptoms such as temperature of patients, occurrence of
nausea, lumbar pain, urine pushing, micturition pains and images of Cardiac Single
variables for construction of the model [Irfan Y. Khan, P.H. Zope, S.R. Suralkar
(2013)]. However, neural network models performed well when compared with other
predictive models, while predicting or diagnosing of disease and proved physical and
clinical symptoms were playing a vital role in achieving the model’s accuracy.
studies. First point, the majority of the diagnosis of Leprosy is made through the
considered the symptoms of the current resurvey. And the second point, the study
consists such a large dataset (1,21,904 in second resurvey) with the meager
occurrence of disease (372 in second resurvey). In this situation, the study had
attempted to generate the models (both MLP and RBF) to the diagnosis of leprosy at
second resurvey. In the case of the constructing the model without symptoms, the
variables such as age, gender, previous intake of BCG vaccine (through the presence
equally distributed, the models can predict the leprosy using probability of
114
occurrence of disease (Leprosy), the suggesting network models also could not give
better prediction ( table 5.6 and the table 5.7) in such a large cohort study with rare
occurrence of diseases. Figure 5.4 and figure 5.6 had shown the structure of
Multilayer Perceptron Layer and Radial Basis Function network model, while we
construct the models without symptoms. Instead of this, if we observe the ROC
curve; it gives us a visual display of the Sensitivity and Specificity. It shows 0.67 and
0.611 (table 5.8) predicted pseudo-probabilities of the area under the curve. The
classification tables showed that the predictive models are considerably better at
variables.
5.9 LIMITATION
prediction of any other diseases, the current study unable to predict the
5.10 CONCLUSION
prediction of Leprosy, especially the rare occurrence of disease for large datasets.
This study has revealed that the performance of Multilayer perceptron and Radial
115