You are on page 1of 5

2nd IEEE International Conference on Engineering and Technology (ICETECH), 17th & 18th March 2016, Coimbatore, TN,

India.

Soft Computing Based Expert System for Hepatitis


and Liver Disorders

Dr. R.R.Janghel Dr. Anupam Shukla Kshitij Verma


Assistant Professor Professor Research Associate
Department of IT ICT Department
ABV-IIITM, Gwalior, India
NIT Raipur, India ABV-IIITM, Gwalior, India
dranupamshukla@gmail.com
janghel1310@gmail.com dranupamshukla@gmail.com

Abstract – The main objective of this research work is to develop an of neural networks [10].
expert system for the diagnosis and detection of Hepatitis and liver All the neural network architectures were trained 25 times
disorders based on various Artificial Neural Networks models. In each to obtain an average and accurate estimate of the
this research work Artificial Neural Networks models like Back efficiency and this average was compared with others. The
Propagation Algorithm, Probabilistic Neural Networks, Competitive efficiency was calculated as the percentage of total number of
learning Networks, Learning vector quantization and Elman
Networks have been used for detection and diagnosis of Hepatitis
input pattern correctly classified out of total input patterns
and liver disorders. The various networks developed with the help used for simulation.
of MATLAB. Required data has been chosen from trusty
machine learning data base (UCI). This system in comparison
with other traditional diagnostic systems is faster, more reliable II. LITERATURE REVIEW
and more accurate. One can use this system as a specialist Anupum Shukla et. al proposed a new machine learning
assistant or for training medicine students. technique for large data sets using Fuzzy Neuro Systems [1].
R.R.Janghel et al invented an expert system for Diagnosis of
Key word: ANN, soft computing, liver disorder, Hepatitis. Thyroid Disorders using Artificial Neural Networks [2]. Ritu
Tiwari et al proposed a Knowledge Based Approach for
I. INTRODUCTION Diagnosis of Breast Cancer [3]. Lale Ozyilmaz et al
investigated a system for diagnosis of hepatitis diseases using
Disease diagnosis is one of the most important
MLP and RBF Neural network and got 81.25, 87.5
problems in medicine. Medical diagnostics is quite complex
percentage accuracy respectively [4]. R. Jajoo et al developed
and visual task which is often handled by expert doctors. Two
a system for prediction of hepatitis C using artificial neural
problems are the most common in the field of automatic
network and rule based [5]. M.Neshat et al Fuzzy Expert
diagnostic, the selection of necessary parameter set for right
System Design for Diagnosis of liver disorders [6]. Laercio
diagnostics and forming of steady and powerful algorithm
Brito Gonçalves et al proposed A Novel Neuro-Fuzzy Model
which doesn’t require long time to run. In recent times, neural
for Pattern Classification and Rule Extraction in Databases a
networks have become a widely used method for this purpose.
new neuro-fuzzy model that has been specifically created for
A reasonably good solution to medical problems could be
record classification and rule extraction in databases. The
given by the neural network algorithms. In this work, five
HNFB is based on the Hierarchical Neuro-Fuzzy Binary Space
neural network algorithms have been investigated for
Partitioning Model (HNFB), which embodies a recursive
diagnosis of hepatitis, liver disorders and Iris flower plant
partitioning of the input space, is able to automatically
diseases. All five neural networks were applied to hepatitis
generate its own structure, and allows a greater number of
and liver disorders data.
inputs [7]. H Fukuda et al applied an image analyzing system
We focus on the need of the comparison on accuracies
using an artificial neural network for evaluating the
obtained for specific data sets using various trained neural
parenchymal echo pattern of cirrhotic liver and chronic
networks, to identify the best among them suitable to obtain
hepatitis [8]. Hideneo Abeet et al developing an integrated
the results on those data sets.
time-series data mining environment for medical data mining.
The basic objective was to train those data sets using various
Medical time-series data mining is one of key issues to get
architectures of different neural networks and to identify the
useful clinical knowledge from medical databases [9]. Doina
best suited neural networks for them.
Dragulescu et al proposed an Expert System for Medical
Here we used the Neural Networks Toolbox of MATLAB for
Predictions expert using logical inference and statistical
training and testing the data sets. MATLAB was used because
inference. Both of these were used in this system in order to
it provides all the neural network architectures with easy to
make some predictions regarding the hepatitis diagnosis and
use functions and hence very convenient training and handling
the evolution of an infected patient [11].
978-1-4673-9916-6/16/$31.00 ©2016 IEEE
2nd IEEE International Conference on Engineering and Technology (ICETECH), 17th & 18th March 2016, Coimbatore, TN, India.
Spleen Palpable No, Yes
Spiders No, Yes
III. METHODOLOGY
Ascites No, Yes
In this work, the data sets are first treated for missing values
Varices No, Yes
(if any), and then they are normalized. Now, the data sets are
split into two parts, one for training and the other for Bilirubin 0.39, 0.80, 1.20, 2.00, 3.00, 4.00
simulation and calculation of efficiency which is shown in Alk Phosphate 33, 80, 120, 160, 200, 250
respective tables. Figure 1 depicts this overall methodology of Sgot 13, 100, 200, 300, 400, 500
the work. After splitting, various neural networks are trained
Albumin 2.1, 3.0, 3.8, 4.5, 5.0, 6.0
using the training sets. Finally the average efficiencies of the
best models of various neural networks are compared. Protime 10, 20, 30, 40, 50, 60, 70, 80, 90
Histology No, Yes

Table 2: Class Distribution in training and testing Hepatitis Data Set


Collect data set
No. of No. of
Data Feature Extraction and analysis. instances instances
Total
Processing Removal OR Filling of missing data. having class having class
Data Normalization. 1 (Die) 2 (Live)
Training Set 23 77 100
Testing Set 9 46 55
Simulated Models Total 32 123 155

Training and Testing Phase BUPA Liver Disorder Set: This is a data set of a pattern
classification type of problem. The Data set contains a total of
Diagnosis Decision 345 instances of two different classes shown in table 3. Each
instance consists of 7 different attributes

Class 1 Class 2 Table 3: Class Distribution in training and testing Hepatitis Data Set

No. of No. of
instances instances
Performance Comparison of computational models Total
having class having class
1 2
Training Set 130 170 300
Figure 1: Overall block diagram of the methodology Testing Set 15 30 45
Total 145 200 345

A. Description about Data Sets The first 6 attributes act as inputs. The first 5 variables are all
Hepatitis Domain Data Set: This is a data set of a pattern blood tests which are thought to be sensitive to liver disorders
classification type of problem. The Data set contains a total of that might arise from excessive alcohol consumption. They
155 instances of two different classes. Each instance consists basically determine the mean corpuscular volume (mcv),
of 20 different attributes. The last 19 attributes act as inputs alkaline phosphotase (alkphos), alamine aminotransferase
which is shown in table 1. (sgpt), aspartate aminotransferase (sgot), gamma-glutamyl
transpeptidase (gammagt). The 6th attribute determines the
Table 1: Attributes distribution in Hepatitis Data Set number of half-point equivalents of alcoholic beverages drunk
per day. And 7th attribute is the output used to split these data
Attribute Values sets into two parts as per their disorders.
AGE 10, 20, 30, 40, 50, 60, 70, 80 We used 300 instances of this data set to training our
networks and tested them against the remaining 45 instances
SEX Male, Female
to obtain the efficiencies.
Steroid No, Yes
Antiviral No, Yes B. Back Propagation
Fatigue No, Yes
Back propagation method is the generalization of the Widrow-
Malaise No, Yes
Hoff learning rule to multiple layer networks and nonlinear
Anorexia No, Yes differentiable transfer functions. In this method Input vectors
Liver Big No, Yes and the corresponding target vectors are used to train a
Liver Firm No, Yes network until it approximates a function, associate input
2nd IEEE International Conference on Engineering and Technology (ICETECH), 17th & 18th March 2016, Coimbatore, TN, India.
vectors with specific output vectors or classify input vectors in neighborhoods are adjusted to become even closer to the input
an appropriate way. vector.
Standard back propagation is a gradient descent An LVQ network has a first competitive layer and a
algorithm, i.e., the network weights are moved along the second linear layer. The linear layer does the job of
negative of the gradient of the performance function. The term transforming the competitive layer's classes into target
back propagation refers to the manner in which the gradient is classifications as defined. We refer to the classes learned by
computed for nonlinear multilayer networks. A single iteration the competitive layer as subclasses and the classes of the
of this algorithm can be written as: linear layer as target classes.

xk+1 = xk - αkgk F. Elman Networks

Where xk is a vector of current weights and biases, gk is the Elman networks are two-layer back propagation networks,
current gradient, and αk is the learning rate. The 1st attribute is with addition of a feedback connection from the output of the
the class valued as DIE/LIVE. The original data set contains hidden layer to its input. The Elman network commonly is a
many missing values and hence the missing values are filled two-layer network with feedback from the first-layer output to
using the average value of that attribute for all the instances. the first layer input. This recurrent connection allows the
We used 100 instances of this data set to train our Elman network to both detect and generate time-varying
networks and tested them against the remaining 55 instances patterns.
to obtain the efficiencies. The Elman network has tansig neurons in its hidden
(recurrent) layer, and purelin neurons in its output layer.
C. Probabilistic Neural Networks
a1 (k) = tansig (IW1,1p+LW1,1a1(k-1) + b1)
Probabilistic neural networks are a type of Radial Basis
Function which is used for classification problems. When an a2 (k) = purelin (LW2,1a1(k) + b2)
input is presented the first layer computes the distances from
input vector to the training input vectors and produces a vector This combination is special in that two-layer networks with
whose elements indicate how close the input is to a training these transfer functions can approximate any function with
input. The second layer or Radial Basis Layer adds these arbitrary accuracy. The only requirement is that the hidden
contributions for each class of inputs to produce its net output layer must have enough neurons. More hidden neurons are
as a vector of probabilities. The transfer function for a radial needed as the function being fit increases in complexity.
basis neuron is expressed as:
radbas(n) =
IV. RESULTS AND COMPARISON
Finally a complete transfer function on the output of the
second layer selects the maximum of these probabilities and Performance comparison of various neural networks models
produces a 1 for that class and a 0 for the other classes. on the basis of best average results.

D. Competitive Learning Hepatitis Data Set


Table 4: Final Results for Hepatitis Domain Data Set
Competitive learning modulates the difference between
Computation time
weight at each synapses and the corresponding output signal. Artificial Neural Network Accuracy (%)
( in seconds )
The inputs are provided to the neurons in output layer and the LVQ 92.7273 % 45.8700
one having weights closest to the input outputs 1 and rest Backpropagation 88.0727 % 0.2545
0.The weights of this neuron are adjusted so that it becomes Elman Network 86.1818 % 16.7650
closer to the input. The feed forward network usually Probabilistic Neural
80.0000 % 0.0600
Network
implements an excitatory Hebbian learning rule. Competitive Neural
Classification of instances is unsupervised and is often based 66.3636 % 20.001
Network
on clustering and hence forming groups of all the similar input
vectors based on some similarity measure. BUPA Liver Disorder Set

E. Linear Vector Quantization Table 5: Final Results for Hepatitis Domain Data Set

In a learning vector quantization (LVQ), the input vector is Artificial Neural Network Accuracy (%)
Computation time
compared to the weight vector of each neuron. Neurons ( in seconds )
LVQ 71.1111% 43.2300
having the most closely match of weights with the input are Backpropagation 78.7556 % 0.5912
known as the best match unit (BMU) of the system. The Elman Network 64.1778 % 12.5420
weight vector of the BMU and all the other neurons in its Probabilistic Neural
62.2222 % 0.0661
Network
2nd IEEE International Conference on Engineering and Technology (ICETECH), 17th & 18th March 2016, Coimbatore, TN, India.
There can be an increase in the work efficiency with the help
of other techniques and algorithms.

TABLE 6: OPTIMIZED TRAINED FEATURES OF FEED FORWARD MODEL


BY BPA Acknowledgment
The preferred spelling of the word “acknowledgment” in
Data Goal Learning No.of No.of MSE %age
Rate Neurons epochs Accuracy
America is without an “e” after the “g.” Avoid the stilted
in hidden of expression “one of us (R. B. G.) thanks ...”. Instead, try “R. B.
layer Diagnosis G. thanks...”. Put sponsor acknowledgments in the
unnumbered footnote on the first page.
Testing 0.01 0.001 50 30,000 0.01 88.83
Data(
Hepatitis)
References
Testing 0.01 0.001 50 30,000 0.01 78.75 [1] Rahul Kala, Anupum Shukla, Ritu Tiwari, “Fuzzy Neuro Systems
Data (Liver) for Machine Learning for Large Data Sets”, IEEE IACC 2009,pp
6-7 March 2009.
10
2
Performance is 0.00999983, Goal is 0.01 [2] Anupum Shukla, Ritu Tiwari, Prabhdeep Kaur, R R Janghel,
“Diagnosis of Thyroid Disorders using Artificial Neural
1
Networks”, IEEE IACC 2009, 6-7 March 2009.
10
[3] Anupum Shukla, Ritu Tiwari, Prabhdeep Kaur, “Knowledge Based
Approach for Diagnosis of Breast Cancer”, IEEE IACC 2009, 6-7
Training-Blue Goal-Black

10
0
March 2009.
[4] Lale Ozyilmaz, Tulay Yildrim, “Artificial Neural Networks for
10
-1 diagnosis of Hepatitis disease” on Proceedings of the International
Joint Conference on Neural Networks, 20-24 July 2003.
10
-2
[5] R. Jajoo, D. Mital, S. Haque, S. Srinivan, “Prediction of hepatitis C
using artificial neural network”, ICARCV 2002, 7th International
Conference, 2-5 Dec. 2002.
[6] M.Neshat, M.Yaghobi, M.B.Naghibi, A.Esmaelzadeh, “Fuzzy
-3
10
0 0.5 1 1.5 2 2.5
29684 Epochs
Expert System Design for Diagnosis of liver disorders ,
4
x 10

Figure2: Training curve of the BPA NN model. International Symposium on Knowledge Acquisition and
Modeling, pp 252-256,2008.
[7] Laercio Brito Gonçalves, Marley Maria Bernardes Rebuzzi
Vellasco, Inverted hierarchical “neuro-fuzzy bsp System: A Novel
V. DISCUSSION AND FUTURE WORK Neuro-Fuzzy Model for Pattern Classification and Rule Extraction
in Databases”, IEEE transactions on systems, man, and
cybernetics—part c: applications and reviews, vol. 36, no. 2, march
On the basis of these stated results, we come to the conclusion 2006.
that the best possible architectures for computing operations [8] Hideneo Abe, Hideto Yokoi, Miho Ohsaki, Takahira Yamaguchi,
on these data sets will be as: “Developing an Integrated Time-Series Data Mining Environment
for Medical Data Mining”, 7th IEEE International Conference on
Data Mining, Pages 127-132, 2007.
For Hepatitis Domain Data Set Linear Vector Quantization [9] Doina Drăgulescu, Adriana Albu, “Expert System for Medical
(LVQ) is the best Neural Network known so far. It requires a Predictions”, IEEE pp 123-128, 2007,.
mean time of 45.87 seconds per training and on the other hand [10] MATLAB Tutorial, version 7.0.1., Service Pack 1, September 13
backpropagation method requires just 0.2545 sec per training, 2004.
[11] Dong-Hoi Kim, Jin Kim, Sung Won Cho, Jae Youn Cheong,
but with lesser efficiency. “Chronic Hepatitis Classification Using SNP Data and Data
Hence, both these methods can be applied as per the type of Mining Techniques”, FBIT 2007, pp 81-86, 11-13 Oct. 2007.
application. [12] H Fukuda, M Ebara, A Kobayashi, N Sugiura, M Yoshikawa, H
Saisho, F Kondo, S Yoshino, T Yahagi, “An image analyzing
system using an artificial neural network for evaluating the
For BUPA Liver Disorder Set Backpropagation method using parenchymal echo pattern of cirrhotic liver and chronic hepatitis”,
‘trainlm’ function, with learning rate equal to 0.05 and having IEEE Transactions, Vol 45, Issue 3, Pp396-400,March 1998.
a single neuron in the hidden layer, is the best possible Neural [13] A Ahmadian, A Mostafa, M D Abolhassani, N R Alam,
Network known so far.It requires just 0.5912 seconds to get “Analyzing hepatitis data using ordered information”, Proceedings
of ICSP ’04, International Conference, Vol. 2, 31 Aug.-4 Sept. , pp
trained, and gives an accuracy of 78.7556 %. 971-974,2004.
[14] Ying Yang, Tuan-fa Qin, Jian-ning Jiang, Zhi-ming Liu,
“Distributed Medical Information System for Chronic Viral
Table 6: Best Neural Networks model for various data sets Hepatitis”, ISISE ’08 International Symposium, Vol. 2, 20-22 Dec.
, pp 559-562,2008.
Disease/ Computation [15] K Ogawa, M Fukushima, K Kubota, N Hisa, “Computer-aided
Artificial Neural Accuracy
Problems time diagnostic system for diffuse liver diseases with ultrasonography
Network Models (%)
( in seconds ) by neural networks”, IEEE Transactions, Vol. 45, Issue 6, Part 2, ,
Hepatitis LVQ Neural Network 92.7273 45.8700 pp 3069- 3074,Dec. 1998.
Liver Back propagation [16] A K C Wong, T Y Young, P S Liu, “Application of pattern
78.7556 0.5912
Disorder Neural Network recognition techniques to discrete clinical data”, IEEE Conference,
Vol. 15, Part 1, , pp 158-161,Dec 1976.
2nd IEEE International Conference on Engineering and Technology (ICETECH), 17th & 18th March 2016, Coimbatore, TN, India.
[17] E Berry, Zheng Rong Yang, Xikun Wu, “A biology inspired neural
learning algorithm for analysing protein sequences”, 15th IEEE
International Conference, pp 18-25,3-5 Nov. 2003.
[18] C Bustamante-Sa, F F Nobre, ”Forecasting epidemiological time
series with backpropagation neural networks”, Proceedings of the
38th Midwest Symposium, Vol. 2, pp 1365-1368,Aug13-16. 1995.
[19] R.J. Almeida, J.M.C. Sousa, “Comparison of fuzzy clustering
algorithms for classification”, Evolving Fuzzy Systems,
International Symposium,pp112-117 ,Sept. 2006.
[20] Zhi-Hua Zhou, Yuan Jiang, “Medical diagnosis with C4.5 rule
preceded by artificial neural network ensemble”, IEEE
Transactions, Vol. 7 Issue 1, pp37-42, March 2003.
[21] Sangeeta Kakarwal, Mahananda Malkauthekar, Shubhangi Sapkal,
Ratnadeep Deshmukh, “Face Recognition using Fourier Descriptor
and FFNN”, IEEE IACC 2009, 6-7 March 2009.
[22] M Fukushima, K Ogawa, T Kubota, N Hisa, “Quantitative tissue
characterization of diffuse liver diseases from ultrasound images
by neural network”, Nuclear Science Symposium, 1997, IEEE,
Vol. 2, pp 1233-1236, 9-15 Nov. 1997,.
[23] Simon Haykin, Neural Networks: A Comprehensive Foundation,
McMaster University, Hamilton, Ontario, Canada.
[24] B.Yegnanarayana, Artificial Neural Networks.

You might also like