You are on page 1of 7

Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]

IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

Performance Analysis of Chronic Kidney Disease


through Machine Learning Approaches
Minhaz Uddin Emon Al Mahmud Imran Rakibul Islam
Daffodil International University Daffodil International University Daffodil International University
Dhaka, Bangladesh Dhaka, Bangladesh Dhaka, Bangladesh
minhazkhondokar21@gmail.com md.almahmudimran@gmail.com rakibulhassan282@gmail.com
2021 6th International Conference on Inventive Computation Technologies (ICICT) | 978-1-7281-8501-9/21/$31.00 ©2021 IEEE | DOI: 10.1109/ICICT50816.2021.9358491

Maria Sultana Keya Raihana Zannat Ohidujjaman


Daffodil International University Daffodil International University Daffodil International University
Dhaka, Bangladesh Dhaka, Bangladesh Dhaka, Bangladesh
maria.sultana.keya@gmail.com zannat.swe@diu.edu.bd jaman.cse@diu.edu.bd

Abstract—Data mining and machine learning play a vital greater and deeper burdens that developing nations are not
role in health care and also medical information and detection, prepared to cope with. There has been a lack of exposure
Now a day machine learning techniques use awareness of to chronic diseases, chronic kidney disease in particular,
some major health risks such as diabetic prediction, brain
tumor detection, covid 19 detections, and many more. The primarily due to the focus of the global health community on
kidney is the most important organ of our body and if it has infectious diseases, and lack of consciousness. There is a vital
any problem then the impact is more dangerous to our body. need to concentrate on it and the adoption of more inclusive,
Chronic kidney disease (CKD), otherwise referred to as renal cost-effective, and preventive chronic disease strategies by
disease. CKD requires disorders that damage and reduce the developed countries. There are many hospitals which store
capacity of our kidneys to keep us healthy. So, it is required
to be concerned about kidney disease to our very primary the data of chronic kidney disease patients in their database.
stage. We take a few attributes to measure our analysis about Through analyzing these data, various patterns can be found
chronic kidney disease and this attribute is one of the major that will be helpful for decision-making. Using data mining
occurrences of chronic kidney disease. Therefore 8 machine techniques on these data, it is possible to discover many
learning classifier are used to measure analysis using weka tools kinds of knowledge and use this knowledge to predict the
namely: Logistic Regression(LG), Naive Bayes(NB), Multilayer
Perceptron(MLP), Stochastic Gradient Descent(SGD), Adaptive disease. There is a massive amount of people who are affected
Boosting(Adaboost), Bagging, Decision Tree(DT), Random by chronic kidney disease. In Bangladesh maximum are not
Forest(RF) classifier are used. We feature extraction of all concerned about the disease. As a result, the affected people by
attributes using principal component analysis(PCA). We gain the disease are increasing day by day. If people could detect or
the highest accuracy from the Random Forest(RF) and it is predict if they are affected or going to be affected then it could
99% and ROC(receiver operating characteristic) curve value is
also highest from other algorithms. be controlled. People could take the necessary steps to not be
affected. For predicting the disease, predictive can be used.
Keywords—Chronic Kidney Disease, Machine Learning, Pre- There are several methods that can be used like, classification,
diction, PCA, Co-relation Metrics, Random Forest. regression, categorization. Among those, many people think
classification is the best. After accomplishing the research,
I. INTRODUCTION it will help to predict Chronic Kidney disease. People will
Nowadays machine learning is an important field for re- be aware of the disease and their health condition. The main
search. In different areas, different statistical and machine objectives of the research are to predict the disease using a
learning algorithms are implemented. Such as marketing, machine learning algorithm, warning whether a person at risk
health, and medical problems, weather forecasting, the study of the disease or not, and comparing the outcomes of various
of socioeconomic behavior, etc we can use machine learning machine learning algorithms to determine which algorithm
in those fields. In the medical field, many diseases can be provides the best performance. The paper residue is structured
detected by machine or predicted by machine by the appli- as follows: the related works are described in Section II, and
cation of machine learning algorithms. Chronic diseases pose the entire methodology is shown in Section III. The result
an important threat to the global health agenda of the 21st analysis are discussed in Section IV. Finally, conclusion are
century. The rising prevalence of chronic diseases such as represented in Section V.
chronic kidney disease has serious consequences for health and
economic output in developing countries. The rapid increase II. RELATED WORKS
in common risk factors, especially among the poor, such Amirgaliyev et. al. expected the patient’s outcomes from
as diabetes, hypertension, and obesity, would result in even the chronic kidney disease dataset. Medical history, physical

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 713

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

exams, and laboratory samples were the basis of the gathered the diagnosis of Chronic Kidney Disease. On entirely different
dataset. They developed models for this dataset using a support algorithms such as Naive Bayes, Decision Tree, K-Nearest
vector machine algorithm for linear kernel SVM classifiers Neighbor, and Support Vector Machine, the expected results
that were tested to find the best scores for sensitivity, speci- were carried out. The expected outcome suggests that the K-
ficity, and accuracy metrics. They also shown that the highest Nearest Neighbour algorithm delivers a better outcome than
precision of 93.1% [4]. Behind Charleonnan et. al. using the other classification algorithms and generates 98% accuracy
clinical evidence, projected chronic kidney disease. Along [10]. Serpen et. al. The outcome of diagnosis rule extraction
with KNN, SVM, LR, and DT classifiers for predicted CKD, was redicted from patient data for Chronic Kidney Disease
four machine learning techniques were explored. The overall (CKD). To devise a bunch of designation rules for CKD, the
precision was five times that of the four classifiers. It can be C4.5 decision tree rule was added to the patient data. For a
seen from the experimental outcomes that the SVM classifier cumulative patient count of 400, they appropriately categorized
has the best accuracy than the others with 98.3% though 393 patients and wrongly classified seven cases. Predicted ac-
Logistic, Tree of Decision and KNN can achieve average curacy was obtained by the C4.5 algorithm using 3-fold cross-
correctness of 96.55%, 94.8%, and 98.1% respectively [5]. validation 98.25% [11]. Ahmad et. al.The findings predicted
Sinha et. al. Chronic Kidney Disease was predicted using whether or not kidney failure patients had reached a stage of
the Support Vector (SVM) and K-Nearest Neighbor (KNN) progressive kidney disease. The methodology of this thesis
classifiers. The experimental results showed that the efficiency consisted of 2 main phases: simulation of classification and
of the KNN classifier is higher than that of the SVM. Help development of the scheme. Modeling of classification con-
vector machine (SVM) accuracy was 73.75% and K-Nearest sists of gathering information, planning data, grouping data,
Neighbor (KNN) accuracy was 73.75% was 78.75% [6]. classification, extracting rules. It was expected that this device
Sharma et. al. evaluated different classification methods for would help the doctor in determining on the chronic condition
diagnosis of Chronic Kidney Disease (CKD). The dataset of patients with kidney failure with sufficient precision. The
used for this analysis consisted of 400 instances and 24 performance of this structure was 98.34% [12]. Hosseinzadeh
attributes. The authors tested twelve classification techniques et. al. The planned diagnostic model for chronic nephrosis on
by applying them to the CKD results. The predictor parameters the network of things platform. The tools of classification were
used for performance measurement are prognostic accuracy, the Judgment Treaties (J48), SVM, MLP and Naı̈ve Bayes.The
precision, sensitivity and specificity. The findings suggest that experimental data shows that 97% precision, 99% sensitivity
the precision of the decision-tree has been reached of 98.6%, and 95% specificity through the decision tree (J48) classifica-
sensitivity of 97.20, exactness of one, and specificity of one tor relative to the support vector machines (SVM), multilayer
[7]. Khan et. al. Expected machine learning (ML) strategies perceptron (MLP) classifiers and naı̈ve bayes classifiers are
for Chronic Kidney Disease (CKD) Prophecy. The dataset obtained by the implemented data collection with its suggested
that was taken from the UCI ML pool consisted of 400 elect choices. Furthermore, the proposed feature set increases
instances. The ML algorithms included NBTree, J48, Support execution time relative to numerous data sets with entirely
Vector Machine, Logistic Regression, Multi-layer Perceptron, different choices [13].
Naı̈ve Bayes. The findings of the above listed approaches have
been compared with the characterization of the most accurate TABLE I
methodology for classifying CKD and NOT CKD patients with P ERFORMANCE C OMPARISON OF CHRONIC KIDNEY DISEASE FROM
high precision. Exactness of the experimental findings used EXISTING MODELS

95.75% for NB, 96.50% for LR, 97.25% for MLP, 97.75% Author Model Result
for J48, 98.25% for SVM, 98.75% for NBTree [8]. Polat et. Amirgaliyev [4] SVM 93.1%
al. Define selection methods estimated the outcome of the Sinha [6] SVM, KNN 73.75%, 78.75%
Sharma [7] Decision tree 98%
diagnosis of CKD, based on support vector machines. Two Serpen [11] Decision tree 98%
main types of selection strategies, namely wrapper and filter Ahmad [12] Classification 98%
approaches, have been selected to scale down the dimension of modeling
Hosseinzadeh Decision tree 97%
the Chronic kidney disease dataset. The wrapper method was [13] (J48), SVM,
accompanied by classification sub-sets of evaluators of greedy MLP, and Naı̈ve
incremental programs and wrapper subset evaluators with the Bayes strategies
simplest first program. The association sub-set evaluator with
greedy step-by-step program and sub-set evaluator with the
shortest filter approach is the approach to the filter. The
findings revealed that the Help Vector Machine classifier III. M ETHODOLOGY
features a higher accuracy rate when using the filtered subset
evaluator with the simplest first software feature selection This section divided into ten subsections: proposed model,
process (98.5%) within the diagnosis of CKD compared to algorithms description, data descriptions, data prepossessing,
other selected methods [9]. Radha et. al. The classification feature extraction, machine learning classifier, cross validation,
algorithms were expected to be taken into consideration for evaluation metrics, confusion metrics and best model.

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 714

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

A. Proposed Model classifiers and regressors. It needs number of hyper parame-


The procedures for execution are illustrated in this section. ters.
Python and the Scikit-learn library have been used to carry AdaBoost: AdaBoost was originally designed to improve
out the research. In fig 1, generate a diagram for procedure of the efficiency of binary classifiers as an ensemble learning
Proposed model. process (also known as ”meta learning.”) AdaBoost uses an
iterative approach to learn from and improve the errors of
weak classifiers.
Bagging Classifier: The Bagging Classifier is a collection of
Data Input
meta-estimators that fits each base classifier into the random
subsets of the original dataset and then aggregates their indi-
vidual predictions to make the final prediction. It’s commonly
used to reduce a black-box estimator’s variance.
Data Preprocssing Decision Tree Classifier: Tree-like structures are used in
Decision Tree classification. The root nodes means conditions
and child nodes means the class label. Branches of the root
Feature Extraction nodes means effects of the conditions.
Random Forest Classifier: It is a learning algorithm based
on an ensemble tree. A selection of decision trees from the
Machine Learning
Classifier randomly chosen training set subset is the Random Forest
Classifier. To evaluate the final class it takes votes from the
different decision trees [14] [15].
Cross Validation
C. Data Description
Precision ROC
In this paper, 400 chronic kidney disease datasets are
Recall TP rate used from the UCI machine learning repository [16]. In this
Evaluation Metrics
dataset, there are 25 attributes, 24 are predictive variables
F measure TN rate and 1 attribute is decision class. Now we discuss briefly our
all attributes.
Age: This attribute represents the age in years of a person. It
Best classifier is a numeric value and it is a predictive variable.
Bp: This attribute means blood pressure and it is a numeric
value and it is a predictive variable.
Fig. 1. Implementation Procedure of Chronic Kidney Disease
Sg: This attribute means specific gravity and its value is
numeric. It is a predictive variable.
Al: This attribute means albumin. its value is numeric. It is a
B. Algorithms Description predictive variable.
Naive Bayes: The classifiers of Naive Bayes are a group of Su: This attribute means sugar. its value is numeric. It is a
classification algorithms that are logically centered in Bayes. predictive variable.
It’s not a single algorithm, but a family of algorithms where Rbc: This attribute means red blood cells. its value is
they all share a similar definition. nominal. It is a predictive variable.
Logistic regression: Logistics regression is a mathematical Pc: This attribute means pus cell. its value is nominal. It is a
model that, while there are many more complicated extensions, predictive variable.
uses logistical functions in its basic form to model a binary de- Pcc: This attribute means pus cell clumps. its value is
pendent variable. Regression analysis measures logistics model nominal. It is a predictive variable.
parameters by logistic regression (a form of binary regression). Ba: This attribute means bacteria. its value is nominal. It is a
The statistical Binary Logistic Model has a dependent variable predictive variable.
of two possible values, for example pass/fail, suggesting the Bgr: This attribute means blood glucose random. its value is
two values ”0” or ”1”. numeric. It is a predictive variable.
MLP: The multi-layer perceptron MLP classifier is a feed- Bu: This attribute means blood urea. its value is numeric. It
forward artificial neural network model that maps input data is a predictive variable.
sets to a compatible output range. The term MLP is used Sc: This attribute means serum creatinine. its value is
ambiguously, often loosely for any feed-forward ANN, often numeric. It is a predictive variable.
explicitly to refer to networks composed of several layers of Sod: This attribute means sodium. its value is numeric. It is
perceptions. a predictive variable.
SGD Classifier: Stochastic Gradient Descent ( SGD) is very Pot: This attribute means potassium. its value is numeric. It
effective approach to fitting convex loss functions with linear is a predictive variable.

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 715

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

Hemo: This attribute means hemoglobin. its value is numeric. complementary data subset. Comparing and selecting a model
It is a predictive variable. or a given predictive modeling problem is commonly used in
Pcv: This attribute means packed cell volume. its value is applied machine learning because it is simple to grasp, easy
numeric. It is a predictive variable. to implement, and results in ability predictions that usually
Wc: This attribute means white blood cell count. its value is have a smaller bias than other methods. Since small numbers
numeric. It is a predictive variable. are in our data collection, we use k-fold cross validation. The
Rc: This attribute means red blood cell count. its value is data is split into k subsets in K Fold cross validation. Now
numeric. It is a predictive variable. the holdout solution is repeated k times such that every time
Htn: This attribute means hypertension. its value is nominal. the test set/validation is used as one of the k subsets and all
It is a predictive variable. the other k-1 subsets are combined to form a training set. In
Dm: This attribute means diabetes mellitus. its value is all k experiments, the error estimate is summed for maximum
nominal. It is a predictive variable. efficacy of our model. Which can be shown, each data point is
Cad: This attribute means coronary artery disease. its value checked exactly once and is trained in a k-1 times series. As
is nominal. It is a predictive variable. we use most of the fitting data, this decreases bias considerably
Appet: This attribute means appetite. its value is nominal. It and also considerably reduces variance, as most data are also
is a predictive variable. used in validation sets. Another drawback of this approach
Pe: This attribute means pedal edema. its value is nominal. is that the interchange of training sets and evaluation sets.
It is a predictive variable. K = 5 or 10 is normally optimal as a general rule and an
Ane: This attribute means anemia. its value is nominal. It is observational proof, but nothing is set and any meaning can
a predictive variable. be taken. We use k(10) fold cross validation to find out our
Class: It is the response attribute. A person having chronic evaluation metrics.
kidney disease or not having chronic kidney disease are given
here. And this variable is a nominal type. H. Evaluation Metrics
The outcome results of this study are described briefly as
D. Data Preprocessing follows
Precision
In this dataset, total 25 attributes are used, 24 are predictive
P recision or positive predictive value
variables and 1 is response variable. In predictive 24 attributes T RU E P OSIT IV E
some attributes are nominal some attributes are numerical. =
T RU E P OSIT IV E + F ALSE P OSIT IV E
Therefore convert nominal attributes into numerical attributes Recall
T RU E P OSIT IV E
using a mapping function. In this dataset, the nominal value Recall =
contains rbc, appet, pc, pcc, ba, dm, htn, cad, pe, ane this T RU E P OSIT IV E + F ALSE N EGAT IV E
F Measure
attribute we convert this a numeric value using a mapping 2 ∗ TP
F 1score =
function. And now our data set is full of numeric value. 2 ∗ TP + FP + FN
Moreover splitting the dataset into 80% for training and 20%
Accuracy
for testing use. TP + TN
Accuracy =
E. Feature Extraction TP + TN + FP + FN

In this section, from fig 2 shown feature extraction using


principle component analysis. In principle component analysis I. Confusion Metrics
we colour into two type red and green. The red colour This figure 3 represents the confusion matrix structure.
represent CKD and the green colour represent not NOT CKD • TP means true positive = when a label predicted correctly,
F. Machine Learning Classifier we predicted it is ckd and it is exactly the same.
• FP means false positive = when a label predicted falsely,
This study is performed using eight machine learning we predicted it is ckd but it is not ckd.
algorithms to evaluate the CKD or NOT CKD performance. • FN means false negative = when a predicted label is
This algorithms are: Naive Bayes(NB), Logistic missing, we predicted it is not ckd but it is ckd.
Regression(LR), Multilayer Perceptron(MLP), Stochastic • TN means true negative = when a label is correctly
Gradient Descent(SGD), Adaptive Boosting(Adaboost), predicted by another label, we predicted it is not ckd
Bagging, Decision Tree(DT), and Random Forest(RF). and in the dataset it is not ckd.

G. Cross Validation J. Best Model


Cross validation is a technique for testing machine learning The degree of precision of the data set is also calculated to
models by training multiple machine learning techniques on generate accuracy of various types of algorithms and the best
the available input data subsets and evaluating them on the one can be achieved by using a Random Forest classification.

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 716

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

2 component PCA
ckd
notckd
8

4
Principal Component 2

−2

−4

−4 −2 0 2 4 6 8
Principal Component 1

Fig. 2. Principle Component Analysis

IV. RESULT ANALYSIS dient Descent(SGD), Adaptive Boosting(Adaboost), Bagging,


Decision Tree(DT), Random Forest(RF). Moreover the highest
The confusion metrics analysis are shown in table II, value from the random forest(RF) classifier and it accuracy is
therefore confusion metrics analysis is important part of 99%. The second highest accuracy achieved from MLP, SGD,
this research. For achieve good accuracy predicted positive, and Decision Tree classifiers and value are 98%. The lowest
predicted negative is more important attribute. In this paper, accuracy we get from the Naive Bayes(NB) classifier and
Random Forest gives the best outcome from any other value is 95%. Second lowest accuracy we gain from Logistic
classifiers. Regression classifier and it is 96%. This all the result section
of this paper focus to find out best outcome of every classifier
The perform analysis shown in table III using different ma-
chine learning classifiers namely: Naive Bayes(NB), Logistic
Regression(LR), Multilayer Perceptron(MLP), Stochastic Gra-

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 717

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

TABLE II
C ONFUSION METRICS ANALYSIS FOR OUR APPLIED CLASSIFIER

Model Label Predictive Positive Predictive Negative


Naive Bayes CKD 229 21
NOT CKD 1 149
Logistic CKD 239 11
NOT CKD 7 143
MLP CKD 241 9
NOT CKD 1 149
SGD CKD 243 7
NOT CKD 0 150
AdaBoost CKD 243 7
NOT CKD 8 142
Bagging CKD 246 4
NOT CKD 7 143
Decision tree CKD 249 1
NOT CKD 9 141
Random forest CKD 249 1
NOT CKD 3 147

TABLE III
E VALUATION METRICS ANALYSIS OF OUR USES CLASSIFIER

Model Accuracy Label Precision Recall F Score ROC TP FP


Naive Bayes 95% CKD 0.99 0.91 0.95 0.99 0.91 0.00
NOT CKD 0.87 0.99 0.93 0.99 0.08
Logistic 96% CKD 0.97 0.95 0.96 0.98 0.95 0.04
NOT CKD 0.92 0.95 0.94 0.95 0.04
MLP 98% CKD 0.99 0.96 0.98 0.99 0.96 0.00
NOT CKD 0.94 0.99 0.96 0.99 0.03
SGD 98% CKD 1.00 0.97 0.98 0.98 0.97 0.00
NOT CKD 0.95 1.00 0.97 1.00 0.02
AdaBoost 96% CKD 0.96 0.97 0.97 0.99 0.97 0.05
NOT CKD 0.95 0.94 0.95 0.94 0.02
Bagging 97% CKD 0.97 0.98 0.97 0.98 0.98 0.04
NOT CKD 0.97 0.95 0.96 0.95 0.01
Decision tree 98% CKD 0.96 0.99 0.98 0.97 0.99 0.06
NOT CKD 0.99 0.94 0.96 0.94 0.00
Random forest 99% CKD 0.98 0.99 0.99 0.98 0.99 0.02
NOT CKD 0.99 0.98 0.98 0.98 0.00

V. CONCLUSION

Chronic kidney disease(CKD) describes the significant loss


of kidney function, also called chronic kidney failure. Kidneys
filter the wastes and excess blood fluids from the blood
and then excreted in the urine. Dangerous levels of fluid,
electrolytes, and waste can build-up in the body when chronic
kidney disease reaches an advanced stage. CKD raises the risk
of end-stage renal disease (ESRD) and cardiovascular disease,
and other risk factors for heart disease, such as elevated blood
lipids, are also found in individuals with CKD. In individuals
with CKD, the most common cause of death is cardiovascular
disease rather than kidney failure. Kidney disease or chronic
kidney disease is not a minor issue for a person’s health. It
will be a death cause for a person, so everyone needs to be
concerned about his health to prevent this in the very early
stage of his or her life. In this paper, several machine learning
classifiers have been used to find out best accuracy, ROC,
Fig. 3. Confusion Metrics Structure
precision, recall and f measure. However random forest gives
the 99% accuracy and its ROC value is also highest.

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 718

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.
Proceedings of the Sixth International Conference on Inventive Computation Technologies [ICICT 2021]
IEEE Xplore Part Number: CFP21F70-ART; ISBN: 978-1-7281-8501-9

R EFERENCES
[1] Schmidt, M. I., Duncan, B. B., Sharrett, A. R., Lindberg, G., Savage, P. J.,
Offenbacher, S., ... & ARIC investigators. (1999). Markers of inflamma-
tion and prediction of diabetes mellitus in adults (Atherosclerosis Risk in
Communities study): a cohort study. The Lancet, 353(9165), 1649-1652.
[2] Wikipedia contributors. (2020, December 2). Chronic kidney disease. In
Wikipedia, The Free Encyclopedia. Retrieved 09:48, December 3, 2020,
from https://en.wikipedia.org/w/index.php
[3] Wikipedia contributors. (2020, November 21). Coronary arteries. In
Wikipedia, The Free Encyclopedia. Retrieved 09:50, December 3, 2020,
from https://en.wikipedia.org/w/index.php
[4] Amirgaliyev, Y., Shamiluulu, S., Serek, A. (2018, October). Analysis of
chronic kidney disease dataset by applying machine learning methods. In
2018 IEEE 12th International Conference on Application of Information
and Communication Technologies (AICT) (pp. 1-4). IEEE.
[5] Charleonnan, A., Fufaung, T., Niyomwong, T., Chokchueypattanakit, W.,
Suwannawach, S., Ninchawee, N. (2016, October). Predictive analytics
for chronic kidney disease using machine learning techniques. In 2016
Management and Innovation Technology International Conference (MITi-
con) (pp. MIT-80). IEEE.
[6] Sinha, P., Sinha, P. (2015). Comparative study of chronic kidney disease
prediction using KNN and SVM. International Journal of Engineering
Research and Technology, 4(12), 608-12.
[7] Sharma, S., Sharma, V., Sharma, A. (2016). Performance based eval-
uation of various machine learning classification techniques for chronic
kidney disease diagnosis. arXiv preprint arXiv:1606.09581.
[8] Khan, B., Naseem, R., Muhammad, F., Abbas, G., Kim, S. (2020).
An Empirical Evaluation of Machine Learning Techniques for Chronic
Kidney Disease Prophecy. IEEE Access, 8, 55012-55022.
[9] Polat, H., Mehr, H. D., Cetin, A. (2017). Diagnosis of chronic kidney
disease based on support vector machine by feature selection methods.
Journal of medical systems, 41(4), 55.
[10] Radha, N., Ramya, S. (2015). Performance analysis of machine learning
algorithms for predicting chronic kidney disease. Int. J. Comput. Sci. Eng.
Open Access, 3, 72-76.
[11] Serpen, A. A. (2016). Diagnosis rule extraction from patient data for
chronic kidney disease using machine learning. International Journal of
Biomedical and Clinical Engineering (IJBCE), 5(2), 64-72.
[12] Ahmad, M., Tundjungsari, V., Widianti, D., Amalia, P., Rachmawati, U.
A. (2017, November). Diagnostic decision support system of chronic kid-
ney disease using support vector machine. In 2017 Second International
Conference on Informatics and Computing (ICIC) (pp. 1-4). IEEE.
[13] Hosseinzadeh, M., Koohpayehzadeh, J., Bali, A. O., Asghari, P., Souri,
A., Mazaherinezhad, A., ... & Rawassizadeh, R. (2020). A diagnostic
prediction model for chronic kidney disease in internet of things platform.
Multimedia Tools and Applications, 1-18.
[14] M. U. Emon, M. S. Keya, T. I. Meghla, M. M. Rahman, M. S. A.
Mamun and M. S. Kaiser, ”Performance Analysis of Machine Learning
Approaches in Stroke Prediction,” 2020 4th International Conference on
Electronics, Communication and Aerospace Technology (ICECA), Coim-
batore, 2020, pp. 1464-1469, doi: 10.1109/ICECA49313.2020.9297525.
[15] M. A. Islam, M. S. Islam, M. S. Hossen, M. U. Emon, M. S. Keya
and A. Habib, ”Machine Learning based Image Classification of Papaya
Disease Recognition,” 2020 4th International Conference on Electronics,
Communication and Aerospace Technology (ICECA), Coimbatore, 2020,
pp. 1353-1360, doi: 10.1109/ICECA49313.2020.9297570.
[16] Dua, D. and Graff, C. (2019). UCI Machine Learning Repository
http://archive.ics.uci.edu/ml. Irvine, CA: University of California, School
of Information and Computer Science.

978-1-7281-8501-9/21/$31.00 ©2021 IEEE 719

Authorized licensed use limited to: UNIVERSITY OF WESTERN ONTARIO. Downloaded on May 25,2021 at 13:02:02 UTC from IEEE Xplore. Restrictions apply.

You might also like