Professional Documents
Culture Documents
Authorized licensed use limited to: Middlesex University. Downloaded on October 17,2020 at 20:53:41 UTC from IEEE Xplore. Restrictions apply.
2020 17th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON)
n
X
Gini = 1 − p2 (ci ) (3)
i=1
748
Authorized licensed use limited to: Middlesex University. Downloaded on October 17,2020 at 20:53:41 UTC from IEEE Xplore. Restrictions apply.
2020 17th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON)
749
Authorized licensed use limited to: Middlesex University. Downloaded on October 17,2020 at 20:53:41 UTC from IEEE Xplore. Restrictions apply.
2020 17th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON)
TABLE I. C OMPARE AUC, S ENSITIVITY, S PECIFICITY AND ACCURACY OF ALL MODELS , TRAINED AND TESTED BY FOUR SUB - DATASETS FOR THE
PREDICTIONS OF CATEGORICAL CLASSES OF AGG CAG.
AUC Sensitivity Specificity Accuracy
D1 D2 D3 D4 D1 D2 D3 D4 D1 D2 D3 D4 D1 D2 D3 D4
Logistic Regression 0.76 0.71 0.76 0.74 0.53 0.60 0.60 0.67 0.83 0.83 0.75 0.79 0.72 0.74 0.69 0.74
Decision Tree 0.72 0.74 0.75 0.55 0.60 0.73 0.67 0.60 0.83 0.75 0.83 0.50 0.74 0.74 0.77 0.54
SVM 0.76 0.75 0.72 0.69 0.60 0.73 0.67 0.60 0.92 0.83 0.96 0.71 0.79 0.79 0.85 0.67
Random Forest 0.88 0.85 0.85 0.81 0.67 0.67 0.67 0.73 0.88 0.75 0.83 0.75 0.79 0.72 0.77 0.74
AdaBoost 0.68 0.76 0.68 0.69 0.87 0.73 0.60 0.67 0.50 0.79 0.75 0.71 0.64 0.77 0.69 0.69
XGBoost 0.83 0.80 0.82∗ 0.74 0.60 0.73 0.80∗ 0.53 0.88 0.83 0.71∗ 0.79 0.77 0.79 0.74∗ 0.69
Gradient Boosting 0.78 0.79 0.85 0.78 0.73 0.80 0.73 0.73 0.79 0.50 0.79 0.75 0.77 0.62 0.77 0.74
Mean 0.74 0.75 0.75 0.69 0.62 0.70 0.62 0.59 0.76 0.75 0.81 0.74 0.71 0.73 0.74 0.68
that this assumption cannot be applied to some tasks of image boosting process and also has in-built L1 (Lasso Regression)
datasets, in which every pixel is considered training features. and L2 (Ridge Regression) regularization that prevents the
model from overfitting. Moreover, the results suggested that
Therefore, we experimented on seven predictive models,
although including patient characteristics (D1 and D2) did not
i.e., Logistic Regression, Decision Tree, SVM, Random Forest,
harm the results of stenosis predictions, it did not improve the
AdaBoost, XGBoost and Gradient Boosting. For model selec-
performance noticeably either.
tion, our study aims to closely control the behavior of predic-
tive models when optimizing for performance and finding the V. CONCLUSION
right balance between bias and variance. We thus performed
a robust two-level validation for, at the first-level, estimating In this paper, we have evaluated the prediction performance of
the generalization performance and for, at the second level, seven machine learning models for the stenosis classifica-tion.
finding the best hyperparameter settings of models. High-level rMPI features from 4D-MSPECT and patient
characteristics have been employed to train and test classifiers,
We opted for a classic strategy, called “Model selection with a purpose to find whether patient characteristics could
via k-fold cross validation” [9]. That is, in an outer process, enhance the predictions or not. PCA has also been applied to
we used a two-way holdout sampling by randomly splitting reduce noise in the rMPI features, grouped by coronary
dataset into a training set and a test set with a ratio of arteries. We have set up an empirical study using two-level
80 to 20, respectively. In an inner process, we used 10-fold validation for model selection and performance optimization.
cross-validation (i.e., k=10) for performance optimization and Results of our study have suggested that the XGBoost trained by
hyperparameter tuning. Recall that as the most interesting only original high-level rMPI features is the most suitable
metric of our task was AUC, we defined AUC as a key scorer model for predicting the stenosis of coronary arteries.
when optimizing our predictive models. After the inner process
was complete, the best hyperparameters of each model were ACKNOWLEDGMENT
selected to train a specific classifier in the outer process. For This work was a collaboration between KMITL and
model selection and comparison, the predictions of classifiers Rajavithi hospital. Two IRBs have ethically approved this
was validated against unseen test set for an unbiased estimate research with the certificate no. “EC-KMITL 63 035” and
of the generalization performance. “Rajavithi 075/2563”. Many special thanks should also go to
IV. E VALUATION R ESULTS Miss Taratip Narawong and Dr. Tarit Taerakul (MD.) for their
patient guidance and data collection.
Table I shown the experimental results in terms of four
metrics, i.e., AUC, Sensitivity, Specificity and Accuracy. The REFERENCES
results highlighted in bold indicated the best performing mod-
[1] F. A. Mettler and M. J. Guiberteau, “Cardiovascular system,” in
els in each column or sub-dataset (D1, D2, D3, or D4.) Italic Essen-tials of Nuclear Medicine Imaging, 6th ed. Elsevier Health
means the model obtained the highest score in each metric. As Sciences, 2012.
we can see, the best performing model in terms of AUC was [2] P. Slomka, Y. Xu, D. Berman, and G. Germano, “Quantitative
Random Forest. However, it performed relatively worse than analysis of perfusion studies: strengths and pitfalls,” Journal of
other models regarding the Sensitivity, which is a significant nuclear cardiology: official publication of the American Society of
metric indicating that the model lacked an ability to correctly Nuclear Cardiology, vol. 19, no. 2, pp. 338–346, April 2012.
classify positive examples (i.e., actual patients with CAD.) [3] T. M. Mitchell, Machine Learning. New York: McGraw-Hill,
1997.
Therefore, with due care we inspected the results by
considering Sensitivity more important than Specificity (but [4] L. Breiman, “Random forests,” Machine Learning, vol. 45, no. 1,
pp. 5–32, 2001.
still acceptably high) while regarding AUC and Accuracy as
equally important. We accordingly highlighted such four mod- [5] Y. Freund and R. E. Schapire, “A decision-theoretic
generalization of on-line learning and an application to
els by underlining them as candidates for possible deployment boosting,” Journal of Computer and System Sciences, vol. 55, no.
in production. As indicated by asterisk (*), we eventually 1, pp. 119 – 139, 1997.
selected the XGBoost model that was trained by the sub- [6] J. H. Friedman, “Greedy function approximation: A gradient
dataset D3 as the most appropriate classifier for the prediction boosting machine,” Annals of Statistics, vol. 29, pp. 1189–1232,
of the stenosis. Although the XGBoost model with D3 did not 2000.
perform the best in terms of AUC, it still performed quite well [7] T. Chen and C. Guestrin, “Xgboost: A scalable tree boosting
with AUC 0.82; meanwhile, relatively high Sensitivity 0.80 system,” in Proceedings of the 22nd ACM SIGKDD International
and acceptable Specificity 0.71. This was because XGBoost Conference on Knowledge Discovery and Data Mining, 2016, pp.
usually performs a cross-validation at each iteration of its 785–794.
[8] I. Jolliffe, “Principal component analysis,” in International
Encyclopedia of Statistical Science, M. Lovric, Ed. Springer Berlin
Heidelberg, 2011, pp. 1094–1096.
[9] S. Raschka, “Model evaluation, model selection, and algorithm
selection in machine learning,” arXiv preprint arXiv:1811.12808,
2018.
750
Authorized licensed use limited to: Middlesex University. Downloaded on October 17,2020 at 20:53:41 UTC from IEEE Xplore. Restrictions apply.