Professional Documents
Culture Documents
1
Samsun University, Faculty of Engineering, Department of Industrial Engineering, Samsun,
55420, Turkey
2
Gaziantep Islam Science and Technology University, Faculty of Engineering and Natural
Sciences, Department of Industrial Engineering, Gaziantep, 27010, Turkey
Abstract
This article has been accepted for publication and undergone full peer review but has not been
through the copyediting, typesetting, pagination and proofreading process which may lead to
differences between this version and the Version of Record. Please cite this article as doi:
10.1002/for.2747
1. Introduction
It is necessary to have a healthy society for a country's economy to be strong and its economic
growth to continue. The existence of a healthy society is provided by qualified health services
that meet the needs of society. Healthcare expenditures (HCE) or healthcare expenditure per
capita (pcHCE) are important parameters that measure the health economy. In recent years,
HCE has increased globally and is expected to continue growing. Such a trend raises concerns
about the sustainability of health financing. Thus, the concerns about the rapid growth in HCE
and its long-term sustainability have led researchers to investigate the determinants of HCE.
In the literature, many studies have been conducted to identify the determinants affecting HCE
or pcHCE. In most studies, the relationship between HCE and gross domestic product (GDP)
has been extensively investigated. For example, Toor and Butt (2005) examined six factors that
affect the HCE of Pakistan, such as pcGDP, urbanization, literacy rate, crude birth rate, and
foreign aid (Toor & Butt, 2005). Khan et al. (2016) investigated the short-run and long-run
equilibrium dynamic causal relationship between HCE and GDP per capita (pcGDP) within
the time series framework from 1981 to 2014 in Malaysia. The analysis results showed that
pcGDP, population growth, population structure, and technology play important roles in
explaining changes in the HCE (Khan, Razali, & Shafie, 2016). Rodríguez and Nieves Valdés
(2019) investigated the relationship between GDP and HCE for a group of Latin American and
the Caribbean countries and OECD countries for the period 1995–2014. They showed that GDP
does not react to changes in the HCE level in the long-term (Rodríguez & Nieves Valdés,
2019).
The change in the HCE is also closely related to the demographic characteristics, socio-
economic factors, and health resources of a country such as the number of hospitals and
physicians, life expectancy at birth, unemployment rate, and crude birth rate, etc. The
unemployment rate is one of the most critical factors affecting the economic structure of
Accurate estimation of HCE is essential for policy-makers to plan future resources. Besides,
estimating the amount of the budget allocated for healthcare has a significant impact on
government policy and planning (Astolfi, Lorenzoni, & Oderkirk, 2012). In the literature, many
studies have been performed on the forecasting of HCE of different countries using various
techniques from classical to advanced models.
For example, Getzen and Poullier (1992) studied the estimation of the HCE for 19 OECD
countries using data between 1965 and 1979. They estimated the HCE from 1980 to 1987 by
using exponential smoothing, moving average, and autoregressive integrated moving average
(ARIMA) methods. The results of the study showed that multivariate regression models
provide more accurate predictions than time series models (Getzen & Poullier, 1992). Di
Matteo (2010) used determinant regression and growth rate extrapolation techniques to forecast
the HCE of Canada. The results showed that the budget allocated for healthcare by the
Canadian state government would continue to increase in the future, and its share in provincial
GDP will increase (Matteo, 2010). Chaabouni and Abednnadher (2013) estimated the HCE of
Tunisia based on socio-economic and demographic variables using artificial neural network
(ANN) and Autoregressive Distributed Lag (ARDL) models (Chaabouni & Abednnadher,
2013). Zhao (2015) predicted annual HCE using the exponential smoothing, ARIMA, and
Vector Autoregression (VAR) models for the 34 member countries of the OECD. Unlike the
study by Getzen and Poullier (1992), it was observed that simple statistical and time series
models provide better predictions against complex micro panel data models (Zhao, 2015).
Klazoglou and Dritsakis (2018) used the ARIMA model to predict total health spending in the
USA from 1900 to 2017. In their study, they aimed to identify the appropriate model based on
the Box–Jenkins methodology. The results showed that the ARIMA (2,1,0) model was the best
model to forecast the HCE in the USA (Klazoglou & Dritsakis, 2018). Özcan and Tüysüz
(2018) used ARIMA and grey forecasting models for predicting the pcHCE of Turkey. They
applied the Genetic algorithm for training data size and parameter optimization of the grey
forecasting models. The results demonstrated that the optimization of parameters and training
data size together with a rolling mechanism highly improve the forecasting performance of the
grey models (Özcan & Tüysüz, 2018). Zheng et al. (2020) used the ARIMA (3,3,0) model to
predict changes in total HCE in China from 2018 to 2022. The analysis results of the study
As shown in Table 1, the dataset was created using different sources for the years between
1990 and 2018. The pcHCE data provided by the Organization for Economic Co-operation and
Development (OECD) was used as an output variable (OECD, 2020). According to the data of
OECD, pcHCE has recorded the highest increase in 2018 with 15.9% since 2012. Although an
increase in the amount of pcHCE has been observed, some declines have been found in the
twenty-eight years. As shown in Figure 1, the most dramatic decrease in the amount of pcHCE
in Turkey occurred between 1993-1994 and 2014-2015 due to the economic crisis. On the other
hand, the input dataset was constructed by collecting historical values such as pcGDP, life
expectancy at birth (LE), unemployment rate (UR), crude birth rate (CBR), the number of
hospitals (HN), and the number of hospitals physicians (PN). Figure 2 indicates the change in
input parameters depending on the years.
Table 2 shows the Pearson correlation coefficients (R) to determine the degree of the linear
relationship between the input variables and the input variables and the pcHCE in the years
between 1990 and 2018. As shown in Table 2, there is a high degree of correlation between the
In this study, five different AI-based estimation techniques, namely RF, ANN, MLR, SVR, and
RVM, were used for pcHCE prediction of Turkey. An evolutionary genetic algorithm based
feature selection (GAFS) technique was also applied to all models to get relevant features
subset to improve the performance accuracy of the models. All analyses were carried out on
the RapidMiner Studio 9.6 software. The details of the five models are described in the
following sub-paragraphs.
Random Forest (RF) algorithm is an ensemble learning method that combines a large set of
classification and regression trees. The RF starts with many bootstrap samples randomly
selected from the original pre-dataset. A regression tree is fitted to each of the bootstrap
samples. For each node per tree, a small set of input variables is extracted randomly from all
set at a certain rate (Wang, Zhou, Zhu, Dong, & Guo, 2016). The predicted value of an
observation is calculated by taking the average of all trees.
RF is general in many data sets to show the significance values of variables in its analysis
(Archer & Kimes, 2008). It does not include bias in the estimation results obtained. Therefore,
transactions with many independent variables are obtained as unbiased estimates. In other AI
techniques, variables that are more specific and meet certain criteria are selected, while random
variables are included in the algorithm in RF. This enables the results of the random forest to
be more generalizable.
RF is expected to perform better in terms of prediction accuracy than single base learners. Also,
ensemble learning in RF can alleviate the overfitting problem. This makes RF less prone to
overfitting than other machine learning algorithms such as ANN, SVM, etc. Also, since the
Artificial neural network (ANN) is a computational model that mimics the structure, processing
capability, and learning ability of a human brain (Haykin, 1994; Lippmann, 1987). It is one of
the most widely used AI techniques to solve complex problems in engineering, finance, health,
chemistry, etc. (Graupe, 2013; Su et al., 2019; Wang et al., 2019). ANN consists of
interconnected adaptive simple processing elements, also called artificial nodes or neurons
based on brain learning ability (Li, Wang, & Qiu, 2019). It has the ability to store knowledge
and learn and model the complex nonlinear relationship between output and input data (Ceylan
& Bulkan, 2018). This feature makes the ANN model superior to other AI methods. It has
generalization ability so that it can interpret unseen data. Moreover, the important feature of
the ANN model is that it is fault-tolerant. This means that it is capable of processing noisy or
fuzzy information, while it can endure incomplete or missing data. Like all approaches, ANN
has some limitations. Training can be time-consuming depending on the complexity of the
modelled data. As the number of hidden layers needed to capture the properties of the data
increases, so does the time required to complete the training. Overfitting is another problem of
the ANN model, which causes the memorization of training cases and poor performance of the
network.
Multiple Linear Regression (MLR) analysis is one of the most common techniques used for
estimation due to its easy use. MLR is also the most preferred approach to measure the effect
of independent variables on dependent variables. This technique usually refers to an equation
between a single dependent variable and multiple independent variables. The equation contains
a constant number and coefficients for independent variables. If an independent variable has
no effect on the dependent variable, its coefficient will be zero. The main limitation of this
method is its unsuitability in nonlinear problems (Atalan, 2019).
Relevance Vector Machine (RVM) is a kernel-based AI algorithm used for both classification
and regression analysis (Zhou & Dhupia, 2020). RVM is based on Bayesian inference that
provides restricted solutions. Thus, RVM uses an entirely probabilistic framework and brings
the prior probability to the weights of the model. In this model, there is a one-to-one and
dominating-dominated relationship from the hyperparameters to the weights. The
hyperparameters are obtained by repeated iteration, where the posterior distribution of most
weights approaches zero (Kong et al., 2019). It adopts the same functional form of SVM.
Compared to the SVM model, the RVM model can provide probabilistic predictions. Besides,
the most significant advantage of the RVM model is that the high sparseness of RVM can
reduce the number of kernel functions involved in computation, which makes it especially
suitable for online monitoring. The kernel functions of SVM must meet Mercer's state.
However, the selection of kernel function is no longer limited by Mercer’s condition in the
RVM model. Although the RVM model has the above advantages, it has some limitations in
practical applications. The learning procedure of the RVM model is usually much slower than
the SVM since it is a problem of O(N3).
The feature vector plays an essential role in the performance of the models. The relevant feature
is necessary for the training process as it has an informative aspect to improve the model.
However, the irrelevant features are less informative, so their inclusion can negatively affect
the performance of the models. It is the task of the feature selection methods to determine
which features are relevant or irrelevant. Feature selection or input selection methods are used
to select a subset of variables that can efficiently identify input data. In this way, irrelevant and
redundant features that do not contribute or decrease the accuracy of the predictive model are
determined and removed from the dataset (Srinivasa Murthy & Koolagudi, 2018).
In this study, the feature selection task was performed using a Genetic algorithm (Gómez and
Quesada, 2017; Rachmani et al., 2019). The genetic algorithm (GA) is an important heuristic
algorithm inspired by the natural evolution procedure. It generates solutions to optimization
problems based on the mechanics of natural genetics and biological evolution. The steps of the
GAFS method is depicted in Figure 3 (Huang & Wang, 2006; Welikala et al., 2015).
The first step is to create the individuals in the population. Each individual (chromosome) in
the population represents a candidate solution to the feature subset selection problem. The
number of genes is the total number of features in the data set. Let m be the total number of
features. Each individual is represented by a binary vector of dimension m. Each positive gen
value of ‘1’ means that the corresponding feature is included in the model; otherwise (if ‘0’),
the feature is not selected. A random population is then generated that represents different
points in the search space. Fitness function, which gives the quality of the produced member
of the population, is assigned to each individual. After performing fitness assignment, the
selection operator chooses the individuals that will recombine for the next generation.
The selection mechanism mimics the principle of survival-of-the-fittest mechanism in nature.
Thus, the selection operator selects the individuals according to their pre-assigned fitness
levels, and selected ones enter the mating pool. After the selection process, crossover and
mutation operators are applied to the selected individuals. The crossover operator, also called
recombination, randomly selects two individuals and recombines their features to generate new
offsprings for the new population. Afterward, for each offspring, its fitness function is
calculated. The crossover operator can produce offsprings that are very similar to parents. This
leads to a new generation with low diversity. The mutation operator overcomes this problem
by randomly changing the value of some features in the offsprings. Afterward, for each
offspring, its fitness function is calculated. The fitness assignment, selection, crossover, and
∑𝑛𝑖=1(𝑝𝑗 − 𝑎𝑗 )2
𝑅𝑀𝑆𝐸 = √ (3)
𝑛
𝑛
1 𝑝𝑗 − 𝑎𝑗
𝑀𝐴𝑃𝐸 = ∑ | | × 100 % (4)
𝑛 𝑎𝑖
𝑖=1
2
2
∑𝑛𝑖=1[(𝑝𝑗 − 𝑎𝑗 )]
𝑅 =1− 2 (5)
∑𝑛𝑖=1[(𝑎𝑗 − 𝑎̅)]
where 𝑎𝑗 is the actual value and 𝑝𝑗 is the predicted value for the 𝑗𝑡ℎ value, 𝑎̅ and 𝑝̅ are the
average of the actual and predicted values, respectively. Also, n is the total number of data
points. The higher values of R2 and the lower values of MAD, RMSE, and MAPE mean better
performance of the developed model. Each model was divided with a percentage split strategy
(70% for training and 30% for testing), which is a widespread tool to predict the performance
of a model. The testing stage was carried out to evaluate the prediction precision and the
generalization capability of the developed models. Table 3 depicts the performance metric
values for all predictive models. Here all the features (pcGDP, UR, CBR, PN, HN, and LE) in
the dataset were used as inputs for modeling at all stages, including training and testing. The
values of R2 were in the range of 0.9569-0.9969 in the training dataset and 0.9385-0.9874 in
the testing dataset. The highest R2 (0.9874) value and the lowest MAD (29.2356), RMSE
(42.2965), and MAPE (4.6082) values in the test data were obtained by the RF model, which
implies that the model was successful in predicting pcHCE. Although the ANN and RVM
models performed close to the RF model during the testing phase, the performance of the MLR
model was rather low compared to other models. The nonlinear structure of the data makes it
difficult to achieve high accuracy by traditional regression methods such as MLR. In such
3.4. Discussion
The objective of this study is to employ accurate and robust AI-based models to estimate the
pcHCE of Turkey accurately. For this purpose, the GAFS method, a useful technique for
dimension reduction to increase the performance of the models, was used as a feature selection
technique in this study. Compared to AI-based estimation models, optimized models with the
GAFS method provided better predictive performance in the dataset. As seen in Table 6, the
performance measures were synthesized by using overall performance measure; 𝑆𝐼 =
1 𝑃𝑖 −𝑃𝑚𝑖𝑛,𝑖
∑𝑚
𝑖=1 ( ), where 𝑃𝑖 is the value of the 𝑖𝑡ℎ average performance measurement,
𝑚 𝑃𝑚𝑎𝑥,𝑖 −𝑃𝑚𝑖𝑛,𝑖
𝑃𝑚𝑖𝑛,𝑖 is the minimum value of the 𝑖𝑡ℎ performance measure, 𝑃𝑚𝑎𝑥,𝑖 is the maximum value of
4. Conclusion
There are many parameters used in evaluating the healthcare systems of countries. Just as the
economic developments of countries can be expressed with pcGDP, the development of health
systems can also be measured by the amount of pcHCE. In this study, six important factors
affecting pcHCE, namely pcGDP, life expectancy at birth, the unemployment rate, crude birth
rate, the number of physicians and hospitals, were used as inputs. The different AI-based
forecasting models such as RF, ANN, MLR, SVR, and RVM were used for the prediction of
pcHCE of Turkey. An evolutionary genetic algorithm was used as the feature selection method
to improve the prediction performance of the AI models. A comprehensive comparison
demonstrated that the combination of the GAFS method and AI-based prediction models has
strong predictive power compared to basic models that do not use the selection method. Among
all approaches, the GAFS-RF algorithm outperformed in predicting the pcHCE of Turkey. The
results of the study can be an important resource for the future planning of decision-makers in
health to take the necessary measures and determine sustainable policies. In addition, the
authors believe that comparing other evolutionary approaches such as particle swarm
optimization (PSO) and ant colony optimization (ACO) to select appropriate features would be
a good subject for future studies.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
Akca, N., Sönmez, S., & Yılmaz, A. (2017). Determinants of health expenditure in OECD countries:
A decision tree model. Pakistan Journal of Medical Sciences, 33(6), 2–6.
https://doi.org/10.12669/pjms.336.13300
Archer, K. J., & Kimes, R. V. (2008). Empirical characterization of random forest variable
importance measures. Computational Statistics & Data Analysis, 52(4), 2249–2260.
https://doi.org/10.1016/J.CSDA.2007.08.015
Astolfi, R., Lorenzoni, L., & Oderkirk, J. (2012). Informing policymakers about future health
spending: A comparative analysis of forecasting methods in OECD countries. Health Policy,
107(1), 1–10. https://doi.org/10.1016/j.healthpol.2012.05.001
Boachie, M. K., Mensah, I. O., Sobiesuo, P., Immurana, M., Iddrisu, A.-A., & Kyei-Brobbey, I.
(2014). Determinants of public health expenditure in Ghana: a cointegration analysis. Ournal of
Behavioural Economics, Finance, Entrepreneurship, Accounting and Transport, 2(2), 35–40.
https://doi.org/10.12691/jbe-2-2-1
Braendle, T., & Colombier, C. (2016). What drives public health care expenditure growth? Evidence
from Swiss cantons, 1970–2012. Health Policy, 120(9), 1051–1060.
https://doi.org/10.1016/J.HEALTHPOL.2016.07.009
Ceylan, Z, & Bulkan, S. (2018). Forecasting PM10 levels using ANN and MLR: A case study for
Sakarya City. Global Nest Journal, 20(2), 281–290.
Ceylan, Z. (2020a). Assessment of agricultural energy consumption of Turkey by MLR and Bayesian
optimized SVR and GPR models. Journal of Forecasting.
Ceylan, Z., Bulkan, S., & Elevli, S. (2020). Prediction of medical waste generation using SVR, GM (
1, 1) and ARIMA models : a case study for megacity Istanbul.
Chaabouni, S., & Abednnadher, C. (2013). Modelling and forecasting of Tunisia's health expenditures
using artificial neural network and ARDL models -.
Clemente, J., Lázaro-Alquézar, A., & Montañés, A. (2019). US state health expenditure convergence:
A revisited analysis. Economic Modelling, 83, 210–220.
https://doi.org/10.1016/j.econmod.2019.02.011
Cortes, C., & Vapnik, V. (1995). Support-vector networks. Machine Learning, 20(3), 273–297.
Di Matteo, L., & Cantarero-Prieto, D. (2018). The Determinants of Public Health Expenditures:
Comparing Canada and Spain, (87800). Retrieved from https://mpra.ub.uni-
muenchen.de/87800/1/MPRA_paper_87800.pdf
Getzen, T. E., & Poullier, J.-P. (1992). International health spending forecasts: Concepts and
evaluation. Social Science & Medicine, 34(9), 1057–1068. https://doi.org/10.1016/0277-
9536(92)90136-E
Graupe, D. (2013). Principles of Artificial Neural Networks (Vol. 7). WORLD SCIENTIFIC.
Haykin, S. (1994). Neural networks: A Comprehensive Foundation. Neural Networks. Prentice Hall
PTR.
Jaba, E., Balan, C. B., & Robu, I.-B. (2014). The Relationship between Life Expectancy at Birth and
Health Expenditures Estimated by a Cross-country and Time-series Analysis. Procedia
Economics and Finance, 15(14), 108–114. https://doi.org/10.1016/s2212-5671(14)00454-7
Jiang, H., Rusuli, Y., Amuti, T., & He, Q. (2019). Quantitative assessment of soil salinity using multi-
source remote sensing data based on the support vector machine and artificial neural network.
International Journal of Remote Sensing. https://doi.org/10.1080/01431161.2018.1513180
Khan, H. N., Razali, R. B., & Shafie, A. B. (2016). Modeling determinants of health expenditures in
Malaysia: Evidence from time series analysis. Frontiers in Pharmacology, 7(MAR), 1–7.
https://doi.org/10.3389/fphar.2016.00069
Klazoglou, P., & Dritsakis, N. (2018). Modeling and Forecasting of US Health Expenditures Using
ARIMA Models. In N. Tsounis & A. Vlachvei (Eds.), Advances in Panel Data Analysis in
Applied Economic Research (pp. 457–472). Cham: Springer International Publishing.
Kong, D., Chen, Y., Li, N., Duan, C., Lu, L., & Chen, D. (2019). Relevance vector machine for tool
wear prediction. Mechanical Systems and Signal Processing, 127, 573–594.
https://doi.org/10.1016/j.ymssp.2019.03.023
Li, F., Wang, Z., & Qiu, J. (2019). Long‐term streamflow forecasting using artificial neural network
based on preprocessing technique. Journal of Forecasting, 38(3), 192–206.
https://doi.org/10.1002/for.2564
Lippmann, R. P. (1987). An Introduction to Computing with Neural Nets. IEEE ASSP Magazine.
https://doi.org/10.1109/MASSP.1987.1165576
Marinakis, Y., Dounias, G., & Jantzen, J. (2009). Pap smear diagnosis using a hybrid intelligent
scheme focusing on genetic algorithm based feature selection and nearest neighbor
classification. Computers in Biology and Medicine.
https://doi.org/10.1016/j.compbiomed.2008.11.006
Matteo, L. Di. (2010). The sustainability of public health expenditures : evidence from the Canadian
federation, 569–584. https://doi.org/10.1007/s10198-009-0214-x
Özcan, T., & Tüysüz, F. (2018). Healthcare Expenditure Prediction in Turkey by Using Genetic
Algorithm Based Grey Forecasting Models. In C. Kahraman & Y. I. Topcu (Eds.), Operations
Research Applications in Health Care Management (pp. 159–190). Cham: Springer
Pichon-Riviere, A., Augustovski, F., Garcia Marti, S., & Caporale, J. (2015). The Efficiency Path: An
Estimation of Cost-Effectiveness Thresholds for 185 Countries Based on Per Capita Health
Expenditures and Life Expectancy. Value in Health, 18(7), A695–A696.
https://doi.org/10.1016/J.JVAL.2015.09.2592
Rachmani, E., Hsu, C. Y., Nurjanah, N., Chang, P. W., Shidik, G. F., Noersasongko, E., … Lin, M. C.
(2019). Developing an Indonesia’s health literacy short-form survey questionnaire (HLS-EU-
SQ10-IDN) using the feature selection and genetic algorithm. Computer Methods and Programs
in Biomedicine. https://doi.org/10.1016/j.cmpb.2019.105047
Ray, D., & Linden, M. (2020). Health expenditure, longevity, and child mortality: dynamic panel data
approach with global data. International Journal of Health Economics and Management, 20(1),
99–119. https://doi.org/10.1007/s10754-019-09272-z
Rodríguez, A. F., & Nieves Valdés, M. (2019). Health care expenditures and GDP in Latin American
and OECD countries: a comparison using a panel cointegration approach. International Journal
of Health Economics and Management, 19(2), 115–153. https://doi.org/10.1007/s10754-018-
9250-3
Srinivasa Murthy, Y. V., & Koolagudi, S. G. (2018). Classification of vocal and non-vocal segments
in audio clips using genetic algorithm-based feature selection (GAFS). Expert Systems with
Applications. https://doi.org/10.1016/j.eswa.2018.04.005
Su, Y., Wang, Z., Jin, S., Shen, W., Ren, J., & Eden, M. R. (2019). An architecture of deep learning in
QSPR modeling for the prediction of critical properties using molecular signatures. AIChE
Journal, 65(9), 1–11. https://doi.org/10.1002/aic.16678
Toor, I. A., & Butt, M. S. (2005). Determinants of health care expenditure in Pakistan. Pakistan
Economic and Social Review, (1), 133–150.
Wang, L., Zhou, X., Zhu, X., Dong, Z., & Guo, W. (2016). Estimation of biomass in wheat using
random forest regression algorithm and remote sensing data. The Crop Journal, 4(3), 212–219.
https://doi.org/10.1016/j.cj.2016.01.008
Wang, Z., Su, Y., Shen, W., Jin, S., Clark, J. H., Ren, J., & Zhang, X. (2019). Predictive deep learning
models for environmental properties: the direct calculation of octanol–water partition
coefficients from molecular graphs. Green Chemistry, 21(16), 4555–4565.
https://doi.org/10.1039/C9GC01968E
Zheng, A., Fang, Q., Zhu, Y., Jiang, C., Jin, F., & Wang, X. (2020). An application of ARIMA model
for predicting total health expenditure in China from 1978-2022. Journal of Global Health,
10(1), 1–8. https://doi.org/10.7189/jogh.10.010803
Zhou, S., & Dhupia, J. S. (2020). Online adaptive water management fault diagnosis of PEMFC based
on orthogonal linear discriminant analysis and relevance vector machine. International Journal
of Hydrogen Energy. https://doi.org/10.1016/j.ijhydene.2019.12.193
Model
Dataset Measure
RF ANN MLR SVR RVM
MAD 13.5405 28.0721 51.7192 32.4688 28.7808
RMSE 18.0620 34.9361 66.4252 45.3277 36.6990
Training
MAPE (%) 2.58020 5.54380 11.4017 6.73610 5.82480
R2 0.99690 0.98810 0.95690 0.98070 0.98680
MAD 29.2356 39.6744 51.7120 40.4495 39.9719
RMSE 42.2965 51.2992 89.6839 68.2814 54.5336
Test
MAPE (%) 4.60820 5.74930 9.11650 6.59810 6.48720
R2 0.98740 0.98180 0.93850 0.97170 0.98110
(a)
1
Performance Measure (R2)
0.99
0.98
0.97
0.96
0.95
0.94
0.93
RF ANN MLR SVR RVM
Single Model Optimized Model
(b)
Figure 4. The performance of single and optimized models a) in the training set b) in the
testing set