You are on page 1of 11

1

Table of Contents
1) Introduction......................................................2
2) Aims and Objectives.......................................2
2.1) Problem.................................................3
2.2) Solution.................................................3
3) Related Work...................................................3
3.1) Literature Review..................................3
4) Method.............................................................3
4.1) Artificial Neural Network.....................3
4.1.1) ANN Experiment 1............................4
4.1.2) Comparison between the confusion matrix and error rates with Random Forest Tree
4.1.3) ANN Experiment 2............................5
4.1.4) PCA....................................................5
4.1.5) PCA on ANN.....................................5
4.2) Gradient Boosting..................................6
4.2.1) Gradient Boosting Experiment 1.......6
4.3) Difference between ANN and Gradient Boosting 7
4.4) Advantages of ANN and Gradient Boosting 7
4.5) Dis Advantages of ANN and Gradient Boosting 7
5) Result and Discussion......................................8
5.1) Artificial Neural Network.....................8
5.2) Performing PCA on ANN.....................9
5.3) PCA on ANN Experiment 1..................9
5.4) Gradient Boosting Experiment 1.........10
6) Conclusion and Future Work.........................10
7) REFRENCES.................................................10
2

Artificial Intelligence Applications & Algorithms


parameter tweaking, in contrast to other statistical
Abstract-- This research paper provides details learning algorithms that typically give equal
about the techniques that are being used to accuracy (e.g., neural networks and support vector
evaluate and solve a problem using Artificial machines). The method is particularly resistant to
Intelligence. This report provides with the dirty data and may be used to solve classification
information related to the algorithms that are or regression problems with a wide range of
going to be used to solve the problem. This response distributions. The squared-error and
paper provides with the comparison and exponential error loss functions are two plausible
working of the two algorithms. This report loss functions that are commonly used in
offers an overview of the use of artificial neural
networks in financial market price forecasting. regression and classification problems,
The goal of this research is to assess the respectively (Guelman, 2012).
possibilities of utilizing artificial neural networks Gradient Boosting (GB) is an iterative procedure
to forecast the financial system, which has that results in a highly accurate prediction rule by
been discussed in a number of recent merging basic parameterized functions with "poor"
publications. It will include research and performance (high prediction error). GB offers
implementation recommendations as well as interpretable findings with minimum data
references. This research also explains about preparation and parameter tweaking, in contrast to
the gradient boosting that is being used in other statistical learning algorithms that typically
banking sector for financial risk management.
Keywords: Artificial Neural Network, Gradient Boosting, give equal accuracy (e.g., neural networks and
Random Forest Tree, Data Mining, Artificial Intelligence, support vector machines). The method is
Algorithms. particularly resistant to dirty data and may be used
to solve classification or regression problems with
1) INTRODUCTION a wide range of response distributions. The
squared-error and exponential error loss functions
T his article explains two different algorithms
that are being used in this research paper and
the problem that they are going to be used for and
are two plausible loss functions that are commonly
used in regression and classification problems,
respectively (Guelman, 2012).
the solution for that problem. The two algorithms Other loss functions, on the other hand, may be
are used in this research paper. The first algorithm more appropriate in some situations. When the
is Artificial Neural Network and the second Bayes error rate is not close to zero or the target
algorithm is Gradient Boosting. Artificial neural classes are mislabeled, binomial deviance, for
networks (ANNs) are massively parallel-distributed example, is substantially more robust than
computers that have a natural inclination for storing exponential loss under noisy situations. Similarly,
experience and learning new things and putting it when the error distribution is long-tailed or when
to use. The solution is straightforward. The there are "outliers" in the data, the performance of
paradigm's innovative structure is a key component squared error drops considerably. In such
of it. a data storage and processing system It is instances, other functions, such as absolute error or
made up of a large number of closely connected Huber loss, are more suited. For these alternative
processing components (neurons) that work loss function settings and a specific weak learner.
together to solve a problem. ANNs learn by The gradient boosting strategy is applicable to any
practice. An ANN is tailored for a specific goal, differentiable loss function and solves the issue in
such as pattern recognition or data classification, two phases. The first step estimates using least-
through a learning process. Learning in biological squares to fit a weak learner h(x;a) to the loss
systems involves modifications to the synaptic function's negative gradient (i.e., the "pseudo-
connections that connect neurons. Furthermore, residuals"). In the second step, the optimal value of
current research reveals that an ANN is employed in bt given h is computed (x;at) (Guelman, 2012).
a number of industries; however, none of them, to
our knowledge, use an ANN to predict customer
behavior in banks. This gives us motivation to keep 2) AIMS AND OBJECTIVES
working in this sector (Ogwueleka, Misra, Colomo- The goal is to utilize an AI system to tackle a
Palacios and Fernandez, 2012). Gradient Boosting banking sector problem. The goal is to use the
(GB) is an iterative procedure that results in a
highly accurate prediction rule by merging basic Artificial Neural Network algorithm and the
parameterized functions with "poor" performance Gradient Boosting algorithm to tackle the following
(high prediction error). GB offers interpretable problem.
findings with minimum data preparation and
3

2.1) Problem particularly computers (Hall, Muljawan, Suprayogi,


By calling a customer personally, an item is given & Moorena, 2009).
away as part of a campaign. As part of an The methods for extracting rules from a neural
advertising campaign plan, all enrolled customers
are contacted. In any event, such a rapid approach network differ. Steiner et al. (2006) use the
is useless because the vast majority of called NeuroRule extraction approach to analyses a credit
customers are uninterested in the things offered. A risk data collection. Angelini et al. (2008)
huge number of calls will be squandered under demonstrate the utility of neural networks in credit
these conditions, adding to the mission's woes. risk applications, particularly as black-box non-
However, a few customers may not be on the call
list, and the mission supervisor may overlook linear systems to be utilized in conjunction with
prospective buyers. traditional grading and selection methods. In order
2.2) Solution to anticipate credit risk in banks, a mix of
A Guide Marketing Campaign's purpose is to reach methodologies is sometimes utilized. Pacelli and
the projected number of consumers with registered Azzollini (2011) believe that neural network models
profiles by phoning them and using their combined with linear approaches have provided
information to direct them to the things being
offered. We'll run numerous experiments using further support. The use of contemporary indicators
artificial neural networks and gradient boosting to in addition to classic financial ratio indicators
analyses the data and get the best results. We'll improves prediction accuracy significantly (Atiya,
receive the greatest results and solve our problem 2001).
quickly if we use these methods.
4) METHOD
3) RELATED WORK
Artificial Neural Networks, Gradient Boosting,
There have been a lot of improvements done in this Bayesian classification, rule-based techniques,
sector but still it needs to be improved. In this paper memory-based learning, and support vector
the previous work is considered as the benchmark. machines are all examples of classification
3.1) Literature Review algorithms that are extensively utilized in a number
In contrast to market-wide competition metrics like of applications and have shown their accuracy and
the Herfindahl index and concentration ratios, Boyd correctness (Wu et al., 2014). Artificial Neural
and Runkle (1993) connected signs of bank failure Network and Gradient Boosting are two of the
to bank size. Because the size variable is thought to techniques being developed.
be related to market power, the results of these 4.1) Artificial Neural Network
research are at the very least indicative of bank The neural network model is constructed in such a
failure. They discovered that the size of the bank way that when a collection of inputs is applied, the
had no bearing on the likelihood of collapse desired set of outputs is generated. By feeding it
(Carmona, Climent and Momparler, 2019). patterns and allowing it to alter its weights
depending on a learning rule, the neural network is
Credit rating is a technical aspect in determining educated. The decision-making process of the
credit risk (Khashman, 2010). The purpose of credit ANN is divided into two stages: learning and
rating is to divide applicants into two groups: those reasoning. Before making a decision, the ANN is
with excellent credit and those with bad credit trained using previous evaluation instances.
(Ghodselahi & Amirmadhi, 2011). A type of Backpropagation is used in the algorithm
(Ogwueleka, Misra, Colomo-Palacios and
universal approximation is multilayer feedforward Fernandez, 2012). ANNs are information
networks (Hornik, Stinchcombe, & White, 1989). processing systems whose structure and function
One of the most fundamental and crucial elements are inspired by the cognitive processes and
of ANN models is the idea of "universal organizational structure of neurobiological
approximation." These models offer a high level of systems. The basic components of these networks
are neurons, which are densely coupled processing
predictability (Steiner, Neto, Soma, Shimizu, & dements that work independently in parallel.
Nievola, 2006). This means that the networks may Neural pathways are used to transfer messages
adapt to random and unknown functional forms from one neuron to another, and their strength
with a degree of accuracy that can be set freely. varies. These neurons store information and build
Because of their excellent pattern detection, non- meaningful patterns by strengthening their
interactions. A neuron fires and sends the stimulus
parametric models are well-known in many fields,
4

to nearby neurons when it receives a certain are first classified. Geographic location, product
number of impulses and their aggregate surpasses price variety, overall price level, overall quality of
a certain threshold value (Witkowska, 1999).
Artificial neural networks are commonly used to items, time of waiting for payment, service attitude,
handle pattern classification problems (ANNs). A and shopping environment were all included as
case in point is bankruptcy prediction, in which input factors to the network. For the output variable,
neural networks are used to identify whether or not there is only one possible measurement value.
a corporation is insolvent. Evaluation of the firm's (2012) Ogwueleka, Misra, Colomo-Palacios, and
creditworthiness is related to the bankruptcy
prediction issue since the bank makes the Fernandez ANNs are commonly used because of
decision after researching the firm's financial their modelling capabilities. ANN is a versatile,
situation (Witkowska, 1999). Throughout the non-parametric technique based on biological
1990s, Poland saw a surge in the number of neural circuits. The multilayer perceptron is an
commercial banks. These banks analyses and effective classification tool that is widely used in
determine a company's creditworthiness using
several forms of analysis. practice.
In ANN models, the input and output parameters Neural networks are used by banks to analyses
are the same as in linear models. The three basic credit and loan applications and determine the
components of these models are the input data possibility of default. ANN approaches for default
layer, hidden layer(s), and output measure(s) layer. prediction, which first appeared in 1990, generally
Each of these levels has nodes, which are connected
to nodes in nearby layers (s). The transformation used multilayer networks and achieved higher Type
function and the weighted summation functions I and Type II classification accuracy than MDA.
may both be found in the hidden layer (s). Both of When MDA, LR, and decision trees were used to
these functions link the data values in the input to forecast bank and thrift failure, the multilayer
the measurements in the output. The weighted network fared the best. The bank failure prediction
summation function is often used in a
feedforward/back propagation neural network based on KOHONEN's self-organizing maps was
model (Nazari and Alidadi, 2022). accurate (Li and Ma, 2010).
4.1.1) ANN Experiment 1
An artificial neural (ANN) is a piece of man-made
brainpower designed to mimic the functioning of
the human mind. ANNs are made up of caring units,
which include information sources and yields. The
data sources are what the ANN uses to provide the
best yield.
We altered the values of the following parameters to
acquire the best outcome while designing a solution
with the ANN Algorithm.
Figure 1: A simplified neural network model  Probability
I used to be
A simple neural network is shown in Figure 1. A
neural network is an interconnected collection of  Number of hidden neurons
nodes comprised of neurons in an input layer, one A multilayer network is one that incorporates buried
or more hidden layers, and an output layer, as neurons. In general, the number of hidden neurons,
shown in Figure 1. Nh, is proportional to the degree of complexity of
The input pattern is internally represented by the the class boundaries. As the class boundaries get
middle layer, which is placed in the center of the more intricate, the number of buried neurons should
input layer. Patterns are recognized at this layer by be increased. To avoid the problem of overfitting,
distilling the input's attributes and then passing the number of hidden neurons, Nh, should be
them on to the output layer. The output layer's maintained as low as possible.
choice on categorization for the input patterns is the During the studies, it was experimented with
result of the output layer. As a result, the distilling various hidden neuron values in order to get the
character layer is referred to as the intermediate highest level of accuracy. The number of neurons
layer. In this study, the input and output variables
5

was set to 2 in the parameter since it was obtaining 4.1.4) PCA


more accurate results during the experiments. First and foremost, we aim to create an informative
collection that provides critical information in a
machine-readable style. Loading and reviewing Iris
data benchmark are two more advancements. An
Figure 2: Experiment 1 ANN model is trained and tested. The data includes
estimates collected from 150 iris flowers of three
 Number of iterations species, as represented in the iris data benchmark.
The number of iterations is the number of times the
data sets in the training are repeated. Using three
iterations, it was able to get good outcomes during
the experiments. It used variable rep with the value
3 to indicate the iterations.

Figure 3: Experiment 1 Number of Iterations

4.1.2) Comparison between the confusion


matrix and error rates with Random Figure 7: PCA iris Values
Forest Tree
For the time being, the data loading and training are
complete. The neural network must now be trained
using one of four different input variables. The
variable size is set to 5 and SoftMax is set to TRUE.
The maximum number of iterations is 100.

Figure 4: Difference between ANN and RFT Error Rates Figure 8: PCA Plot

4.1.3) ANN Experiment 2 4.1.5) PCA on ANN


Values used for the experiment are shown below in PCA on training and validation sets is suggested for
the figure 5. more precise results. PCA is a broad-based starting
construction that is usually lost in subtitles.
Plot before performing PCA on ANN
Figure 5: Experiment 2 Input Values

Figure 6: Probability of Experiment 2


6

on in this paper, gradient boosting trees will be


referred to as "Gradient Boosting" (Guelman,
2012). Extreme Gradient Boosting, which can
improve the accuracy of bank failure predictions
and help banks avoid financial crises (Carmona,
Climent and Momparler, 2019). To forecast bank
failure, Extreme Gradient Boosting (XG Boost) is a
machine learning algorithm. Single decision trees
provide a lot of advantages, but too much gradient
boosting leads to a model with hundreds or
thousands of trees. The final model is difficult to
Figure 9: Plot before PCA on ANN
grasp because of its complexity. XGBoost, on the
other hand, does not require the same level of care
4.2) Gradient Boosting as a black box. Although XGBoost models are
Gradient boosting approach solves the problem in complex, they may be summarized in ways that
two steps and is applicable to any differentiable loss provide deep insight, and their predictive capability
function. The first stage estimates by fitting a weak outperforms most standard techniques. XGBoost is
learner h(x;a) to the loss function's negative a newly updated distributed gradient boosting
gradient (i.e., the "pseudo-residuals") using least- library. XGBoost
squares. The ideal value of bt given h is calculated (http://dmlc.cs.washington.edu/xgboost.html,
in the second phase (x;at) (Guelman, 2012). accessed January 2018) was created by Tianqi
Chen, a PhD student at the University of
Washington. After winning many events, XGBoost
became well-known in the machine learning
community (Carmona, Climent and Momparler,
2019).
4.2.1) Gradient Boosting Experiment 1
Gradient boosting is a machine learning approach
that may be used to address problems like
regression and classification. It gives back a
prediction model in the form of an ensemble of
weak prediction models, which are frequently
decision trees. A gaussian loss function is used in a
Figure 10: Algorithm for Gradient Boosting gradient boosted model. There were 10,000
iterations. There were 13 predictors, and 13 of them
Because the negative gradient in line 3 is merely the had a non-zero impact.
residuals for squared-error loss, the approach is
reduced to regular least-squares boosting in this
case. With absolute error loss, the residuals have a
negative gradient. Regardless of the loss function,
Line 4 applies least-squares.
Although boosting isn't restricted to trees, we'll
focus on the case in which the weak learners
represent a "small" regression tree because they've
been proved to be a good representation for weak
learners h(x;a) in the context of boosting. The
process described above is referred to as gradient
boosting trees in this case, and the parameters Figure 11: Gradient Boosting Influence
reflect the split variables, their split values, and the
fitted values at each tree's terminal node. From here
7

4.3) Difference between ANN and Gradient ii. Lots of versatility - can optimize on several
Boosting loss functions and offers numerous hyper parameter
Boosting is a type of feature choice. Consider adjustment possibilities, making the function fit
Adaboost, which is utilized in the Viola-Jones face quite adaptable.
detector. There are many characteristics at first, but iii. There is no need for data pre-processing; it
include them all in the final classifier might be frequently works well with category and numerical
time-consuming and inaccurate. Thus, boosting information as is.
picks features by first locating the best features and iv. Imputation is not necessary to handle
then increasing the weights of misclassified data missing data.
points such that the following round of feature (Gradient Boosting for Classification |
selection concentrates on them. The feature Paperspace Blog, n.d.)
detectors in this case are regarded as weak 4.5) Dis Advantages of ANN and Gradient
classifiers. Boosting
By linearly combining weak classifiers, a strong Dis-Advantages of ANN
classifier may be generated. Neural networks, on i. Because neural networks are a "black box,"
the other hand, are more difficult to understand and their capacity to clearly identify possible
train. It is difficult to establish the importance of causal linkages is restricted.
characteristics for example, in ANNs. As a result, ii. Neural network models may be more
neural networks are similar to learning black boxes. challenging to utilize in the field since they
It is difficult to extract meaningful simpler data need more processing resources.
models from them once they have learned. They iii. Neural network models are prone to
also learn significantly differently from boosting in overfitting. The creation of neural network
that they use a gradient-based learning method to models is empirical, and many
try to minimize a cost function on all training sets. methodological concerns have to be
The behavior of the total system is determined by resolved.
the learned weights and biases, and it is difficult to from (Tu, 1996)
derive meaningful models from such Dis-Advantages of Gradient Boosting
representations ( What Is the Difference between i. Gradient Boosting  Models will continue to
Boosting and Artificial Neural Networks in Data improve in order to reduce all mistakes. This
Mining? - Quora, n.d.). can lead to overfitting by exaggerating
4.4) Advantages of ANN and Gradient outliers.
Boosting ii. Computationally costly - frequently
Except for the differences between them they have necessitates a large number of trees (>1000),
some advantages as well which can be time and memory consuming.
Advantages of Artificial Neural Network iii. Because of the approach's considerable
i. The development of neural network models flexibility, there are numerous parameters
necessitates less formal statistical training. that interact and substantially impact its
ii. Complex nonlinear interactions between behavior (number of iterations, tree depth,
independent and dependent variables can be regularization parameters, etc.). During
detected implicitly by neural network tuning, this necessitates a huge grid search.
models. iv. Less interpretive in nature, yet this may be
iii. All conceivable interactions between easily remedied with a variety of methods.
predictor variables may be detected using (Gradient Boosting for Classification |
neural network models. Paperspace Blog, n.d.)
iv. Various training techniques may be used to
build neural networks.
from (Tu, 1996)
Advantages of Gradient Boosting
i. Frequently gives exceptional forecasting
accuracy.
8

5) RESULT AND DISCUSSION


Results of all the experiments done above are
discussed here.
5.1) Artificial Neural Network
Result for Experiment 1 for ANN is given below in
the figures 12, 13 and 14 :

Figure 14: Third Iteration Results

During this Experiment 2 for ANN, different values


were tried as an argument for above variable the
results are shown below in figure 15, 16, 17, 18, 19:

Figure 12: First Iteration Results

Figure 15: Result for Experiment 2

Figure 13: Second Iteration Results

Figure 16: Experiment 2 steps


9

Figure 17: Experiment 2 steps Figure 19: Experiment 2 steps

5.2) Performing PCA on ANN


Plot after performing PCA on ANN

Figure 20: Plot after performing PCA on ANN

The objective was to lower both statistical error


estimates. Because a larger number of neurons
works better on sophisticated and straightforward
non-detachable input, various hidden neurons
Figure 18: Experiment 2 steps improve classifier operation.
 The three iris species are represented by the
color red, green, and blue.
5.3) PCA on ANN Experiment 1
To do PCA on ANN in experiment 1, it created
indexes for training (50%) and validation (50%)
data. The following experiment yielded the
following results.
10

implement ANN. Several factors, including the kind


of dataset, determine the method that is used. We
also searched for unusual patterns in the data that
may alter the model's conclusions in these tests. In
our trials, we found that boosting resulted in an
effective decrease in prediction error. In future
studies, alternative strategies will be used to reach
100 percent accuracy in Artificial Neural Network
and Gradient Boosting.

Figure 21: ANN Experiment 1 before PCA


7) REFRENCES
Ogwueleka, F.N., Misra, S., Colomo‐Palacios, R. and
Fernandez, L., 2015. Neural network and classification approach
in identifying customer behavior in the banking sector: A case
study of an international bank. Human factors and ergonomics
in manufacturing & service industries, 25(1), pp.28-42.
Witkowska, D., 1999. Applying artificial neural networks to
bank-decision simulations. International Advances in Economic
Research, 5(3), pp.350-368.
Nazari, M. and Alidadi, M., 2013. Measuring credit risk of bank
customers using artificial neural network. Journal of
Management Research, 5(2), p.17.
Khemakhem, S., Said, F.B. and Boujelbene, Y., 2018. Credit
risk assessment for unbalanced datasets based on data mining,
Figure 22: ANN Experiment 1 after PCA artificial neural network and support vector machines. Journal
of Modelling in Management.
Li, Y. and Ma, W., 2010, October. Applications of artificial
5.4) Gradient Boosting Experiment 1 neural networks in financial economics: a survey. In 2010
The result of gradient boosting experiment is given International symposium on computational intelligence and
in the figure below: design (Vol. 1, pp. 211-214). IEEE.
Guelman, L., 2012. Gradient boosting trees for auto insurance
loss cost modeling and prediction. Expert Systems with
Applications, 39(3), pp.3659-3667.
Carmona, P., Climent, F. and Momparler, A., 2019. Predicting
failure in the US banking sector: An extreme gradient boosting
approach. International Review of Economics & Finance, 61,
pp.304-323.
Khashman, A. (2010). Neural networks for credit risk
evaluation: Investigation of different
neural models and learning schemes. Expert Syst. Appl., 37(9),
6233-6239. http://dx.doi.org/10.1016/j.eswa.2010.02.101
Ghodselahi, A., & Amirmadhi, A. (2011). Application of
Artificial Intelligence Techniques
Figure 23: Result of Gradient Boosting Experiment 1 for Credit Risk Evaluation. International Journal of Modeling
and Optimization, 1(3), 243-249.
http://dx.doi.org/10.7763/IJMO.2011.V1.43
6) CONCLUSION AND FUTURE WORK Hornik, K., Stinchcombe, M., & White, H. (1989). Multilayer
feedforward networks are universal approximators. Neural
To summaries, all of the objectives were completed Networks, 2(5), 359-366. http://dx.doi.org/10.1016/0893-
in order to save money for the bank. Based on the 6080(89)90020-8
criteria mentioned above, the solution was created Steiner, M. T. A., Neto, P. J. S., Soma, N. Y., Shimizu, T., &
utilizing two unique data mining algorithms. The Nievola, J. C. (2006). Using Neural Network Rule Extraction for
Credit-Risk Evaluation. International Journal of Computer
two approaches are Artificial Neural Network and Science and Network Security, 6(5A), 6-16.
Gradient Boosting. The approach used is Hall, M. J. B., Muljawan, D., Suprayogi, & Moorena, L. (2009).
determined by the false positive and false negative Using the artificial neural network to assess bank credit risk: a
statistics. The accuracy of ANN is 90% and the case study of Indonesia. Applied Financial Economics, 19(22),
1825-1846. http://dx.doi.org/10.1080/09603100903018760
error rate is 9%. It is simpler to develop RF than to
11

Steiner, M. T. A., Neto, P. J. S., Soma, N. Y., Shimizu, T., &


Nievola, J. C. (2006). Using Neural Network Rule Extraction for
Credit-Risk Evaluation. International Journal of Computer
Science and Network Security, 6(5A), 6-16.
Angelini, E., di Tollo, G., & Roli, A. (2008). A neural network
approach for credit risk evaluation. The Quarterly Review of
Economics and Finance, 48(4), 733-755.
http://dx.doi.org/10.1016/j.qref.2007.04.001
Pacelli, V., & Azzollini, M. (2011). An Artificial Neural
Network Approach for Credit Risk Management. Journal of
Intelligent Learning Systems and Applications, 3(2), 103-112.
http://dx.doi.org/10.4236/jilsa.2011.32012
Atiya, A. F. (2001). Bankruptcy prediction for credit risk using
neural networks: A survey and new results. IEEE Transactions
on Neural Networks, 12(4), 929-935.
http://dx.doi.org/10.1109/72.935101
Wu, X. et al. (2014) “Data mining with big data,” IEEE
Transactions on Knowledge and Data Engineering, 26(1), pp.
97–107. doi:10.1109/TKDE.2013.109.
What is the difference between boosting and artificial neural
networks in data mining? - Quora. (n.d.). Retrieved January 18,
2022, from https://www.quora.com/What-is-the-difference-
between-boosting-and-artificial-neural-networks-in-data-mining
Gradient Boosting for Classification | Paperspace Blog. (n.d.).
Retrieved January 18, 2022, from
https://blog.paperspace.com/gradient-boosting-for-classification/
Tu, J. v. (1996). Advantages and disadvantages of using
artificial neural networks versus logistic regression for
predicting medical outcomes. Journal of Clinical Epidemiology,
49(11), 1225–1231. https://doi.org/10.1016/S0895-
4356(96)00002-9

You might also like