Professional Documents
Culture Documents
Brain Tumor Classification Using Pretrained Convolutional Neural Networks
Brain Tumor Classification Using Pretrained Convolutional Neural Networks
Abstract— Nowadays, deep learning methods have fulfilled They obtained 90.9% accuracy in glioma detection and 94.2%
great image classification tasks. Convolutional neural networks in grades calculation. A CNN deep learning model was
for classifying brain images into four classes: no tumor, glioma, introduced by Sultan et al. in [7] to classify different BT types.
meningioma, and pituitary are introduced in this paper. The They used open access datasets to classify BTs in (no tumor,
used dataset is public and has 3064 MRI brain images from 233 glioma, meningioma, and pituitary). The proposed
patients, also we have added 980 images with no tumor. A architecture obtained 96.13% accuracy. Different pretrained
comparison between pretrained AlexNet, GoogleNet, networks investigations were presented in [8] by Mehrotra et
DenseNet201, and ResNet101 has been done using our dataset. al. for enhancing BT classification in MRIs. The research
Our results show an accuracy of 98.76% obtained with fine-
group used a PT - CNN for BT classification in two classes
tuned ResNet101.
benign and malign. They achieved an accuracy of 99.04%
Keywords— brain tumor, deep learning, convolutional using a finetuned AlexNet.
neural network, pretrained network, classification, accuracy.
III. CONVOLUTIONAL NEURONAL NETWORKS
I. INTRODUCTION CNNs are a class of deep neural networks and consist of
Brain is a complex organ within the physical body, it neurons that have weights and biases. Each neuron obtains the
weighs around 1.4 kilograms, and it is working with 100 information from some inputs, then calculates a scalar product
billion neurons or nerve cells. Each brain neuron is connected, and freely pursues it with a nonlinearity. CNN architectures
through synapses, with thousands or maybe tens of thousands have images as inputs. This network design solves the image
of others neurons. A brain can form up to a million new classification task best.
connections for every second of our lives. An uncontrolled In recent years CNN is a frequently adopted architecture
division of cells generates a brain tumor (BT). In the vicinity that can execute complicated operations using convolutional
or inside the brain an abnormal cells group is created. This filters [2], [9], [10]. A CNN architecture implements a
cells group can affect the correct brain activity and destroy sequence of feedforward layers with convolutional filters and
healthy cells [1]. pooling layers. This network architecture utilizes several fully
Today, almost every clinic uses the World Health connected layers, after the last pooling layer, that convert
Organization (WHO) classification to detect BT. According to bidimensional feature maps from the previous layers into an
WHO, there are more than 120 BT types. Cell origin and unidimensional vector for classification [11]. Not requiring
behaviour help classify BTs, from the least aggressive feature extraction is a CNN architecture advantage. A CNN
(benign) to the most aggressive (malignant). development from scratch is time consuming and
complicated. It is necessary to apply a large labelled dataset
In recent years neural networks have grown very fast. The for training before the model is ready for classification. CNNs
implementation of accurate methods in training neural are using relatively little preprocessing compared to other
networks has become extremely important in diverse health algorithms for image classification tasks. The network uses
informatics areas such as medical image analysis, filters that in classic algorithms were built manually. A great
bioinformatics, and medical informatics [2]. advantage is that CNNs are independent of previous
knowledge and human effort in designing the features.
II. PREVIOUS REFFERENCES
Convolutional neural networks (CNNs) gained popularity IV. PERFORMANCE PARAMETERS
only in 2012, although they were initially introduced to the The proposed BT detection and classification framework
public in 1998 by LeCun [3]. In 2012 Krizhevsky [4] designed calculates four significant metric equations, that check the
a winning CNN for the ImageNet competition that performed classifier effectiveness [7].
considerably better than previous state of the art models. After Accuracy measures the total number of predictions a
the ImageNet contest from 2012, the CNN’s importance model gets right.
raised in image classification, and the computer vision Precision evaluates how good the model is in predicting
community adopted these network architectures. positive labels.
In [5] Paul et al. used oblique BT images for CNN training, Recall calculates the percentage of actual positives a
and they introduced new methods for detection and model identifies correctly.
classification implementing CNNs. They obtained 91.43% F1 Score is the weighted average considering precision
accuracy. In [6] Anaraki et al. suggest a network to emphasize and recall.
BTs in MRIs considering Genetic Algorithms (GA) and CNN.
Authorized licensed use limited to: Institute of Technology (Nirma University). Downloaded on October 17,2021 at 16:55:40 UTC from IEEE Xplore. Restrictions apply.
Accuracy = (TP + TN) / (TP + TN + FP + FN) (1) C. GoogleNet
GoogleNet [12] is between the first architectures that
Precision = TP / (TP + FP) (2) recommended the inception block that significantly reduced
the number of trainable parameters in the network. The
Recall = TP / (TP + FN) (3) inception module operates with a parallel combination like
a 1 × 1 convolutional filter added before the 3 × 3, and 5 ×
F1 = 2 * (Precision * Recall) / (Precision + Recall) (4) 5 convolutional blocks. Also, an alongside pooling layer is
True positive (TP) predicts a positive class that is positive. implemented for dimension reduction, as shown in Fig.3 [15].
True negative (TN) predicts a negative class that is negative. D. ResNet101
False Negatives (FN), also named Type II error, predicts a
negative class that is positive. False Positives (FP), also named ResNet architecture [14] introduced by Microsoft is
Type I error, predicts a positive class that is negative. showed in Fig.4.
The fundamental concept is to increase the number of
V. PRETRAINED NETS layers introducing a residual connection (with an identity
A. AlexNet layer). This layer is connected directly with the next layer
improving the learning process.
The AlexNet architecture is presented in Fig.1. AlexNet is
a CNN designed by Alex Krizhevsky [4], which contains eight VI. METHODOLOGY
layers after the input layer, firstly five layers are convolutional
layers and then the next three layers, last in the architecture, Our proposed methodology is based on the pretrained
are fully connected layers with 4096 neurons each. CNN’s architectures AlexNet, GoogleNet, DenseNet201, and
Resnet101 for classification. A specified classifier is detecting
B. DenseNet201 the BT in brain MRIs. This methodology is implemented in
The DenseNet architecture is shown in Fig.2 and contains MATLAB and follows the below steps:
four Dense Blocks (Dx). First an input layer, then transition Step 1: MRI brain dataset acquisition.
layers (Tx), and finally the global average pooling (GAP) Step 2: Extract the images.
layer. The transition layers have a batch normalization layer, Step 3: Resize the images to fit the CNN input layer.
a 1×1 convolutional layer, and a 2×2 average pooling layer Step 4: Import pretrained networks.
with a two stride. GAP is comparable to the classic pooling Step 5: Final layers replacement to get better precision and
approach, but GAP has a more severe feature map reduction. accuracy for classification.
In conclusion, GAP layer is minimizing the whole image into Step 6: Conduct experiments, compare, observe and
a single-digit [13]. decide which network has the best results.
Authorized licensed use limited to: Institute of Technology (Nirma University). Downloaded on October 17,2021 at 16:55:40 UTC from IEEE Xplore. Restrictions apply.
Fig. 3. GoogleNet architecture [15].
Because each pretrained network expects an adequate Google Last-2 Pretrained architecture FCL 1000 FCL four with bias two
Net Last 1000 classes layer 4 classes layer
image input size, we need to resize all these images to fit the
input layer of the pretrained network. TABLE II FINE TUNING OF PARAMETERS
C. Final Layers Preparation for Pretrained Nets Parameter Values
It is mandatory to download the necessary add-on called net [“Googlenet","Alexnet","Densenet201",” Resnet101"]
initialLearnRate [1e-4,1e-3]
Deep Learning Toolbox Model for AlexNet, GoogleNet, epochs [6,7,8]
DenseNet201, and ResNet101 and to import the pretrained miniBatchSize [8,10]
solver ["sgdm" "adam" "rmsprop"]
Authorized licensed use limited to: Institute of Technology (Nirma University). Downloaded on October 17,2021 at 16:55:40 UTC from IEEE Xplore. Restrictions apply.
Fig. 5. Training and validation results for AlexNet, GoogleNet, DenseNet201, and ResNet101.
Authorized licensed use limited to: Institute of Technology (Nirma University). Downloaded on October 17,2021 at 16:55:40 UTC from IEEE Xplore. Restrictions apply.