You are on page 1of 5

Childhood Medulloblastoma Classification

Using EfficientNets

Chandra Mohan Bhuma


ECE Department
Bapatla Engineering College
Bapatla, India
2020 IEEE Bombay Section Signature Conference (IBSSC) | 978-1-7281-8993-2/20/$31.00 ©2020 IEEE | DOI: 10.1109/IBSSC51096.2020.9332175

chandrabhuma@gmail.com

Ramanjaneyulu Kongara
ECE Department
PVP Siddhartha Institute of
Technology
Vijayawada, India
kongara.raman@gmail.com

Abstract—In this work, a deep learning methodology for In general, histological slides are analysed using
accurate classification of histological images of the patients architectural level and cell level. The magnification factor
suffering from childhood medulloblastoma is proposed. Pre for the cell level is higher than architecture level. More
trained EfficientNets trained on the ImageNet dataset are details can be obtained from the cell level inspection. As the
considered in this work. Features are extracted from the
problem is basically a classification problem, the task may
average pooling layer of EfficienNets and are given to an error
correcting output code classifier. Ensemble prediction from the be just to categorize the tumour as normal (benign) or
selected pre-trained EfficientNets is employed. For the multi abnormal (malignant). This is a binary classification
class classification, the proposed approach is able to predict problem. However, in multiclass classification, the
with a mean classification accuracy of 98.78% for 10x level requirement is to identify the sub type of malignancy. There
images and 95.67% for the 100x level images for an 80% train are four subtypes in CMB i.e., Nodular, Classic, Large and
and 20% test split. The peak classification accuracy is 100% Desmoplastic.
for both binary and multiclass case at cell level and
architectural level. For the binary classification with same CAD techniques are popular in histological images
split, 100% mean classification accuracy is achieved even classification Kather et al. [2]. Categorizing MB into
without ensemble prediction. The results are compared with
anaplastic and non-anaplastic was done by Lai et al. [3], Roa
an existing work on a similar dataset and the significant
et al. [4] and Galaro et al. [5]. Das et al. [1] used texture,
improvement is demonstrated with the experimental
simulations. color and morphological features to identify the sub type of
childhood MB. Das et al. [6] analysed the statistical
Keywords— Childhood Medulloblastoma, Classification, significance of CMB and its sub types using MANOVA
Efficientnets, Ensemble Learning, ECOC classifier (Multivariate analysis of variance).

I. INTRODUCTION Instead of using traditional feature engineering, deep


learning architectures can be used for accurately classifying
Childhood Medulloblastoma (CMB) is a type of the childhood MB subtypes. In recent past, deep learning
malignant brain tumor developed in the cerebellum of architectures i.e., Convolutional Neural Network (CNN) and
children [1]. Typical symptoms of medulloblastoma include its variants are able to dominate not only in the classification
headache, vomiting, and problems with vision. By using problems but also in object detection, segmentation, and
MRI (Magnetic Resonance Imaging) and CT (Computerized regression. Inspired by the works of Das et al., in the
Tomography) scans, the brain tumor can be detected. In classification childhood MB, a deep learning approach which
addition to chemotherapy, medulloblastoma can be treated by improves the classification accuracy in identifying the
using surgery, or radiation therapy. With correct diagnosis childhood MB subtype is presented. EfficientNets are very
and proper treatment, 90% of the medulloblastoma cases can recent CNN architectures which showed improved
be cured. performance on ‘ImageNet’ dataset. Hence, features
The histological samples from the biopsy can be used to extracted from the pre-trained EfficientNets and the
identify MB sub types. Visual inspection is required and predictions from their ensembles are considered in this work.
expertise is also needed in the identification of type of MB This paper is organized as follows. A brief review of
sub type from the histological images. As the treatment EfficientNets is presented in section II. Ensemble learning
largely depends on the type of tumour, identification of the and various techniques in ensemble learning are presented in
type of tumour plays a key role. In this regard, CAD section III. Details of the CMB database and the proposed
(Computer Assisted Diagnosis) plays an important role.

64on June 18,2022 at 05:50:12 UTC from IEEE Xplore. Restrictions apply.
Authorized licensed use limited to: Guangdong Univ of Tech. Downloaded
algorithm are given in section IV. Experimental results are EfficientNetB2 342 260x260 1408
given in section V. Conclusions are given in section VI. EfficientNetB3 387 300x300 1536

II. EFFICIENTNETS EfficientNetB4 477 380x380 1792

EfficientNets are scaled Convolutional neural networks EfficientNetB5 579 456x456 2048
(CNNs). Number of layers in CNNs are initially fixed. They EfficientNetB6 669 528x528 2304
are scaled up by increasing the number of layers and deeper
features can be extracted from the images. In general, EfficientNetB7 816 600x600 2560
improvement in classification accuracy is expected with the
increase in number of layers. ResNet architecture was tested III. ENSEMBLE LEARNING
on ‘ImageNet’ dataset with 18 layers to 200 layers. Densenet Ensemble learning methods are the methods where in,
architecture was tested on 121, 169 and 201 layers. Model various machine learning models are combined and the
scaling offers flexibility and also compatibility with the prediction is based on the combined model. Hence, it is an
devices on which the model is deployed. This scaling can be algorithm based on Meta learning. The combined predictions
done in depth wise, width wise, and resolution wise. Instead are more accurate than the single model prediction. The key
of randomly increasing and assigning the resolutions, a factor in the ensemble methods is the selection of base
systematic study and approach was given by M.Tan and Q V models. In addition, individual base models must be accurate
Le. [7]. Arbitrary choosing of width, depth and resolutions is in terms of classification accuracy. They should have
time consuming and significant improvement is not diversity either in the model architecture, or in the type of
guaranteed. Scaling coefficients were used in systematically classifier, or in the hyper parameters of the models [8]. If
scaling the baseline CNNs in the work of [7]. They diversity is maintained amongst the models, the way the data
developed [7] EfficientNets, which offered superior is visible to the model is different. Hence, if one model fails
accuracies on ‘ImageNet’ dataset. Balancing all the to understand the nature of the image, at least the other model
dimensions is the key in EfficientNets. EfficientNet might do better. There are many ensemble techniques. The
architecture uses mobile inverted bottleneck convolution simple approaches are majority voting (hard voting),
(MBConv). With this scaling approach, EfficientNets are averaging and weighed averaging the posterior probabilities.
available from EfficientNetB0 to EfficientNetB7 with In majority voting, the class labels from each model or
varying input resolutions. Some part of basic architecture of classifier are computed. From the number of models chosen,
an EfficientNetB0 is shown below in Fig 1. the class label having highest votes is considered as the final
prediction.
224x224x3 Assume that there are N models or classifiers and C
classes. Decision of a model is chosen
Conv3x3
as d n,c  {0,1} , n 1,...N , c 1,.....C, . When a classifier or a
112x112x32
model predicts the labels of the test data, it is assigned [9]
MBConv1,3x3 with a value 1. Choose a class c * if,
112x112x16 N N

MBConv6,3x3 ¦
n 1
d n,c* max c ¦d
n 1
n,c

56x56x24
Different variants have been attempted [9] in the voting
MBConv6,3x3 classification. In unanimous voting, a class label is assigned
56x56x24 if all the classifiers agree unanimously. In simple majority
case, if class labels are predicted by at least one more than
MBConv6,5x5 half the number of classifiers. In this work, the strategy
28x28x40 chosen is that, a class having highest number of votes is the
winner. In addition, posterior probabilities can also be
MBConv6,5x5
calculated from the classifiers and the sum of the
probabilities of each class is calculated. Class having the
highest sum probability across all the models is considered as
Fig. 1 A sub section of baseline EfficientNetB0 architecture
the final class. Equal weightage is given for each class
Number of layers varies from EfficientNetB0 to B7 and probability. However, in weighed averaging, each assigned
the input resolutions are also different. All the images must probability is weighed with a fraction and total sum is
be resized to these resolutions. Features are extracted from computed. In addition to the techniques discussed above,
an average pooling layer which is the last pooling layer. The Bagging, Boosting, Blending, and Stacking are some popular
feature vector sizes are given below. It is as low as 1280 and ensembling techniques.
goes up to 2560. EfficientNetB7 is approximately 3.5 times
deeper than EfficientNetB0. Some architectural details are IV. DATASET AND THE PROPOSED ALGORITHM
given below in TABLE I. The database of the Childhood Medulloblastoma [10]
TABLE I BRIEF ARCHITECTURAL DETAILS OF EFFICIENTNETS consists of two folders having the images acquired at
architecture level and cell level. The database images have
Description No.of layers Input FeatureVector normal tissue samples and CMB cell samples. There are
Resolution Size only two images in the Nodular category at cell level. For
EfficientNetB0 240 224x224 1280
10x and 100x, there are 5 classes. It is a multiclass
classification problem. The classes are imbalanced. When
EfficientNetB1 342 240x240 1280 we want to classify them as normal and abnormal category,

65on June 18,2022 at 05:50:12 UTC from IEEE Xplore. Restrictions apply.
Authorized licensed use limited to: Guangdong Univ of Tech. Downloaded
there are 152 images in the abnormal and 50 in the normal 4. Three EffcientNets are selected and the trained
class for 10x. In 100x level, there are 25 images in normal features from each EfficientNet are given
and 128 abnormal classes. This is dealt as a binary independently to a multiclass ECOC (Error
classification problem. All the images in the database are in Correcting Output Code) classifier. This is the
.jpg format and are 24 bit color images. Number of images training phase. Now three models are available.
in each class are depicted in the below TABLE II. Sample 5. Similarly test image features are extracted from
images from each class are shown in Fig 2. As the EfficientNets.
EfficientNets have different input resolutions, pre- 6. By applying the test features, class labels are
processing step is necessary while extracting the features obtained from the ECOC classifier.
from EfficientNets. 7. A majority voting decision is taken from all the
TABLE II CHILDHOOD MEDULLOBLASTOMA DATASET DETAILS three models for each test image feature as
described in section III. Final class labels are
Description 10x 100x predicted from this majority voting.
Classic 59 82 Classification accuracy for the test set is
Desmoplastic 42 20 calculated.
Nodular 23 02 8. Steps 2 to 7 are repeated 10 times by randomly
Large cell 28 24
Normal 50 25
shuffling the train and test and mean
Total 202 153 classification accuracy is calculated.

The proposed algorithm is shown in in Fig 3. Selection


Class 10x 100x
of three EfficientNets, base classifiers in ECOC, and hyper
parameter selection in step 4 is purely arbitrary. However,
proper selection of appropriate three EfficientNets is done to
Classic improve the mean classification accuracy for 10 repeats of
train and test splits. They are represented as Efx, Efy and
Efz in Fig. 3. The entire process is done for the binary
classification task also. In Fig. 3, Ef0 to Ef7 indicates
EfficientNetB0 to EfficientNetB7.
Image Database
Desmoplastic

Train/Test
Split

LargeCell
Ef0 Ef1 Ef2 Ef3 1 Ef4 Ef5 Ef6 Ef7

…… ……

Feature Extraction
Nodular Last Average
Selection of Pooling Layer
EfficientNets

Efx features Efy features Efz features

Normal

ECOC Classifier ECOC Classifier ECOC Classifier

Fig. 2 Sample images from each class of the CMB dataset [10]

The proposed algorithm consists of the following steps. Ensemble


10 times repeat Prediction
1. All the images are resized to the resolutions of
the EfficientNets.
2. The dataset is split into train and test with 80% Mean Classification
train and 20% ratio. Accuracy
3. Features are extracted from the last pooling
layer (average pooling layer) of the Efficient
Fig. 3 Proposed Algorithm
Nets

66on June 18,2022 at 05:50:12 UTC from IEEE Xplore. Restrictions apply.
Authorized licensed use limited to: Guangdong Univ of Tech. Downloaded
V. EXPERIMENTAL SIMULATIONS AND RESULTS binarycomplete

All the simulations for extracting the features from the Distance=seuclidean
EfficientNetB7 KNN 89.67
EfficientNets are carried using Keras 2.3.0 in Google Colab Coding: onevsone
Environment without using GPU. Features are calculated at
once and they are stored in .mat files in MATLAB R2019b.
Ensemble prediction is done using the shuffling, splitting the Ensemble Mean Accuracy 95.67
.mat files of EfficientNets as it is computationally expensive
to compute the features each time for all the images. There
are only two images in the 100x nodular category. Hence,
they are not included in the multiclass classification of 100x The combination of EfficientNetB3,B0,B7 worked well for
images. However, in the binary classification of 100x 10x level images, with the bases learners being SVM
images, they are included. (Support Vector Machine) offering a mean classification
accuracy of 98.78%. However, for 100x level images,
With an aim of improving the mean classification ensemble of EfficientNetB4, B0, B7 did a fair job offering
accuracy, three EfficientNets are selected at random 95.67% mean classification accuracy. When the train and
manually. Other hyper parameters tuned are test data are shuffled for each repeat, the classification
‘KernelFunction’, ‘KernelScale’, in the Support Vector accuracy varies for each network and in turn, the mean
Machine (SVM) classifier, and the ‘Distance’ metric in the accuracy also varies. For 10x level images, over 10
K Nearest Neighborhood (KNN) classifier. For the 10x iterations, the classification accuracy varies as shown in the
level and 100x level images, the details of the chosen pre- Fig. 4. In Fig. 5, the classification accuracies for the 100x
trained networks and the classifiers are given below in level images are shown.
TABLES III and IV.
TABLE III DETAILS OF THE NETWORKS AND THE CLASSIFIERS FOR 10X
LEVEL IMAGES

Architectural Level (10x) Multiclass Classification


Mean
Base Classification
Network Options Accuracy of
Classifier of each Network
ECOC
(%)
Kernel Scale=2
Linear Kernel
EfficientNetB3 SVM 96.34
Coding:
onevsone
Kernel Scale=2
Linear Kernel Fig. 4 Range of classification accuracies for the 10x images
EfficientNetB0 SVM 97.32
Coding:
onevsone
Kernel Scale=2
Linear Kernel
EfficientNetB7 SVM 95.85
Coding:
onevsone
Ensemble Mean Accuracy 98.78

TABLE IV DETAILS OF THE NETWORKS AND THE CLASSIFIERS FOR


100X LEVEL IMAGES
Cell Level (100x) Multiclass Classification
Mean
Classification
Base
Network Options Accuracy of
Classifier each Network
Fig. 5 Range of classification accuracies for the 100x images
(%)
Kernel Scale=3 From Fig. 4, it can be inferred that the ensemble prediction
assures us the combined model is more general not having
EfficientNetB4 SVM Linear Kernel 93.00 any over fit problem. Another observation from Fig. 4 is
Coding: onevsall that, for certain iterations, the classification accuracy is
Kernel Scale=3 100%. For binary classification of 10x and 100x images as
EfficientNetB0 SVM
normal and abnormal, EfficientNets are able to classify them
Linear Kernel 90.67 effectively. For 10x images, mean classification accuracy of
Coding: EfficientNetB1, B2, and B3 is 100%. However, for 100x

67on June 18,2022 at 05:50:12 UTC from IEEE Xplore. Restrictions apply.
Authorized licensed use limited to: Guangdong Univ of Tech. Downloaded
images, EfficientnetB2 and B7 are able to offer 100% mean binary classification, the classification accuracy is 100% for
classification accuracy. Hence, ensemble of the a 80% train and 20% test split even without ensemble
EfficientNets is not required for binary classification. prediction. As EfficientNets have different resolutions,
different number of layers, the improved accuracy due to
Authors of [6] have also used similar dataset but the ensemble prediction is attributed to the diversity in the
number of images is different. They have used traditional EfficientNet models chosen.
feature i.e., texture, color and shape and applied MANOVA
on the features to identify the significant features. Compared REFERENCES
to the work of [6], the proposed ensemble of EfficientNets
gives good mean classification accuracy. Various % train [1] Das D, Mahanta BL, Ahmed S, Baishya KB, Haque I, “Study on
contribution of biological interpretable and computer-aided features
and test splits and the corresponding mean classification towards the classification of childhood medulloblastoma cells”. J
accuracies obtained are shown in TABLE V. Hyper Med Syst 42:1–12, 2018
parameter tuning is not done in obtaining the results shown [2] Kather NJ, Cleo-AronWeis Bianconi M, Melcher MS, Schad RL,
in TABLE V. Even the set of EfficientNets is also not Timo-Gaiser Marx A, Zöllner FG, “Multi-class texture analysis in
colorectal cancer histology”. Sci Rep.,2016
changed. This is reflected in 90% train and 10% test split [3] Lai Y, Viswanatha S, Baccon J, Ellison D, Judkins AR, Madabhushi
case, as there is no improvement compared to 80%train and A, “A texture-based classifier to discriminate anaplastic from non-
20% test case. But for the cell level case, 90% train and 10% anaplastic medulloblastoma”. In: Proceedings of IEEE 37th Annual
test case is better than 80% train and 20% test case. Northeast Bioengineering Conference Troy, NY, USA, 2011
[4] Roa AC, Arévalo J, Judkins A, Madabhushi A, González F, “A
method for medulloblastoma tumor differentiation based on
TABLE V CLASSIFICATION ACCURACIES WITH VARIOUS TRAIN AND convolutional neural networks and transfer learning”, In:Proceedings
TEST SPLITS FOR MULTICLASS PROBLEM of 11th International Symposium on Medical Information Processing
Mean and Analysis, Ecuador, vol 9681,2015.
Magnification Train% Test% Classification [5] Galaro J, Judkins AR, Ellison D, Baccon J, Madabhushi A, “An
Accuracy (%) integrated texton and bag of words classifier for identifying anaplastic
medulloblastomas”. In: Proceedings of Annual International
80 20 98.78
Conference of the IEEE Engineering in Medicine Biology Society
10x 20 80 92.17 EMBS, Boston, USA, pp 3343–3346, 2011
Architectural 90 10 98.17
Level [6] Das, D., Mahanta, L.B., Ahmed, S. et al. A study on MANOVA as an
10 90 82.53
effective feature reduction technique in classification of childhood
50 50 95.50
medulloblastoma and its subtypes. Network Modelling Analysis in
80 20 95.67
Health Informatics and Bioinformatics 9, 16 , 2020 .
20 80 80.91 [7] Mingxing Tan and Quoc V. Le, “EfficientNet: Rethinking Model
100x
90 10 96.43 Scaling for Convolutional Neural Networks”, International
Cell Level
10 90 68.75 Conference on Machine Learning, 2019, arXiv:1905/11946
50 50 91.07 [8] Chollet Francois, “Deep Learning with Python”, Manning
Publications, 2017, ISBN-10:9781617294433
The database chosen here is a recent one [10] and hence [9] Cha Jhang, Yunqian Ma, “Ensemble Machine Learning-Methods and
Applications”, Springer Publications, ISBN-978-1-4419-9325-0.
direct comparison is not possible. However, the authors of [10] Daisy Das, Dr.Lipi B.Mahanta, “Childhood Medulloblastoma
[6] have classified the CMB subtypes and the dataset was Microscopic Images”, IEEE Dataport, 2020.
made publicly available with some modifications [10]. The
comparison is given in TABLE VI. Results reported by the
authors of [6] in TABLE VI are after feature reduction.

TABLE VI COMPARISON OF THE PROPOSED WORK WITH AN EXISTING


WORK
Authors Level Classification Accuracy
65.21% Multiclass
Das et al., [6] Architecture
100% Binary
83.6% Multiclass
Cell
92.6% Binary
98.78% Multiclass
Proposed Architecture
100% Binary
95.67% Multiclass
Cell
100% Binary

VI. CONCLUSIONS
In this work, an algorithm is proposed for multiclass
classification of histological images of Childhood
Medulloblastoma into sub types. Binary classification of
normal tissue cells and abnormal ones is also presented. For
multiclass classification, mean classification accuracies (10
repeats) achieved with the ensemble of EfficientNets is
98.78% at architectural level and 95.67% at cell level. For

68on June 18,2022 at 05:50:12 UTC from IEEE Xplore. Restrictions apply.
Authorized licensed use limited to: Guangdong Univ of Tech. Downloaded

You might also like