Professional Documents
Culture Documents
CLASSIFICATION
BY
Meherol Hasan
ID: 192-15-13234
Supervised By
Co-Supervised By
Assistant Professor
Department of CSE
Daffodil International University
BOARD OF EXAMINERS
Supervised by:
Co-Supervised by:
Assistant Professor
Department of CSE
Daffodil International University
Submitted by:
Meherol Hasan
ID: 192-15-13234
Department of CSE
Daffodil International University
First, I express my heartiest thanks and gratefulness to Almighty Allah for His divine
blessing which makes me possible to complete the final year project/internship
successfully.
I would like to express my heartiest gratitude to Dr. Shayekh Hayder Noori, Head,
Department of CSE, for his kind help to finish my project and also to other faculty
members and the staffs of CSE department of Daffodil International University.
Finally, I must acknowledge with due respect the constant support and patients of my
parents.
TABLE OF CONTENTS
Declaration iii
Acknowledgments iv
Abstract v
1.2 Motivation 2
2.5 Challenges 9
4.4 Summary 39
6.2 Conclusion 46
REFERENCES 48-50
LIST OF FIGURES
FIGURES PAGE NO
LIST OF TABLES
TABLES PAGE NO
1.1 Introduction
Stomach cancer, commonly known as colon cancer, is a prevalent and deadly disease,
ranking as the third major factor in deaths due to cancer globally [2]. Molecular research
has revealed different subtypes of colon cancer with varying clinical outcomes. One such
classification is based on microsatellite instability (MSI) status, where MSI-high colon
cancer shows a distinct response to chemotherapy and improved survival compared to
microsatellite-stable (MSS) cancer [3]. At present, pathologists are relied upon for
visually diagnosing colon cancer, which can lead to subjectivity and diagnostic errors due
to the time-consuming nature of the process [4]. However, deep learning techniques can
potentially revolutionize medical image analysis by providing automated and efficient
solutions. In recent years, the healthcare sector has conducted significant research into
deep learning (DL) technology, particularly in the field of cancer detection [5]. Deep
learning has proven to be a reliable method for analyzing medical images and accurately
classifying colon cancer MSS vs. MSI from histological images [6]. In this study, I am
examining how effective it is to use deep learning to classify colon cancer MSS and MSI
based on histological image analysis. By training and evaluating deep learning models on
a comprehensive dataset of colon cancer patients, I aim to assess their classification
performance, robustness, and generalizability through rigorous testing and validation.
This research has the potential to impact clinical practice by aiding clinicians in making
informed treatment decisions, leading to improved patient care and outcomes. By
advancing the medical image analysis field, I aspire to contribute to enhanced diagnoses,
personalized treatment approaches, and improved patient outcomes in the battle against
colon cancer. In particular, I am examining how transfer learning techniques can enhance
the proficiency of my deep learning models. Transfer learning lets me leverage pre-
trained models on large datasets to extract relevant features from histological images,
improving classification accuracy. My results demonstrate the effectiveness of transfer
learning in classifying colon cancer MSS vs. MSI, with my models achieving high
accuracy and robustness. Furthermore, I have investigated the interpretability of my
1.2 Motivation
Colon cancer (GC) is a highly heterogeneous and prevalent malignancy, ranking as the
second leading cause of cancer-related deaths worldwide and particularly prevalent in
East Asia. The disease can be classified into different molecular subtypes based on
genomic features, such as microsatellite instability (MSI) or microsatellite stability
(MSS). MSI is characterized by a high mutation rate due to DNA mismatch repair
(MMR) gene defects, while MSS is associated with chromosomal instability and a low
mutation rate. The MSI and MSS subtypes of GC exhibit distinct clinical and biological
characteristics, influencing prognosis, chemotherapy response, immune infiltration, and
tumor mutation burden. Accurate classification of GC based on MSI/MSS status is
essential for enhancing diagnosis, prognosis, and treatment strategies. However, existing
methods for MSI/MSS classification face challenges including cost, invasiveness,
variability, and low sensitivity. Although histological images of GC tissue samples are
routinely obtained, they are not extensively utilized for MSI/MSS classification. Recent
studies have demonstrated the potential of deep learning techniques in analyzing
histological images to extract relevant features and patterns for MSI/MSS classification.
Deep learning, a branch of machine learning utilizing artificial neural networks, can learn
from vast amounts of data and perform complex tasks such as image recognition, natural
language processing, and speech synthesis. In the biomedical field, deep learning has
shown successful applications in cancer detection, diagnosis, prognosis, and treatment. I
chose this topic because I have a strong interest in applying deep learning techniques to
biomedical problems, especially cancer. I have a background in computer science and
2.1 Preliminaries/Terminologies
Colon cancer is a complex disease with many variations, making it difficult to diagnose
and treat. Researchers are exploring new ways to better understand and manage this
disease, including the use of deep learning to classify colon cancer subtypes. This
approach has the potential to provide more personalized treatment options for patients.
Colon cancer is known for its heterogeneity in terms of its microenvironment, genome
instability, and oncogenic signatures. Despite this, there is still a lack of classification that
combines these features. A recent analysis by The Cancer Genome Atlas (TCGA) has
categorized colon cancer into four molecular subtypes: Epstein–Barr virus (EBV)-
positive, microsatellite instability (MSI), genomically stable, and chromosomal instable
tumors. Researchers have used deep learning to detect subtypes of colon cancer that are
sensitive to immunotherapy using histologic images. A new framework for cancer
classification, called deep cancer subtype classification (DeepCC), has also been
developed based on deep learning of functional spectra. In this chapter, I will dive into
the background of colon cancer subtype classification and explore the role of deep
learning in this field.
2.5 Challenges
This study encountered some research challenges that are described below:
a) Data Collection: One of this study's major challenges was obtaining sufficient
histological images for classifying MSI and MSS gastrointestinal cancer. In my country,
collecting this colon cancer data from any medical center was very hard, as they either
did not have the data or did not want to share it for research purposes. Therefore, I had to
look for alternative sources to gather the needed data. Online platforms such as Kaggle
was very helpful in providing me with the histological images for this research project.
Despite my difficulties and limitations, these online sources offered a rich source of colon
b) Data Quality: Another challenge of this study was ensuring the collected data's
quality for classification. Depending on the source, scanner, and stain used, histological
images may vary in quality, resolution, format, and annotation. Some images may need to
be corrected, completed, or mislabeled due to human or technical errors. These issues
may affect the performance and reliability of the deep learning model. Therefore, I had to
properly preprocess the images for classification, which involved converting the images
to a common format and size, removing noise and artifacts, enhancing contrast and
brightness, and verifying labels.
c) Select Deep Learning Approach: This research aimed to establish the most favorable
option for classifying MSI and MSS gastrointestinal cancer using histological images.
Learning through deep learning is a highly effective technique for complex patterns and
features from histological images and provides accurate and efficient classification results
[4]. However, there are many different deep learning techniques that have been proposed
for various medical image analysis tasks, such as convolutional neural networks (CNNs),
recurrent neural networks (RNNs), generative adversarial networks (GANs), and
transformers. Each technique has advantages and disadvantages regarding accuracy,
efficiency, interpretability, and generalization. Therefore, I had to compare different deep
learning techniques and select the one that best suited the task and the data.
d) Accuracy Improvement: One of the final challenges of this study is to improve the
overall performance of the chosen deep learning model and select the best model for the
task. I improved the model’s performance by adjusting hyperparameters like learning
rate, batch size, number of layers and filters, and activation function. I have also used
data augmentation techniques like rotation, flipping, cropping, and scaling and
regularization techniques like dropout, batch normalization, and weight decay.
Additionally, I incorporated domain knowledge such as clinical features or molecular
markers. Choosing the appropriate model is of utmost importance for a given task. I
confidently evaluated the model’s performance on different metrics, including accuracy,
precision, recall, F1-score, and AUC. I conducted statistical tests or used confidence
CHAPTER 3
RESEARCH METHODOLOGY
3.1 Introduction
Colon cancer is a widespread and fatal illness that causes significant rates of illness and
death worldwide. Properly identifying the subtypes of colon cancer is critical for
determining the most effective treatment options and predicting patient outcomes.
Histological analysis has recently become a valuable tool for characterizing colon tumors
and determining their microsatellite instability (MSI) status, a molecular characteristic
associated with distinct biological behaviors and clinical implications in colon cancer.
However, identifying MSI status through traditional tissue examination can be difficult
and subjective, highlighting the need for more objective and efficient methods. In my
research, we propose a new approach that uses images of tissue samples to differentiate
between MSI and microsatellite-stable (MSS) tumors in colon cancer. By harnessing the
power of deep learning and image analysis techniques, we aim to develop a robust and
accurate classification model to assist in the diagnosis and treatment of colon cancer
patients. My research process consists of four main stages: data collection, image
preparation, model selection, and result analysis. In the data collection stage, we obtain
tissue sample images from a well-curated repository and carefully select subsets of colon
cancer patients with known MSI and MSS tumors. These images serve as the foundation
for subsequent analysis. In the image preparation stage, we apply various techniques such
as resizing, color normalization, data augmentation, and noise reduction to standardize
the images and improve their quality. Next, in the model selection stage, I explored
different deep-learning models to find the most suitable one for my classification task. I
aim to use transfer learning, a powerful technique that allows me to adapt pre-trained
models to my specific problem domain. By leveraging the learned representations from
Histological Images: The primary data source for this study is a collection of
histological images obtained from the Zenodo repository. These images represent colon
cancer patients with known MSI and MSS tumors. The images are preprocessed and
labeled according to the microsatellite status of the tumors.
Image Pre-processing Tools: Image pre-processing plays a crucial role in enhancing the
quality and standardizing the input data for analysis. Various tools and techniques are
employed for tasks such as resizing the images to a standardized dimension, color
normalization to remove variations, data augmentation for increasing dataset diversity,
and noise reduction to improve image clarity.
Convolutional Neural Networks (CNNs): CNNs are a class of deep learning models
specifically designed for image analysis tasks. These networks consist of multiple layers,
including convolutional, pooling, normalization, and fully connected layers. CNNs excel
at learning hierarchical representations and extracting features from images. In this study,
CNNs are trained and fine-tuned using transfer learning, where pre-trained models are
adapted to the specific colon cancer classification task.
Transfer Learning Models: To determine the most effective transfer learning model for
the classification task, several pre-trained models are evaluated. The models considered
include VGG16, VGG19, MobileNet, MobileNetV2, InceptionV3, ResNet50,
DenseNet201, and EfficientNetB2. Each model has been previously trained on large-
scale datasets and has demonstrated high accuracy in image recognition tasks.
Evaluation Metrics: To assess the performance of the classification model, standard
evaluation metrics such as accuracy, precision, recall, and F1 score are employed. These
metrics provide quantitative measures of the model's ability to correctly classify MSI and
MSS tumors based on the histological images.
3.6.1 MoibleNet
MobileNet is a class of efficient models designed for mobile and embedded vision
applications. These models use depth-wise separable convolutions to create lightweight
deep neural networks with a streamlined architecture. The MobileNet model comes in
different sizes. The standard version has 4.2 million parameters, while smaller versions
have 1.32 million. The MobileNet model comprises 27 convolutional layers that consist
of 13 depth-wise convolutional layers, one average pool layer, one fully connected layer,
and one softmax layer [21].
3.6.2 MobileNetV2
3.6.3 VGG16
VGG16 is a deep learning model for image recognition developed in 2014 by the
University of Oxford's K. Simonyan and A. Zisserman. It has 16 layers of convolutions
[23] that can learn to recognize and categorize various objects in an image. It can also
generate captions for images, detect and segment objects, and classify images. It can also
transfer its learned features to other neural networks for different tasks. VGG16 is widely
regarded as a highly effective model for image recognition and has demonstrated
impressive performance on the ImageNet challenge, achieving a remarkably low error
rate of 7.3%.
3.6.4 VGG19
VGG19 is a deep-learning model for image recognition and classification created by
Karen Simonyan and Andrew Zisserman in 2014. It belongs to the Visual Geometry
Group (VGG) network family and is the 19th model in the series. VGG19 has 19 layers
[24] that can handle various computer vision tasks. VGG19 has a simple but effective
structure that consists of five blocks of convolutions and three layers of fully connected
neurons. The blocks of convolutions have multiple layers of convolutions with non-linear
activations, pooling layers, and batch normalization layers. A max-pooling layer with a
stride of 2 is placed after each set of convolutions. The fully connected layers have 4096,
4096, and 1000 neurons each. The result produced by the VGG19 model is a vector with
a thousand dimensions that predicts the class of an image.
3.6.5 InceptionV3
3.6.6 RestNet50
ResNet50 is a powerful and widely recognized convolutional neural network (CNN)
architecture that has made significant contributions to the field of computer vision.
Introduced by Microsoft Research, ResNet50 is known for its deep structure, enabling it
to effectively learn complex representations from images. What sets ResNet50 apart is its
use of residual connections, also known as skip connections, which alleviate the
vanishing gradient problem. By introducing these connections, the network can
efficiently propagate information from earlier layers to later layers, allowing for the
successful training of very deep models. ResNet50 consists of 50 layers, including
convolutional layers, pooling layers, fully connected layers, and shortcut connections.
The core building blocks of ResNet50 are residual blocks, which contain multiple
convolutional layers. These blocks enable the network to learn and refine increasingly
abstract features as the information passes through the layers. The skip connections in
ResNet50 enable the network to learn residual mappings, allowing for easier optimization
and improved gradient flow during training. This architectural innovation has been
instrumental in training deeper neural networks more effectively and has contributed to
breakthroughs in various computer vision tasks such as image classification, object
detection, and semantic segmentation. ResNet50's remarkable performance and accuracy
have been demonstrated in competitions such as the ImageNet challenge, where it has
achieved state-of-the-art results. Due to its strong performance and robustness, ResNet50
has become a popular choice for image recognition tasks and serves as a foundation for
many subsequent CNN architectures.
3.6.7 DenseNet201
3.6.8 EfficientNetB2
EfficientNetB2 is a convolutional neural network that was designed specifically to
achieve high accuracy and efficiency for image recognition and classification tasks. It is
part of the EfficientNet family of models developed using neural architecture search and
scaling techniques [26]. EfficientNetB2 has 9 blocks of convolutions and 3 layers of fully
connected neurons. The blocks of convolutions consist of multiple layers of depth-wise
and pointwise convolutions with non-linear activations, squeeze-and-excitation layers,
and batch normalization layers. A dropout layer and a stride of 2 max-pooling layers
follow each block of convolutions. Each connected layer has 1408, 1408, and 1000
neurons. The EfficientNetB2 model generates a 1000-dimensional vector that predicts the
class of images. EfficientNetB2 has fewer parameters and a faster training speed than
previous models, such as VGG19 and InceptionV3 [26].
with Adam optimizer where SoftMax was used as the activation function for all the
architectures shown in equation (2).
f i¿
CHAPTER 4
EXPERIMENTAL RESULTS AND DISCUSSION
TP+TN
Accuracy= (3)
TP+TN + FP+ FN
TP
Precision= (4)
TP+ FP
TP
Recall= (5)
TP+ FN
precision × recall
F 1 Score=2 × (6)
precision+recall
This study's proposed model can classify and detect colon cancer based on CNN, which
is a transfer learning (TL) model. A dataset comprising 100,570 preprocessed colon
cancer images is utilized for classification purposes. Eight CNN transfer learning models,
namely VGG16, VGG19, MobileNet, MobileNetV2, InceptionV3, EfficientNetB2,
performance in terms of accuracy, completion time, and data loss. To ensure efficient
processing and computation, all experiments are conducted on the Kaggle platform using
a dedicated GPU. The models are trained for 10-12 epochs, employing a custom callback
function called LRA, which dynamically adjusts the learning rate during training based
on the training accuracy and validation loss. The implementation of this callback function
provides significant advantages, including more efficient training and improved
performance. Through the development and evaluation of these CNN-based transfer
The table shows the comparison of different models in terms of accuracy, auc, recall,
precision, and f1-score. These metrics measure how well the models can predict the
4.4 Summary
6.2 Conclusion
This study addressed the classification of colon cancer subtypes, specifically MSS
(Microsatellite Stability) and MSIMUT (Microsatellite Instability). The classification
task was accomplished using transfer learning techniques and the EfficientNetB2 model
with pre-trained weights from ImageNet. The dataset consisted of 100,570 images, with
50285 images from the MSS class and 50285 from the MSIMUT class. My experiments
show that the suggested method is effective. The test results show that the model had an
impressive accuracy rate of 98.29%, with AUC of 99.80%, indicating excellent
discrimination power. The evaluation metrics of precision, recall, and f1-score
[1] Sinicrope, F. A., Foster, N. R., Thibodeau, S. N., Marsoni, S., Monges, G.,, "DNA mismatch repair
status and colon cancer recurrence and survival in," JNCI: Journal of the National Cancer Institute,
vol. 103, no. 11, pp. 863-875, 2011.
[2] Bray, F., Ferlay, J., Soerjomataram, I., Siegel, R. L., Torre, L. A., & Jemal, A., "Global cancer
statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185
countries," CA: A Cancer Journal for Clinicians, vol. 68, no. 6, pp. 394-424, 2018.
[3] Ratti, M., Lampis, A., Hahne, J. C., Passalacqua, R., & Valeri, N., "Microsatellite instability in colon
cancer: molecular bases, clinical perspectives, and new treatment approaches," Cellular and
Molecular Life Sciences, vol. 75, no. 22, pp. 4151-4162, 2018.
[4] Kather, J. N., Pearson, A. T., Halama, N., Jäger, D., Krause, J., Loosen, S. H., Marx, A., Boor, P.,
Tacke, F., Neumann, U. P., Grabsch, H. I., Yoshikawa, T., Brenner, H., Chang-Claude, J.,
Hoffmeister, M., Trautwein, C., & Luedde T., "Deep learning can predict microsatellite instability
directly from histology in gastrointestinal cancer," Nature Medicine, vol. 25, no. 7, pp. 1054-1056,
2019.
[6] Zhang, Y., Wang, X., Liu, J., & Liang, C., "Deep learning for colon cancer classification based on
histopathological images," BMC Bioinformatics, vol. 21, no. 1, p. 546, 2020.
[7] Cristescu, R., Lee, J., Nebozhyn, M., Kim, K.-M., Ting, J. C., Wong, S. S., Liu, J., Yue, Y. G., Wang,
J., Yu, K., Ye, X.-S., Do I.-G., Liu, S., Gong, L., Fu, J., Jin, J.-G., Choi, M.-G., Sohn T.S., Lee J.H.,
Bae J.M., Kim S.T., Park S.H., Sohn I., Jung S, "Molecular analysis of colon cancer identifies
subtypes associated with distinct clinical outcomes," Nature Medicine, vol. 21, no. 5, pp. 449-456,
2015.
[8] Iizuka O., Horie H., Morikawa T., Maeda H., Abe H., Kudo S.E. & Mori Y., "Deep learning models
for histopathological classification of colon and colonic epithelial tumors," Scientific Reports, vol. 10,
no. 1, p. 1504, 2020.
[9] Qin Y., Li Z.Y., Zhang X.F. & Li Z.Y., "Artificial intelligence in the imaging of colon cancer: current
applications and future direction," Frontiers in Oncology, vol. 11, pp. 631686-631687, 2021.
[10] J.-H., Lee J.-H. & Kim, "A review of the application of deep learning in medical image analysis,"
[11] Loganathan, Zabiha Khan and R, "Transfer Learning Based Classification of MSI and MSS
Gastrointestinal Cancer," International Journal of Health Sciences, vol. 6, no. S1, pp. 1857-1872,
2022.
[12] CH Sai Venkatesh, Caleb Meriga, M.G.V.L Geethika, T Lakshmi Gayatri, V.B.K.L Aruna, "Modified
ResNet Model for MSI and MSS Classification of Gastrointestinal Cancer," arXiv, 2022.
[13] Shuxuan Fan, Xubin Li, Xiaonan Cui, Lei Zheng, Xiaoyi Ren, Wenjuan Ma, Zhaoxiang Ye,
"Computed Tomography-Based Radiomic Features Could Potentially Predict Microsatellite
Instability Status in Stage II Colorectal Cancer: A Preliminary Study," Academic Radiology, vol. 26,
no. 12, pp. 1633-1640, 2019.
[14] Rikiya Yamashita, Jin Long, Teri Longacre, Lan Peng, Gerald Berry, Brock Martin, John Higgins,
Daniel L Rubin, Jeanne Shen, "Deep learning model for the prediction of microsatellite instability in
colorectal cancer: a diagnostic study," The Lancet Oncology, vol. 22, no. 1, pp. 132-141, 2021.
[15] Hüseyin ERİKÇİ, Ziynet PAMUK, "VGG16 classification of microsatellite instability in colorectal
cancer by deep learning," International Refereed Journal of Engineering and Sciences, 2022.
[16] Franceska Dedeurwaerdere, Kathleen BM Claes, Jo Van Dorpe, Isabelle Rottiers, Joni Van der
Meulen, Joke Breyne, Koen Swaerts, Geert Martens, "Comparison of microsatellite instability
detection by immunohistochemistry and molecular techniques in colorectal and endometrial cancer,"
Scientific Reports, vol. 11, 2021.
[17] Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton, "Imagenet classification with deep
convolutional neural networks," Communications of the ACM, vol. 60, no. 6, p. 84–90, 2017.
[18] Karen Simonyan, Andrew Zisserman, "Very Deep Convolutional Networks for Large-Scale Image
Recognition," arXiv, 2015.
[19] Yuqing Gao, Khalid M. Mosalam, "Deep Transfer Learning for Image-Based Structural Damage
Recognition," Computer-Aided Civil and Infrastructure Engineering, vol. 33, no. 9, pp. 748-768,
2018.
[20] Larsen-Freeman, Diane, "Transfer of Learning Transformed," Language Learning, vol. 63, no. s1, pp.
107-129, 2013.
[21] Andrew G. Howard, Menglong Zhu, Bo Chen, Dmitry Kalenichenko, Weijun Wang, Tobias Weyand,
Marco Andreetto, Hartwig Adam, "MobileNets: Efficient Convolutional Neural Networks for Mobile
Vision Applications," arXiv, 2017.
[22] Mark Sandler, Andrew Howard, Menglong Zhu, Andrey Zhmoginov, Liang-Chieh Chen,
"MobileNetV2: Inverted Residuals and Linear Bottlenecks," arXiv, 2018.
[23] Hussam Qassim, Abhishek Verma, "Compressed residual-VGG16 CNN model for big data places
image recognition," 2018 IEEE 8th Annual Computing and Communication Workshop and
Conference (CCWC), 2018.
[25] Cheng Wang, Delei Chen, Hao Lin, Xuebo Liu, "Pulmonary Image Classification Based on
Inception-v3 Transfer Learning Model," IEEE Access, vol. 7, pp. 146533 - 146541, 2019.
[26] Mingxing Tan, Quoc V. Le, "EfficientNet: Rethinking Model Scaling for Convolutional Neural
Networks," arXiv, 2019.
[27] Mingxing Tan, Quoc V. Le, "EfficientNet: Rethinking Model Scaling for Convolutional Neural
Networks," arXiv, 2019.
[28] Alex Krizhevsky, Ilya Sutskever, Geoffrey E. Hinton, "Imagenet classification with deep
convolutional neural networks," Communications of the ACM, vol. 60, no. 6, p. 84–90, 2017.