Professional Documents
Culture Documents
Abstract—Diabetic Retinopathy (DR) is a prevalent eye which can delay treatment & potentially lead to incorrect
condition affecting individuals with diabetes, potentially leading to management.
vision loss and blindness if left untreated. Manual screening by
ophthalmologists for DR is a time-consuming process. This project To aid in the early detection of DR, fundus cameras are
aims to address this challenge by leveraging Deep Learning (DL), a commonly used by doctors to capture images of the veins &
subset of Artificial Intelligence (AI), to analyze and classify nerves located behind the retina. Unfortunately, the initial stage of
different stages of DR. A model was trained on a large dataset of DR often exhibits no noticeable signs, making it extremely
more than 35,000 high-resolution fundus images to automatically challenging to identify the disease at an early phase. In this
detect and categorize DR stages. The dataset used in this study,
project, we leverage Convolutional Neural Network (CNN)
titled "Diabetic Retinopathy (resized)," is available on Kaggle. The
DR stages are classified as 0, 1, 2, 3, and 4. Patient fundus eye
architectures, specifically designed for deep learning, to improve
images serve as input for this study. The trained model utilizes the the accuracy & efficiency of DR stage detection. By utilizing
MobileNet v2 architecture, a state-of-the-art convolutional neural these AI models, patients can receive timely treatment &
network (CNN) design optimized for efficient deployment on mobile intervention.
and embedded devices, to extract features from the fundus images.
The final output is generated using an activation function. This The dataset used for this project is sourced from EyePACS, a
paper discusses the utilization of DL techniques, the CNN free platform for retinopathy screening, & is available on Kaggle
architecture, and the analysis of fundus images to diagnose and under the name "Diabetic Retinopathy (resized)". The dataset
classify DR stages, aiming to enhance the accuracy and efficiency comprises retinal images that have been resized to a standard
of DR screening. dimension of 1024x1024 pixels, with larger images proportionally
scaled down. It is important to note that the dataset includes
Keywords: Deep Learning, Diabetic Retinopathy (DR), CNN images captured by various models & types of cameras, which
Architecture, Fundus Images, MobileNet. may impact the visual appearance, especially in distinguishing
between left & right eyes. Some images are displayed
anatomically, with the macula on the left & the optic nerve on the
right for the right eye.
I. INTRODUCTION
DR is a serious complication arising from diabetes, causing Recent advancements & research in the field of artificial
substantial harm to the retina and potentially leading to vision intelligence, particularly in deep learning, have demonstrated its
impairment. This condition primarily impacts the blood vessels effectiveness in uncovering hidden patterns & features within
situated in the retinal tissue, resulting in fluid leakage and the medical image analysis tasks. Deep learning models have shown
distortion of visual perception. Within the spectrum of visual promising results in disease classification & decision support,
impairments, including conditions like cataracts and glaucoma, thereby enhancing patient care.
Diabetic Retinopathy (DR) stands out as a highly prevalent
disorder. The classification of this disease comprises five distinct [1]-[3] cite the efficacy of deep learning models in medical
stages denoted as 0, 1, 2, 3, and 4. each with its distinct symptoms image analysis tasks.
& characteristics. However, it is challenging for doctors to [4] emphasizes the potential of AI models, specifically deep
determine the stage of DR from normal retinal images alone. learning, to improve patient care through disease classification &
Existing diagnostic methods are inefficient & time-consuming, decision support.
performance of deep learning-based DR detection & classification
By utilizing deep learning techniques & analyzing fundus systems. Overall, the continued development & refinement of
images, this paper aims to overcome the limitations of current DR these automated systems hold great promise for improving the
diagnostic methods, improve the accuracy of DR stage diagnosis & management of diabetic retinopathy. [5]
classification, & ultimately enhance patient care in the
management of this debilitating disease. The research focuses on developing an automated system for
grading diabetic retinopathy (DR) using deep learning techniques
applied to color fundus images. Skip-connected networks,
II. LITERATURE REVIEW particularly the ResNet50-based network, demonstrated superior
performance in classifying DR stages, especially when dealing
Automated screening systems have revolutionized the with small features. The proposed system has the potential to
diagnosis process by significantly reducing the time required for significantly reduce diagnosis time, benefiting diabetic patients'
determining diagnoses, resulting in cost & effort savings for quality of life. Further improvements can be made by expanding
ophthalmologists & timely treatment for patients. In the context the classification to a more detailed 5-class problem,
of diabetic retinopathy (DR), automated systems play a crucial implementing ensemble approaches, utilizing Convolutional
role in detecting the disease at an early stage. The classification of Neural Networks (CNNs) for tasks like optic disc removal &
DR stages is based on the specific type of lesions observed on the blood vessel extraction, & exploring CNN-based image semantic
retina. This article presents a comprehensive review of the latest segmentation for the detection of challenging cases such as cotton
automated systems for detecting & classifying diabetic wool natured exudates. These enhancements will contribute to
retinopathy, focusing on the utilization of deep learning more accurate & precise DR grading, improving patient care &
techniques. Additionally, it describes the commonly used publicly treatment outcomes. The research showcases the effectiveness of
available fundus DR datasets & provides a brief explanation of deep learning in automating DR classification & highlights the
deep learning techniques. Convolutional Neural Networks potential for continued advancements in this field to further
(CNNs) have emerged as the preferred choice for classifying & enhance the performance & applicability of the automated
detecting DR images due to their efficiency. The review also system. [6]
explores various useful techniques that can be employed to detect Diabetic Retinopathy (DR) poses a significant threat to
& classify DR using deep learning methods. [5] individuals with diabetes, with surveys suggesting a 30% chance
of developing the condition. DR encompasses various stages,
Automated systems have transformed the field of ranging from mild to severe & ultimately progressing to
ophthalmology by streamlining the process of diagnosing diabetic Proliferative Diabetic Retinopathy (PDR), which can result in
retinopathy (DR). These systems offer immense benefits by vision impairment & even blindness if not detected early. Manual
reducing the time & effort required for diagnosis, while also diagnosis of DR through fundus image analysis requires extensive
enabling early detection & timely treatment for patients. Diabetic training & is a time-consuming & challenging task. In order to
retinopathy stages are determined based on the specific types of tackle this problem, researchers have introduced computer vision
retinal lesions present. In recent years, deep learning techniques techniques for the automated detection and classification of
have gained significant attention in the development of automated Diabetic Retinopathy (DR). This study specifically focuses on
systems for DR detection & classification. This review aims to classifying all stages of DR, with a particular emphasis on the
provide a comprehensive overview of these state-of-the-art early stages that pose significant challenges for existing models.
systems & their use of deep learning methodologies. It also sheds The authors present a deep learning ensemble framework based
light on the publicly available fundus DR datasets commonly on Convolutional Neural Networks (CNNs) to identify and
used in these studies. Moreover, the fundamental principles of categorize the various stages of DR using color fundus images.
deep learning techniques are briefly explained, with a focus on The model is trained and evaluated using the largest publicly
Convolutional Neural Networks (CNNs) as the primary choice for available fundus image dataset from Kaggle. The results indicate
DR classification & detection. Additionally, this review discusses that the proposed ensemble model surpasses existing methods and
various useful techniques that can enhance the accuracy & demonstrates accurate detection of all DR stages.. To further
effectiveness of DR detection & classification using deep learning improve the accuracy of early-stage detection, future work
approaches. [5] involves training specific models for each stage & leveraging
ensemble techniques to enhance the performance in identifying
Automated systems utilizing deep learning techniques have early stages of DR. By addressing the limitations of existing
emerged as powerful tools for the detection & classification of models & achieving better performance, this research contributes
diabetic retinopathy. By significantly reducing the time required to the advancement of automated DR detection systems,
for diagnosis, these systems contribute to cost savings & enable potentially leading to earlier interventions & improved patient
timely treatment for patients. Convolutional Neural Networks outcomes. The proposed ensemble-based deep learning approach
(CNNs) have proven to be efficient & effective in analyzing DR shows promise in overcoming the challenges associated with DR
images, making them the preferred choice for classification & detection, bringing us closer to more efficient & accurate
detection tasks. This comprehensive review has provided insights diagnosis in the early stages of the disease.[7]
into the recent advancements in automated DR detection & The aim of this study is to propose a novel architecture called
classification using deep learning techniques. Additionally, it has ResNet-34/DR, which is based on a deep Convolutional Neural
highlighted the publicly available fundus DR datasets & Network (CNN) & transfer learning, for the classification of
discussed various useful techniques that can further enhance the Diabetic Retinopathy (DR) into four classes using publicly
available Kaggle datasets. Previous attempts using the Vgg-19 & images serve as valuable documentation for diagnosing &
Xception architectures to classify DR stages encountered detecting the presence of diabetic retinopathy.
challenges related to high loss values caused by overfitting.
However, by leveraging transfer learning & fine-tuning on the To enable the training and testing of the CNN architecture,
pre-trained ResNet-34 network, the proposed architecture the dataset is accompanied by two CSV files: "train.csv" and
demonstrates exceptional performance on the color fundus image "test.csv." The "train.csv" file includes details such as the
dataset. Additionally, data augmentation techniques were names of fundus eye images and their corresponding severity
employed to increase the training samples, particularly for the levels (classes), while the "test.csv" file includes only the eye
lower classes. The training methodology employed in this
image names, as the trained CNN architecture employs images
research has led to significant advancements in the classification
for conducting testing. [11]. This dataset provides a substantial
of DR, yielding improved results compared to previous
approaches. This study contributes to the field by presenting a amount of data for training the models & evaluating their
robust architecture that addresses the challenges of DR performance on detecting & classifying diabetic retinopathy
classification, paving the way for enhanced diagnosis & treatment stages.
of this condition. [8]
The author presents a straightforward network architecture that
enables the development of highly efficient models suitable for
mobile applications. The architecture's basic unit possesses key
properties that make it well-suited for memory-efficient inference
& leverages standard operations found in all neural frameworks.
In experiments conducted on the ImageNet dataset, the proposed
architecture surpasses and the network achieves superior
performance at different performance levels compared to
real-time detectors on the COCO dataset for object detection
tasks. It outperforms them in terms of accuracy and model
complexity. It is worth mentioning that when integrated with the
SSDLite detection module, the network exhibits remarkable
results.. The proposed architecture requires significantly less
computation (20 times less) & fewer parameters (10 times less)
compared to the YOLOv2 model. On a theoretical level, the
convolutional block introduced in this work exhibits a unique
property that separates the network's expressiveness, represented
by the expansion layers, from its capacity, encoded by bottleneck
inputs. This novel characteristic opens up promising avenues for
future research, enabling the exploration of the architecture's
expressiveness & capacity in a decoupled manner. Overall, this
work contributes to the field of efficient network design &
highlights potential directions for further investigation. [9]
Fig 1: Sample Test Images
Figure 3 presents the tally of each class, illustrating the A. CNN Architecture:
distribution of images across different categories within the
dataset.
The CNN model comprises of two primary stages: feature
extraction and classification [1]. During the feature extraction
stage, relevant information and distinguishing characteristics are
extracted from the images., while the classification phase utilizes
these extracted features to categorize the images based on the
target variable of the problem at hand.
Convolution layer +
Activating function+
Fig.3: Image Class Distribution
Pooling layer+
Consequently, the figure below displays the cumulative count
of images assigned exclusively to two distinct classes. Fully Linked Layer
which means we move the filter by one column at a time. The
stride determines the shift in columns or rows when applying the
filter.
3) Pooling Layer
The convolution layer applies a filter to the input image to Two commonly used types of pooling, Max Pooling and
extract or identify its characteristics. Multiple filtering processes Average Pooling, are frequently employed. The operational
are performed to create a feature map that aids in classifying the process of Max Pooling is visually illustrated in the figure below.
input image. To better understand this, let's consider an example In this example, we perform pooling on the obtained feature map
using a 2D input image with standardized pixels. using a 2x2 window and a stride of 2. The maximum value within
each highlighted region is extracted, resulting in a new version of
the input image with a reduced size of 2x2. Consequently, the
In the given illustration, we have an input image of size 66 &
application of pooling reduces the overall size of the feature map.
we place a 3x3 filter on it to identify specific features. It is
important to note that multiple similar filters are actually used in
practice to extract information from the image. 4) Fully Connected Layer
The filter is then shifted by a single column in the next stage, We have completed the Feature Extraction operations up to that
as depicted in the figure. In this example, we choose a stride of 1, stage; the subsequent step is Classification. The input image is
categorised in a label utilising the completely connected layer training methods and expanding its applicability to various
(which is what we're dealing with in ANN). This layer establishes domains and activities, in future research projects. We can open
a connection between the output layer and the data obtained from up new opportunities for deploying potent computer vision
the preceding steps, namely the convolutional and pooling layers. models on mobile and embedded devices by continuing to
Its primary function is to assign the appropriate label to the input innovate and improve MobileNetV2's capabilities. This will
data. ultimately advance the development of deep learning for
resource-constrained situations.
The image below demonstrates the whole approach for
constructing a CNN model.
1) Architecture of MobileNet v1
VI.CONCLUSION
Given the significant impact of Diabetic Retinopathy on
individuals affected by diabetes, the manual identification of this
Fig.12: Accuracy Plot disease requires a substantial amount of time. To address this
issue, we opted to employ MobileNet v2, a cutting-edge
convolutional neural network renowned for its exceptional
performance. By utilizing this network architecture, we
developed a model for the automated detection of Diabetic
Retinopathy. Our proposed approach achieved a comparative
accuracy of 0.9311, demonstrating its effectiveness in accurately
identifying this condition.
REFERENCES
[1] “Deep Learning Approach to Diabetic Retinopathy Detection”
Borys Tymchenko, Philip Marchenko, Dmitry Spodarets
arXiv:2003.02261
[2] Anuj Jain, Arnav Jalui, Jahanvi Jasani, Yash Lahoti, Ruhina
Karani. "Deep Learning for Detection & Severity
Classification of Diabetic Retinopathy", 2019 1st International
Conference on Innovations in Information & Communication
Technology (ICIICT), 2019
[3] Cai L, Gao J, Zhao D. A review of the application of deep
learning in medical image classification and segmentation.
Fig.13: Loss Plot
Ann Transl Med. 2020;8(11):713.
doi:10.21037/atm.2020.02.44
[4] Suganyadevi, S., Seethalakshmi, V. & Balasamy, K. A review
on deep learning in medical image analysis. Int J Multimed Info
Retr 11, 19–38 (2022).
https://doi.org/10.1007/s13735-021-00218-1
[5] Wejdan L. Alyoubi, Wafaa M. Shalash, Maysoon F. Abulkhair,
Diabetic retinopathy detection through deep learning
techniques: A review, Informatics in Medicine Unlocked,
Volume 20, 2020, 100377,
ISSN 2352-9148, https://doi.org/10.1016/j.imu.2020.100377.
[6] T.R. Athira, Jyothisha J Nair, Diabetic Retinopathy Grading
From Color Fundus Images: An Autotuned Deep Learning
Approach, Procedia Computer Science, Volume 218, 2023,
Pages 1055-1066, ISSN 1877-0509,
https://doi.org/10.1016/j.procs.2023.01.085.
[7] S. Qummar et al., "A Deep Learning Ensemble Approach for images," 2016 International Conference on Signal Processing
Diabetic Retinopathy Detection," in IEEE Access, vol. 7, pp. and Communications (SPCOM), Bangalore, India, 2016, pp.
150530-150539, 2019, doi: 10.1109/ACCESS.2019.2947484. 1-4, doi: 10.1109/SPCOM.2016.7746702.
[8] “RESNET-34/DR: A RESIDUAL CONVOLUTIONAL [12] R. Sarki, K. Ahmed, H. Wang, and Y. Zhang(2020), “Automatic
NEURAL NETWORK FOR THE DIAGNOSIS OF Detection of Diabetic Eye Disease through Deep Learning
DIABETIC RETINOPATHY” Noor M. Al-Moosawi M. Using Fundus Images: A Survey,” IEEE Access, vol. 8, pp.
Al-Moosawi, Raidah Salim Khudeyer 151133–151149, doi: 10.1109/ACCESS.2020.3015258.
[9] “MobileNetV2: Inverted Residuals & Linear Bottlenecks”
Mark S&ler, Andrew Howard, Menglong Zhu, Andrey
Zhmoginov, Liang-Chieh Chen
[10] Kele Xu, Dawei Feng, & Haibo Mi, “Deep Convolutional
Neural Network-Based Early Automated Detection of Diabetic
Retinopathy Using Fundus Image,” Received: 10 November
2017; Accepted: 22 November 2017; Published: 23 November
2017.
[11] S. K. Vengalil, N. Sinha, S. S. S. Kruthiventi and R. V. Babu,
"Customizing CNNs for blood vessel segmentation from fundus
.