You are on page 1of 9

Biomedical Engineering: Applications, Basis and Communications, Vol. 31, No.

5 (2019) 1950034 (9 pages)


DOI: 10.4015/S1016237219500340

DIAGNOSIS OF BREAST CANCER USING


MULTISCALE CONVOLUTIONAL NEURAL
NETWORK
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

Homayoon Yektaei*,§, Mohammad Manthouri† and Faezeh Farivar‡


*Department
of Biomedical Engineering
Islamic Azad University
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

Tehran North Branch, Tehran, Iran



Department of Electrical and Electronic Engineering
Shahed University, Tehran, Iran

Department of Mechatronics Engineering
Science and Research Branch
Islamic Azad University
Tehran, Iran

Accepted 17 April 2019


Published 6 June 2019

ABSTRACT
The diagnosis of benign and malignant breast cancer is a challenging issue today. Breast cancer is the most common
cancer that women su®er from. The sooner the cancer is detected, the easier and more successful it is to treat it. The
most common diagnostic method is the mammography of a simple radiographic picture of the chest. The use of image
processing techniques and identifying patterns in the detection of breast cancer from mammographic images reduces
human errors in detecting tumors, and speeds up diagnosis time. Arti¯cial Neural Network (ANN) has been widely
used to detect breast cancer, and has signi¯cantly reduced the percentage of errors. Therefore, in this paper, Convo-
lution as Neural Network (CNN), which is the most e®ective method, is used for the detection of various types of
cancers. This study presents a Multiscale Convolutional Neural Network (MCNN) approach for the classi¯cation of
tumors. Based on the structure of MCNN, which presents mammography picture to several deep CNN with di®erent
size and resolutions, the classical handcrafted features extraction step is avoided. The proposed approach gives better
classi¯cation rates than the classical state-of-the-art methods allowing a safer Computer-Aided Diagnosis of pleural
cancer. This study reaches the diagnosis accuracy of 97  0:3 using multiscale convolution technique which reveals the
e±cient proposed method.

Keywords: Breast cancer; Feature extraction; Multiscale convolutional neural network.

INTRODUCTION a pathologist, and with the ever-increasing amount of


Computer-Aided Diagnosis (CAD) is an increasing do- virtual biomedical data, CAD has now become a crucial
main through image processing.1,2 Indeed, automatic point for enabling safer and faster diagnosis.3 Cancer,
salient feature extraction by a computer can save time of being a serious threat to human life, is a combination of

§
Corresponding author: Homayoon Yektaei, Department of Biomedical Engineering, Islamic Azad University, Tehran North
Branch, Tehran, Iran. E-mail: Hyektai1994@gmail.com

1950034-1
H. Yektaei, M. Manthouri & F. Farivar
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

Fig. 1 New cases of breast cancer for women and the number of women dying in the last twelve years.4
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

diseases, and more speci¯cally unwanted and abnormal signi¯cantly increased accuracy in the cases of breast
growth of the cells of the human body is known as tumor. Inter-dataset training and evaluation using
cancer. Cancer can attack any part of the body and can multiple datasets. (g) Generalization: This demonstrates
then be distributed to any part of the body. Among all generalization of the method to new data, from separate
type of cancers, breast cancer (BC) is the most common datasets and graders, and captured from di®erent devi-
one among women. Statistics show the increasing rate of ces. CNN are known for robustness to small changes in
BC each year. Figure 1 shows the number of females inputs because of °exibility. Also, they do not require
recently facing BC, as well as the number of females who any speci¯c feature extraction step.5,6 The proposed
have died in Australia since 2007. This ¯gure shows that MCNN architecture relies on several deep neural net-
more and more females are newly facing BC, and the works that alternate convolutional and pooling layers.
number of females dying of it has also increased each These deep neural networks belong to a broad class of
year. This is the situation of Australia (population models generally termed multi-stage architectures.7 This
20–25 million), but it can be used as a symbol of the BC architecture of convolution interlaced with pooling lay-
situation of the whole world.5 ers is dedicated to the automatic feature extraction. The
Four crucial and markable indexes include: (a) Speed ¯nal classi¯cation is performed by some classical fully
and autonomation: This results in a fast method connected layers stacked on top.3
requiring no user input. (b) Independence: The method This research presents multi-scale CNN for diagnosis
is not dependent on other techniques succeeding such as of breast cancer. Due to the abilities of CNN and the
segmentation or detecting other landmarks. (c) No creation of images in the desired size and magnitudes
handcrafted features: Since features do not need to be with the help of a multiscale CNN, the accuracy of the
manually de¯ned, we avoid the di±culty encountered by diagnosis is increased. To show the e±ciency of
conventional machine learning algorithms in identifying the proposed method, comparison is make with other
the best feature set that represents the data. It also methods.40,41 The results reveal the signi¯cant achieve-
removes the requirement of a skilled technician to ment of the suggested technique.
identify such features manually which takes a consider- The remainder of this paper is organized as follows:
able amount of time and can produce subjective results, Second-section describes the proposed methodology.
particularly with a large dataset. (d) Accurate simulta- The results of this research are described in third section.
neous detection: We detect more than one position Fourth section is discussion part for analytical results of
simultaneously, retaining high accuracy for each. the paper. And the paper is concluded in ¯fth section.
(e) Robustness: The method is robust in the sense that it
continues to work well even on poor quality images. We
develop a multiscale approach to convolutional neural MATERIALS AND METHODS
networks (CNN) to focus on the region of interest.
Materials
(f) Improved Accuracy: This approach allows the
method to focus on the region of interest, removing re- Database
dundant background data from consideration and fa- The study is based on MIAS,8 referring to the patient's
cilitating re¯nement of the localization. This results in imaging center. Pictures of 1024 samples of 322 patients,

1950034-2
Diagnosis of Breast Cancer using Multiscale Convolutional Neural Network

Fig. 2 Image pre-processing stages.


by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

including benign and malignant breast cancer, are la-


beled on this site.9 The pro¯le of 322 samples of cancer
patients in this dataset was collected in 1994.
Pre-Processing
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

This project aims to detect breast cancer in breast tis-


sue, so this research needs to ignore the processing of the Fig. 3 The general architecture of the proposed. MCNN approach is
image from the extra tissue of the breast, and only use in this way we examined the mammography method, which, based on
our shape, increases the composition of the coulomb layer and creates a
the areas of our 1024 images to become patches, zoom in multi-scale convolution, the same as the method. The pyramid once
or out and resize when needed (Fig. 2). the convolution on a large image is once again smaller and smaller,
therefore, it is better for us to be able to determine the number of our
Deep CNN architecture convulsions because we want to identify the endocervical or malignant
In contrast to conventional shallow classi¯ers, such as mass of the mass.
neural networks and support vector machines, for which
a feature extraction step is essential, hierarchies of sig-
ni¯cant features are learned by deep learning algorithms classi¯cation results than a linear combination whose
directly from the raw input data. Recently, deep con- weights learned over a cross-validation set.18 However,
volutional neural nets (convents) have succeeded in in Ref. 17 there was no reason to weigh the classi¯ers
improving many computer vision applications such as since they act at the same resolution and only di®er from
image classi¯cation,10 object recognition,11 and key the distortions applied to the training set. In our case, a
point localization.12 Also, some impressive results have reasonable hypothesis is that the classi¯er at the lowest
been seen in biomedical applications such as neuronal resolution may be less salient than the one at full reso-
membrane segmentation13,14 and other applications.15,16 lution. Final outputs are computed as a linear combi-
A typical CNN comprises one or more convolutional nation of the outputs of the N CNNs. The overall
layers alternated with pooling layers (subsampling lay- architecture of the MCNN is shown in Fig. 3.
ers) and then followed by fully connected layers (FC) and Convolution layer
¯nally a classi¯cation/regression layer. CNNs can be In these layers, the CNN network uses di®erent kernels
considered as special forms of feedforward multilayer to convolve the input image and also the middle map
perceptron neural networks (MLPs). However, the feature, and this kind of map features di®erent mappings
number of parameters that need to be tuned is reduced to than what you see in Fig. 4. Performing convolution
a level that becomes tractable for the current computing operations has three advantages.19
power. For example, in convolution layers, a limited
number of convolutional kernels is needed. (1) The weight sharing mechanism in each feature map
will reduce the number of parameters.
Multiscale CNN (MCNN) (2) The local connection recognizes the relationship
It is well accepted that human insight is a multi-di- between neighboring pixels.
mensional process. In this work, N CNNs have been (3) Provides change and stability to change location.
created with di®erent sizes of breast cancer tumor ima-
ges. A given input pattern is rescaled N-times to ¯t Due to the bene¯ts provided by the operation, convo-
the breast cancer tumor of the CNNs. The question of lution used some of the renowned research papers to
optimizing the outputs of classi¯ers is a recurrent ques- replace fully connected layers to increase the speed of
tion in the pattern recognition ¯eld. For handwritten the learning process.20,21 One of the interesting meth-
digits, it is shown17 that a simple average gives better ods for managing the layers of the convoys is Network

1950034-3
H. Yektaei, M. Manthouri & F. Farivar

Pooling layers
A pooling layer is usually located after a convolutional
layer and can be used to reduce the size of the feature
maps and network parameters. Its overall goal is to re-
duce the size of the data and reduce the volume of the
calculations. Like the convoys, the pooling layers
are stable (stable) to change neighborhoods in their
calculations. The pooling layer is placed after the
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

convolutional layers, and the data size is halved or


smaller. Implementing the Pooling layer using the max
pooling and the Average pooling (Average) function are
the most common implementations. In Fig. 5, you can
see a sample of the process of max pooling. Using A max
pooling ¯lter of 2  2 and 2 stride sizes, and an 8  8
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

feature map creates an output of 4  4.22,23


Fig. 4 Convolution layer.26–28
Given an input volume of size Ni  Ni  Di,
max-pooling window size R  R, and the stride para-
meters, the number of neurons in the output volumes
No  No  Do is calculated by the formula (1)24:
Ni  R
N0 ¼ þ 1; D0 ¼ D1 : ð1Þ
S

Fully connected layer


This usually represents the ¯nal layers of a deep neural
network architecture. Each node in the fully connected
layer is completely connected to all of the nodes in the
previous layer and the weights of these links are speci¯c
to each node. The number of neurons in the fully con-
Fig. 5 Max pooling operation.29,30 nected layers is considered as a hyper-parameter to be
empirically chosen (Fig. 6).25
in Network (NIN), in which the main idea is to replace the
convoys with a small perceptron neural network, which
consists of several fully connected layers with nonlinear Experiments
activation functions. In this way, linear ¯lters are In this section, some related works on diagnosing breast
replaced by nonlinear neural networks. This method cancer are discussed. In all cases, the focus is on reducing
produces good results in image categorization (Fig. 4). the error and increasing the accuracy of tumor detection.

Fig. 6 Block diagram of a CNN.

1950034-4
Diagnosis of Breast Cancer using Multiscale Convolutional Neural Network

Bermain Sahir et al. used arti¯cial neuronal classi¯ca- Pak et al. also utilized Non-subsampled CT for
tion of convolution to classify the masses and natural breast-image (MIAS dataset) classi¯cation and obtained
tissue of the breast. Initially, they captured the area of 91.43% mean Accuracy and 6.42% mean False Positive
interest (ROI) and placed on the average in a subset. Rate (FPR).32
Second, Gradient di®erence (GLDS) and Spot Gray This research has proposed a new approach to CNN
Dependence (SGLD) characteristics from di®erent in detecting breast cancer using the minimias data to
regions are calculated, and ¯nally, the calculated char- maximize accuracy. This proposed method is multiscale
acteristics were introduced as inputs to the convolu- convolution. Also, the proposed approach has yielded
tional classi¯cation.26
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

promising results and outperforms conventional neural


In Ref. 27, the diagnosis of breast cancer in networks, which demonstrates that deep learning tech-
DCE-MRI was investigated using a combination of niques will be able to detect the strong and accurate
CNN, which included a new computer diagnostic (CAD) diagnosis of benign and malignant lumps.
system in contrast magnetic contrast resonance imaging
(DCE-MRI) Panasonic appears. The CAD system is
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

based on a combination of neurotransmitter (ME-CNN)


designed to distinguish between optic and malignant RESULTS
tumors of the breast. ME-CNN is a modular, image- Initially, this research considered the mammography
based group that can split large-dimensional image images from the relevant data (MIAS) and made them
space through concurrent and competitive learning of its smaller patches. We made all our photos in 50, 50, 60,
modules. The proposed system in our database evalu- 60, 70, 70, 80, 80, 90, 90 and 90, respectively, and then
ated 112 DCE-MRI studies including rigid breast masses used convolutional polysaccharides, respectively. Multi-
using a wide range of classi¯ed measures. The ME-CNN variate images in di®erent sizes and zooms to the con-
model, which consists of three CNN experts and a can- volutions were taken at the end of the survey to show
nula-shifting network, has a precision of 96.39%, a sen- that it is optimistic or malignant. In the following, we
sitivity of 97.33% and an 87.94% gain. Experimental will use some of the images in which the textures and
results also show that there are competing functions, details are clear.
in contrast to the three existing methods, of a single- In the data, they write a series of photographs and
integer and two-way convolutional groups. The pro- have extra areas or have no resolution, so the so-called
posed ME-CNN model can be an e®ective tool for radi- variable form must be double. So that disturbing areas
ologists to analyze breast DCE-MRI images.27 do not a®ect our detection. Because our photos are pairs,
Most image categorization is based on CNN method left breast and right chest were all in one direction. The
based on global features extraction techniques. Recent- mirror method is used to create the chest part (Fig. 7).
ly, researchers have also shown an interest in how local In each image, MCNN has a series of intruder points
features can be utilized with the CNN model for data and a series of ideal or ROI points, so it can use the same
classi¯cation. Both global and local features have been polygonal method to separate it from a 6-dot on the
used by Rezaeilouyeh et al.28 for Histopathological image (Fig. 8).
image classi¯cation. For local features extraction, To remove the area associated with the background,
shearlet converters and transmitters have been used to gray levels of less than 21% are not considered the highest
obtain a precision of 86  0:00. To extract local features, gray level. An example of a histogram of mammograms, in
Sharma et al.29 used the GLCM method, GLDM, and which gray levels of less than 21% are observed, are shown
then fed local attributes to a CNN model for mammo- in Fig. 9. The histogram is prescreened and used to opti-
gram image classi¯cation and obtained a precision of mize and eliminate light and to equalize the brightness
75.33% for the classi¯cation of adipose tissues. range of the image that aims to enhance clarity.
For local feature extraction, the authors utilized the The average of all gray levels of residual pixels in each
Shearlet transform and obtained an accuracy of 86   window is considered as the threshold level of T and is
3.00%. For local feature extraction, both global and local obtained from Eq. (2).
features have been used by Jiao et al.30 for mammogram P
i0:2Imax hðiÞ  i
image classi¯cation and they obtained 96.70% accuracy. T ¼ P : ð2Þ
i0:2Imax hðiÞ
Kooi et al.31 utilized both global features and hand-
crafted features for Mammogram image classi¯cation. In in which i represents the gray level, hðiÞ the histogram of
their experiment, they also utilized the transfer learning the selected Imax window represents the highest gray
method.4 level in the image. CLAHE algorithm, one of the image

1950034-5
H. Yektaei, M. Manthouri & F. Farivar

processing methods, is used to optimize images and


uniformize the histogram.

Networks Architectures
Construct 5 di®erent CNNs that act at di®erent reso-
lutions. The size of the images at full resolution is
80  80. This size is successively divided by a factor
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

Fig. 7 Mirror method.


(Fig. 10).
Then the original image is given to MATLAB and
from the ideal regions, like the canvassing method, we
give a number of patches that vary in size and shape.
Since we want patches that give us useful and accurate
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

information, we used File Extraction. Our coding ar-


rangement is CPCP, and we use the SoftMax method for
the classi¯cation in the fully connected layer (SVM
method can also be used). Formula (3) was used to
calculate the system's ERROR:
Fig. 8 Polygonal method. 1 X
mean square: Lðy; y~Þ ¼ ðyi  y~i Þ2 : ð3Þ
N i

TEST and TRAIN using Cross Validation, and included


80% to 20%.
At the end of the steps, as in Table 1, we were able to
recognize the accuracy, sensitivity and speci¯city, and to
conclude that the larger our patches (like the pyramid)
the higher precision conveyed. For this reason, we used a
multiplicity.

DISCUSSION
As the literature shows, di®erent methods and techni-
ques for categorizing images in di®erent types of breast
data have been used. However, the state-of-the-art
image classi¯cation technique of the CNN has put its
Fig. 9 A histogram of mammogram images. strong footprint in the image-analysis ¯eld, especially

Fig. 10 The architecture of the CNN80.

1950034-6
Diagnosis of Breast Cancer using Multiscale Convolutional Neural Network

Table 1. The Properties and Percentages of the Multi-Level


Convolution Method.

CNN1 CNN2 CNN3 CNN4 CNN5 CNN-VOTE


ACC 84.17 87.21 92.07 88.66 92.99 95.05
SEN 86.52 87.92 91.98 87.72 94.44 95.87
SPE 85.42 87.78 93.40 86.93 92.39 94.70
PER 84.94 97.95 92.10 90.76 91.31 97.33
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

the image-classi¯cation ¯eld. Though the \AlexNet" Fotin et al. also utilized the CNN method for tomo-
model proposed by Krizhevsky has gained new mo- synthesis image classi¯cation and obtained an Area
mentum in the CNN research ¯eld, a CNN model was Under the Curve (AUC) value of 0.93. Transfer learning
¯rst utilized by Fukushima et al.33 who proposed the is another important concept of the CNN method,
\Necognitron" model that recognizes stimulus patterns. which allows the model to derive attributes from the
For the mammogram image classi¯cation, Wu et al. beginning and not use a weight sharing concept to teach
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

¯rst utilized the CNN model.34 Though little work had a model. This method is helpful when the database
been done on the CNN model until the end of the 20th contains fewer images. Jiang et al.38 utilized a transfer
century, this model is only obtained from the AlexNet learning method for Mammogram image classi¯cation
model. Advanced engineering techniques have been and obtained an AUC of 0.88. Before utilizing it in a
used by research groups such as the Visual Geometry CNN model, they performed a preprocessing operation
Group and Google, which have modeled the VGG-16, to enhance the images. Suzuki et al.39 also used the
VGG-19 and GoogleNet models. Arevalo et al.35 clas- bene¯t of transfer learning techniques to train their
si¯ed benign and malignant lesions using the CNN model to classify mammogram images and obtained
model, and this experiment was performed on 766 sensitivity of 89.9%. They performed their experiment
mammogram images, where 426 images contain benign with just 198 images. Most image categorization is
and 310 malignant lesions. Before classifying the data, based on the CNN method based on global features
they utilized preprocessing techniques to increase the extraction techniques. Recently, researchers have also
image enhancement and obtained a 0:82  0:03 Re- shown an interest in how local features can be utilized
ceiver Operating Characteristic (ROC) value. Google- with the CNN model for data classi¯cation. Both
Net and AlexNet methods have been utilized by Zejmo global and local features have been used by Rezaei-
et al.36 for the classi¯cation of cytological specimens louyeh et al.28 for Histopathological image classi¯cation.
into benign and malignant classes. The best accuracy For local feature extraction, the authors used the
obtained when they utilized the GoogleNet model was Shearlet transform and obtained an accuracy of
83.00%. Qiu et al.37 used the CNN method to extract 86  3:00%. For local feature extraction, Sharma et al.29
global features for Mammogram image classi¯cation used the GLCM, GLDM methods and then fed the local
and obtained an average achieved accuracy of 71.40%. features to a CNN model for the Mammogram image

Table 2. Summary of Discussion.

Researcher Methods Accuracy AUC


Krizhevski Alexnet  

Fukushima et al. Necognitron  

Wu et al. utilized the CNN model  

Arevalo et al. classi¯ed benign and malignant lesions using the CNN model 0:82  0:03 

Zejmo et al. GoogleNet and AlexNet methods have been utilized 83.00%. 

Qiu et al. used the CNN method to extract global features for Mammogram image classi¯cation 71.40% 

Fotin et al. utilized the CNN method for tomosynthesis image classi¯cation  0.93
Jiang et al. utilized a transfer learning method for Mammogram image classi¯cation  0.88
Suzuki et al. used the bene¯t of transfer learning techniques to train their model to classify mammogram  89.9%.
images
Rezaeilouyeh et al. for Histopathological image classi¯cation 86  3:00% 

Sharma et al. ] used the GLCM, GLDM methods and then fed the local features to a CNN model 75.33% 

Jiao et al. Both global and local features have been used for mammogram image classi¯cation 96.70% 

Kooi et al. utilized both the global features and the manipulated features were used to classify the 
 

mammography image

1950034-7
H. Yektaei, M. Manthouri & F. Farivar

classi¯cation, obtaining 75.33% accuracy for the fatty REFERENCES


and dense tissue classi¯cation. Both global and local 1. Malek J, Sebri A, Mabrouk S, Torki K, Tourki, R,
features have been used by Jiao et al.30 for mammogram Automated breast cancer diagnosis based on GVF-Snake
image classi¯cation, and they obtained 96.70% accura- segmentation, wavelet features extraction and fuzzy
cy. Kooi et al.31 utilized both the global features and the classi¯cation, J Signal Process Syst 49:66, 2009.
manipulated features to classify the mammography 2. Puls J, Dytch H, Roman, M, Automated screening for
cervical cancer using image processing techniques, in Proc
image. In their testing, they also used the transfer
1980 FPS Users Group Meeting, 1980.
learning method (Table 2). 3. Buyssens P, Elmoataz A, Lezoray, O, Multiscale con-
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

A new approach based on a multi-dimensional deep volutional neural networks for vision–based classi¯cation
learning technique is suggested for benign diagnosis of of cells, in Asian Conf Computer Vision, Springer, Vol.
lump in breast mammary imaging. Designed by CNNs 342, p. 352, 2012.
4. Nahid A-A, Kong Y, Histopathological breast-image
by discovering sophisticated images of breasts without
classi¯cation using local and frequency domains by con-
human supervision, the performance of our proposed volutional neural network, Information 9:19, 2018.
system can improve the competitive approach and play 5. Nebauer, C, Evaluation of convolutional neural networks
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

to a much higher degree of precision. In this section, we for visual recognition, IEEE Trans Neur Networks
reviewed a few examples of convolutional work in this 685:696, 1998.
¯eld. In all of the work that worked on the masses and 6. LeCun Y, Bottou L, Bengio Y, Ha®ner, P, Gradient-based
learning applied to document recognition, Proc IEEE
used the Mia data, the method of coupling of multi-
2278:2324, 1998.
variate convolution has had the best performance. 7. Hubel DH, Wiesel TN, Receptive ¯elds, binocular inter-
action and functional architecture in the cat's visual cor-
tex, J Physiol 106:154, 1962.
CONCLUSION 8. Suckling J, Parker J, Dance D, Astley S, Hutt I, Boggis C,
Ricketts I, Stamatakis E, Cerneaz N, Kok S, The mam-
This paper describes a breast cancer diagnostic system mographic image analysis society digital mammogram
using image and network processing techniques. This database, in Int Congress Series Exerpta Medica, Vol.
375, p. 378, 1994.
work is based on MCNN. In the proposed method, before
9. http://peipa.essex.ac.uk/info/mias.html/5 oct 2018.
the canvassing, the processing of the mammography 10. Krizhevsky A, Sutskever I, Hinton GE, Imagenet classi-
image of the surplus areas, the pectoral muscle and the ¯cation with deep convolutional neural networks, in Adv
labels (if any) are removed, and since working on the Neur Inf Process Syst 1097:1105, 2012.
chest, we are forced to mirror the images to pair them. 11. Girshick R, Donahue J, Darrell T, Malik, J, Rich feature
This research improves the quality of mammogram hierarchies for accurate object detection and semantic
segmentation, in Proc IEEE Conf Computer Vision and
images using the CLAHE algorithm. Depending on the
Pattern Recognition 580:587, 2014.
breasts, we split the image into 5 di®erent sizes. Using 12. Long JL, Zhang N, Darrell T, Do convnets learn corre-
the multi-dimensional method, each image is convoluted spondence?, in Adv Neur Inf Process Syst 1601:1609, 2014.
to a di®erent size, and the number of convolutions to get 13. Ciresan D, Giusti A, Gambardella LM, Schmidhuber J,
the result. Using convolutional multiscale method, the Deep neural networks segment neuronal membranes in
electron microscopy images, Adv Neur Inf Process Syst
mean accuracy of the proposed method was 97  0:3%,
2843:2851, 2012.
the sensitivity of detection, 95.9%, speci¯city of 94.8% 14. Cernazanu-Glavan C, Holban S, Segmentation of bone
and the detection error of 3%. The results indicate that structure in X-ray images using convolutional neural
preprocessing operations on mammographic images network, Adv Electr Comput Eng 87:94, 2013.
signi¯cantly increase the speed and precision of the 15. Li S, Chan AB, 3d human pose estimation from monocular
system. At the end of the process, MCNN will be able to images with deep convolutional neural network, in Asian
Conf Computer Vision, Springer, Vol. 332, p. 347, 2014.
detect and, as far as possible, to improve the accuracy,
16. Levi G, Hassner T, Age and gender classi¯cation using
sensitivity, speci¯city, and accuracy. To conclude, the convolutional neural networks, in Proc IEEE Conf
larger our patches (like the pyramid) would be, the Computer Vision and Pattern Recognition Workshops,
higher the precision conveyed. If we increase the number Vol. 34, p. 42, 2015.
of convolutions from 5 to 7 and even more, our precision 17. Meier U, Ciresan DC, Gambardella LM, Schmidhuber J,
Better digit recognition with a committee of simple neural
will also increase in the multiscale convolution method.
nets, in Int Conf IEEE, Document Analysis and Recog-
In the further work, the purpose of this study is to in- nition (ICDAR), Vol. 1250, p. 1254, 2011.
crease the convolutional layers, which, as a result of the 18. Ueda N, Optimal linear combination of neural networks
investigation, increases the layer thickness, in a multi- for improving classi¯cation performance, IEEE Trans
scale CNN, accuracy also increases. Pattern Anal Mach Intell 207:215, 2000.

1950034-8
Diagnosis of Breast Cancer using Multiscale Convolutional Neural Network

19. Zeiler MD, Hierarchical Convolutional Deep Learning in of mammographic lesions, Med Image Anal 303:312,
Computer Vision, New York University, 2013. 2017.
20. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov 32. Pak F, Kanan HR, Alikhassi A, Breast cancer detection
D, Erhan D, Vanhoucke V, Rabinovich A, Going deeper and classi¯cation in digital mammography based on Non-
with convolutions, in Proc IEEE Conf Computer Vision Subsampled Contourlet Transform (NSCT) and Super
and Pattern Recognition, Vol. 1, p. 9, 2015. Resolution, Comput Methods Programs Biomed 89:107,
21. Oquab M, Bottou L, Laptev I, Sivic J, Is object localiza- 2015.
tion for free?-weakly-supervised learning with convolu- 33. Fukushima K, Miyake S, Ito T, Neocognitron: A neural
tional neural networks, in Proc IEEE Conf Computer network model for a mechanism of visual pattern recog-
Vision and Pattern Recognition, Vol. 685, p. 694, 2015. nition, IEEE Trans Syst Man Cybernetics 826:834, 1983.
by UNIVERSITY OF NEW ENGLAND on 09/21/19. Re-use and distribution is strictly not permitted, except for Open Access articles.

22. Scherer D, Müller A, Behnke S, Evaluation of pooling 34. Wu CY, Lo S-CB, Freedman MT, Hasegawa A, Zuurbier
operations in convolutional architectures for object rec- RA, Mun SK, Classi¯cation of microcalci¯cations in
ognition, in Arti¯cial Neural Networks–ICANN 2010, radiographs of pathological specimen for the diagnosis of
Springer, Vol. 92, p. 101, 2010. breast cancer, in Medical Imaging: Image Process Int Soc
23. Cireşan DC, Meier U, Masci J, Gambardella LM, Opt Photonics 630:642, 1994.
Schmidhuber J, High-performance neural networks for 35. Arevalo J, Gonz alez FA, Ramos-Pollan R, Oliveira JL,
Biomed. Eng. Appl. Basis Commun. 2019.31. Downloaded from www.worldscientific.com

visual object classi¯cation, arXiv:1102.0183. Lopez MAG, Representation learning for mammography
24. Huang FJ, Boureau Y-L, LeCun Y, Unsupervised learning mass lesion classi¯cation with convolutional neural net-
of invariant feature hierarchies with applications to object works, Computer Methods Programs Biomed 248:257,
recognition, in IEEE Conf Computer Vision and Pattern 2016.
Recognition, CVPR'07, Vol. 1, p. 8, 2007. 36. Żejmo M, Kowal M, Korbicz J, Monczak R, Classi¯cation
25. Al-Bander B, Al-Nuaimy W, Williams BM, Zheng Y, of breast cancer cytological specimen using convolutional
Multiscale sequential convolutional neural networks for neural network, in J Phys Conf Ser 12:60, 2017.
simultaneous detection of fovea and optic disc, Biomed 37. Qiu Y, Wang Y, Yan S, Tan M, Cheng S, Liu H, Zheng B,
Signal Process Control 91:101, 2018. An initial investigation on developing a new method to
26. Satapathy SC, Udgata SK, Biswal BN, in Proc Int Conf predict short-term breast cancer risk based on deep
Frontiers of Intelligent Computing: Theory and Applica- learning technology, in Med Imag Comput-Aided Diag-
tions (FICTA), Springer Science & Business Media, 2012. nosis: Int Soc Opt Photonics 97:521, 2016.
27. Rasti R, Teshnehlab M, Phung SL, Breast cancer diag- 38. Jiang F, Liu H, Yu S, Xie Y, Breast mass lesion classi¯-
nosis in DCE-MRI using mixture ensemble of convolu- cation in mammograms by transfer learning, in Proc 5th
tional neural networks, Pattern Recog 381:390, 2017. Int Conf Bioinformatics and Computational Biology,
28. Rezaeilouyeh H, Mollahosseini A, Mahoor MH, Micro- ACM, Vol. 59, p. 62, 2017.
scopic medical image classi¯cation framework via deep 39. Suzuki S, Zhang X, Homma N, Ichiji K, Sugita N,
learning and shearlet transform, J Med Imag 44:501, Kawasumi Y, Ishibashi T, Yoshizawa M, Mass detection
2016. using deep convolutional neural network for mammo-
29. Sharma K, Preet B, Classi¯cation of mammogram images graphic computer-aided diagnosis, in 55th Annual Conf
by using CNN classi¯er, in 2016 Int Conf IEEE, Advances IEEE Society of Instrument and Control Engineers of
in Computing, Communications and Informatics Japan (SICE), pp. 1382–1386, 2016.
(ICACCI), Vol. 2743, p. 2749, 2016. 40. Huai L and Ray Liu KJ, Multi-modular neural network
30. Jiao Z, Gao X, Wang Y, Li J, A deep feature based for breast cancer detection, Adv Signal Process Technol
framework for breast masses classi¯cation, Neuro- Soft Comput 140:158, 2000.
computing 221:231, 2016. 41. SC Fok, Ng EYK and Tai K, Early detection and visu-
31. Kooi T, Litjens G, van Ginneken B, Gubern-Merida alization of breast tumor with thermogram and neural
A, Sanchez CI, Mann R, den Heeten A, Karssemeijer network, J Mech Med Biol 185–195, 2002.
N, Large scale deep learning for computer aided detection

1950034-9

You might also like