Professional Documents
Culture Documents
Biomedical Signal Processing and Control: Sameena Pathan, K. Gopalakrishna Prabhu, P.C. Siddalingaswamy
Biomedical Signal Processing and Control: Sameena Pathan, K. Gopalakrishna Prabhu, P.C. Siddalingaswamy
a r t i c l e i n f o a b s t r a c t
Article history: A clinically oriented Computer-Aided Diagnostic (CAD) system is of prime importance for the diagnosis
Received 17 September 2018 of melanoma, since the deadly disease is associated with high morbidity and mortality. Unfortunately,
Received in revised form 16 January 2019 the development of CAD tools is hampered by several issues, such as (i) smooth boundaries between the
Accepted 9 February 2019
lesion and the surrounding skin, (ii) subtlety of features between the melanoma and non-melanoma skin
lesions, and (iii) lack of reproducibility of CAD systems due to complexity. The proposed system aims to
Keywords:
address the aforementioned issues. First, the lesion regions are localized by incorporating chroma based
Benign lesions
deformable models. Second, the lesion patterns are analyzed to detect various dermoscopic criteria.
Dermoscopy
Malignant lesions
Further, a robust ensemble architecture is developed using dynamic classifier selection techniques to
Melanocytic nevi detect malignancy. Quantitative analysis is performed on two diverse datasets (ISBI and PH2) achieving
Pigment network an accuracy of 88% and 97%, sensitivity of 95% and 97% and specificity of 82% and 100% for ISBI and PH2
Shape datasets respectively.
Color © 2019 Elsevier Ltd. All rights reserved.
Texture
1. Introduction benign Pigmented Skin Lesions (PSL) [4]. The construction of the
classifier mainly depends on the features selected based on the
The malignant tumor due to abnormal growth of melanocytes analysis strategy.
is known as melanoma. It can originate in any part of the body A complete CAD tool for melanoma diagnosis mainly requires 4
and is the most lethal form of skin cancer [1]. Although, melanoma major steps: lesion segmentation, feature extraction, selection and
accounts for less than 20% of all the skin cancer cases, it has a lesion classification. Although, these steps are performed sequen-
high mortality rate. The prevalence of cutaneous melanoma has tially, the major concern is that the segmentation accuracy has a
increased rapidly over the past 30 years in the Caucasian popula- major decisive influence on the lesion diagnosis. Irrespective of the
tion [2]. Dermoscopy is an in vivo assessment technique performed surrounding skin, the Region of Interest (ROI) characteristics pre-
using the dermoscope for observation of the structures inside the dominantly effects the lesion classification [5]. However, due to
epidermis and dermis. It is an essential tool for dermatologists, the smooth boundaries between the lesion and normal skin, hair
plastic surgeons attempting early diagnosis of skin related disor- artifacts, and variabilities in skin texture, the accuracy of classifier
ders [3]. However, dermoscopic analysis suffers from significant is adversely affected. Artifacts that might change the morphology
interobserver disagreement as small observations go unnoticed. of the dermoscopic image, thereby hindering development of CAD
The investigation of computerized dermoscopy pictures can be system for lesion diagnosis, need to be eliminated. The Fig. 1 pro-
performed using various methodologies, either determined by the vides an illustration of the dermoscopic image analysis.
restorative master viewpoint, or driven by computerized knowl-
edge. Computer aided diagnostic process involves extraction of
huge number of low-level elements from the dermoscopic images 1.1. Related work
which allow the discrimination between malignant melanoma and
Several approaches have been proposed to design CAD systems
for melanoma diagnosis [6–17]. These approaches can be divided
∗ Corresponding author.
into 2 categories based on the features used for predicting the lesion
E-mail addresses: sameena.pathan.k@gmail.com (S. Pathan),
type (i) Dermoscopic based approaches and, (ii) Pattern recognition
gk.prabhu@manipal.edu (K. Gopalakrishna Prabhu), pcs.swamy@manipal.edu based approaches. The CAD tools that mimic the visual percep-
(P.C. Siddalingaswamy). tion of the dermatologist belong the former category. This involves
https://doi.org/10.1016/j.bspc.2019.02.013
1746-8094/© 2019 Elsevier Ltd. All rights reserved.
60 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72
the extraction of dermoscopic structures such as pigment net- the total dermoscopic score for classification of skin lesions in con-
work, blue-white veil, dots and globules etc. On the other hand, junction with Geodesic Active Contours (GAC). However, stopping
the extraction of subtle information from the region of interest criteria constraints resulted in misclassification. In the recent years,
belong to the latter category. Among the significant works reported several researchers have proposed deep learning architectures for
in literature, an interesting work was reported by Celebi et al. [6], skin lesion classification [15–16]. Yu et al. [15] employed deep
wherein color and texture features were extracted from the seg- networks in conjunction with residual networks of more than 50
mented lesion area for classification. Similarly, Schaefer et al. [7] layers for segmentation and classification of melanoma skin lesions.
used an ensemble of color, shape and texture features. A classi- Although, residual networks were used to overcome the network
fication accuracy of 83.4% was achieved by Barata et al. [8] by degradation problem, the rate of correct classification of melanoma
employing color constancy models. The method proposed in [8] lesions was reported to be 50.7%. Codella et al. [16] employed an
concentrates mainly on the clinical aspects of color in dermoscopic ensemble of machine learning and deep learning frameworks. Fully
images, while missing out the important role of shape features. convolution U-net is used for obtaining segmented masks, further
Pennisi et al. [9] proposed an automatic skin lesion segmen- hand-crafted features, sparse coding techniques and deep resid-
tation algorithm by employing Denulay triangulation. Although, ual networks are used for melanoma recognition. Although, deep
the segmentation approach produced better results in case of learning architectures have reportedly improved the classification
benign lesions, the segmentation accuracy was relatively lesser accuracies by learning from a large amount of data, optimizing the
for malignant lesions. Further, a classification sensitivity of 93.5% network parameters to reduce the computational complexity is a
and specificity of 87.1% was obtained by extracting geometri- challenging issue. Despite the significant advances, there exists a
cal and color features. A classification accuracy of 90.3% was margin scope in improvement for lesion segmentation and classi-
achieved by Maglogiannis et al. [10] by extracting features from fication.
dots/globules. The dots/globules were segmented using multi-
resolution approach. However, the classification accuracy could
be enhanced by combining the dots specific features with region
1.2. Contributions
based descriptors. A recent study by Barata et al. [11] employed
color and texture features while missing out the most important
The proposed system focusses on the development of a dermo-
shape features. Tejeddin et al. [12] classified malignant lesions from
scopic inspired framework for distinguishing benign and malignant
benign using a set of shape, color and texture features extracted
skin lesions. The proposed method differs from the existing related
from the lesion’s peripheral region. A pseudo-polar space was used
works reported in literature [6–17] in the following ways.
for mapping of peripheral region pixels. Jukic et al. [13] employed
tensor decomposition for analyzing color data from autofluores-
cent RGB images, since color of the lesion is an important feature
• The system takes into account the global and dermoscopic fea-
used for discriminating malignant melanocytic lesions. Oliveira
et al. [14] build an ensemble classification model using feature sub- tures for differentiating the pigmented skin lesions.
• The chroma based deformable model takes into account the
set selection from shape, color and texture features. However, the
method was computationally intensive and lacked pre-processing chrominance properties of the lesion, thus is robust in the pres-
and ROI extraction algorithms. In contrast to the handcrafted fea- ence of background variabilities.
• An ensemble of dynamic classifier architectures employed have
ture extraction approaches, deep learning based approaches aim
to automatically extract features in a hierarchical fashion. Kasmi proven to achieve better accuracy in identifying benign and
et al. [17] implemented the ABCD dermoscopic rule for computing malignant lesions in contrast to the computationally intensive
deep learning approaches.
S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72 61
The organization of the paper is as follows. In Section 2 the function ∅ (x) , the unknown variable C is replaced by ∅ (x) and the
overview of the proposed system is presented. The subsections in energy function is re-written as given in (2).
the Section 2 briefly describe the methodology used for building a
CAD system. The results obtained are briefed in Section 3 followed E = 1 (I (x) − c1)2 H(∅ (x))dx
by conclusion in Section 4.
2. Materials and methods + 2 (I (x) − c2)2 (1 − H(∅ (x))dx + ı (∅ (x)) |∇ ∅ (x) |dx
1
2
z
2.1. Dataset
H (∅) = 1+ arctan (3)
2 ε
Two datasets namely PH2 [18] and ISBI 2016 [19] were used
for the extraction of features and classification. The PH2 datasets and
consists of 200 pigmented skin lesions, (160 benign and 40 malig- 1 ε
nant lesions) and the ISBI dataset consists of 1279 pigmented skin ı (∅) =
ε2 + z 2
lesions, (1033 benign lesions and 246 malignant lesions).
Here, c1 and c2 are obtained by differentiating equation (2) with
2.2. Pre-processing and lesion localization (∅ (x)) fixed.
respect to c1 and c2 keeping
H(∅)I(x) (1−H(∅))I(x)
Correspondingly, c1 = and c2 =
The presence of hair greatly affects the accuracy of the segmen- H(∅) (1−H(∅))
tation algorithm. Thus, the primary step for development of a CAD The proposed energy function is defined as given in (4)
tool is the detection and exclusion of dermoscopic hair. Scalar or
vector dermoscopic images may be used for pre-processing. Sev- E Total = E Global + E Chroma (4)
eral hair detection and exclusion methods have been discussed in
the past [20–22]. Moreover, these techniques were designed based E Global is calculated from the grayscale dermoscopic image.
on the assumption that the color of the hair is much darker than the E Chroma is calculated from the CIE L*a*b color space using the
skin and lesion. However, most of the lesions are brown or black chrominance component of the dermoscopic image I(x, y). The CIE
in color due to the localization of melanin in the upper and lower L*a*b is more convenient than the tristimulus values with respect
epidermis [23]. This signifies the fact that, there exists a need to to its conceptual relationship to visual perception, and additionally,
include attributes specific to the properties of dermoscopic hair, it also provides the means to measure the differences between any
in the hair detection algorithm. In order to address this issue, the two colors. The color difference between the colors is calculated
hair detection procedure takes into account the width, magnitude using co-ordinate geometry as given in (5).
and direction of the hair shafts. As these properties are unique to 1
the dermoscopic hair, overlap between the lesion features and hair 2 2 2
E = [L + a + b ] 2 (5)
features is eliminated. Directional Gaussian filters are used iden-
tify the hair artifacts. A group of 16 directional filters were used for where, L is the lightness component, a and b indicate the chroma.
detection of hair using the luminance component of the perceptu- The lightness attribute L gives the measure of grey-scale from black
ally uniform CIE L*a*b color space. A detailed explanation of the hair to white. As the perceived color of the object mainly depends on
detection method used is given in [24]. the nature of the illuminating light, in some circumstances the dif-
The lesions are localized using the chroma based deformable ferences between the two colors may be considered in terms of
models. The performance of the geometric deformable models the differences in chroma (a, b). The Fig. 2 provides an illustration
mainly relies on the initial conditions used and the evolution of the of color space selection. The skin lesions should be recognizable
speed function. Color plays a very important role in dermoscopy, by their color dissimilarities from normal skin irrespective of the
since the color of melanin mainly depends on extent of the local- capturing device. Thus, in the proposed method the chroma com-
ization in the skin. Thus, the segmentation approach is proposed by ponent given in (6) is considered for lesion segmentation.
exploiting the aforementioned domain knowledge of skin lesions
by considering the chroma component, rather than the conven- C (x, y) = a2 + b2 (6)
tional RGB channels.
The speed function is defined to segment the lesions in der- Assuming that the intensity values in the lesion region follow a
moscopic images with variabilities in intensities. The basic idea of Gaussian distribution, C(x, y) is represented as given in (7)
Chan-Vese is to partition the given image I (x) into foreground and
background [25]. The deformation of the contour towards the lesion − (C (x, y) − L )2
CG (x, y) = exp (7)
boundaries is through minimization of the energy function E. The 2L 2
Chan-Vese energy function is given as (1).
The statistical values L , L are the mean and standard devi-
E = 1 (I (x) − c1)2 dx + 2 (I (x) − c2)2 dx + length (C) ation of the lesion region respectively. The statistical values L ,
and L are computed approximately by binarizing the dermoscopic
image by performing clustering in the chrominance color space. The
+ vArea (C) (1)
Euclidean distance is used as a metric to perform the classification
where, 1 , 2 , , v are fixed parameters as given in [18]. Using of pixels. Let (a0 , b0 ) and (a1 , b1 ) bethe pixel values of the ran-
the level set to represent curve C, the zero level set of a Lipschitz domly selected centroids, and ai , bj be the corresponding pixel
62 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72
d(H(∅)
= ı (∅)
d∅
dı (∅ (x)) |∇ ∅ (x) | ∇∅
= div
d∅ ∇ ∅
2
E Chroma = ı (∅) − 1 ( Gradient(C G (x, y)) − c1)
∇∅
+2 ( Gradient(C G (x, y)) − c2) + div
2
(10)
∇ ∅
Fig. 2. Illustration of color space selection.
∂ C G (x,y) 2 ∂C 2
where, Gradient( C G (x, y)) = + G (x,y)
∂x ∂y
∂ C G (x,y) ∂ C G (x,y)
and are the gradients in x and y direction
∂x ∂y
The div |∇ ∅
∇ ∅|
known as curvature is given by (11).
2 2
∇∅ fyy fx + fxx fy − 2fx fy fxy
div = (11)
∇ ∅ 2 2
3
(fx + fy ) 2
f x (x, y) = (f (x + 1, y) − f (x − 1, y))*0.5
f xx (x, y) = f (x + 1, y) + f (x − 1, y) − 2f (x, y)
f xy (x, y) = (f (x + 1, y + 1) − f (x + 1, y − 1) − f (x − 1, y + 1)
Fig. 3. Illustration of clustering.
+f (x − 1, y − 1)) *0.25
values of C (x, y), then using the following rule the dermoscopic
image I(x, y) is binarized.
The curve is evolved as given in (12).
0 if d0 (a, b) ≤ d1 (a, b)
I ’ (x, y) = (8) ∅n+1 = ∅n + deltaT * E Total (12)
1 , otherwise
The value of deltaT known as step size as given by Chan Vese
should be between 0.1 to 0.9. The curve evolution is stopped when
where, d0 (a, b) = (ai − a0 ) − (bi − b0 ) and
the difference between the last two iterations is less than the
d1 (a, b) = (ai − a1 ) − (bi − b1 ) . The Fig. 3 illustrates the step size as given in (13), or when maximum iterations have been
binary image obtained after clustering using the above method. achieved.
Further, the binarized image is fused with the dermoscopic n+1 n
image to calculate the statistical values of L , and L . Using L , ∅ − ∅ ≤ deltaT (13)
and L , C G (x, y) is computed as explained below. The evolution
of the initial contour is attracted towards the lesion boundaries by The lesions segmented are further subjected to extraction of
taking the gradient of the Gaussian distribution C G (x, y). Thus, the features. The Fig. 4 provides an illustration of results of lesion seg-
total energy function is the summation of energy from the grayscale mentation. A mean segmentation overlap error of 11.5% and 7.2%
image and from the gradient of the Gaussian distribution C G (x, y). was obtained for ISBI 2016 and PH2 datasets respectively.
S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72 63
PL 2
CI = (14)
4AL
P L and AL indicate the perimeter and area of the lesion. Thus a total
of 3 shape features were computed.
Fig. 6. Illustration of Pigment Network detection (a) Original Image (b) Pigment
network detected.
2.3.3. Texture features A non-parametric test based on Wilcoxon Rank Sum statistics
Tamura’s features characterize coarse texture, inhomogeneous [29] is conducted to test the difference in median values for the
contrast and irregularity in the texture patterns [27]. Since malig- features extracted for benign and malignant classes. A statistical
nant lesions exhibit the aforementioned texture patterns, Tamura’s significance p ≤ 0.05, is used to test the null hypothesis.
features are computed for the region of interest.
Discrete Wavelet Transform (DWT) features – The energy and Ho. The extracted features for benign and malignant lesions have
entropy measures from the coefficients obtained by DWT are equal medians.
computed for each of the Haar wavelet sub-bands obtained by The null hypothesis is tested against the alternative that they are
a-three-level decomposition for RGB, HSV, CIEL*a*b and CIEL*u*v not at 5% significance level. Thus, a set of 401 p-values were com-
color spaces. Since it is a three level decomposition, 10 wavelet puted for PH2 dataset. The features having p ≤ 0.05 were selected.
energy co-efficients and one entropy co-efficient are obtained for Among the 401, features extracted, 296 features had p-values <0.05
each channel, thus resulting in 132 DWT features. hence, these features were found to be statistically significant. The
GLCM features – Gray-level co-occurrence matrix (GLCM) is a sta- p-values were computed for the features extracted from the PH2
tistical approach which considers the spatial relationship between dataset, and the same selected features (296 features) were used
pixels. From the normalized co-occurrence matrix, 12 statistical from the ISBI datasets for classification. It can be observed from
measures were extracted from the image. These measures are con- the ISBI dataset, that 81% samples are benign and 19% samples
trast, dissimilarity, energy, entropy, homogeneity, sum average, are malignant. Since the dataset is imbalanced, with fewer number
sum entropy, difference entropy, difference variance, sum variance, of diseased cases, an oversampling technique termed as Adaptive
auto-correlation, and correlation [28]. These features were calcu- Synthetic Approach (ADAYSN) is used to synthetically generate
lated for CIEL*a*b, RGB, CIEL*u*v and HSV color spaces thus resulting samples from the minority class. ADASYN techniques are basically
in 12 × 3 × 4 = 144 features. Thus, a total of 279 texture features applied to medical problems, wherein the number of diseased cases
were computed. are comparatively low relative to the non-diseased cases [30]. The
S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72 65
Bold values indicate the best results for each of the methods used.
si = xi + (xi − xzi ) * (17)
Classification forms the last block of a computer aided diagnostic Bold values indicate the best results for each of the methods used.
system. The literature reports various classifiers used for diagno-
sis of skin lesions [31]. Three classifiers namely SVM, ensemble of Among the 401 features, features such as contrast, direction-
decision stump trees and ensemble of AdaBoost classifiers with ality, and color variance for RGB channel and color scores for light
dynamic selection of classifiers technique are used. In ensemble brown, white and blue-grey were omitted. Since contrast was mea-
approach, a number of classifiers are tested, and the most appropri- sured as a part of GLCM texture feature, it was omitted from
ate one is chosen for a given test sample. Different classifiers usually Tamura texture feature. Similarly, color variance was also a part
make different errors on different samples, which means that, by of statistical values computed from RGB color space and hence
combining classifiers, we can put together an ensemble that makes it was omitted. Further, the computation of texture directional-
more accurate decisions [32]. In order to have classifiers with dif- ity was computationally intensive, hence it was omitted. A close
ferent errors, it is advisable to create diverse classifiers and group look at the color scores indicated that, light brown was the most
them into what is known as an Ensemble of Classifiers (EoC). If one common color present in the skin lesions irrespective of the type
classifier from an EoC can correctly classify a given pattern, then lesion (benign/malignant). Similarly, white and blue-grey colors
this EoC is considered to be able to classify this pattern. Intuitively, were found to be very rare. Thus, these three scores were ignored.
the more diverse the EoC, the better the classification. The major This lead to the reduction of 13 features from a set of 401 features
objective of the classification is to select those classifiers which leading to 388 features.
might be capable to correctly classify a given pattern. This is done
through a dynamic fashion, since different patterns might require
3. Results and discussions
different ensembles of classifiers. Thus, the method is known as
dynamic ensemble selection. The advantage of dynamic ensemble
3.1. Results of classification using SVM
selection is that we distribute the risk of over-generalization by
choosing a group of classifiers instead of one individual classifier
The following two set-ups were used for evaluating the classifi-
for a test pattern.
cation ability for the features extracted.
In the proposed work, two ensemble models are created. (i) The
Set-up 1: A hold out set of 20% was used for testing and 80%
first model is built by creating an ensemble of decision stump trees
was used for training the SVM classifier. The testing process was
by bagging, (ii) The second model is built by creating an ensem-
carried out in three iterations, with a different hold-out set used
ble of AdaBoost classifiers built using weak learners by bagging.
during each iteration. The results reported are the average per-
The dynamic ensemble selection methods such as Overall Local
formance values of the classifiers for the three iterations. The
Accuracy (OLA), Local Class Accuracy (LCA), A-Priori, A-Posterior,
performance metrics are sensitivity, specificity and accuracy as
KNORA-E and KNORA-U [33] are employed for the selection of the
given in (18)–(20). Sensitivity (SE) indicates the rate of correct clas-
classifiers for both the models.
sification of malignant lesions. Specificity (SP) indicates the rate of
First Model Decision Stump Trees – The number of decision
correct classification of benign lesions. Accuracy (ACC) indicates the
stump trees are decided based on the number of input features.
overall correct classification rate of benign and malignant lesions.
Thus, an ensemble of 401 decision stump trees are built using bag-
Table 1 and 2 depicts the classification performance on PH2 and
ging. The maximum depth of the tree is two. Once an ensemble of
ISBI datasets.
decision stump trees are built using bagging, the aforementioned
dynamic selection methods are used to select an ensemble of classi- TP
SE = (18)
fiers for a given test pattern. Further, the chosen classifier is applied (FN + TP)
to predict the label for the given test pattern from the ensemble of TN
classifiers selected. SP = (19)
(FP + TN)
Second Model - Similar to the first model, the decision stump
trees are replaced by AdaBoost of weak learners and further the TN + TP
ACC = (20)
dynamic selection models are used to predict a test label. (FN + FP + TN + TP)
The two models are applied to the features extracted from PH2 True Positive (TP): The classifier correctly predicts that the lesion is
and ISBI 2016 databases. The following section provides the results malignant.True Negative (TN): The classifier correctly predicts that
obtained using the aforementioned classification techniques. The the lesion is benign.False Positive (FP): The classifier predicts that
classification techniques are applied to 3 feature sets: the lesion is benign when it is malignant.False Negative (FN): The
classifier predicts that the lesion is malignant when it is benign.
i) 401 features extracted. Set-up 2 – To estimate the general performance of the classifica-
ii) 296 features that were found to be statistically significant tion model k-fold cross validation scheme with k = 10 is applied for
iii) 388 features. each set of the data. Each dataset is divided into ten equal sub-parts,
66 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72
Table 3 Table 5
Performance parameters for PH2 dataset using k-fold cross validation. Performance parameters for evaluating the generalization ability.
Bold values indicate the best results for each of the methods used.
Table 6
Performance parameters for evaluating the generalization ability using k-fold cross
Table 4 validation.
Performance parameters for ISBI dataset using k-fold cross validation.
Approach SP SE ACC
No. of features SP SE ACC i) 0.76 0.81 0.79
401 0.72 0.85 0.78 ii) 0.85 1 0.93
388 0.84 0.72 0.78
296 0.71 0.82 0.76
Bold values indicate the best results for each of the methods used.
(ii) The classifier is trained on ISBI dataset and tested on PH2
dataset.
Table 7
Performance parameters obtained using Decision Stump trees on ISBI dataset.
Normal 0.75 0.79 0.77 0.58 0.79 0.68 0.65 0.74 0.70
Ensemble 0.78 0.88 0.83 0.59 0.79 0.69 0.66 0.74 0.70
OLA 0.78 0.87 0.83 0.70 0.88 0.79 0.68 0.84 0.76
LCA 0.69 0.90 0.80 0.70 0.89 0.79 0.68 0.86 0.77
A-Priori 0.71 0.85 0.78 0.61 0.87 0.74 0.66 0.83 0.75
A-Posterior 0.51 0.97 0.74 0.61 0.93 0.76 0.63 0.89 0.76
KNORA-E 0.79 0.93 0.86 0.70 0.91 0.80 0.73 0.87 0.80
KNORA-U 0.79 0.90 0.85 0.58 0.82 0.70 0.63 0.80 0.72
Bold values indicate the best results for each of the methods used.
Table 8
Performance parameters obtained using Decision Stump trees on ISBI dataset using k-fold cross-validation.
Normal 0.52 0.82 0.67 0.67 0.63 0.65 0.66 0.67 0.66
Ensemble 0.53 0.82 0.67 0.61 0.73 0.67 0.65 0.70 0.68
OLA 0.59 0.95 0.77 0.67 0.95 0.81 0.66 0.93 0.79
LCA 0.61 0.95 0.78 0.69 0.95 0.90 0.70 0.94 0.81
A-Priori 0.56 0.94 0.75 0.62 0.94 0.78 0.66 0.91 0.78
A-Posterior 0.54 0.98 0.76 0.58 0.99 0.78 0.60 0.97 0.78
KNORA-E 0.61 0.95 0.78 0.70 0.96 0.83 0.78 0.95 0.82
KNORA-U 0.54 0.88 0.71 0.61 0.91 0.76 0.65 0.85 0.74
Bold values indicate the best results for each of the methods used.
Table 9
Performance parameters obtained using AdaBoost on ISBI dataset.
Normal 0.75 0.79 0.77 0.81 0.83 0.82 0.74 0.78 0.76
Ensemble 0.78 0.88 0.83 0.80 0.84 0.82 0.80 0.83 0.82
OLA 0.69 0.87 0.83 0.74 0.88 0.81 0.79 0.86 0.83
LCA 0.71 0.90 0.80 0.71 0.91 0.81 0.74 0.92 0.83
A-Priori 0.51 0.85 0.78 0.73 0.83 0.78 0.73 0.77 0.75
A-Posterior 0.79 0.97 0.74 0.57 0.98 0.77 0.57 0.96 0.77
KNORA-E 0.79 0.93 0.86 0.82 0.95 0.88 0.82 0.91 0.87
KNORA-U 0.79 0.90 0.85 0.81 0.87 0.84 0.81 0.88 0.84
Bold values indicate the best results for each of the methods used.
Table 10
Performance parameters obtained using AdaBoost on ISBI dataset using k-fold cross validation.
Normal 0.76 0.79 0.77 0.74 0.80 0.77 0.73 0.77 0.65
Ensemble 0.79 0.85 0.85 0.79 0.84 0.81 0.79 0.83 0.81
OLA 0.75 0.91 0.83 0.76 0.90 0.82 0.75 0.97 0.82
LCA 0.73 0.90 0.84 0.73 0.96 0.83 0.74 0.94 0.84
A-Priori 0.74 0.85 0.80 0.75 0.81 0.78 0.74 0.85 0.79
A-Posterior 0.57 0.98 0.78 0.58 0.99 0.78 0.59 0.98 0.78
KNORA-E 0.81 0.93 0.87 0.82 0.93 0.96 0.82 0.93 0.87
KNORA-U 0.79 0.87 0.83 0.79 0.87 0.83 0.79 0.78 0.83
Bold values indicate the best results for each of the methods used.
Table 11
Performance parameters obtained using decision stump trees on PH2 dataset.
Normal 0.88 0.88 0.88 0.97 0.80 0.93 0.90 1.00 0.93
Ensemble 0.91 0.75 0.88 1.00 0.80 0.95 1.00 0.80 0.95
OLA 0.88 0.62 0.82 0.90 1.00 0.93 0.97 0.60 0.88
LCA 0.88 0.62 0.82 0.93 0.90 0.93 0.97 0.70 0.90
A-Priori 0.91 0.88 0.90 0.90 0.80 0.88 0.90 0.70 0.85
A-Posterior 0.88 0.75 0.85 0.97 0.80 0.93 0.97 0.80 0.93
KNORA-E 0.88 0.88 0.88 0.97 1.00 0.97 1.00 0.80 0.95
KNORA-U 0.91 0.75 0.88 1.00 0.80 0.95 1.00 0.80 0.95
Bold values indicate the best results for each of the methods used.
68 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72
Table 12
Performance parameters obtained using decision stump trees on PH2 dataset using k-fold cross-validation.
Normal 0.92 0.59 0.85 0.95 0.67 0.90 0.95 0.72 0.90
Ensemble 0.96 0.70 0.91 0.97 0.67 0.91 0.96 0.65 0.90
OLA 0.87 0.80 0.85 0.95 0.77 0.91 0.94 0.75 0.81
LCA 0.92 0.75 0.88 0.94 0.77 0.91 0.93 0.76 0.90
A-Priori 0.90 0.65 0.86 0.88 0.62 0.84 0.92 0.80 0.90
A-Posterior 0.91 0.72 0.88 0.93 0.72 0.89 0.93 0.80 0.91
KNORA-E 0.93 0.77 0.90 0.85 0.77 0.91 0.93 0.77 0.90
KNORA-U 0.96 0.70 0.91 0.97 0.67 0.91 0.96 0.75 0.92
Bold values indicate the best results for each of the methods used.
Table 13
Performance parameters obtained using AdaBoost on PH2 dataset.
Normal 1.00 1.00 1.00 0.97 1.00 0.97 0.97 0.80 0.93
Ensemble 0.94 0.75 0.90 0.97 0.80 0.93 1.00 0.80 0.95
OLA 1.00 0.75 0.95 0.93 1.00 0.95 0.97 0.90 0.95
LCA 0.97 0.75 0.93 0.93 1.00 0.95 0.97 0.90 0.95
A-Priori 0.94 0.75 0.90 0.97 1.00 0.97 1.00 1.00 1.00
A-Posterior 0.91 0.62 0.85 1.00 0.70 0.93 0.97 0.70 0.90
KNORA-E 0.94 0.75 0.90 1.00 0.90 0.97 1.00 0.90 0.97
KNORA-U 0.94 0.75 0.90 1.00 0.80 0.95 1.00 0.80 0.95
Bold values indicate the best results for each of the methods used.
Table 14
Performance parameters obtained using AdaBoost on PH2 dataset using k-fold cross validation.
Normal 0.98 0.70 0.92 0.94 0.75 0.90 0.95 0.85 0.92
Ensemble 0.98 0.75 0.93 0.97 0.80 0.84 0.96 0.77 0.92
OLA 0.96 0.70 0.91 0.95 0.85 0.83 0.95 0.80 0.92
LCA 0.96 0.70 0.91 0.95 0.82 0.92 0.95 0.77 0.92
A-Priori 0.94 0.72 0.90 0.96 0.77 0.92 0.93 0.97 0.88
A-Posterior 0.96 0.70 0.91 0.95 0.77 0.89 0.97 0.70 0.92
KNORA-E 0.98 0.80 0.93 0.97 0.82 0.84 0.95 0.72 0.92
KNORA-U 0.98 0.75 0.93 0.97 0.80 0.93 0.93 0.77 0.92
Bold values indicate the best results for each of the methods used.
Table 15
Summary of the results reported in Tables 7 to 14.
Normal 1 1 1
Ensemble 1 1 1 1 1
OLA 1 1
LCA 1 2 1
A-Priori 1 1 1
A-Posterior 1
KNORA-E 2 3 3 3 3 2
KNORA-U 2 1 2 1 1 1
Table 16
Performance parameters for evaluating the generalization ability using approach (i).
Bold values indicate the best results for each of the methods used.
S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72 69
Table 17
Performance parameters for evaluating the generalization ability for k-fold cross-validation using approach (i).
Bold values indicate the best results for each of the methods used.
Table 18
Performance parameters for evaluating the generalization ability using approach (ii).
Bold values indicate the best results for each of the methods used.
Table 19
Performance parameters for evaluating the generalization ability for k-fold cross validation using approach (ii).
Bold values indicate the best results for each of the methods used.
Fig. 8. ROC curve for (a) Decision Stump trees (b) Ensemble of AdaBoost with KNORE E dynamic selection for ISBI dataset.
70 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72
Fig. 9. ROC curve for (a) Decision Stump trees (b) Ensemble of AdaBoost with KNORE E dynamic selection on PH2 dataset.
Fig. 10. Illustration of frequency of best performance for benign and malignant lesion classification (a) Using Decision Stump Trees (b) Using AdaBoost of weak learners.
Fig. 11. ROC curve for (a) Decision Stump trees (b) Ensemble of AdaBoost with KNORE E dynamic selection for approach (i).
Fig. 12. ROC curve for (a) Decision Stump trees (b) Ensemble of AdaBoost with KNORE E dynamic selection for approach (ii).
Fig. 13. Frequency of Best Performance (a) Based on dynamic selection of ensemble classifiers (b) Based on Features.
3.3. Transfer learning paradigm tion net was passed to an Average pooling layer followed by a 2
neuron Softmax output. This output was trained on 400 Cat & Dog
A transfer learning paradigm approach is adopted to compare Images obtained from Kaggle [36], with the Xception model made
the efficacy of the proposed model with that of the convolution non-trainable. The trained model was tested on 256 × 256 cropped
neural network model. The proposed model takes into account der- ISBI skin lesion images. The Xception network resulted in a test
moscopic inspired features for designing the classification model, accuracy of 80.67%. It has been observed that the proposed model
whereas the transfer learning strategy is built using a pre-trained performs better in comparison to the Xception model built using
Xception Network that was trained on the Image Net dataset [35]. the transfer learning strategy.
Keras Library on top of Tensorflow was used. Output from the Xcep-
72 S. Pathan, K. Gopalakrishna Prabhu and P.C. Siddalingaswamy / Biomedical Signal Processing and Control 51 (2019) 59–72