You are on page 1of 14

applied

sciences
Article
Classification of Heart Sound Signal Using
Multiple Features
Yaseen, Gui-Young Son and Soonil Kwon *
Department of Digital Contents, Sejong University, Seoul 05006, Korea; gikian86@gmail.com (Y.);
sgy1017@naver.com (G.-Y.S.)
* Correspondence: skwon@sejong.edu; Tel.: +82-2-3408-3847

Received: 10 October 2018; Accepted: 20 November 2018; Published: 22 November 2018 

Abstract: Cardiac disorders are critical and must be diagnosed in the early stage using routine
auscultation examination with high precision. Cardiac auscultation is a technique to analyze and
listen to heart sound using electronic stethoscope, an electronic stethoscope is a device which provides
the digital recording of the heart sound called phonocardiogram (PCG). This PCG signal carries
useful information about the functionality and status of the heart and hence several signal processing
and machine learning technique can be applied to study and diagnose heart disorders. Based on
PCG signal, the heart sound signal can be classified to two main categories i.e., normal and abnormal
categories. We have created database of 5 categories of heart sound signal (PCG signals) from various
sources which contains one normal and 4 are abnormal categories. This study proposes an improved,
automatic classification algorithm for cardiac disorder by heart sound signal. We extract features
from phonocardiogram signal and then process those features using machine learning techniques
for classification. In features extraction, we have used Mel Frequency Cepstral Coefficient (MFCCs)
and Discrete Wavelets Transform (DWT) features from the heart sound signal, and for learning and
classification we have used support vector machine (SVM), deep neural network (DNN) and centroid
displacement based k nearest neighbor. To improve the results and classification accuracy, we have
combined MFCCs and DWT features for training and classification using SVM and DWT. From our
experiments it has been clear that results can be greatly improved when Mel Frequency Cepstral
Coefficient and Discrete Wavelets Transform features are fused together and used for classification
via support vector machine, deep neural network and k-neareast neighbor(KNN). The methodology
discussed in this paper can be used to diagnose heart disorders in patients up to 97% accuracy.
The code and dataset can be accessed at “https://github.com/yaseen21khan/Classification-of-Heart-
Sound-Signal-Using-Multiple-Features-/blob/master/README.md”.

Keywords: heart sound signal classification; discrete wavelets transform; mel frequency
cepstral coefficient

1. Introduction
Cardiovascular system is a perpetual source of data that sanctions soothsaying or distinguishing
among cardiovascular diseases. External constrains can lead to cardiac diseases that can cause sudden
heart failure [1]. Cardiovascular diseases are censorious and must be detected with no time delay [2].
Heart diseases may be identified by elucidating the cardiac sound data. The heart sound signal
characteristics may vary with respect to different kinds of heart diseases. A huge difference in the
pattern can be found between a normal heart sound signal and abnormal heart sound signal as their
PCG signal varies from each other with respect to time, amplitude, intensity, homogeneity, spectral
content, etc [3].

Appl. Sci. 2018, 8, 2344; doi:10.3390/app8122344 www.mdpi.com/journal/applsci


Appl. Sci. 2018, 8, 2344 2 of 14
Appl. Sci. 2018, 8, x FOR PEER REVIEW 2 of 14

Cardiovascular auscultation
Cardiovascular auscultation is is the
the principal
principal yet yet simple
simple diagnosing
diagnosing method method used used toto assess
assess andand
analyzethe
analyze theoperation
operationand andfunctionality
functionalityof ofthe
theheart.
heart. ItIt is
is aa technique
technique of of listening
listening to to heart
heart sound
sound withwith
stethoscope. The
stethoscope. The main
main source
source of of the
the generation
generation of of heart
heart sound
sound is is due
due to
to unsteady
unsteady moment
moment of of blood
blood
called blood
called blood turbulence.
turbulence. The opening
opening and and closing
closing of of atrioventricular
atrioventricular values, values, mitral
mitral andand tricuspid
tricuspid
valuescauses
values causesdifferential
differentialblood
bloodpressure
pressureand andhighhighacceleration
accelerationand andretardation
retardationofofblood bloodflowflow[4]. The
[4]. The
auscultation of cardiac disorders is carried out using an electronic stethoscope
auscultation of cardiac disorders is carried out using an electronic stethoscope which is indeed cost which is indeed cost
effectiveand
effective andnon-invasive
non-invasive approach
approach [3]. [3].
The The
digital digital
recordingrecording
of heart ofsound
heartwith soundthe with
help of the help of
electronic
electronic stethoscope is
stethoscope is called PCG [4]. called PCG [4].
Once the
Once the heart
heart sound
sound is is obtained,
obtained, itit cancan be be classified
classified via via computer
computer aided aided software
software techniques,
techniques,
these techniques
these techniques needneed more
more accurately
accurately defined
defined heart
heart sound
sound cycle for feature feature extraction process [3].
Manydifferent
Many differentautomated
automated classifying
classifying approaches
approaches have have been including
been used, used, including
artificial artificial neural
neural network
networkand
(ANN) (ANN)
hidden and hiddenmodel
Markov Markov modelMultilayer
(HMM). (HMM). Multilayer
perceptron-back perceptron-back
propagationpropagation
(MLP-BP),
(MLP-BP),
Wavelet Wavelet and
coefficients coefficients and neural
neural networks werenetworks
used for were used forof classification
classification heart sound signal.of heart Butsound
they
signal. But they have often resulted in lower accuracy
have often resulted in lower accuracy due to segmentation error [2]. due to segmentation error [2].
Toachieve
To achievebetter
betteraccuracy,
accuracy, in in this
this work,
work, wewe propose
propose SVM, SVM, centroid
centroid displacement
displacement based based
KNNKNN and
and DNN based classification algorithm using discrete wavelets
DNN based classification algorithm using discrete wavelets transform (DWT) and (MFCC) features. transform (DWT) and (MFCC)
features.
The The recognition
recognition accuracy can accuracy can be increased
be increased for both noisyfor both andnoisy
clearand
speech clearsignals,
speechusing
signals, using
discrete
discrete wavelets
wavelets transform transform fused together
fused together with MFCCs with MFCCsfeatures features
[5,6]. We [5,6]. We extracted
extracted MFCCs MFCCs
featuresfeatures
and
and discrete wavelets transform features for heart sound signals
discrete wavelets transform features for heart sound signals (normal and abnormal) and classified (normal and abnormal) and
classified
them usingthem using
support support
vector vectordeep
machine, machine,
neuraldeep network neuralandnetwork
centroid and centroid displacement
displacement based KNN,
based KNN, the highest accuracy we
the highest accuracy we achieved so far is 97.9%. achieved so far is 97.9%.
The beginning
The beginning of of this
this paper
paper describes
describes somesome background
background knowledgeknowledge about about heart
heart sound
sound signal
signal
processing and
processing and categories
categories of of heart
heart disorders
disorders in in brief
briefdetail,
detail,afterwards
afterwards the themethodology
methodology of of features
features
(MFCCs and DWT) and classifiers we used is discussed. In the
(MFCCs and DWT) and classifiers we used is discussed. In the experiment section we talk about experiment section we talk about
classificationof
classification offeatures,
features,detail
detailabout
aboutclassifiers,
classifiers,thethetools
toolswe weused,
used,and and database
database in in detail.
detail. InIn the
the end,
end,
wehave
we havediscussed
discussedour ourresults
resultsandand performance
performance evaluation
evaluation usingusing accuracy
accuracy and and averaged
averaged F1 score
F1 score and
and the last paragraph concludes
the last paragraph concludes the paper. the paper.

2.
2. Background
Background
The
Thehuman
humanheartheartisiscomprised
comprisedofoffourfourchambers.
chambers. TwoTwoof them
of themareare
called atrias
called andand
atrias theythey
make the
make
upper portion
the upper of theofheart
portion whilewhile
the heart remaining two chambers
remaining are called
two chambers ventricles
are called and they
ventricles andmake
theylower
make
portion of heart,
lower portion ofand
heart,blood
and enters
blood the heart
enters thethrough atrias and
heart through exit
atrias andthrough ventricles.
exit through Normal
ventricles. heart
Normal
sounds (Figure(Figure
heart sounds 1a is an1aexample of normal
is an example heart sound
of normal heartsignal)
soundare generated
signal) by closing
are generated byand opening
closing and
of the valves
opening of theof valves
heart. The heartThe
of heart. sound
heartsignals
soundproduced, are directly
signals produced, are related
directlyto opening
related and closing
to opening and
of the valves,
closing of theblood
valves,flow andflow
blood viscosity. ThereforeTherefore
and viscosity. during exercise
duringorexercise
other activities
or other which increases
activities which
the heart rate,
increases bloodrate,
the heart flowblood
throughflowthethrough
valves is increased
the valves isand hence heart
increased soundheart
and hence signal’s intensity
sound is
signal’s
increased and
intensity is during and
increased shock (low shock
during blood (low
flow)blood
the sound
flow) signal’s
the soundintensity
signal’sisintensity
decreased [2]. Figure[2].
is decreased 1f
also shows
Figure the shows
1 f also spectrum theof a PCG signal.
spectrum of a PCG signal.

(a) (b)
Figure 1. Cont.
Appl. Sci. 2018, 8, 2344 3 of 14
Appl. Sci. 2018, 8, x FOR PEER REVIEW 3 of 14

(c) (d)

(e) (f)
1. (a)
Figure 1. (a)AAnormal
normalheart
heartsound
sound signal; (b)(b)
signal; Murmur
Murmur in systole (MVP);
in systole (c) Mitral
(MVP); Regurgitation
(c) Mitral (MR);
Regurgitation
(d) Mitral
(MR); Stenosis
(d) Mitral (MS); (e) Aortic
Stenosis(MS); Stenosis
(e) Aortic (AS); (f)
Stenosis Spectrum
(AS); of a PCG
(f) Spectrum of asignal.
PCG signal.

The
The movements
movements of of heart valves generate
heart valves generate audible
audible soundssounds of of frequency
frequency rangerange lower
lower thanthan 22 kHz,
kHz,
commonly
commonly refer refertotoasas“lub-dub”
“lub-dub” TheThe “lub” is the
“lub” is first
the portion in the in
first portion heart
thesound
heart signal
soundand is denoted
signal and is
by (S1), it is generated from the closing of mitral and tricuspid valve.
denoted by (S1), it is generated from the closing of mitral and tricuspid valve. One complete cardiacOne complete cardiac cycle is the
heart
cycle is the heart sound signal wave which starts from S1 and ends to the start of next S1 and it as
sound signal wave which starts from S1 and ends to the start of next S1 and it is described is
one heartbeat.
described as oneThe duration,The
heartbeat. pitch, and shape
duration, pitch,ofand the heart
shapesound
of the shows us detail
heart sound showsabout us the different
detail about
conditions
the different of conditions
heart [4]. Closing
of heart of[4].
mitral valves
Closing ofismitral
followed by closing
valves of tricuspid
is followed by closingvalve of and usually
tricuspid the
valve
delay between
and usually thethis operation
delay between is this
20 tooperation
30 ms. Due is 20 totothe30left
ms.ventricle
Due to the contraction first,contraction
left ventricle the mitral valvefirst,
component occurs first followed by tricuspid valve component in the
the mitral valve component occurs first followed by tricuspid valve component in the signal. If the signal. If the duration between
these
duration twobetween
sound components
these two sound is in between
components 100 isto in200 ms, it is100
between called a split
to 200 ms, itand its frequency
is called a split andrangeits
lies from 40range
frequency to 200 Hzfrom
lies and40 it is
Hz considered
to 200 Hz fatal and it if is
theconsidered
delay is abovefatal 30 msdelay
if the [7]. “Dub”,
is above is generated
30 ms [7].
(the
“Dub”,second heart sound
is generated (thecomponent
second heart denoted
soundby “S2”), when
component the aortic
denoted vales and
by “S2”), when pulmonary
the aortic valves
vales andare
closed. S1 is usually of longer time period and lower frequency than
pulmonary valves are closed. S1 is usually of longer time period and lower frequency than S2 which S2 which has shorter duration
but
has higher
shorterfrequency
duration but ranges (50 to
higher 250 Hz).ranges
frequency A2 is another
(50 to 250 heart
Hz).sound
A2 issignal
anothercomponent
heart sound generated
signal
due to movement of Aortic valve and P2 is another heart sound
component generated due to movement of Aortic valve and P2 is another heart sound signal signal component generated due to
movement
componentof pulmonary
generated duevalve. Aortic pressure
to movement is highervalve.
of pulmonary compared
Aortictopressure
pulmonary pressure
is higher hence the
compared to
A2 component appears before P2 in the heart sound signal.
pulmonary pressure hence the A2 component appears before P2 in the heart sound signal. TheThe presence of cardiac diseases such as
pulmonary
presence of stenosis and atrial
cardiac diseases suchseptal defect can be
as pulmonary identified
stenosis by analyzing
and atrial closely
septal defect canthe besplit between
identified by
aortic
analyzing component
closely theA2 andsplit pulmonary
between aortic component
component P2 and A2 their intensities. component
and pulmonary During deep P2emotional
and their
and excitement
intensities. Duringstate,
deepthe emotional
duration between A2 and state,
and excitement P2 may thebeduration
longer than usual,A2
between causing
and P2longmayand be
wide
longersplitting in between
than usual, causing thelong
components.
and wideThis state isin
splitting considered
between fatal if the delay between
the components. This state these is
component
considered fatal is longer
if thethan
delay30 between
ms [7]. these component is longer than 30 ms [7].
Beside “lub-dub”, some some noisy
noisy signals
signals may may be be present
present in the heart sounds called called murmurs.
murmurs.
Murmurs,
Murmurs, both bothnormal
normaland andfatal, areare
fatal, continuous
continuous vibrations
vibrationsproduced
produced due due
to thetoirregular
the irregularflow of blood
flow of
in cardiovascular
blood system. system.
in cardiovascular Murmurs can be ofcan
Murmurs twobe types
of two“normal
typesmurmurs” and “abnormal
“normal murmurs” murmurs”,
and “abnormal
normal
murmurs”, murmursnormal are murmurs
usually present in heartpresent
are usually sound signal in heartcomponent
sound signalof infants, children of
component andinfants,
adults
during
childrenexercise
and adultsalso during
in women (during
exercise alsopregnancy),
in women (during this type of murmurthis
pregnancy), cantype
be detected
of murmur in the
canfirst
be
heart sound. Abnormal murmurs are usually present in heart patience
detected in the first heart sound. Abnormal murmurs are usually present in heart patience which which indicate a heart valve
defect
indicate called stenosis
a heart valve (squeezed
defect calledheart stenosis
valves) and regurgitation
(squeezed heart (bleeding
valves) and heart valves) [7]. Murmurs
regurgitation (bleeding
heart valves) [7]. Murmurs are the unusual sound present in heart sound cycle of an abnormal heart
sound which indicates some abnormalities. Based on the murmur position in the heart cycle, they
Appl. Sci. 2018, 8, 2344 4 of 14

are the unusual sound present in heart sound cycle of an abnormal heart sound which indicates some
abnormalities. Based on the murmur position in the heart cycle, they can be called systolic murmurs,
diastolic murmurs and continuous murmurs [4]. These murmurs sounds and clicks are a key points to
identify cardiovascular disease [7]. Murmurs in the heart sound can be discovered using stethoscope,
echocardiography or phonocardiography.
Murmurs are classified as continuous murmurs, systolic murmurs and diastolic murmurs.
Systolic murmurs are present during systole and they are generated during the ventricles contraction
(ventricular ejection). In the heart sound component they are present between S1 and S2 hence are
called as systolic murmurs. Based on their types, these systolic murmurs can be called as either
ejection murmurs (atrial septal defect, pulmonary stenosis, or aortic stenosis) Figure 1e, or regurgitant
murmurs (ventricular septal defect, tricuspid regurgitation, mitral regurgitation or mitral valve
prolapse) Figure 1c.
Diastolic murmurs are created during diastole (after systole), when the ventricles relax. Diastolic
murmurs are present between the second and first heart sound portion, this type of murmur is usually
due to mitral stenosis (MS) or aortic regurgitation (AR) by as shown in Figure 1d. Mitral valve
prolapse (MVP) is a disease where the murmur sound is present in between the systole section as
shown in Figure 1b. The murmur of AR is high pitched and the murmur of AS is low pitched.
In mitral regurgitation (MR), systolic component S1 is either soft or buried or absent, and the diastolic
component S2 is widely split. In mitral stenosis, the murmur is low pitched and rumbling and is
present along the diastolic component. In mitral valve prolapse, the murmur can be found throughout
S1compnent. The sound signal of VSD and MR are mostly alike [2,7].
Summarizing the above discussion, we can see that heart sound signal can be acquired using and
electronic stethoscope, in each complete cycle of heart sound signal we have S1–S4 intervals, S3 and
S4 are rare heart sounds and are not normally audible but can be shown on the graphical recording
i.e., phonocardiogram [4]. S3 and S4 intervals are called murmur sound and the heart sound signal
which carries murmurs are called abnormal heart sounds and can be classified according to murmur
position in the signal. In this study automatic classification of heart sound signal is carried out using
five categories, one normal category and four abnormal categories. These abnormal categories are
aortic stenosis (AS), mitral regurgitation (MR), mitral stenosis (MS) and MVP (murmur exist in the
systole interval). Figure 1 shows graphical representation of theses heart sound signals categorically.
The classification of heart sound signal can be carried out by several machine learning classifiers
available for biomedical signal processing. One of them is SVM which can be used for classification of
heart sounds. SVM is machine learning, classification and recognition technique that totally works
on statistical learning and theorems, the classification accuracy of SVM is much more efficient than
conventional classification techniques [8]. Another classifier that recently gained attention is DNN,
DNN acoustic models have high performance in speech processing as well as other bio medical
signal processing [9]. In this paper, our study shows that these classifiers (SVM and DNN also
centroid displacement based KNN) have good performance for heart sound signal classification.
The performance of Convolutional Neural Network is optimized for image classification, so this
method is not suitable for the feature set we used in this paper. We also used Recurrent Neural
Network in the pilot test, but our DNN method showed better results [10,11].

3. Methodology
The heart sound signal contains useful data about the working and health status of the heart,
the heart sound signal can be processed using signal processing approach to diagnose various heart
diseases before the condition of the heart get worse. Hence, various signal processing techniques can be
applied to process the PCG signal. The stages involved in processing and diagnosing the heart sound
signal are, the acquisition of heart sound signal, noise removal, sampling the PCG signal at a specific
frequency rate, feature extraction, training and classification. As shown in the following Figure 2.
Appl. Sci. 2018, 8, 2344 5 of 14
Appl. Sci. 2018, 8, x FOR PEER REVIEW 5 of 14

(a) (b)

Figure 2. (a) Proposed heart sound


Figure 2. sound signal
signal classification
classification algorithm
algorithm using SVM; (b)
(b) Proposed
Proposed heart
heart
sound
sound signal
signal classification
classification algorithm
algorithm using
using DNN.
DNN.

From heart sound


From heart sound signal
signal wewe extract
extracttwo
twodifferent
differenttypes
typesof offeatures:
features: MFCCs
MFCCs andand
DWT.DWT.To
To classify these features we used SVM, centroid displacement based KNN and DNN
classify these features we used SVM, centroid displacement based KNN and DNN as classifiers. For as classifiers.
For performance
performance evaluation
evaluation of features
of features using
using these
these classifiers,we
classifiers, wecarried
carriedout
out our
our experiment
experiment and
and
evaluated the averaged F1 score and accuracy for all of the three features (MFCCs,
evaluated the averaged F1 score and accuracy for all of the three features (MFCCs, DWT and a DWT and
acombination
combinationofofthese)
these)using
usingthe
thethree
threeclassifies
classifies(SVM,
(SVM,DNN DNNand andcentroid
centroidbased
basedKNN).
KNN).

3.1.
3.1. Mel
Mel Frequency
Frequency Cepstral
Cepstral Coefficients
Coefficients (MFCCs)
(MFCCs)
We extracted features as a set of measured,
We extracted features as a set of measured, non-redundant
non-redundant and derived valuesvalues
and derived from heart
fromsignal
heart
which has been used in our study. As we used two types of features, MFCCs,
signal which has been used in our study. As we used two types of features, MFCCs, DWT and theDWT and the combination
of both, MFCCs
combination features
of both, are enormously
MFCCs features are used in signal processing
enormously used in signaland processing
recognition.and They were first
recognition.
calculated and used in speech analysis by Davis and Mermelstein in the
They were first calculated and used in speech analysis by Davis and Mermelstein in the 1980’s. Human can recognize
1980’s.
small
Human changes in pitch of
can recognize sound
small signalinatpitch
changes loweroffrequencies
sound signal andatthe MFCC
lower scale is linear
frequencies and the forMFCC
those
frequencies
scale is linear for those frequencies having range <1 kHz. The power aspect of MFCCs is that they the
having range <1 kHz. The power aspect of MFCCs is that they are able to represent are
signal information efficiently. The MFCCs feature are almost similar to the log filter
able to represent the signal information efficiently. The MFCCs feature are almost similar to the log bank energies
and
filterinclude the mel and
bank energies scaleinclude
which makes
the melitscale
an exact
which copy on smaller
makes scalecopy
it an exact of what humansscale
on smaller understand,
of what
the frequency in mel scale is given by Equation (1):
humans understand, the frequency in mel scale is given by Equation (1):
f
Mel ( f ) = 2595 log (1+ ) (1)
(1)
Mel (f) = 2595 log (1 + 700 )

In MFCC algorithm, using Hamming window an audio signal is reshaped to smaller windows
In MFCC algorithm, using Hamming window an audio signal is reshaped to smaller windows
hence splitting of signal to frames is done. Using Fast Fourier Transform, spectrum is computed for
hence splitting of signal to frames is done. Using Fast Fourier Transform, spectrum is computed for
each frame, using filter bank each spectrum is being weighted. Finally using Logarithm and Discrete
each frame, using filter bank each spectrum is being weighted. Finally using Logarithm and Discrete
Cosine Transform, the MFCC vector is computed [12,13]. Compared to other features, MFCCs have
Cosine Transform, the MFCC vector is computed [12,13]. Compared to other features, MFCCs have
good performance against noisy signal when used in speech signal processing hence can be used in
good performance against noisy signal when used in speech signal processing hence can be used in
biomedical signal processing [12]. During feature extraction, we resampled each signal frequency
biomedical signal processing [12]. During feature extraction, we resampled each signal frequency to
to 8 KHz, the length of the each signal’s feature extracted is 19 and the length of frame in each
8 KHz, the length of the each signal’s feature extracted is 19 and the length of frame in each sample is
sample is 240 and the step size is 80. The following diagram represents the extraction of MFCCs
240 and the step size is 80. The following diagram represents the extraction of MFCCs features
features Figure 3a.
Figure 3a.
Appl.
Appl.Sci. 2018,
Sci. 8, 8,
2018, 2344
x FOR PEER REVIEW 6 of 1414
6 of

(a) (b)
Figure3.3.(a)
Figure (a)MFCCs
MFCCscomputation;
computation;(b)
(b)Decomposition
Decompositiontree
treeofofwavelets
wavelets(3rd
(3rd level).
level).

3.2.
3.2.Discrete
DiscreteWavelets
WaveletsTransform
Transform(DWT)
(DWT)
DWT
DWTis is mathematical
mathematical tooltool
for decomposing
for decomposing data in a top
data indown
a topfashion. DWT represent
down fashion. a function a
DWT represent
infunction
terms ofina terms
rough of overall
a roughform, and aform,
overall wideandrange of details.
a wide range Despite
of details. of the requirement
Despite and type
of the requirement
ofand
function i.e., signals images etc. DWT offers sublime methodology and
type of function i.e., signals images etc. DWT offers sublime methodology and technique for technique for representing
the amount of the
representing detail present
amount [14]. Wavelets
of detail perform
present [14]. and offer
Wavelets performscaleand based
offeranalysis for aanalysis
scale based given data.for a
Agiven
wide range of applications and usage has been found for these wavelets
data. A wide range of applications and usage has been found for these wavelets including including signal processing,
mathematics and numerical
signal processing, analysis
mathematics and and, for its better
numerical performance
analysis and, for its in signals and image processing
better performance in signalsitandis
considered an alternative to Fast Fourier Transform as DWT provide
image processing it is considered an alternative to Fast Fourier Transform as DWT provide time-frequency representation [15].
When there is a need
time-frequency for processing
representation andWhen
[15]. analyzing
therenonis stationary
a need for tool, DWT can be
processing andused [16]. Study
analyzing non
shows
stationary tool, DWT can be used [16]. Study shows that discrete wavelets transformso
that discrete wavelets transform have high performance in speech signal processing far and
have high
especially
performance whenincombined with MFCCs,
speech signal processingthe so
performance is boosted
far and especially further
when [5].
combined with MFCCs, the
DWT can be defined
performance is boosted further [5].as a short wave which carries its energy and information condensed in time.
They are DWT limited
can be in space
defined andaspossessing
a short wave an average value approaching
which carries its energy and to zero. Each signal
information containsin
condensed
two components
time. i.e., lowinfrequency
They are limited space andand high frequency
possessing an average components. Low frequency
value approaching to zero.components
Each signal
mostly contain maximum information part, and high frequency
contains two components i.e., low frequency and high frequency components. Low frequencycomponents convey quality. That is
why in DWT estimation and details are being rated. The estimation
components mostly contain maximum information part, and high frequency components convey is of the high-scale, low-frequency
components
quality. That of the signalinand
is why DWT the Details
estimationare the
andlow-scale,
details are high-frequency
being rated.components.
The estimationIn Figure
is of 3b,
the
ithigh-scale,
is shown that down sampling a given signal can be divided into
low-frequency components of the signal and the Details are the low-scale, its component at level one it has
cA1, cD1 approximated
high-frequency and detailed
components. In Figure components and atthat
3b, it is shown level two sampling
down it has cA2a and
given cD2 estimated
signal can be
and detailed components. Down sampling ensures the absence of
divided into its component at level one it has cA1, cD1 approximated and detailed components redundancy in a given data [17].
and
Mathematically,
at level two it DWT has cA2 canand
be expressed in the and
cD2 estimated following
detailedEquation (2) as: Down sampling ensures the
components.
absence of redundancy in a given data [17].
Z ∞ Mathematically,
1
DWT can be expressed in the following
t − 2x y
Equation (2) as: DWT ( x, y) = Z (t) p ϕ( )dt (2)
−∞ |2 x | 2x
DWT (x, y) = 𝑍(𝑡) 𝜑( ) dt (2)
The above equation can be used to extract more useful | | data from audio signal using DWT [18].

In caseThe
of DWT
abovethe length of
equation canfeature vector
be used is 24. more useful data from audio signal using DWT [18].
to extract
In case of DWT the length of feature vector is 24.
3.3. Support Vector Machine (SVM)
3.3.Its
Support Vector
working Machine
is based (SVM) Learning Theory (SLT) and structure risk minimization principal
on Statistics
(SRMP), it is considered a unique method in machine learning for signal processing. It has unique
Its working is based on Statistics Learning Theory (SLT) and structure risk minimization
performance in solving nonlinear learning problems and sample learning problems [19,20]. Based on
principal (SRMP), it is considered a unique method in machine learning for signal processing. It has
optimization technology SVM is a unique tool in solving machine learning challenges. It has been first
unique performance in solving nonlinear learning problems and sample learning problems [19,20].
used by Vapnik in 1990’s and now days they are used in pattern recognition in image processing and
Based on optimization technology SVM is a unique tool in solving machine learning challenges. It
has been first used by Vapnik in 1990’s and now days they are used in pattern recognition in image
Appl. Sci. 2018, 8, 2344 7 of 14

signal processing. Based on SLT and SRMP, it can solve many complex structures, process big data
appropriate to over learning [20]. The use of SVM has been started as hyper plane classifier and is
useful in a situation where data is required to be separated linearly [21].
Suppose having sample Si and its sorts Ti , expressed as (Si , Ti ), S ∈ Rd , Ti ∈ (1, −1), i = 1, . . . , N.
The d is dimension number of input space. For standard SVM, its classifying border or edge is ||w2 || .
Making the classifying borer or edge to maximum is equal to making the ||w||2 to minimum. Therefore,
the optimization problem of making the classifying edge to maximum can be expressed in the follow
quadratic programming problem:

min 1 N
wbε J (w, ε) = kwk2 + C
2 ∑ i =1 ε i (3)

Subject to
Ti wT ϕ(Si ) + b ≥ 1 − ε I
 

i = 1, ..., N (4)
ε i ≥ 0, i = 1, ..., N
where εi ≥ 0 (i = 1, . . . , N ) is flexible variable, which ensures classification validity under linear
non-separable case, and parameter C is a positive real constant which determines penalties to
approximation errors, a higher value of C is similar as to assigning a higher penalty to errors.
The function ϕ is a non-linear mapping function, by which the non-linear problem can be mapped as
linear problem of a high dimension space. In the transforming space, we can get optimal hyper plane.
The SVM algorithm can discern right classification to sample by solving above quadratic programming
problem [21]. We choose nonlinear kernel (quadratic SVM) model for our experiments, in that case
SVM needs another parameter gamma along with C to be optimized for better results. Both of these
values greatly affect the classification decision of SVM, larger values of gamma brings classifier closer
to training accuracy and lower values make it away from it [22]. Similarly lower values of C are more
suitable in bringing decision function closer to training accuracy and vice versa. In our experiments
we achieved good training accuracy using C values in range of 0.0003 to 0.0001 and gamma value
of log103 .

3.4. Deep Neural Network (DNN)


DNN is a new AI algorithm for machine learning and has been widely used in many research and
developing fields. DNN is inspired bio natural object’s visual mechanism where it have many layers
of neural network, it process data from their light carrying organs to center of the brain, replicating
that mechanism, in DNN the process is done layer by layer where the information are carried out from
one layer to another, sequentially extracting edge features, shape feature, part feature, and eventually
forms abstract concept [23]. A DNN is a network of neurons in which they are grouped together
in multiple layers, those layers are interconnected together in several sequential layers and neurons
accepts the neuron activation from previous layer and calculate weighted sum multiply it with neuron
and process the output to next layer and so on. The neurons in each layer of network combined execute
a complex nonlinear mapping from the beginning to the last. This mapping causes the network to
learn from neuron back propagation techniques using weighted sum of neurons. Figure 4a shows
a simple neural network where there is an input to collection of many hidden layers and a final output.
Appl.
Appl.Sci. 2018,8,8,2344
Sci.2018, x FOR PEER REVIEW 8 8ofof1414

(a) (b)
Figure 4.4. (a)(a)Neural
Figure Neural network
network composed
composed of many
of many interrelated
interrelated neurons;
neurons; (b) Data(b) Data collection
collection procedure
procedure
outline. outline.

For
Forour
ourDNN
DNNmodel
modelwe wekept
keptbatch
batchsize
sizeequal
equaltoto100,
100,three
threehidden
hiddenlayers
layersand
and2000
2000epochs
epochsforfor
each
each training phase. We input the data to layers in chunks and the total numbers of chunkswere
training phase. We input the data to layers in chunks and the total numbers of chunks were
calculated
calculatedfrom
fromthe
thetotal
totalnumber
numberofoffeatures
featureslength
lengthdivided
dividedby by40,
40,asaswe
weselected
selectedthe
thechunk
chunksize
sizetotobe
be
40 for optimum results.
40 for optimum results.
4. Experiments
4. Experiments
We collected and arranged data in the form of database, which consist of two sets of data:
We collected and arranged data in the form of database, which consist of two sets of data:
normal set and abnormal set, the whole data in database (normal and abnormal) have five categories,
normal set and abnormal set, the whole data in database (normal and abnormal) have five
one normal category (N) and four abnormal categories, the four abnormal categories were: aortic
categories, one normal category (N) and four abnormal categories, the four abnormal categories
stenosis (AS) mitral stenosis (MS) mitral regurgitation (MR) mitral valve prolapse (MVP), details of
were: aortic stenosis (AS) mitral stenosis (MS) mitral regurgitation (MR) mitral valve prolapse
data is shown in Table 1, there complete detail (w.r.t to category) were discussed in background section.
(MVP), details of data is shown in Table 1, there complete detail (w.r.t to category) were discussed in
The total numbers of audio files were 1000 for normal and abnormal categories (200 audio files/per
background section. The total numbers of audio files were 1000 for normal and abnormal categories
category), the files are in .wav format. Our data collection procedure outline is shown in Figure 4b,
(200 audio files/per category), the files are in .wav format. Our data collection procedure outline is
in which we first selected cardiac disorder category, collected data then applied data standardization
shown in Figure 4b, in which we first selected cardiac disorder category, collected data then applied
after filtering [Appendix A].
data standardization after filtering [Appendix A].
Table 1. Dataset detail.
Table 1. Dataset detail.
Classes Number of Samples Per Class
Classes Number of Samples Per Class
Normal N 200
Normal N 200
AS AS 200 200
MR 200
Abnormal MR 200
Abnormal MS 200
MS MVP 200 200
MVP 200
Total 1000
Total 1000

After
Aftergathering
gatheringthe the data,
data, the experiments we
the experiments we performed,
performed,can canbebecategorized
categorizedintointonine
ninetypes,
types,
as
asdiscussed
discussedininmethodology
methodologysection,
section,wewedid
didfirst
first3 3experiments
experimentsusing
usingSVM
SVMclassifier
classifierand
andthree
threetypes
typesof
offeatures:
features:MFCCs,
MFCCs,DWT DWTand andMFCCs
MFCCs++DWT DWT (fused
(fused features),
features), second
second three
three experiments
experimentsusingusingDNN DNN
and
and last 3 experiments using centroid displacement based KNN, each of these features weretrained
last 3 experiments using centroid displacement based KNN, each of these features were trained
and
andclassified
classifiedusing
usingeacheachclassifier
classifierindividually.
individually.In Incase
caseofofcentroid
centroiddisplacement
displacementbased basedKNN
KNNwe we
choose the value of k = 2.5 and that worked well for our dataset
choose the value of k = 2.5 and that worked well for our dataset training. [24].training. [24].
For
Fortraining
trainingand
and classification
classification ofof MFCCs
MFCCsfeatures
featuresusing
usingSVM
SVMthetheparameters
parameters andand arguments
arguments we
we used while extracting MFCCs features were as follows: sampling
used while extracting MFCCs features were as follows: sampling frequency 8000 Hz, feature frequency 8000 Hz, feature
vector
vector
having having
lengthlength
19 and 19 and
frameframe
sizesize in sample
in sample 240240 withstep
with stepsize
sizeof
of 80.Similarly,
80.Similarly, for
for training
trainingandand
classification of DWT features using SVM the parameters and arguments we used
classification of DWT features using SVM the parameters and arguments we used while extracting while extracting DWT
features were aswere
DWT features follows: sampling
as follows: frequency
sampling 8000 Hz,8000
frequency and Hz,
length
andoflength
featureofvector 24.vector
feature Using24.
fivefold
Using
cross validation technique, we classified our data using SVM and DNN. From
fivefold cross validation technique, we classified our data using SVM and DNN. From dataset, dataset, 20% of sample
20%
data in each class was used for testing and eighty percent of sample data in each
of sample data in each class was used for testing and eighty percent of sample data in each class for class for training,
Appl. Sci. 2018, 8, 2344 9 of 14
Appl. Sci. 2018, 8, x FOR PEER REVIEW 9 of 14

the performance
training, of each machine
the performance of eachlearning
machinealgorithm
learning with different
algorithm features
with were
different recorded
features with
were respect
recorded
to their given accuracies.
with respect to their given accuracies.
Referring to
Referring to Figure
Figure 5,
5, the
the data
data training phase has
training phase has been
been shown, where data
shown, where data and
and features
features are
are
distributed clearly shows that in case of DWT, the data is spread more across its region in
distributed clearly shows that in case of DWT, the data is spread more across its region in plan, as it plan,
as it represents
represents bothboth
lower lower and higher
and higher frequencies
frequencies onaxes
on its its axes hence
hence better
better performance
performance evaluation.
evaluation.

(a) (b)

(c)
Figure 5. (a) MFCCs Features data pattern; (b) DWT Features data pattern; (c) DWT plus MFCCs
Features data pattern.

Figure 6 shows confusion matrices for features type and the training model, as explained in the
footer section
section of
of this
this figure,
figure,a;
a;d;
d;ggbelongs
belongsto
toMFCCs
MFCCsfeatures,
features,b,b,e,e,hhbelongs
belongstotoDWT
DWTfeatures
featuresand
andc,c,f,
if,belongs to combined features (MFCCs plus DWT).
i belongs to combined features (MFCCs plus DWT).
The heart sound signal preprocessing and classification is performed by three different methods
and classifiers. In case of any speech and signal processing, study shows that features extracted with
wavelets transform have greater performance and when those features are used in combination with
MFCCs, the accuracy goes even higher [5,6,25].
Appl. Sci.Appl.
2018, 8,2018,
Sci. x FOR PEER REVIEW
8, 2344 10 10 of 14
of 14

Figure 6. Cont.
Appl. Sci. 2018, 8, 2344 11 of 14
Appl. Sci. 2018, 8, x FOR PEER REVIEW

Figure 6. (a) Confusion matrix


Figurefor6.MFCCs Featuresmatrix
(a) Confusion and Centroid
for MFCCsdisplacement
Featuresbased
and KNN classifier;
Centroid displacement based KNN c
(b) Confusion matrix for DWT Features and Centroid displacement based KNN classifier; (c) Confusion
(b) Confusion matrix for DWT Features and Centroid displacement based KNN class
matrix for MFCCs plus DWT Features and Centroid displacement based KNN classifier; (d) Confusion
Confusion matrix for MFCCs plus DWT Features and Centroid displacement based KNN c
matrix for MFCCs Features and DNN classifier; (e) Confusion matrix for DWT Features and DNN
(d) Confusion
classifier; (f) Confusion matrix matrix
for MFCCs plus DWTfor MFCCs
Features andFeatures and (g)
DNN classifier; DNN classifier;
Confusion matrix(e) Confusion matrix f
for MFCCs Features andFeatures and DNN
SVM classifier; classifier;
(h) Confusion (f) for
matrix Confusion matrix
DWT Features andfor
SVMMFCCs plus DWT Features an
classifier;
classifier;
(i) Confusion matrix for MFCCs plus(g)
DWTConfusion matrix
Features and SVMfor MFCCs Features and SVM classifier; (h) Confusion m
classifier.
DWT Features and SVM classifier; (i) Confusion matrix for MFCCs plus DWT Features a
5. Results and Discussion
classifier.
In our experiments we classified our extracted features from heart sound signal database
(including normal and abnormal category)
The heart usingsignal
sound three different classifiers,
preprocessing SVM,
and DNN and centroid
classification is performed by thr
based KNN. Table 2 shows the baseline
methods results obtained
and classifiers. usingofMFCCs
In case featuresand
any speech and their
signalcorresponding
processing, study shows
accuracies. In order toextracted
improve the accuracy,
with we found
wavelets out some
transform optimum
have greatervalues for feature extraction
performance and when those features
such as sampling frequency and feature vector length on which the resulting accuracies were higher
combination with MFCCs, the accuracy goes even higher [5,6,25].
than before and the results are shown in Table 3. For better accuracy, new features (DWT) were taken
into consideration for our experiments, after experimenting and training our data with DWT we
5. Results and Discussion
have achieved results shown in Table 4. We can see from Tables 3 and 4 that there is improvement in
the accuracy but we considered In our experiments
to combine those we classified
features together ourandextracted
train usingfeatures
SVM andfrom DNN, heart sound sign
after performing experiments and finding some optimal values for sampling frequency,
(including normal and abnormal category) using three different classifiers, SVM, DNNwe arrived
at the conclusion thatbased
we haveKNN. much better
Table output
2 shows than using those features
the baseline resultsseparately,
obtained and using
we can MFCCs feature
see the results from Table 5. Unfortunately DNN couldn’t perform well as compared
corresponding accuracies. In order to improve the accuracy, we found to SVM and theout some optimu
reason for that is we have very limited amount of data available and we know that DNN requires
feature extraction such as sampling frequency and feature vector length on which t
thousands of files and features for training and classification, so in case if sufficient amount of data
accuracies were higher than before and the results are shown in Table 3. For better ac
were available, DNN could have performed higher than SVM .
features
The highest accuracy (DWT)from
we achieved werethose
taken into consideration
experiments was from fusedfor our experiments,
(combined) featuresafter
we experimenting
our features
classified; these combined data with haveDWTneverwe have
been achieved
used results shown
for classification of heart in Table
sound 4. We
signal can see from Table
before.
Tables 3–6 summarizesthere is improvement
the output, in the in
and classification accuracy
terms ofbut we considered
accuracy averaged f1to combine
score those features toget
sensitivity
using SVM
and specificity of our experiment in and DNN,
detail. after
Several performing
experiments wereexperiments
performed with and finding
MFCCs andsome
DWT optimal values
frequency,
with variant frequencies, we arrived
variant Cepstral at theand
coefficient conclusion that we have
their corresponding muchaccuracy
changing better output
was than using th
recorded. From our experiments
separately,Tables
and we 3–5 summarize
can see thethe last and
results finalTable
from results5.ofUnfortunately
MFCCs, DWT and DNN couldn’t per
combined features (MFCCs and DWT) respectively.
compared to SVM and the reason for that is we have very limited amount of data avai
know that DNN requires thousands of files and features for training and classification
Table 2. Baseline results with MFCCs features.
sufficient amount of data were available, DNN could have performed higher than SVM
ClassifierThe highest accuracy we Features
achieved from those Accuracy
experiments was from fused (combin
we classified; these combined features have never been
SVM MFCC 87.2%used for classification of heart
DNN MFCC 82.3%
before. Tables 3–6 summarizes the output, and classification in terms of accuracy avera
Centroid displacement base KkNN MFCC 72.2%
sensitivity and specificity of our experiment in detail. Several experiments were per
MFCCs and DWT with variant frequencies, variant Cepstral coefficient and their co
changing accuracy was recorded. From our experiments Tables 3–5 summarize the l
results of MFCCs, DWT and combined features (MFCCs and DWT) respectively.

Table 2. Baseline results with MFCCs features.


Appl. Sci. 2018, 8, 2344 12 of 14

Table 3. Improved Result based on MFCCs Features.

KNN
MFCCs SVM DNN
(Centroid Displacement Based)
Accuracy 80.2% 91.6% 86.5%
Average F1 Score 92.5% 86.0% 96.9%

Table 4. Result based on DWT Features.

KNN
DWT SVM DNN
(Centroid Displacement Based)
Accuracy 91.8% 92.3% 87.8%
Average F1 Score 98.3% 99.1% 95.6%

Table 5. Result based on MFCCs and DWT Features.

KNN
MFCCs + DWT SVM DNN
(Centroid Displacement Based)
Accuracy 97.4% 97.9% 92.1%
Average F1 Score 99.2% 99.7% 98.3%

Table 6. Sensitivity and Specificity calculated for each classifier performance based on the Figure 6
confusion matrix.

Classifiers Features Sensitivity Specificity


MFCCs 81.98% 93.5%
Centroid displacement based KNN DWT 92.0% 97.9%
MFCCs + DWT 97.6% 98.8%
MFCCs 86.8% 95.1%
DNN DWT 91.6% 97.4%
MFCCs + DWT 94.5% 98.2%
MFCCs 87.3% 96.6%
SVM DWT 92.3% 98.4%
MFCCs + DWT 98.2% 99.4%

Depending on the size of dataset and type of features in the classification experiment different
results can be achieved with variant accuracies. Using fivefold cross validation the evaluation results
are achieved as shown in Table 5. It is observed that the overall maximum accuracy of 97.9% (F1 score
of 99.7%) has been achieved using combine features of discrete wavelet transform and MFCCs,
and training them via SVM classifier. In case of centroid displacement based KNN the highest accuracy
achieved is 97.4% (F1 score of 99.2%).
Similarly, in case of DNN, the maximum accuracy of 89.3% has been achieved when we used
MFCCs Features combined with DWT features, keeping the parameters for all the features same as
mentioned in above paragraph, we trained our deep neural network with features (MFCCs and DWT),
in deep neural network we kept batch size of 100, three hidden layers and one output layer and 2000
epochs. In case of DNN too, we can observe from our results table that using combined features of
MFCCs and DWT, we can get higher results by efficiently utilizing the information present in the heart
sound signal. The improved accuracy is due to usage of completely different domain signals (MFCCs
and DWT) together.
Appl. Sci. 2018, 8, 2344 13 of 14

6. Conclusions
PCG signals carries information about the functioning of heart valves during heartbeat, hence
these signals are very important in diagnosing heart problems at early stage. To detect heart problems
with great precision, the new features (MFCCs and DWT) were used and have been proposed in this
study to detect 4 types of abnormal heart diseases category and one normal category. The results that
have been obtained shows clear performance superiority and can help physicians in the early detection
of heart disease during auscultation examination. This work can be greatly improved if the dataset is
prepared on a larger scale and also if the data features are managed in a more innovative way, then the
performance can be more fruitful. Some new features can also be introduced to asses and analyses
heart sound signal more accurately.

Author Contributions: The authors contributed equally to this work. Conceptualization, visualization, writing
and preparing the manuscript is done by Y., reviewing, supervision, project administration, validation and
funding was done by S.K., Software related work, data creation, project administration was the task of G.Y.S.
Funding: This research was funded by the Ministry of SMEs and Startups (No. S2442956). The APC was funded
by Sejong University.
Acknowledgments: This work was supported by the Ministry of SMEs and Startups (No. S2442956),
the development of intelligent stethoscope sounds analysis engine and interactive smart stethoscope
service platform.
Conflicts of Interest: The authors declare no conflict of interest.

Appendix A
The data was collected from random sources, such as books (Auscultation skills CD, Heart sound
made easy) and websites (48 different websites provided the data including Washington, Texas, 3M,
and Michigan and so on). After conducting efficient cross check test, we excluded files having extreme
noise, data was sampled to 8000 Hz frequency rate and was converted to mono channel, 3 period heart
sound signal, data sampling, conversion and editing was done in cool edit software.

References
1. Bolea, J.; Laguna, P.; Caiani, E.G.; Almeida, R. Heart rate and ventricular repolarization variabilities
interactions modification by microgravity simulation during head-down bed rest test. In Proceedings of
the 2013 IEEE 26th International Symposium onComputer-Based Medical Systems (CBMS), Porto, Portugal,
22–23 June 2013; IEEE: New York, NY, USA, 2013; pp. 552–553.
2. Kwak, C.; Kwon, O.-W. Cardiac disorder classification by heart sound signals using murmur likelihood and
hidden markov model state likelihood. IET Signal Process. 2012, 6, 326–334. [CrossRef]
3. Mondal, A.; Kumar, A.K.; Bhattacharya, P.; Saha, G. Boundary estimation of cardiac events s1 and s2 based
on hilbert transform and adaptive thresholding approach. In Proceedings of the 2013 Indian Conference
on Medical Informatics and Telemedicine (ICMIT), Kharagpur, India, 28–30 March 2013; IEEE: New York,
NY, USA, 2013; pp. 43–47.
4. Randhawa, S.K.; Singh, M. Classification of heart sound signals using multi-modal features. Procedia
Computer Science 2015, 58, 165–171. [CrossRef]
5. Anusuya, M.; Katti, S. Comparison of different speech feature extraction techniques with and without
wavelet transform to kannada speech recognition. Int. J. Comput. Appl. 2011, 26, 19–24. [CrossRef]
6. Turner, C.; Joseph, A. A wavelet packet and mel-frequency cepstral coefficients-based feature extraction
method for speaker identification. Procedia Comput. Sci. 2015, 61, 416–421. [CrossRef]
7. Thiyagaraja, S.R.; Dantu, R.; Shrestha, P.L.; Chitnis, A.; Thompson, M.A.; Anumandla, P.T.; Sarma, T.;
Dantu, S. A novel heart-mobile interface for detection and classification of heart sounds. Biomed. Signal
Process. Control 2018, 45, 313–324. [CrossRef]
8. Wu, J.-B.; Zhou, S.; Wu, Z.; Wu, X.-M. Research on the method of characteristic extraction and classification of
phonocardiogram. In Proceedings of the 2012 International Conference on Systems and Informatics (ICSAI),
Yantai, China, 19–20 May 2012; IEEE: New York, NY, USA, 2012; pp. 1732–1735.
Appl. Sci. 2018, 8, 2344 14 of 14

9. Maas, A.L.; Qi, P.; Xie, Z.; Hannun, A.Y.; Lengerich, C.T.; Jurafsky, D.; Ng, A.Y. Building dnn acoustic models
for large vocabulary speech recognition. Comput. Speech Lang. 2017, 41, 195–213. [CrossRef]
10. Nanni, L.; Aguiar, R.L.; Costa, Y.M.; Brahnam, S.; Silla, C.N., Jr.; Brattin, R.L.; Zhao, Z. Bird and whale species
identification using sound images. IET Comput. Vis. 2017, 12, 178–184. [CrossRef]
11. Nanni, L.; Costa, Y.M.; Aguiar, R.L.; Silla, C.N., Jr.; Brahnam, S. Ensemble of deep learning, visual and
acoustic features for music genre classification. J. New Music Res. 2018, 1–15. [CrossRef]
12. Nair, A.P.; Krishnan, S.; Saquib, Z. Mfcc based noise reduction in asr using kalman filtering. In Proceedings
of the Conference on Advances in Signal Processing (CASP), Pune, India, 9–11 June 2016; IEEE: New York,
NY, USA, 2016; pp. 474–478.
13. Almisreb, A.A.; Abidin, A.F.; Tahir, N.M. Comparison of speech features for arabic phonemes recognition
system based malay speaker. In Proceedings of the 2014 IEEE Conference on Systems, Process and Control
(ICSPC), Kuala Lumpur, Malaysia, 12–14 December 2014; IEEE: New York, NY, USA, 2014; pp. 79–83.
14. DeRose, E.J.S.T.D.; Salesin, D.H. Wavelets for computer graphics: A primer part 1 y. Way 1995, 6, 1.
15. Ghazali, K.H.; Mansor, M.F.; Mustafa, M.M.; Hussain, A. Feature extraction technique using discrete
wavelet transform for image classification. In Proceedings of the 5th Student Conference on Research and
Development SCOReD 2007, Selangor, Malaysia, 11–12 December 2007; IEEE: New York, NY, USA, 2007;
pp. 1–4.
16. Fahmy, M.M. Palmprint recognition based on mel frequency cepstral coefficients feature extraction.
Ain Shams Eng. J. 2010, 1, 39–47. [CrossRef]
17. Sharma, R.; Kesarwani, A.; Mathur, P.M. A novel compression algorithm using dwt. In Proceedings of
the 2014 Annual IEEE India Conference (INDICON), Pune, India, 11–13 December 2014; IEEE: New York,
NY, USA, 2014; pp. 1–4.
18. Li, M.; Chen, W.; Zhang, T. Classification of epilepsy eeg signals using dwt-based envelope analysis and
neural network ensemble. Biomed. Signal Process. Control 2017, 31, 357–365. [CrossRef]
19. Zhong, Y.-C.; Li, F. Model identification study on micro robot mobile in liquid based on support vector
machine. In Proceedings of the 3rd IEEE International Conference onNano/Micro Engineered and Molecular
Systems, NEMS 2008, Sanya, China, 6–9 January 2008; IEEE: New York, NY, USA, 2008; pp. 55–59.
20. Lu, P.; Xu, D.-P.; Liu, Y.-B. Study of fault diagnosis model based on multi-class wavelet support vector
machines. In Proceedings of the 2005 International Conference on Machine Learning and Cybernetics,
Guangzhou, China, 18–21 August 2005; IEEE: New York, NY, USA; pp. 4319–4321.
21. Yang, K.-H.; Zhao, L.-L. Application of the improved support vector machine on vehicle recognition.
In Proceedings of the 2008 International Conference on Machine Learning and Cybernetics, Kunming, China,
12–15 July 2008; IEEE: New York, NY, USA, 2008; pp. 2785–2789.
22. Chapelle, O.; Vapnik, V.; Bousquet, O.; Mukherjee, S. Choosing multiple parameters for support vector
machines. Mach. Learn. 2002, 46, 131–159. [CrossRef]
23. Yi, H.; Sun, S.; Duan, X.; Chen, Z. A study on deep neural networks framework. In Proceedings of the 2016
IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference
(IMCEC), Xi’an, China, 3–5 October 2016; IEEE: New York, NY, USA, 2016; pp. 1519–1522.
24. Nguyen, B.P.; Tay, W.-L.; Chui, C.-K. Robust biometric recognition from palm depth images for gloved hands.
IEEE Trans. Hum.-Mach. Syst. 2015, 45, 799–804. [CrossRef]
25. Janse, P.V.; Magre, S.B.; Kurzekar, P.K.; Deshmukh, R. A comparative study between mfcc and dwt feature
extraction technique. Int. J. Eng. Res. Technol. 2014, 3, 3124–3127.

© 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).

You might also like