You are on page 1of 13

See discussions, stats, and author profiles for this publication at: https://www.researchgate.

net/publication/226091040

Support vector machines models for surface roughness prediction in CNC


turning of AISI 304 austenitic stainless steel

Article  in  Journal of Intelligent Manufacturing · June 2012


DOI: 10.1007/s10845-010-0415-2

CITATIONS READS

93 583

2 authors, including:

Sami Ekici
Firat University
44 PUBLICATIONS   616 CITATIONS   

SEE PROFILE

Some of the authors of this publication are also working on these related projects:

Analyzing Breast Cancer using Thermography View project

All content following this page was uploaded by Sami Ekici on 06 June 2014.

The user has requested enhancement of the downloaded file.


J Intell Manuf
DOI 10.1007/s10845-010-0415-2

Support vector machines models for surface roughness prediction


in CNC turning of AISI 304 austenitic stainless steel
Ulaş Çaydaş · Sami Ekici

Received: 8 September 2009 / Accepted: 3 May 2010


© Springer Science+Business Media, LLC 2010

Abstract In the present investigation, three different type a widely used index of product quality and in most cases a
of support vector machines (SVMs) tools such as least square technical requirement for mechanical products. Hence,
SVM (LS-SVM), Spider SVM and SVM-KM and an artifi- achieving the desired surface quality is of great importance
cial neural network (ANN) model were developed to estimate for the functional behaviour of a part (Benardos et al. 2003).
the surface roughness values of AISI 304 austenitic stain- Surface roughness is the result of the process parameters such
less steel in CNC turning operation. In the development of as tool geometry (i.e. nose radius, edge geometry, rake angle,
predictive models, turning parameters of cutting speed, feed tool tip radius, chamfer thick, etc.), cutting conditions (i.e.
rate and depth of cut were considered as model variables. feed rate, cutting speed, depth of cut, etc.) and workpiece
For this purpose, a three-level full factorial design of experi- properties (Umbrello et al. 2008).
ments (DOE) method was used to collect surface roughness Nowadays, many papers have been published about the
values. A feedforward neural network based on backpropa- prediction of surface quality in turning process. Researchers
gation algorithm was a multilayered architecture made up of have approached the problem from different points of view
15 hidden neurons placed between input and output layers. and using different artificial intelligence techniques. Durmuş
The prediction results showed that the all used SVMs results Karayel (Karayel 2008) presented a neural network model
were better than ANN with high correlations between the for prediction and control of surface rougness in a computer
prediction and experimentally measured values. numerically controlled (CNC) lathe. A feedforward multi-
layered neural network was developed and the neural net-
Keywords Surface roughness · Support vector machines · work model was trained using the scaled conjugate gradient
AISI 304 machining algorithm (SCGA), which is a type of back-propagation. The
predicted surface roughness from the model was very close
to the values measured experimentally with a high correla-
Introduction tion coefficient. Additionally, the feed rate parameter was
found the most dominant factor on surface roughness, while
Turning is the primary operation in most of the production the effect of depth of cut was not regular and had a var-
processes in the industry. The turning operation produces the iable character. Muthukrishnan and Davim (2009) studied
components, which have critical features that require spe- the surface roughness of Al-SiC (20p) composites in turn-
cific surface finish (Davim et al. 2008). Surface roughness is ing operation by using coarse grade polycrystalline diamond
(PCD) insert under different cutting conditions. A multilay-
ered perception ANN model has been constructed with back-
U. Çaydaş (B)
Technical Education Faculty, Department of Manufacturing, propagation algorithm using the input parameters of depth
University of Firat, 23119 Elazig, Turkey of cut, cutting speed and feed. Additionally, the analysis
e-mail: ucaydas@firat.edu.tr of variance (ANOVA) technique was adopted to study and
it was shown that the statistical influence of feed rate on
S. Ekici
Technical Education Faculty, Department of Electric, surface roughness was the highest (51%), right after the
University of Firat, 23119 Elazig, Turkey depth of cut (30%) and the cutting speed (12%). The result

123
J Intell Manuf

of ANN model showed close matching between the model in ultra precison diamond turning of Al 6061/SiCp metal
output and directly measured surface roughness. Grzesik matrix composite. Wong and Hamouda (2003) introduced
and Brol (2003) made an assesment on surface quality of a new type of artificial neuron in the design of neural net-
turned, ground and honed specimens by using three differ- work for turning process, namely the product neuron, which
ent approaches, i.e. statistical, fractal and ANNs. It has been had multiplication instead of summation. The back-propa-
shown that the results of modeling techniques showed good gation algorithm was used to train the network. Comparison
results. Surjya and Chakraborty (2005) developed a method- among neural network prediction, fuzzy logic predicton and
ology for the prediction of surface roughness in turning using handbook data was made. Nalbant et al. (2009) presented an
back-propagation neural networks. A good number of exper- experimental investigation of the effects of uncoated, PVD-
iments have been conducted on mild steel workpiece using coated and CVD-coated cemented carbide inserts and cutting
HSS as the cutting tool material. Optimum network archi- parameters on surface roughness of AISI 1030 steel in CNC
tecture has been found out based on the mean square error turning operation and its prediction using ANNs. The ANN
and convergence rate. Finally, the presented network model results and experimental values were compared by statisti-
was efficient to predict the surface roughness in an accept- cal error analyzing method. The absolute fraction of vari-
able error rate. Bağcı and Işık (2006) carried out orthogonal ance value was obtained as 0.99985 within an acceptable
cutting tests on unindirectional glassfibre reinforced plastics accuracy. Paiva et al. (2007) presented a hybrid approach,
(GFRP), using cermet tools. Depth of cut, feed rate and cut- combining RSM and principal component analysis (PCA) to
ting speed parameters were varied. A statistical three level optimize multiple correlated responses in a turning process.
full factorial experimental design technique was designed to The mixed ceramic tool life, processing cost per piece, cut-
collect surface roughness data. An ANN and response surface ting time, the total turning cycle time, surface roughness and
models were developed for surface roughness prediction. material removal rate were the main outputs under investiga-
Good agreement was observed between the predicted models tion. It was shown that the multiresponse optimization was
and experimental measurements. Özel et al. (2007) investi- achieved at a cutting speed of 238 m/ min, with a feed rate
gated the surface finishing and tool flank wear in turning of of 0.08 mm/rev and a depth of cut of 0.32 mm settings in
AISI D2 steels (60HRC) using ceramic wiper (multi-radii) turning of AISI 52100 hardened steel. An attempt has also
design inserts. Surface roughness values as low as 0.18– been made by Gaitonde et al. (2009) to analyze the effects
0.20 µm were achieved. Multiple linear regression models of depth of cut and machining time on machinability aspects
and neural network models were developed for predicting such as machining force, power, specific cutting force, sur-
surface roughness and tool flank wear. The results showed face roughness and tool wear using second order mathemat-
that neural network models were suitable for predicting the ical models during turning of high chromium AISI D2 cold
outputs for a range of cutting conditions and could be utilized work tool steel with CC650, CC650WG and GC6050WH
in intelligent process planning for hard turning. Suresh et al. ceramic inserts. From parametric analysis, it was revealed
(2002) developed a surface roughness prediction model for that the CC650WG wiper insert performed better with ref-
machining mild steel, using response surface methodology erence to surface roughness and tool wear. Basheer et al.
(RSM). A simple three level (34 ) factorial design of exper- (2008) presented an experimental work on the analysis of
iments was adopted for experimentation. A second order machined surface quality on Al/SiCp composites leading to
mathematical model, in terms of machining parameters, was an ANN-based model to predict the surface roughness. The
developed for surface roughness prediction using RSM. The model showed high coefficient of correlation of R = 0.977
factor effects of the individual process parameters have been and mean absolute error of 10.4% between the actual and
determined. An attempt also has been made to optimize predicted values. The best surface quality was obtained at
the surface roughness prediction model using Genetic Algo- lowest value of feed rate, the smaller particle size and the
rithms (GA) to optimize the objective function. Minimum largest tool-nose radius parameter combinations. Ezugwu
and maximum surface rıughness values and their respective et al. (2005) used a three layered feed forward backprop-
optimal machining conditions have also been determined by agation neural network for the analysis and prediction of
GA program. Lee and Tarng (2001) used a computer vision the tangential force, axial force, spindle motor power con-
technique to inspect surface roughness of S45C steel under sumption, machined surface roughness, average flank wear,
a variation of turning operation. They used digital camera to maximum flank wear and nose wear during high-speed turn-
acquire a surface image of the workpiece and then the feature ing of nickel-based, Inconel 718 alloy using CNMG 120412
of the surface image. A polinomal network using a self-orga- inserts. Correlation coefficients between the model predic-
nizing adaptive modeling method was applied to contruct tions and experimental values were ranging from 0.6595 for
the relationships between the features of the surface image cutting force to 0.9976 for nose wear prediction. Al-Ahmari
and the actual surface roughness. Roy (2006) developed a (2007) compared the performance of multiple linear regres-
GA trained fuzzy expert system for predict the surface finish sion analysis (RA), RSM and computational neural networks

123
J Intell Manuf

(CNN) models in prediction of tool life, cutting force and (2006) presented an experimental work in dry turning of
surface roughness in turning austenitic AISI 302 stainless austenitic stainless steels (AISI 304 and AISI 316) using
steels. Descriptive statistics and hypothesis tests were con- CVD multilayer coated cemented carbide tools. The influ-
ducted to study and it was found that the CNN models were ences of cutting speed, cutting tool coating top layer and
better than RA and RSM models. Özel and Karpat (2005) workpiece material were investigated on the machined
utilized neural network modeling to predict surface rough- surface roughness and the cutting forces. The tool wear
ness and tool flank wear over the machining time for veriety characteristics were also examined under scanning electron
of cutting conditions in finish hard turning of hardened AISI microscope (SEM). Korkut et al. (2004) investigated the
H-13 steels using cubic boron nitride (CBN) tools. Regres- effect of cutting speed on tool wear and surface roughness in
sion models were also developed in order to capture process turning of AISI 304 austenitic stainless steel. Correlation was
specific parameters. Predictive ANN models were found to made between the tool wear/surface roughness and the chips
be capable of better predictions for machining outputs than obtained at the cutting speeds of 120, 150 and 180 m/ min.
regression model. Apart from the introduction, it is possible to find no
As can be seen from the literature, artificial intelligence attempt on modeling the process parameters in turning of
tools like GA, ANN, regression modeling and fuzzy logic are AISI 304 austenitic stainless steels from the machinability
found to be extremely used in modeling of surface finishing literature of these steels. The objective of this paper is to use
of turned parts. Although a broad literature survey has been the SVM for estimate surface roughness in turning AISI 304
conducted in turning of different materials under different austenitic stainless steel. Three different toolboxes namely,
turning parameters, little attention has been paid to turning of least square SVM (LS-SVM), Spider and SVM-KM based on
AISI 304 austenitic stainless steels. Additionally, the predict- clustering by k-means were used in modeling surface rough-
ing results of used techniques seem to have some disadvan- ness. Additionally, an ANN model has also been used to
tages, i.e. low convergence rate, obvious “over fitting”, and model surface roughness to compare the results with SVMs.
especially poor generalisation when few samples are avail-
able (Shi and Gindy 2007). In this context, this study becomes
Experimental setup and cutting conditions
necessary to model the surface roughness during machining
AISI 304 material by using support vector machines (SVM),
Turning experiments were conducted in dry machining con-
an effective artificial iltelligence tool, different from the novel
ditions on a JOHNFORD TC-35 lathe machine with Fanuc
literature. SVM based on statistical learning theory is a new
18-T CNC control and programmable tailstock and maximum
achievement in the field of data-driven modeling and imple-
spindle speed of 3.500 rpm and a 15 kW drive motor. Figure 1
mented succesfully in classification, regression and function
shows the CNC lathe machine where the turning is operated.
estimation (Kwok 1999; Cao and Tay 2003).

Cutting tools and workpiece materials


Austenitic stainless steel machining
The cutting tools used were commercial grade cemented
The AISI 300 series of austenitic stainless steels represent carbide inserts produced by Kennametal with the geometry
the largest group of steels in use of total (O’Sullivan and of CNMG 120408. The inserts were recommended for
Cotterell 2002). However, their machinability is more dif-
ficult than the other alloy steels due to the reasons such
as having low heat conductivity, built-up edge (BUE) ten-
dency and high work hardening. Many attemps have been
made to improve the machinability of austenitic stainless
steels. Xavior and Adithan (2009) used three types of cut-
ting fluids, namely coconut oil, emulsion and neat cutting
oil (immiscible with water) to determine their effects on
tool wear and surface roughness during turning of AISI
304 with carbide tools. Tekiner and Yeşilyurt (2004) deter-
mined the best cutting parameters in the AISI 304 austen-
itic stainless steel machining by considering the acoustic
emission during the cutting process. Tool flank wear, BUE,
chip formations, surface roughness of machined parts and
machine power consumptions were investigated and used
for determine the optimal cutting speed and feed rate. Ciftci Fig. 1 Orthogonal cutting of AISI 304 austenitic stainless steel

123
J Intell Manuf

Table 1 Chemical composition of AISI 304 austenitic stainless steel Table 3 Turning parameters and their factor levels
Elements Composition (%wt.) Symbol Turning parameter Level 1 Level 2 Level 3

C 0.0452 A Cutting speed (m/ min) 30 60 90


Mn 1.78 B Feed rate ( mm/rev) 0.15 0.25 0.35
Si 0.32 C Depth of cut ( mm) 0.5 1 1.5
S 0.019
P 0.0397
Cr 18.35 Table 4 Experimental results
Ni 8.9 Exp. Feed rate Cutting Depth of Surface
Mo 0.456 No ( mm/dev) speed cut ( mm) roughness
Nb 0.0312 (m/dak) (µm)

Cu 0.768 1 0.10 60 0.5 3.48


Ti 0.0001 2 0.10 60 1.0 2.54
V 0.06 3 0.10 60 1.5 3.01
Fe Balance 4 0.10 120 0.5 2.89
5 0.10 120 1 2.15
6 0.10 120 1.5 2.66
Table 2 Mechanical and physical properties of tested material
7 0.10 180 0.5 1.12
Density (×1000 kg/m3 ) 8
8 0.10 180 1 1.35
Poisson’s ratio 0.28
9 0.10 180 1.5 0.94
Elastic modulus (GPa) 194
10 0.10 60 0.5 3.27
Tensile strength ( MPa) 515
11 0.10 60 1 3.86
Yield strength ( MPa) 206
12 0.10 60 1.5 2.65
Elongation (%) 41
13 0.10 120 0.5 2.99
Hardness (HRB) 87
14 0.10 120 1 2.34
Thermal expansion (10−6 /◦ C) 17.4
15 0.10 120 1.5 1.63
Thermal conductivity (W/m-k) 15.8
16 0.10 180 0.5 1.21
Electric resistivity (10−9 -m) 720 17 0.10 180 1 0.86
18 0.10 180 1.5 1.45
19 0.10 60 0.5 6.21
machining austenitic stainless steels by Kennametal. The 20 0.10 60 1 4.83
inserts were clamped mechanically on a rigit tool holder. 21 0.10 60 1.5 5.36
As far as possible, the tests were carried out in accordance 22 0.10 120 0.5 3.17
with ISO 3685. The workpiece material was AISI 304 au- 23 0.10 120 1 2.69
stenitic stainless steel. This steel finds its applications in air 24 0.10 120 1.5 2.31
craft fittings, aerospace componenets, such as shipbuilding, 25 0.10 180 0.5 2.45
shafts, valves, special screws, crygenic vessels and compo-
26 0.10 180 1 2.35
nents for severe chemical environments (Xavior and Adithan
27 0.10 180 1.5 2.97
2009). The workpiece specimens were 120 mm long and
40 mm in diameter. Table 1 shows the chemical composition
of the machined workpiece material while the physical and
mechanical properties are listed in Table 2. and tested by using a three-level full factorial experimental
design method to determine surface roughness of samples. In
Turning parameters and experimental design this study, three independent variables, such as cutting speed,
feed rate and depth of cut were used as turning parameters.
The design of experiments (DOE) method is an important The machining parameters and their factor levels are shown
step in selection of cutting parameters and their levels within in Table 3. Surface roughness measurements were carried out
an effective and suitable experimental run. The correct choise on the machined surfaces using a Mitutoyo SJ-211 portable
of the experimental design has a large influence on the accu- device. Three measurements were made on the each surface
racy and the construction cost of the approximations. The 27 averaged for an accurate reading. Table 4 shows the experi-
different combination of turning parameters were determined mental results.

123
J Intell Manuf

H1
Support vector machines
H
SV

SVM is a very useful soft computing method based on H2 w

statistical learning theory presented by Vladimir Vapnik


(Salat and Osowski 2004; Vapnik 1998). SVM try to Class-2

construct a separating hyperplane maximizing the margin


between two data sets according to their classes which 2
SV m=
have been previously mapped to a high dimensional space w
(Thukaram 2005). To determine the margin, two parallel
W
hyperplanes are constructed on each side of the separating SV
hyperplane. An optimal separation is achieved by the hyper- m
Class-1 w ⋅ x + b = +1
plane which has the largest distance to the neighboring data
points of the both classes, because the larger the margin the w⋅x +b = 0

better the generalization error of the classifier is obtained. w ⋅ x + b = −1

SVM has been used in many pattern recognition and regres- Fig. 2 Optimal separating hyper plane
sion estimation problems and has been applied to the prob-
lems of dependency estimation, forecasting, and constructing
intelligent machines (Ekici et al. 2009). geometrical margin is found to be w−2 (Dash et al. 2007).
Considering noise with the slack variables ξi and error
SVM for classification problems penalty C, the optimal hyperplane can be found by solving
the following convex quadratic optimization problem;
SVMs have the potential to handle very large feature spaces, minimize
because the training of SVM is carried out so that the dimen- M (3)
1
2 w + C i=1 ξi
2
sion of classified vectors does not have as a distinct influence
on the performance of SVM as it has on the performance subject to
of conventional classifiers. That is why it is noticed to be  
yi w T xi + b ≥ 1 − ξi for i = 1, 2, . . . , M (4)
especially efficient in large classification problems (Vapnik
1998). Also, SVM-based classifiers are claimed to have good ξi ≥ 0 for all i
generalization properties compared to conventional classi- where ξi is the distance between the margin and the exam-
fiers, because in training the SVM classifier, the so-called ples xi lying on the wrong side of the margin. Applying the
structural misclassification risk is to be minimized, whereas Kuhn-Tucker conditions into the equivalent Lagrange dual
traditional classifiers are usually trained so that the empiri- problem;
cal risk is minimized (Dash et al. 2007). The mathematical maximize
expression of the SVM classifier can be defined as below.
Let n-dimensional input xi (i = 1, 2, ..., M), M be the 
M
1 
M
W (α) = αi − αi αk yi yk xiT xk (5)
number of samples belong to class-1 or class-2, and asso- 2
i=1 i,k=0
ciated labels be yi = 1 for class-1 and yi = − 1 for class-
2 respectively. For linearly separable data a hyperplane subjects to
f (x) = 0 which separates the data can be determined as, 
M


n yi αi = 0, C ≥ αi ≥ 0, i = 1, 2, . . . , M (6)
f (x) = w T x + b = wjxj + b = 0 (1) i=1
j=1 The number of variables in the dual problem gives the num-
where w is an n-dimensional vector and b is a scalar. The ber of training data. Denoting the optimal solution of the
vector w and the scalar b determine the position of the sep- dual problem with α ∗ and w ∗ the equality conditions in (2)
arating hyperplane. This separating hyperplane satisfies the holds for the training input-output pair (xi , yi ) only if the
constraints f (xi ) ≥ 1 if yi = 1 and f (xi ) ≥ −1 if yi = −1 associated α ∗ = 0. In this case, the training example xi is
and this results in called as support vector (SV). The number of SVs is con-
  siderably lower than the number of training samples making
yi f (xi )=yi w T xi + b ≥ +1 for i = 1, 2, . . . , M (2) SVM computationally very efficient (Dash et al. 2007). The
value of the optimal bias b∗ is found from the geometry,
The separating hyperplane that creates the maximum dis-
1  
tance between the plane and the nearest data is called as b∗ = − yi αi∗ s1T xi + s2T xi (7)
the optimal separating hyperplane as shown in Fig. 2. The 2
SV s

123
J Intell Manuf

where s1 and s2 are the arbitrary SVs for class-1 and 2005; Smola et al. 1998; Hearst 1998). A version of a SVM
class-2 respectively. Only the samples associated with SVs for regression was proposed in 1996 by Vladimir Vapnik,
are summed because the other elements of the optimal Harris Drucker, Chris Burges, Linda Kaufman and Alex
Lagrange multiplier α ∗ are equal to zero. The final decision Smola (Drucker et al. 1997). This method is called support
function is given by vector regression (SVR). The model produced by support
 vector classification (as described above) only depends on
f (x) = αi yi xiT x + b∗ (8) a subset of the training data, because the cost function for
SV s
building the model does not care about training points that
The unknown data sample x is then classified as, lie beyond the margin. Analogously, the model produced by
 support vector regression only depends on a subset of the
Class-1, if f (x) ≥ 0
x∈ (9) training data, because the cost function for building the model
Class-2, otherwise
ignores any training data that are close (within a threshold ε)
The nonlinear classification problems can also be solved by to the model prediction.
using SVM applying a kernel function. The classified data is In SVM regression, the learning task is transformed to the
mapped onto a high-dimensional feature space where the lin- minimization of the error function, defined through the so
ear classification is possible. Using a nonlinear vector func- called ε-insensitive loss function which controls the accu-
tion, racy of the regressor. The variables measure the cost of the
errors on the training points. These are zero for all points that
φ(x) = φ1 (x), φ2 (x), . . . , φm (x), m  n (10)
are inside the band.
to map the n-dimensional input vector x into the L ε (d, y(x)) ;
m-dimensional feature space, the linear decision function in L ε (d, y(x)) = {|d − y(x)| − ε , for |d − y(x)| ≥ ε
dual form is given by, 0, for |d − y(x)| < ε

f (x) = αi yi φ T (xi ) φ(x) (11) (14)
SV s
where ε is the assumed accuracy, d is the destination, x is the
Notice that in (11) as well as (4), the inner products are used. input vector which includes cutting speeds, feed rates and
A function that returns a dot product of the feature space depths of cut, and y(x) is the actual output of the network
mapping of the original data points is called a kernel func- (surface roughness). The actual output of the SVM network
tion K (x, z) = φ T (x)φ(x). The learning in the feature space is defined by
does not require the inner products where a kernel function
is applied. Using a kernel function, the decision function can 
K
y(x) = ω j ϕ j (x) = w T ϕ(x) (15)
be written as,
j=0

f (x) = αi∗ yi K (xi , x) (12) where w = [w0 , w1 , . . . wk ]T is the weight vector and
SV s
ϕ(x) = [ϕ0 (x), ϕ1 (x), . . . , ϕk (x)]T the basis function vec-
There are different kernel functions used in the literature. tor. The learning task is defined as the minimization problem
Mercer’s theorem states that any symmetric positive-defi- of the error function E (Salat and Osowski 2004).
nite matrix can be regard as a kernel matrix. In this paper,
1
p
Gaussian radial basis kernel function which gives the best E= L ε (di , y (xi )) (16)
results is selected. The radial basis kernel function is defined p
i=1
as,

at the upper bound on the weight vector, w2 < C0 , where
|x − z|2 C0 is a user-specified constant and p is the number of learning
K (x, z) = exp − (13)
2σ 2 data pairs (xi , di ). Introducing the slack variables ξi and ξi∗ ,
the learning problem can be redefined as the minimization of
where σ is the width of the Gaussian function and it will
the cost function
be called as “kernel parameter” in this paper. The detailed p
information about the SVM can be found in Vapnik (1998)     1
∗ ∗
φ w, ξ, ξ = C ξi + ξi + wT w (17)
and Vojtech and Hlavac (2004). 2
i=1

SVM for regression problems at the following functional and boundary constraints:
di − w T ϕ (xi ) ≤ ε + ξi
Although SVM is used for classification problems, it is w T ϕ (xi ) − di ≤ ε + ξi∗ (18)
also an excellent tool for regression problems (Thukaram ξi , ξi∗ ≥ 0

123
J Intell Manuf

The two variables ε and C are free parameters that control Artificial neural networks
the dimension of the approximating function. Both must be
selected by the user (Salat and Osowski 2004; Burges 1998; ANNs are very efficient on adaptation and learning and for
Zhang 2004). The optimization problem is solved by the this reason they are used as modeling tools in a number of
introduction of the Lagrangian function and Lagrange mul- applications. An ANN is made of three types of layers: an
tipliers αi , αi∗ (i = 1, 2, ....., p) responsible for functional input layer which accepts the input variables. Hidden layers
constraints. The minimization of the Lagrangian function has has some number of neurons, and an output layer made of
been transformed to the so-called dual problem, one neuron that in the case examined herein gives the sur-
 p
   p
  face roughness Ra. Hidden and output layers are composed

max di αi − αi − ε αi + αi∗ of some number of neurons that perform a specific nonlinear
i=1 i=1 function. The neurons of one layer are interconnected to the
1    
p p
 neurons of the pre and after layers through weighted links.
± αi − αi∗ α j − α ∗j K xi , x j (19) Each neuron of the hidden and output layers is offset by a
2
i=1 j=1
threshold value (Sharma et al. 2008).
at the constraints The mathematical model of an artificial neuron’s behavior
 p
  is the simplification of the biological brain neuron as shown
αi − αi∗ ≥ 0 in Fig. 3. Various inputs x(n) to the network multiplied by
i=1 (20) weights w(n) are sent to a neuron. Performing accumulation
0 ≤ αi ≤ C,
and threshold, the neuron sums the weighted inputs, passes
0 ≤ αi∗ ≤ C the result through a non-linear transfer function and provides
    an output Yi (Zhang et al. 2007; Chang et al. 2010)
where K xi , x j = ϕ T (xi ) ϕ x j is a inner-product kernel n−1

defined in accordance with Mercer’s theorem on the basis of 


the learning data set. Note that the dual problem belongs to Yi = f wi xi − θ (23)
the quadratic programming optimization tasks with respect i=0

to the Lagrange multipliers. The solution of it is relatively where the inputs of xi in this study corresponds to cutting
easy and leads to the global minimum. After solving the dual speed, feed rate, and depth of cut; θ is the internal threshold
problem, the optimum solution for the vector w is given by or offset of a neuron; and f is the non-linear transfer function.
The most commonly used f is defined by the sigmoid logistic

Ns
 
∗ function as:
w= αi0 − α0i ϕ (xi ) (21)
i=1 1
f (x) = (24)
where Ns = K is the number of so-called support vectors 1 + e−x
(equal to the number of nonzero Lagrange multipliers). The the weights are dynamically updated using the BP algorithm.
network output signal y(x) can be expressed through the The difference between the target output and actual output
Lagrange multipliers and kernel function K (x, xi ) (Salat and (learning error) for a sample p is (Tosun and Ozler 2002)
Osowski 2004);
1 
K
2

Ns
  Ep = d pk − o pk (25)

y(x) = αi0 − α0i K (x, xi ) + ω0 (22) 2
k=1
i=1
where d pk and o pk are the desired and calculated output for
and does not need to know the explicit form of the nonlin-
kth output, respectively. K denotes the number of neuron in
ear basis function ϕ(x). The most known kernel functions
output of network. The average error for whole system is
used in practice are radial (Gaussian), polynomial, spline,
obtained by (Çaydaş and Hasçalık 2008):
or sigmoidal functions (Salat and Osowski 2004). The most
important is the choice of coefficients σ (kernel parameter)
and C for obtaining the optimum network. The constant C is X0
the weight, determining the balance between the complex- W1
n −1
ity of the network, characterized by the weight vector w and X1 Output
∑ wi xi − θ
Transfer
Input W2 function
the error of approximation, measured by the slack variables i =1
ξi (i = 1, 2, . . . , p). For the normalized input signals, the Wn-1
X2
value of C is usually adjusted as much bigger than 1. In this
paper, these parameters will be adjusted by testing a lot of
different values to be found the best results. Fig. 3 The behavior of an artificial neuron

123
J Intell Manuf

1 
P K
2
Ep = d pk − o pk (26)
2
p=1 k=1

where P is the total number of instances. For the purpose


of minimizing E p , the weights of the inter-connections are
adjusted during the training procedure until the expected
error is achieved. To adjust the weights of the networks, the
process starts at the output neuron and works backward to the
hidden layer. The weights in BP based on the delta learning
rule can be expressed as follows:
w new = w old + w (27) Fig. 4 Grid search for LS-SVM

∂Ep
w = −η out j (28) 5
∂w 4.5

Surface Roughness
where out j is the j th neuron output, η is the learning rate 4
parameter controlling stability and rate of convergence of 3.5
the network, which is a constant between 0 and 1. Once the 3
weights of all the links of the network are decided, the deci- 2.5
sion mechanism is then developed. 2
1.5
1 Estimated
Experimental
Estimation of surface roughness using SVM regression 0.5
1 2 3 4 5 6 7 8 9
and ANN
Number of Experiment
To estimate the surface roughness, 27 experiments were per- Fig. 5 Comparison of experimental and estimated results of LS-SVM
formed firstly. In each experiment; cutting speed, feed rate
and depth of cut were changed and the surface roughness was
recorded. The cutting speed, feed rate and depth of cut were The data set obtained from the experiments was divided
selected as 30-60-90, 0.15-0.25-0.35 and 0.5-1-1.15 respec- the two groups for training and testing of SVM. One data
tively. Three parameters including cutting speed, feed rate set obtained from 18 experiments was used for training and
and depth of cut were selected as the inputs (xi ) of SVM the rest of them (9 experiment) which are not introduced to
while the surface roughness was the output (yi ). the network before were used for testing of SVM. The grid
All of the computer implementations were performed by search results of LS-SVM are shown in Fig. 4.
using MATLAB software (Demuth et al. 2006). In the lit- As can be shown in Fig. 4, the better results are obtained
erature, there are several SVM toolboxes for MATLAB. while σ > 0.5 and C > 120. After grid search, the optimum
To evaluate the performances of different SVM toolboxes, parameters are found as σ = 7.7 and C = 670 when the sum
least square SVM (LS-SVM) (Suykens et al. 2002), of average error is equal to 0.2997. The comparison of real
Spider (Weston et al. 2006) and SVM-KM (Canu et al. 2005) and estimated surface roughness results which are obtained
based on clustering by k-means were experimented. Gauss- from the best SVM structure constructed by the optimum
ian radial basis function which was used most often in the free parameters are shown in Fig. 5.
literature was selected as the kernel function. As mentioned The grid search results of Spider are shown in Fig. 6. The
above, the free parameters σ and C for the kernel function better results are obtained from the values of σ > 0.5 and
should be adjusted by users to obtain the best SVM outputs. C > 40. The minimum sum of average error is 0.2748 by
Grid search is the simplest way to determine the values for adjusting σ > 3.9 and C > 900. The comparison of the
parameters σ and C. By setting the upper and lower bounds experimental and estimated results is illustrated in Fig. 7.
(search interval) and the jumping interval in the search, we Finally, a grid search implementation was performed for
can find sets of values for parameters σ and C which lead to the SVM-KM. The grid search results for SVM-KM are illus-
the highest accuracy rate of classification in this interval. In trated in Fig. 8. As can be shown in Fig. 8, better test results
this study, search intervals were selected as [0 10] for kernel are obtained by adjusting σ > 0.4 and C > 10. The best
parameter σ and [0 1000] for the regularization parameter C. results are obtained from the testing when the free param-
The jumping intervals were adjusted as 0.1 and 10 for σ and eters are selected as σ = 0.5 and C = 30. By using these
C respectively. parameter values, the sum of average error is equal to 0.2832.

123
J Intell Manuf

5
4.5

Surface Roughness
4
3.5
3
2.5
2
1.5
1 Estimated
Experimental
0.5
1 2 3 4 5 6 7 8 9
Number of Experiment
Fig. 6 Grid search for spider
Fig. 9 Comparison of experimental and estimated results of SVM-KM
5
4.5 4.5
Surface Roughness

4
4
3.5
3 3.5
2.5
3

Average error
2
1.5 2.5
1 Estimated
Experimental
2
0.5
1 2 3 4 5 6 7 8 9 1.5
Number of Experiment
1
Fig. 7 Comparison of experimental and estimated results of spider
0.5

0
0 5 10 15 20 25
Number of hidden layer neurons

Fig. 10 Performance of ANNs with different hidden layer neurons

one input, one hidden and one output layer. The network
architecture has 18 inputs, and 1 output neurons. Number
of hidden layer neurons is varied from 1 to 25. The perfor-
mances of these networks which have different number of
neurons in hidden layers are shown in Fig. 10. As shown in
Fig. 8 Grid search for SVM-KM
Fig. 10, the best test results are obtained from the network
which has 15 hidden neurons in the hidden layer. Hyperbolic
tangent sigmoid function is selected as the transfer function
The comparison results of the experimental and estimated for input and hidden layer. Linear transfer function is used in
data belonging to surface roughness are shown in Fig. 9. the output layer. The network is trained by using the Leven-
All of the grid search applications show that the selec- berg-Marquardt training function. After 1000 epochs, train-
tion of the optimum parameters is very important for the ing of the network is stopped and the testing results of this
performance of SVM. It can be said that the selection of ker- network are shown in Fig. 11.
nel parameter as σ ≥ 0.4 and C > 1 is very effective to be It is observed from the Fig. 11 that results of ANN are
obtained the best performance of SVM. By adjusting the opti- worse than the results of SVMs. On the other hand, the train-
mal free parameters, LS-SVM, Spider and SVM-KM give ing time of ANN is very long (122.93 s) whereas the time
the similar results which have the acceptable error tolerance. needed for the training of SVM is very short (only a few sec-
To investigate the performance of artificial neural networks onds). The average errors and elapsed times for each SVMs
(ANN), the same data set is used as inputs of ANN. A multi and ANNs are shown in Table 5.
layered feed forward neural network with the back-propaga- The testing results of all SVMs and ANNs are compared
tion algorithm is employed. The multi layered network has with the experimental results as can be shown in Table 6. To

123
J Intell Manuf

6
Table 6 Comparison of SVMs and ANN results with experimental
findings
5
Surface Roughness

Experimental LS-SVM Spider SVM-KM ANN


4
2.5400 2.7519 3.2399 2.9430 2.9435
3 2.1500 2.2457 2.5437 2.6151 2.7495
1.3500 1.1037 1.0326 1.0177 1.0359
2
3.8600 3.5411 3.8731 3.6891 2.9905
1 2.3400 2.2062 2.6080 2.5215 2.2527
Estimated
Experimental 0.8600 1.6884 1.5769 1.5380 1.3245
0
1 2 3 4 5 6 7 8 9 4.8300 4.7266 4.8229 4.8105 5.7824
Number of Experiment 2.6900 2.1497 2.7047 2.4034 3.0646
Fig. 11 Comparison of experimental and estimated results of ANN 2.3500 2.5687 2.3941 2.3714 2.7639

Table 5 Elapsed time and average errors for SVMs and ANN
Statistical paired t test
LS-SVM Spider SVM-KM ANN

Average error 0.2997 0.2748 0.2832 0.3275


In order to validate the modeling results of SVM and ANN
with the experimental results, the statistical paired t test was
Elapsed time (s) 0.45 0.42 0.25 122.93
employed. SPSS package program was used for this purpose.
The results of test are summarized in Table 7. As can be seen
from this table, all pairs are significance at 0.001 levels with
compare the results clearly, a post-process operation is per-
high correlation coefficients. The most correlated pair was
formed by using ‘postreg’ command of MATLAB. Postreg
pair 2 (spider SVM). Pairs 3 (SVM-KM), 1 (LSSVM) and 4
command post-processes the network training set by per-
(ANN) were followed.
forming a linear regression between each element of the net-
work response and the corresponding target. Where R is the
regression value (R = 1 means perfect correlation) (Sharma Concluding remarks
et al. 2008). The post-regression results are illustrated in the
Fig. 12. The post-regression results show that the R values In this study, the experimental observations were incorpo-
of the SVMs are better than the R value of the ANN. rated into the SVMs models for orthogonal cutting of AISI

Outputs vs. Targets (T), R=0.94496 Outputs vs. Targets (T), R=0.95715
6 6
Y=(0.84)T+(0.42)

Y=(0.91)T+(0.44)

data points data points


best linear fit
best linear fit 4
4 Y=T
Y=T

2 2

0 0
0 1 2 3 4 5 0 1 2 3 4 5
LS-SVM Spider

Outputs vs. Targets, R=0.95602 Outputs vs. Targets (T), R=0.91451


6 6
Y=(0.89)T+(0.39)

data points data points


Y=(1)T+(0.13)

best linear fit


best linear fit
4 4 Y=T
Y=T

2 2

0 0
0 1 2 3 4 5 0 1 2 3 4 5
SVM-KM ANN

Fig. 12 Post-regression results

123
J Intell Manuf

Table 7 Paired samples t-test transmission line using support vector machine. IEEE Transac-
tions on Power Delivery, 22, 67–73.
N Correlation Mean t Sig. Davim, J. P., Gaitonde, V. N., & Karnik, S. R. (2008). Investiga-
tion into the effect of cutting conditions on surface roughness
Pair 1 EXP & LSSVM 9 .945 −.0013333 −.010 .000
in turning of free machining steel by ANNmodels. Journal of
Pair 2 EXP & SPIDER 9 .957 −.2028778 −1.754 .000 Materials Processing Technology, 205, 16–23.
Pair 3 EXP & SVMKM 9 .956 −.1044111 −.888 .000 Demuth, H., Beale, M., & Hagan, M. (2006). Neural network toolbox
Pair 4 EXP & ANN 9 .915 −.2152778 −1.179 .001 5, user’s guide (pp. 9–16). MA: The Mathworks, Inc.
Drucker, H., Burges, C. J. C., Kaufman, L., Smola, A., & Vapnik, V.
(1997). Support vector regression machines. Advances in Neural
Information Processing Systems 9, NIPS 1996, 155–161, MIT
Press.
304 austenitic stainles steels. Three different SVMs mod- Ekici, S., Yıldırım, S., & Poyraz, M. (2009). A pattern recogni-
els were developed to predict surface roughness after the tion application for distance protection. Journal of The Fac-
ulty of Engineering and Architecture of Gazi University 24,1,
turning process. Additionally, a multi layered feedforward
51–56
backpropagation ANN model was also used for the same Ezugwu, E. O., Fadare, D. A., Bonney, J., Da Silva, R. B., & Sales, W.
purpose. Spider SVM was found the most predictive model F. (2005). Modeling the correlation between cutting and process
with R of 0.95715, while the SVM-KM (R = 0.95602) was parameters in high-speed machining of Inconel 718 alloy using an
artificial neural network. International Journal of Machine Tools
the second accuracy model and the LS-SVM was the least
and Manufacture, 45, 1375–1385.
(R = 0.94496). All the used SVM predictions were better Gaitonde, V. N., Karnik, S. R., Figueira, L., & Davim, J. P. (2009)
than ANN results (R = 0.91451). The ANN model involves Machinability investigations in hard turning of AISI D2 cold
more computationally time than SVMs’s. work tool steel with conventional and wiper ceramic inserts.
International Journal of Refractory Metals and Hard Materials
(in press).
Grzesik, W., & Brol, S. (2003). Hybrid approach to surface rough-
ness evaluation in multistage machining processes. Journal of
References Materials Processing Technology, 134, 265–272.
Hearst, M. A. (1998). Support vector machines. IEEE Intelligent
Al-Ahmari, A. M. A. (2007). Predictive machinability models for a Systems 18–21.
selected hard material in turning operations. Journal of Materials Karayel, D. (2008). Prediction and control of surface roughness in
Processing Technology, 190, 305–311. CNC lathe using artificial neural network. Journal of Materials
Bağcı, E., & Işık, B. (2006). Investigation of surface roughness in Processing Technology.
turning unidirectional GFRP composites by using RS methodol- Korkut, I., Kasap, M., Ciftci, I., & Seker, U. (2004). Determination
ogy and ANN. International Journal of Advanced Manufacturing of optimum cutting parameters during machining of AISI 304
Technology, 31, 10–17. austenitic stainless steel. Materials and Design, 25, 303–305.
Basheer, A. C., Dabade, U. A., Joshi, S. S., Bhanuprasad, V. V., & Gad- Kwok, J. (1999). Moderating the outputs of support vector machine
re, V. M. (2008). Modeling of surface roughness in precison classifier. IEEE Transactions on Neural Networks, 10, 1018–1031.
machining of metal matrix composites using ANN. Journal of Lee, B. Y., & Tarng, Y.S. (2001). Surface roughness inspection by
Materials Processing Technology, 197, 439–444. computer vision in turning operations. International Journal of
Benardos, P. G., & Vosniakos, G. C. (2003). Predicting surface rough- Machine Tools and Manufacture, 41, 1251–1263.
ness in machining: A review. International Journal of Machine Muthukrishnan, N., & Davim, J. P. (2009). Optimization of machin-
Tools and Manufacture, 43, 833–844. ing parameters of Al/SiC-MMC with ANOVA and ANN anal-
Burges, C. J. C. (1998). A tutorial on support vector machines for pat- ysis. Journal of Materials Processing Technology, 209, 225–232.
tern recognition. Data Mining and Knowledge Discovery, 2, 121– Nalbant, M., Gökkaya, H., Toktaş, I., & Sur, G. (2009). The exper-
167. imental investigation of the effects of uncoated, PVD- and
Canu, S., Grandvalet, Y., Guigue, V., & Rakotomamonjy, A. (2005). CVD-coated cemented carbide inserts and cutting parameters
SVM and Kernel Methods Matlab Toolbox, Perception Systèmes on surface roughness in CNC turning and its prediction using
et Information, INSA de Rouen, Rouen, France. artificial neural network. Robotics and Computer Integrated Man-
Cao, L. J., & Tay, F. E. H. (2003). Support vector machine with ufacturing, 25, 211–223.
adaptive parameters in financial time series forecasting. IEEE O’Sullivan, D., & Cotterell, M. (2002). Machinability of austenitic
Transactions on Neural Networks, 14, 1506–1518. stainless steel SS 303. Journal of Materials Processing Technol-
Çaydaş, U., & Hasçalık, A. (2008). A study on surface roughness ogy, 124, 153–159.
in abrasive waterjet machining process using artificial neural Özel, T., & Karpat, Y. (2005). Predictive modeling of surface rough-
networks and regression analysis method. Journal of Materials ness and tool wear in hard turning using regression and neural
Processing Technology, 202, 574–582. network. International Journal of Machine Tools and Manufac-
Chang, P. C., Lin, J. J., & Dzan, W. Y. (2010). Forecasting of manu- ture, 45, 467–479.
facturing cost in mobile phone products by case-based reasoning Özel, T., Karpat, Y., Figueira, L., & Davim, J. P. (2007). Modeling
and artificial neural network models. Journal of Intelligent Man- of surface finish and tool flank wear in turning of AISI D2
ufacturing. doi:10.1007/s10845-010-0390-7. steel with ceramic wiper inserts. Journal of Materials Processing
Ciftci, I. (2006). Machining of austenitic stainless steels using CVD Technology, 189, 192–198.
multi-layer coated cemented carbide tools. Tribology Interna- Paiva, A. P., Ferreira, J. R., & Balestrassi, P. P. (2007). A multi-
tional, 39, 565–569. variate hybrid approach applied to AISI 52100 hardened steel
Dash, P. K., Samantaray, S. R., & Ganapati, P. (2007). Fault classifica- turning optimization. Journal of Materials Processing Technol-
tion and section identification of an advanced series-compensated ogy, 189, 26–35.

123
J Intell Manuf

Roy, S. S. (2006). Design of genetic-fuzzy expert system for predicting Tosun, N., & Ozler, L. (2002). A study of tool life in hot machin-
surface finish in ultra-precison diamond turning of metal matrix ing using artificial neural networks and regression analysis
composite. Journal of Materials Processing Technology, 173, 337– method. Journal of Materials Processing Technology, 124, 99–
344. 104.
Salat, R., & Osowski, S. (2004). Accurate fault location in the power Umbrello, D., Ambrogio, G., Filice, L., & Shivpuri, R. (2008). A
transmission line using support vector machine approach. IEEE hybrid finite element method—Artificial neural network approach
Transactions on Power Systems, 19, 879–886. for predicting residual stress and the optimal cutting conditions
Sharma, V. S., Dhiman, S., Schgal, R., & Sharma, S. K. (2008). Estima- during hard turning of AISI 52100 bearing steel. Materials and
tion of cutting forces and surface roughness for hard turning using Design, 29, 873–883.
neural networks. Journal of Intelligent Manufacturing, 19, 473– Vapnik, V. (1998). The support vector method of function estima-
483. tion. In J. Suykens, J. Vandewalle (Eds.), Nonlinear modeling:
Shi, D., & Gindy, N. N. (2007). Tool wear predictive model based Advanced black-ox techniques (pp. 55–86). Dordrecht: Kluwer
on least squares support vector machines. Mechanical Systems Vojtech, F., & Hlavac, V. (2004). Statistical pattern recognition tool-
and Signal Processing, 21, 1799–1814. box for MATLAB (SPRTOOL). User’s Guide; http://cmp.felk.
Smola, A. J., & Schölkopf, B. (1998). A tutorial on support vec- cvut.cz/cmp/cmpsoftware.html.
tor regression. Technical Report NC2-TR-1998-030, ESPRIT Weston, J., Elisseeff, A., Bakır, G., & Sinz, F. (2006). The spider
Working Group in Neural and Computational Learning. software package. http://www.kyb.tuebingen.mpg.de/bs/people/
Suresh, P. V. S., Rao, P. V., & Deshmukh, S. G. (2002). A genetic spider.
algorithmic approach for optimization of surface roughness pre- Wong, S. V., & Hamouda, A. M. S. (2003). Machinability data rep-
diction model. International Journal of Machine Tools and Man- resentation with artificial neural network. Journal of Materials
ufacture, 42, 675–680. Processing Technology, 138, 538–544.
Surjya, K. P., & Chakraborty, D. (2005). Surface roughness prediction Xavior, M. A., & Adithan, M. (2009). Determining the influance of
in turning using artificial neural network. Neural Computing and cutting fluids on tool wear and surface roughness during turn-
Applications, 14, 319–324. ing of AISI 304 austenitic stainless steel. Journal of Materials
Suykens, J. A. K., Van Gestel, T., De Brabanter, J., De Moor, B., & Processing Technology, 209, 900–909.
Vandewalle, J. (2002). Least squares support vector machines. Zhang, J. H. R. (2004). A new algorithm of improving fault loca-
World Scientific, Singapore, (ISBN 981-238-151-1). tion based on SVM. Eighth IEE International Conference on
Tekiner, Z., & Yeşilyurt, S. (2004). Investigation of the cutting param- Developments in Power System Protection, 1, 204–207.
eters depending on process sound during turning of AISI 304 Zhang, J. Z., Chen, J. C., & Kirby, E. D. (2007). The development of an
austenitic stainless steel. Materials and Design, 25, 507–513. in-process surface roughness adaptive control system in turning
Thukaram, D., Khincha, H. P., & Vijaynarasimha, H. P. (2005). Arti- operations. Journal of Intelligent Manufacturing, 18, 301–311.
ficial neural network and support vector machine approach for
locating faults in radial distribution systems. IEEE Transactions
on Power Delivery, 20, 710–721.

123

View publication stats

You might also like