You are on page 1of 5

Indian Journal of Fibre & Textile Research

Vol. 39, March 2014, pp. 55-59

Performance of SVM based on PUK kernel in comparison to SVM based on RBF


kernel in prediction of yarn tenacity
Khalid AA Abakar & Chongwen Yua
College of Textiles, Donghua University, Shanghai 201620, People's Republic of China

Received 25 May 2012; revised received and accepted 23 January 2013

A new kernel function of SVM based on the Pearson VII function has been applied and compared with the commonly
applied kernel functions, i.e. the polynomial and radial basis function (RBF), to predict yarn tenacity. It is found that the
SVM model based on Pearson VII kernel function (PUK) shows the same applicability, suitability, performance in
prediction of yarn tenacity as against SVM based RBF kernel. The comparison with the ANN model shows that the two
SVM models give a similar predictive performance than ANN model.

Keywords: Artificial neural network, Pearson VII kernel function (PUK) kernel, Radial basis function kernel, Support
vector machines, Yarn properties

1 Introduction linear form. Some popular kernels are linear,


Modeling of yarn properties and the relationship polynomial, Gaussian (radial basis function (RBF))
between fibre and yarn properties are the most and Sigmoid kernel. The particular choice of a kernel
fascinating areas of research in the domain of textile function to map the non-linear input space into a
engineering. Researchers have made a lot of efforts to linear feature space depends highly on the nature of
formulate mathematical, statistical, empirical and the data, i.e. which kind of underlying relationship
intelligent models to predict various yarn properties, needs to be estimated to relate the input data with the
such as yarn strength and yarn unevenness. In recent desired output property. Because the nature of the
years, there has been growing use of soft computing data is usually unknown, the best mapping function
tools or intelligent models like artificial neural must be determined experimentally by applying and
network (ANN)1-3, fuzzy logic4-5 and genetic validating various kernel functions yielding the
algorithm or genetic programming (GP)6 to predict highest generalization performance. Therefore, it will
various yarn properties and to optimize the process. be very useful to search a kind of universal kernel,
Most recently, support vector machines (SVMs) is which can be adaptive for many types of data through
one of the best machine learning or data mining adjusting the kernel parameters.
techniques of knowledge discovery that aims at Pearson VII Universal Kernel was applied as a
extracting the information from databases. The kernel function of SVM10, and referred to as PUK.
suitability of using SVM in the prediction of cotton The Pearson VII function has excellent flexibility and
yarn properties to know the accuracy of prediction is possibility to change easily, from a Gaussian into a
studied by many reviews7-9. Lorentzian peak shape and more by adapting its
They found that like ANN model, the SVM model parameters. So, it is possible to use the Pearson VII
is able to predict with a reasonably good accuracy in function as a generic kernel which can replace the
most cases. However, one of the main reasons for the earlier mentioned set of kernel functions.
popularity of SVM is its ability to model complex
In this work, attempt has been made to study the
nonlinear relationships by selecting a suitable kernel
nonlinear regression problem of fibre/ yarn properties
function. Briefly, the kernel function transforms the
relationships, and to investigate the applicability,
input space into a high dimensional feature space
suitability and performance of SVM based on PUK
where non-linear relationships can be represented in a
kernel in comparison to the commonly applied SVM
a
Corresponding author. based on the RBF kernel to predict yarn tenacity from
E-mail: yucw@dhu.edu.cn cotton fibre properties.
56 INDIAN J FIBRE TEXT RES, MARCH 2014

2 Materials and Methods where ( x ) is the features of the input variables,


while w and b are the coefficients and they estimated
2.1 Data Set by minimizing
The cotton fibre properties and corresponding yarn
tenacity data were collected from the published 1 N
1

2
E ( w) = C yi f ( xi , w) + w (2)
review11, using the case-based reasoning technique to N i =1 2
predict yarn properties. A total of twenty-five
different cotton samples were collected and measured 1 N
by a high volume instrument (HVI). The fibre
The first term yi f ( xi , w) is the
N i =1
C
properties used as input data were fibre strength
empirical risk (error) and can be given by the
(g/tex), fibre length (inch), elongation (%), trash
insensitive loss function, as shown bwlow:
content (Cnt), length uniformity (%), and micronaire
(M). The yarn tenacity (cN/tex) was used as the 0 if x <
target. A set of twenty-five samples was used to train x := (3)
and test the models. x else

2.2 Introduction to Support Vector Machines 1 2


The term w is the regularization term and is the
Support vector machines (SVMs) introduced by 2
Vapnik12 are machine learning methods based on measure of the function fitness. The value of C ,
statistic theory. It was proposed as a novel approach being the a regularization constant , determines the
for solving classification problems. Additionally, with trade-off between the empirical risk and the
the introduction of Vapniks -insensitive loss regularization term.
function, the SVM has been extended to solve a In order to obtain the estimations w and b , Eq. (2)
nonlinear regression- estimation problem, called the is transformed into following Eq. (4) as optimal
SVM for regression (SVR)13,14. function, by introducing the positive slack variables
Recently, the SVR has been applied to various i and i* as follows:
fields such as optimal control, time-series prediction
and interval regression analysis15-17. However, to n
1
Minimize E ( w) = C ( i + i* ) +
2
obtain better performance, some parameters (called w
i =1 2
hyperparameters) must be selected carefully in the
SVM or SVR18,19. The hyperparameters include the Subjected to
kernel parameters, the epsilon value in Vapniks -
insensitive loss function, and the regularization yi f ( xi , w) + i
constant (The complexity parameter C). *
f ( xi , w) yi + i (4)
Basic comprehensive description of the concept *
underlying SVR modeling can be given as i , i 0
considering a data set {( x1 , y1 ), ( xi , yl )} ,
Slack variables i and i* can be introduced when
where x R d is the dimensional input space and
data cant be estimated by the function f under the
y R is the corresponding output and it can be
precise . Introducing Lagrange multipliers and
written as a regression function y = f ( x ).
according to Karush-Kuhn-Tucker conditions, Eq. (4)
The basic idea of support vector regression is to can be transformed into the following form:
map the input vector x into high dimensional feature
space by nonlinear mapping function and then to Minimize:
perform linear regression in the feature space. N l
However, the correlation between input and output LP ( * , ) = ( i* + i ) + yi ( i* i ) +
can be written as the following formula: i =1 i =1
(5)
1 N N

f ( x ) = w. ( x ) + b (1)

2 i =1 j =1
( i* i )( *j i ) K ( xi , x j )
ABAKAR & YU: SVM BASED ON PUK KERNEL IN COMPARISON TO SVM BASED ON RBF KERNEL 57

Subjected to: of commonly applied kernel functions, i.e. the linear,


N
polynomial, Gaussian and Sigmoid kernels. Adopting
(
i =1
*
i i ) = 0,
(6)
Pearson VII function as kernel function, it might
avoid the case that SVM cant match data well if the
* kind of kernel function of SVM was chosen wrongly.
0 , i C , i = 1, 2,..., N
i The pearson VII kernel function of multi-dimensional
input space is given by the following formula:
In Eq. (6), i , i* are Lagrange multipliers. The
model output is given below: 2
K ( xi , x j ) = 1/[1 + (2 xi x j 2(1/ ) 1 / )2 ]
N (11)
f ( x, ) = ( i* i )K ( xi , x j ) + b. (7)
i =1 There are no examples of application of PUK kernel
methods available in textile engineering related
where K(.,.) is the kernel function. Kernel function
research despite the many advantages that it
can have various forms but the two widely used
offers.According to Eq. (10), Pearson VII universal
kernel for real -valued data are polynomial and
kernel function, referred to as PUK in this study, is
Gaussian radial basis function (RBF) kernel. The
used as SVM kernel function and also compared to
polynomial kernel of degree d is defined as
the commonly used kernel the polynomial and
K ( xi , x j ) = ( xi .x j + k ) d (8) Gaussian radial basis (RBF) and ANN.

2.3 Artificial Neural Network


where k is the constant. The kernel with d = 1 is the
Artificial neural network (ANN) is suitable for
linear kernel function. The second very widely used
modeling nonlinear relationship. It is a powerful data
kernel is the Gaussian radial basis function (RBF)
modeling tool that is able to capture and represent any
kernel defined by
kind of input output relationships. The theory of
K ( xi , x j ) = exp( xi , x j / 2 ) 2 (9) ANN and its application in textile engineering exactly
in spinning process and modeling yarn properties
where > 0 is a parameter that controls the width of studies have extensively been discussed. The
the Gaussian. It plays a similar role as the degree of ANN employed in this study is a three layer back-
the polynomial kernel in controlling the flexibility of propagation (BP) multilayer perceptron network, the
the resulting classifier. input layer, one hidden layer, and the output layer.
Pearson VII universal kernel (PUK) is the other More details on the theory and applications of the
type of kernel function that can be used in support ANN can be found in a number of publications1-3.
vector machines and based on Ustun et al.10. The
2.4 Implementation
general form of the Pearson VII function for curve
The parameters of support vector machine for
fitting purposes is given by the following relationship:
regression (SVMR) such as the complexity parameter
C, the values of - insensitive loss function , and the
f ( x) = H /[1 + (2( x x0 ) 2(1/ ) 1 / ) 2 ] kernel parameters such as degree d of polynomial
(10)
kernel, width of RBF kernel function , and PUK
where H is the peak height at the center x0 of the kernel parameters ( and ) were optimized by using
peak, and x represents the independent variable. The grid search approach, in the training set data using
parameters and control the half-width (also 10-fold cross validation. This grid search tries values
named Pearson width) and the tailing factor of the of each parameter across the specified search range
peak. The main reason to use the Pearson VII function using geometric steps. This approach tries to search
for curve fitting is its flexibility to change, by varying best possible values using each parameters across
the parameters and . specified chosen search range, hence the best
Pearson VII function is adopted as an alternative performance results were reported with good
generic kernel function in this study. It might serve as prediction accuracy.
a kind of universal kernel which can replace To evaluate the prediction performance of each
(by selecting the appropriate parameter setting) the set algorithm, a 10-fold cross validation technique is
58 INDIAN J FIBRE TEXT RES, MARCH 2014

used. In 10-fold cross validation, the initial dataset is C=160.84, = 0.02, = 85.0, and = 9.1, which give
randomly partitioned into 10 mutually exclusive the smallest value of RMSE.
subsets or folds D1, D2, , D10, each of The initial architecture of the ANN selected has
approximately equal size. The training and testing are used all the eight variables in input layer with four
performed 10 times. In iteration i, partition Di is neurons in the hidden layer, selected by the auto
reserved as the test set and the remaining partition are builtin function and one output neuron. The
collectively used to train the model. In this method, optimization of the BP multilayer perceptron model is
each data point is used for the same number of times done with 10-fold cross validation. After several
for training and once for testing. Then the error trials, the optimal learning rate of 0.01 and a
estimates are calculated and then averaged. The errors momentum coefficient of 0.3 are determined, and the
that were used as an indicator of the predictive number of epochs is selected as 300.
performance of the models were root mean-squared After the completion of model development or
error (RMSE), relative error (RE %), and correlation training, all the models based on SVM and ANN are
coefficient (R). For implementation and to carry out subjected to the unseen testing data set. Statistical
our experiments, the SVM and ANN models were parameters such as root mean-squared error (RMSE),
executed by using rapid miner software program. relative error (RE %), and correlation coefficient (R)
are used to compare the predictive power of the SVM-
3 Results and Discussion based and ANN based models. Results are
3.1 Models Performance Comparison summarized in Table1.
The goal of this part of research is to compare the It can be seen from Table 1 that the largest values
prediction results provided by SVM based on of RMSE and RE are provided by SVM based on
polynomial, Gaussian radial basis (RBF) and PUK polynomial kernel. This indicates that the
kernels function as well as the ANN model. The generalization performance of SVM based on
optimization of SVM parameters is performed by polynomial kernel is bad. Therefore, this model does
using grid search approach in the training set data not fit the data well. The RMSE of SVM based on
using 10-fold cross validation, and depending on RBK kernel is bound to be lower than that of SVM
smallest RMSR error we selected the optimal based on PUK kernel. The RE and R values provided
parameters of the model. The regularization constant by both SVM models based on RBK and PUK kernels
( the complexity parameter C ), parameters of the are the same.
- insensitive loss function, kernel parameters of Therefore, the RMSE values of both SVM models
the degree d of polynomial kernel , width of RBF based on RBK and PUK kernels are lower than that of
kernel , and PUK kernel parameters and were ANN model. The RE and R values are found to be
optimized. The optimal parameters of SVM based on close with a little difference. This comparison shows
polynomial kernel are C=1.0, = 0.002, and d = 3.0. that both SVM models based on RBK and PUK
The optimal parameters of SVM based on RBF kernel kernels and ANN model are pretty similar in terms of
are C=160.84, = 0.001, and = 0.04, and the the average errors. The comparison of performance of
optimal parameters of SVM based on PUK kernel are the nonlinear models demonstrates that the SVM
Table 1Comparison of the predictive performance of SVM and ANN model
Experimental data Predicted value using
SVM based SVM based SVM based ANN
polynomial kernel RBF kernel PUK kernel
11.86 11.266 11.155 11.275 11.559
13.52 12.247 12.582 12.616 13.564
18.28 30.042 20.582 18.748 16.705
11.77 6.092 11.278 10.796 9.484
10.82 11.362 10.736 10.794 11.691

Error
RMSE 5.369 0.911 0.646 1.280
RE, % 22.42 5.02 4.31 7.91
R 0.934 0.987 0.989 0.901
ABAKAR & YU: SVM BASED ON PUK KERNEL IN COMPARISON TO SVM BASED ON RBF KERNEL 59

Table 2Anova test model shows that the two SVM models have similar
Source Square DF Mean F Prob prediction performance as ANN model. The result of
sums squares this study will be helpful for further work of feature
Between 0.003 1 0.003 0.060 0.812 selection to know the importance of fibre properties
Residuals 0.821 18 0.046 on yarn tenacity by using SVM based on PUK kernel.
Total 0.824 19
Acknowledgement
based on PUK kernel has the similar ability in
The authors gratefully acknowledge Shanghai
predicting yarn tenacity with SVM based on RBF
Municipal Education Commission and Shanghai
kernel and ANN model.
Education Development Foundation for providing
3.2 Compare SVM Models by Significance Test
Shu Guang project (No. 10GG17).
The two SVM models with RBF and PUK kernel
have been compared to know which of them is more References
1 Majumdar P K & Majumdar A, Text Res J, 74(7) (2004) 652.
accurate and suitable in prediction of yarn tenacity
2 Chattopadhyay R & Guha A, Text Prog, 35(1) (2004) 1.
property, Hence, we applied two cross validation on 3 reyen M E & Grkan P, Fibers Polym, 9(1) (2008) 87.
the training set data using SVM model based on RBF 4 Majumdar A & Ghosh A, J Eng Fibers Fabrics, 3(4) (2008) 61.
kernel and on PUK kernel. The significance test 5 Admuthe L S &. Apte S, Text Res J, 80(9) (2010) 841.
(T test and Anova) is performed on the resulting 6 M. Dayik, Text Res J, 79(11), 963 (2009).
7 Yang J G, Lv Z J & Xiang Q, Yarn properties prediction
performance value to decide whether the difference in using support vector machines: an intelligent reasoning
performance is significant. Table 2 shows that the method. Proceedings, Fourth International Conference on
probability for random values with the same result is Fuzzy Systems and Knowledge Discovery, Haikou, 2007, 1.
0.812, these values are bigger than alpha= 0.050, 8 Ghosh A & Chatterjee P, Fibers Polym, 11(1) (2010) 84.
9 Nurwaha D & Wang X, Fibers Polym, 12(4) (2011) 546.
which indicates a probably difference between the
10 stn B, Melssen W J & Buydens L M C, Chemometrics
actual mean values was not significant, since Intelligent Laboratory Systems, 81(1) (2006) 29.
0.812 > alpha = 0.050. This means that the SVM 11 Cheng Y S J & Cheng K P S, Text Res J, 74(8) (2004) 718.
model based on PUK kernel has the same ability to 12 Vapnik V N, The Nature of Statistical Learning Theory
that of SVM model based on the RBF kernel to map (Springer-Verlag New York), 1995.
13 Vapnik V N, IEEE Transactions on Neural Networks,10(5)
the nonlinear relation between input and output data (1999) 988.
for predicting the yarn tenacity property. Hence, it is 14 Smola A J & Schlkopf B, Statistics Computing, 14(3)
possible to predict the yarn properties by using SVM (2004) 199.
model based on PUK kernel. 15 Suykens J A K, Vandewalle J & Moor B D, Neural
Networks, 14(1) (2001) 23.
16 Mukherjee S, Osuna E & Girosi F, Nonlinear prediction of
4 Conclusion chaotic time series using support vector machines,
It is observed that like SVM based on RBF, the Proceedings, IEEE Workshop on Neural Networks for Signal
SVM model based on PUK kernel shows the same Processing VII (Amelia Island Florida) 1997. 511.
applicability, suitability and performance to map the 17 Jeng J T, Chuang C C & Su S F, Fuzzy Sets Systems, 138(2)
nonlinear relation between input and output data for (2003) 283.
18 Tay F E H & Cao L, Omega, 29(4) (2001) 309.
predicting the yarn tenacity. A comparison of SVM 19 Chuang C C & Su S F, IEEE Transactions on Neural
models based on RBF and PUK kernels with ANN Networks, 13(6) (2002) 1322.

You might also like