You are on page 1of 6

2014 2nd International Conference on Information and Communication Technology (ICoICT)

Texture Feature Extraction Using Co-Occurrence


Matrices of Sub-Band Image For Batik Image
Classification
Agus Eko Minarno , Yuda Munarko , Arrie Kurniawardhani and Fitri Bimantoro Nanik Suciati
Jurusan

Teknik Informatika, Universitas Muhammadiyah Malang, Malang, Indonesia


Email: agoes.minarno@gmail.com and yuda.munarko@gmail.com
Institut Teknologi Sepuluh Nopember, Surabaya, Indonesia
Email: emailkuwardhani@gmail.com, beema.yichi@gmail.com and emailkuwardhani@gmail.com and nanik@if.its.ac.id

AbstractIn this study, we propose a new method to extract


texture features of batik images. The proposed method is called
co-occurrence matrices of sub-band images. This method is
proposed to overcome the problem in classifying batik images that
are acquired randomly from the internet. The problem of those
images is the batik images contain various types of noise, such
as unbalanced brightness, there are folds on fabrics images, the
different size of basic motifs, low contrast, and there is watermark
on the images. This method combines the advantages of gray-level
co-occurrence matrices (GLCM) and discrete wavelet transform
(DWT). First, the original image is decomposed using DWT to
provide sub-band images. Second, GLCM is applied to sub-band
images to extract the texture features. Those features will become
the input for the probabilistic neural network (PNN). The results
show that this method is robust enough to classify batik images.
The maximum accuracy that can be achieved is 72%.
Index Termsbatik, wavelet, gray-level co-occurrence matrices, probabilistic neural network

I. I NTRODUCTION
On October 2nd, 2009, UNESCO acknowledged that batik
is Indonesias cultural heritage. Name of batik is derived
from Javanese language, namely ?gembat? (throw) and ?itik?
(writing some dots on fabric or other materials) [1]. Batik is
an ancient technique for decorating textile [2]. Patterns that
compose batik are called motif. Based on the ornaments and
their structures, the type of batik can be identified by its
basic motif, as listed in Table I [3][4]. Therefore, batik can
be classified based on those motifs.
Several studies have discussed about batik image recognition and classification. Cheong and Loke [5] did recognition and classification colored texture for ?atik?and
?ongket?motifs. They used six multispectral co-occurrence
matrices to extract colored texture based on RGB. The
Tchebichef orthogonal polynomial was used to get the moment coefficients, as texture descriptors, from co-occurrence
matrices. That method gives a good degree of accuracy
in discriminating color texture. Continuing previous study,
Loke and Cheong [6] did feature reduction using Principal
Component Analysis (PCA). They reduced down feature to
2%. The result showed that the appropriate feature reduction
can increase the speed of recognition without significantly

978-1-4799-3580-2/14/$31.00 2014 IEEE

reducing the classification rate. Nurhaida et al [1] did performance comparison on three extraction methods for batik image
recognition. Those methods are Gray Level Co-occurrence
Matrices (GLCM), Canny Edge Detection, and Gabor filters.
The comparisons were conducted to determine which the best
method to recognizing batik images. The result showed that
GLCM is the best method in extracting information from batik
images.
Most studies of batik image recognition use texture feature
as descriptor to distinguish the type of batik. Most of them use
texture feature because texture can be used to represent motifs
in batik images. A texture extraction method can be done either
based on statistical approaches or spectral approaches. GLCM
is the most common texture extraction method that based
on statistical approaches. GLCM have been proved to be an
effective texture descriptor in [1] [7] and [8]. Whereas wavelet
transform is the most common texture extraction method based
on spectral approaches. Discrete wavelet transform (DWT)
have been proved to be a very powerful texture descriptor
used in image analysis [9] [10] [11].
Most studies about batik image recognition use dataset
that have been well prepared for a classification process.
For instance, one class of batik motif may consist of a
number of images that are acquired from one fabric that
captured from different sides. Whereas, in this study, dataset
are acquired randomly from the internet, thus one class of
batik motif contains various fabrics but has the same basic
motif. Moreover, as shown in Fig. 1, dataset acquired randomly
from the internet may contain various types of noise. First, the
image color could be light (high intensity) in one side and dark
(low intensity) in another side. It can be caused by unbalanced
brightness when capturing the images. Second, there are folds
on fabrics. Third, the different size of basic motifs. Fourth,
the low contrast that caused the edges of batik motif could
not be clearly visualized. Lastly, there is watermark on batik
images. Because of these complexity, classification of these
complicated batik images is not the trivial task.
To overcome these problems, the proper extraction method
of texture feature should be selected to achieve the high
accuracy rate of classification rate. In this paper, we propose

249

2014 2nd International Conference on Information and Communication Technology (ICoICT)

Fig. 1: Noise in images that acquired from internet. (a) unbalanced brightness, (b) folding, (c) different scale, (d) low contrast,
(e) watermark
TABLE I: Examples Type of Batik
Motif
Banji

Ceplok

Nitik

Lereng

Semen

Lunglungan
Buketan

Motifs Based on Geometrical


Short Description
Based on the swastika symbol,
a cross motif are connected with
each other at an angle of 90 degrees.
Based on repetitive basic shapes,
such as circle, square, oval, corner
less square, rosette, star and cross
sections of fruit.
Formed by a woven pattern that
composed by dots or short line.

A diagonal design, with a rather


straight line. There is an additional
decoration for the empty space in
a parallel pattern (issn issn).
Non-Geometrical Motifs
Consists of ornaments in the
form of plants, animals, mountains
(mere) or planting pot with free
placement rules.
Similar to the cement but with
fewer number of ornaments and the
absent of Meru
Bouquet

The second dataset is used to compare the proposed method


and the existed methods. As shown in Fig. 3, the second
dataset consists of seven types of batik motif, namely ceplok,
parang, semen, truntum, lereng, lung-lungan and tambal. Each
type consists of twelve batik images. The twelve batik images
are obtained by capturing images from two fabrics, which each
fabric is captured from six different sides. Thus, the second
dataset has 84 images. Size of each image in the second dataset
is 128x128 pixels. From each class, two images are used for
testing, and the rest images are used for training.

Examples
banji bengkok,
guling, kerton

kawung
ceplok
nogosari

picis,
keci,

rengganis,
nitikkrawitan,
tirtateja
alit,
truntum
parang barong,
parangrusak,
udanliris

B. Preprocessing
There are three pre-processing steps for the main dataset.
First, images are resized to 320x320 pixels. Second, the
resized images are converted to the greylevel intensity. Finally,
applying the adaptive histogram equalizer technique on the
converted images. Different from the main dataset, the preprocessing step for the second dataset is just the conversion
to the grey level intensity.

semen gurdo, kasut, yogya

babonangrem,
gragehwalu, lung
klewer, peleman
buket
isen
latar,
snow
white,
buketan
pekalongan

C. Wavelet

a new method to extract texture feature of batik images using


co-occurrence matrices of sub-band images. That method
combining the advantages of GLCM and discrete wavelet
transform, is robust enough to classify batik images that are
acquired randomly from the internet.

Wavelet transformation is utilised to decompose an image


into sub-band images. Sub-band images may consist of a low
frequency image or a high frequency image. From each subband image, a number of texture features can be extracted.
Thus, the texture features can be utilized for recognition and
classification.
The basic concept of wavelet is explained in decomposing
1-D signal. Wavelet decomposes the signal f(x) into sub-band
signals, using scaling and translation function known as the
mother wavelet. Mother wavelet is used for generating the
other window functions. The mother wavelet function is:

II. M ATERIAL AND M ETHOD

xa
1
)
a,b (x) = p (
b
|a|

This study used two types of datasets to measure the


performance of texture feature extraction. The first dataset
was acquired randomly from the internet. This dataset is the
main dataset. As shown in Fig. 2, the main dataset consists of
six types of batik motif, namely banji, ceplok, kawung, nitik,
parang, and lereng. Each type consists of ten batik images.
Thus, the main dataset has 60 images. For each class, three
images are used for testing, and the rest images are used for
training.
Whereas the second dataset was acquired by capturing
fabrics from various sides. This dataset is the second dataset.

Where a and b are the scaling factor and the translation


parameter, related to the location of the window. The mother
wavelet function is applied to the original function of signal
f(x), such as:

A. Dataset

978-1-4799-3580-2/14/$31.00 2014 IEEE

(1)

Z
(W f )(a, b) = {f, a,b (x)} =

f (x). a,b (x)dx

(2)

Image consists of 2-D signal, namely row and column.


Thus to decompose 2-D signal, we need 2-D discrete wavelet
transform (DWT).2-D DWT can be computed using 2-D

250

2014 2nd International Conference on Information and Communication Technology (ICoICT)

Fig. 2: Some of the Main Dataset : (a) banji, (b) ceplok, (c) kawung, (d) nitik, (e) parang, and (f) lereng

Fig. 3: Example Second Dataset : (a) ceplok, (b) parang, (c) semen, (d) truntum, (e) lereng, and (f) lung-lungan

Fig. 4: The 2-D DWT One Level Decomposition

wavelet filters followed by 2-D down sampling operations as


illustrated in Fig. 4 for one level decomposition.
2-D DWT is computed 1-D DWT over image rows and
columns separately. Firstly, filters are convoluted separately
on rows and columns to obtain a 2-D transformation. The 2D wavelet filters are obtained from product of 1-D wavelet
filters, such as:
DLL = lT l

(3)

DLH = lT h

(4)

DHL = hT l

(5)

DHH = hT h

(6)

three detail images HL, LH, and HH. Every sub-band images
contains the information at a specific scale and orientation. LL
corresponds to the approximation image. LH corresponding to
horizontal details contains image information of low horizontal
frequency and high vertical frequency. HL corresponding to
vertical details contains high horizontal frequency and low vertical frequency. HH corresponding to diagonal details contains
high horizontal and high vertical frequencies.
To get another level of decomposition, LL sub-band image
is successively decomposed using 2-D DWT. Thus, wavelet
packet decomposition of an image produces 4i sub-band
images, where ??is the level of decomposition.DWT has
various type of filter bank, such as Haar, Daubechies, Coiflets,
Biorthogonal, and etc. The different filter bank that used can
caused the different wavelet transform function that gives
different range of frequencies
D. Gray Level Co-Occurrence Matrix (GLCM)
GLCM is a square matrix with size LxL, where L is
the number of grey-level in the original image. That matrix
contains the probability value of two pixels with i greylevel intensity and j grey-level intensity, respectively, which
is separated by distance d and direction . Thus, the probability value can be written as P(i, j, d, ). The distance
d can be chosen from 1 to 8 and the direction can be any
one of 0 , 45 , 90 , 135 , 180 , 225 , 270 , 315 . Fig. 5 show
illustration of direction that can be used in GLCM. Whereas,
Fig. 6 show illustration of GLCM.

Where, l and h denote the 1-D low pass and high pass
filter, respectively. After transformation process, each image
gotten from transformation process is downsampled (?2) by a
factor of 2 in each direction. Downsampled operation is used
to compress the size of each of the bands to half.
In each level of decomposition, the 2-D DWT provides
four sub-band images, namely one low pass image LL and

978-1-4799-3580-2/14/$31.00 2014 IEEE

251

Fig. 5: Direction of

2014 2nd International Conference on Information and Communication Technology (ICoICT)

Corr =

L1
X

(i x )(j y )P (i, j, d, )
x y
i,j=0

(11)

Where:
PL1
x = i,j=0 i.P (i, j, d, ),
PL1
y = i,j=0 j.P (i, j, d, ),
PL1
x = i,j=0 (i x )2 .P (i, j, d, ),
PL1
y = i,j=0 (i y )2 .P (i, j, d, )
E. Probabilistic Neural Network (PNN)

Fig. 6: GLCM constructions. (a) Original image, (b) GLCM


indices, (c) Co-occurrence matrix, (d) Probability value

To characterize the texture of an image, some statistical


measures is extracted from GLCM. The most common and relevant statistical measures used in GLCM are Angular Second
Moment, Entropy, Contrast, Correlation and Inverse Difference
Moment [12].
Angular Second Moment (ASM) is also called energy.
Energy measures textural uniformity. If an image is completely
homogeneous, the value of energy will be high. Energy is
calculated by,
L1
X

ASM =

P 2 (i, j, d, )

Probabilistic neural network (PNN) is chosen, because its


speed in operation and accuracy in pattern recognition or classification [13]. PNN is a feed forward neural network based
on Bayes?arzen classification in statistics, which provides a
general solution to the pattern classification problems
The architecture of PNN is shown in Fig. 7. The figure
shows that PNN is composed of four layers, namely input
layer, pattern layer, summation layer, and decision layer.

(7)

i,j=0

Entropy is inversely correlated to energy, which measures


the diversity or randomness of an image. Entropy is calculated
by,
Ent =

L1
X

P (i, j, d, ).logP (i, j, d, )

Fig. 7: PNN Architecture

(8)

i,j=0

Contrast measures of local gray level variation of an image,


which indicates the variance of the gray level. Contrast will
have low value in a smooth image and high value in a coarse
image. Contrast is calculated by,
Cont =

L1
X

(i j)2 .P (i, j, d, )

(9)

The input layer consists of input nodes of feature vector. Thus, input layer does not conduct any computational
operation.The pattern layer consists of neurons equal to the
total number of training dataset. In the pattern layer, the
input pattern is calculated by the following multi-dimensional
Gaussian function with the probability density function (PDF)
based on Parzen window.

i,j=0

Inverse Difference Moment (IDM) is also called homogeneity. Homogeneity will have high value for a low contrast
image. IDM is calculated by,
IDM =

L1
X

P (i, j, d, )
i + |i j|2
i,j=0

(10)

||xi xki ||2


1
exp(
)
ki (x) = p
2 2
(2)d . d

Where xi is feature vector, xk i is training dataset, k indicates the class, and is smoothing parameters.The summation
layer calculated a summation and mean of the pattern layer
output for each class, using

Correlation measures the linear dependency among two


pixels. Correlation is calculated by,

978-1-4799-3580-2/14/$31.00 2014 IEEE

(12)

252

n
1 X
||xi xki ||2
exp(
)
2 2
(2)d n d i=1

gk (x) = p

(13)

2014 2nd International Conference on Information and Communication Technology (ICoICT)

Fig. 8: The Co-Occurrence Matrices Of Sub-Band Images

Where n is a number of training dataset. It calculates the


maximum likelihood of the feature vector being classified into
class k, indirectly.The Competitive layer compares the output
of summation layer for each class, using
C(x) = argmax{gk (x)}, k = 1, 2, ..., m

(14)

Where m is a number of class. Subsequently picks the


maximum probabilities, and produces a 1 for that class and
a 0 for the other classes.
F. Co-Occurrence Matrices of Sub Band Images
The block diagram of our proposed method is shown in
Fig. 8. The proposed method consists of original image decomposition, GLCM construction, and feature extraction.After
original image is pre-processed, that image is decomposed
using 2-D DWT into four different channels of sub-band
images, namely LL, LH, HL, and HH. For comparison, we
decompose original images using four wavelet filters namely
Haar, Daubechies (db4), Coiflet (coif2) and Biorthogonal
(bior2.2). This scheme is done, in order to know which wavelet
filters that appropriate and robust for batik image classification.
Besides that, we decompose original image up to fifth level,
in order to know which level that optimal when represents
information of batik.
Thereafter, GLCM is calculated for each sub-band. In
this study, the co-occurrence matrices are constructed at a
distance of d equals to 1, and at four directions of , namely
0 , 45 , 90 , 135 .Then statistic textural features in each cooccurrence matrix are calculated. The statistic textural features
that we calculated are Angular Second Moment, Entropy, Contrast, Correlation and Inverse Difference Moment.So that, we

978-1-4799-3580-2/14/$31.00 2014 IEEE

have twenty schemes (five level of decomposition is multiplied


by four type of wavelet filter) for one image. Furthermore, each
image has eighty features from multiply four directions of the
co-occurrence matrices by five statistic textural features and
four sub-band images.
III. R ESULT AND D ISCUSSION
Twenty schemes are used to test the performance of our
proposed method. The results of our proposed method are
listed in Table II and Table III which are correlated to the
main dataset and the second dataset respectively. From these
table shows that maximum accuracy of the main dataset
that can achieve is 72%. Whereas maximum accuracy of the
second dataset that can achieve is 100%. Moreover, in most
of schemes the accuracy of the second dataset is higher than
the main dataset. This show that our proposed method is
very robust for fine dataset and robust enough for dataset that
randomly gets from internet. Our proposed method is robust
because the strength of DWT and GLCM is combined to
overcome the weakness of DWT and GLCM. In HH sub-band
from DWT decomposition contains diagonal information in
45 and 135 simultaneously, but it is difficult to distinguish,
whether that the diagonal information is oriented in 45 and
135 . Whereas GLCM can distinguish diagonal information
in 45 and 135 by construct the co-occurrence matrices at
45 and 135 separately. In the other side, GLCM give useful
information, but it is generally too large. Thus decompose
the image into low frequency and high frequency subband
images using DWT can reduce the computational cost without
removing the significant texture features.

253

2014 2nd International Conference on Information and Communication Technology (ICoICT)

In addition, Table IV show that Daubechies is the most


appropriate and robust filter for batik image classification.
It is indicated the choice of a wavelet filter bank is an
important issue, because can affect the quality of classification.
Daubechies filter is robust enough because it preserves edge
information using overlapping windows. Moreover its mother
wavelets are orthogonal and asymmetrical, and their energies
concentrated at the starting point of their support.Table V show
that second level has the highest accuracy. This indicates that
second level is the optimal level to represents information of
batik. This can be analyzed because second level contains
the most significant texture of batik motif. In first level, subband images contain common texture features, which another
classes of motif can have those feature too. It can make
them difficult to classify, because there are pattern similarity
between each class. On the contrary, in second level, several
features are reduced, thus remaining the significant features.
Whereas in third level until fifth level, the features are reduced
overly, thus there are several significant features removed. It
can make them difficult to classify too, because there is not
enough information to classify motifs.
TABLE II: Accuracy results for The Main Dataset
Decomposition

Haar

1st Level
2nd Level
3rd Level
4th Level
5th Level

44
67
39
34
39

Accuracy (%)
Daubechies
(db4)
44
72
33
28
22

Coiflet
(coif2)
44
50
28
28
17

Biorthogonal
(bior2.2)
17
28
50
39
28

Haar

1st Level
2nd Level
3rd Level
4th Level
5th Level

100
86
71
71
29

Accuracy (%)
Daubechies
(db4)
100
79
86
64
64

Coiflet
(coif2)
100
93
71
50
29

Biorthogonal
(bior2.2)
29
100
86
86
64

TABLE IV: The rate accuracy of wavelet filter


Decomposition

Haar

mean

58

Accuracy (%)
Daubechies
(db4)
59

Coiflet
(coif2)
51

Biorthogonal
(bior2.2)
53

IV. C ONCLUSION
This paper proposes a new method to extract texture features
of batik images using co-occurrence matrices of sub-band
images. This method is proposed to overcome the problem
in classifying batik images that acquired randomly from the
internet. Next, the acquired batik images are called the main

978-1-4799-3580-2/14/$31.00 2014 IEEE

Accuracy (%)
Decomposition mean
1st Level
60
2nd Level
72
3rd Level
58
4th Level
50
5th Level
37

dataset. Then, as comparison, there is another dataset that


acquired from capturing fabric in fine condition. Next, this
dataset is called the second dataset. Our proposed method,
which is combining the advantages of GLCM and DWT, is
robust enough to classify batik images in the second dataset.
It shows by the 100% of maximum accuracy. Furthermore, the
classification of batik images in the main dataset performs a
promising result by 72% of maximum accuracy. In addition,
it seems that the most appropriate wavelet filter bank for
batik image classification is Daubechies. Moreover, the most
optimal level of decomposition to represents information of
batik is the second level. For overall results we conclude that
the combination of wavelet texture analysis and GLCM is able
to extract the inherent characteristic in each class of batik motif
effectively.
R EFERENCES

TABLE III: Accuracy results for The Second Dataset


Decomposition

TABLE V: The rate accuracy of wavelet decomposition level

[1] I. Nurhaida, R. Manurung, and A. M. Arymurthy, Performance comparison analysis features extraction methods for batik recognition, in
Advanced Computer Science and Information Systems (ICACSIS), 2012
International Conference on. IEEE, 2012, pp. 207212.
[2] A. Haake, The role of symmetry in javanese batik patterns, Computers
& Mathematics with Applications, vol. 17, no. 4, pp. 815826, 1989.
[3] H. S. Doellah, Batik : Pengaruh Zaman dan Lingkungan. Batik Danar
Hadi Solo, 2002.
[4] F. Kerlogue, The Book of Batik. Archipelago Press, 2004.
[5] M. Cheong and K.-S. Loke, An approach to texture-based image
recognition by deconstructing multispectral co-occurrence matrices using tchebichef orthogonal polynomials, in Pattern Recognition, 2008.
ICPR 2008. 19th International Conference on. IEEE, 2008, pp. 14.
[6] K.-S. Loke and M. Cheong, Efficient textile recognition via decomposition of co-occurrence matrices, in Signal and Image Processing
Applications (ICSIPA), 2009 IEEE International Conference on. IEEE,
2009, pp. 257261.
[7] C. W. D. de Almeida, R. M. de Souza, and A. L. B. Candeias, Texture
classification based on co-occurrence matrix and self-organizing map,
in Systems Man and Cybernetics (SMC), 2010 IEEE International
Conference on. IEEE, 2010, pp. 24872491.
[8] R. W. Conners and C. A. Harlow, A theoretical comparison of texture
algorithms, Pattern Analysis and Machine Intelligence, IEEE Transactions on, no. 3, pp. 204222, 1980.
[9] S. Sidhu and K. Raahemifar, Texture classification using wavelet
transform and support vector machines, in Electrical and Computer
Engineering, 2005. Canadian Conference on. IEEE, 2005, pp. 941
944.
[10] G. Fan and X.-G. Xia, Wavelet-based texture analysis and synthesis
using hidden markov models, Circuits and Systems I: Fundamental
Theory and Applications, IEEE Transactions on, vol. 50, no. 1, pp.
106120, 2003.
[11] S. G. Mallat, A theory for multiresolution signal decomposition: the
wavelet representation, Pattern Analysis and Machine Intelligence,
IEEE Transactions on, vol. 11, no. 7, pp. 674693, 1989.
[12] N. Otsu, A threshold selection method from gray-level histograms,
Automatica, vol. 11, no. 285-296, pp. 2327, 1975.
[13] D. F. Specht, Probabilistic neural networks, Neural networks, vol. 3,
no. 1, pp. 109118, 1990.

254

You might also like