Professional Documents
Culture Documents
Hitesh Jangir3
3Mody University of Science and Technology, CET,
Lakshmangarh, Rajasthan, India
hiteshjangir.cet@modyuniversity.ac.in.
Abstract- This paper proposes a system that classifies textual first, the proposed model checked whether the image is a
images (images that encounter text within) using low-level image document or non document image taking into consideration
features. Image classification and content based image retrieval mean and skewness features only [4]. Then, if the image is
is a growing field in the area of image classification. In this non-document image, then it will be either caption or scene
paper, the approach is based on various low-level image features image. So, secondly, to classify non document images,
including GLCM features like mean, skewness, energy, contrast, energy, contrast and homogeneity features are taken into
homogeneity. Using these various features, the differences account. Then the proposed system performs classification
between images are measured, and then these are used to classify and clustering (data visualization) using J48 decision tree
the textual images by performing classification and clustering
classifier. We have used about 60 images in weka for
techniques on datasets. The proposed method experimented on
60 different textual images to obtain an improved result that was
training and testing using J48 decision tree classifier [4].
not obtained in earlier systems along with classification of
images in three main categories: document, scene and caption.
I. INTRODUCTION
Depending on the feature values, system can classify and Condition i,3 is written as (h> … or h<...)
cluster. At first, the system classifies document images for And (i = 1: M; M is the number of conditions where a pixel
whom mean is highest and then scene and caption are is interpreted as non doc pixel). [9]
classified on the basis of energy and contrast as there is high
contrast in scene images[4].
IV. SIMULATION & EVALUATION
Weka is a multipurpose data mining tool and the The ARFF file prepared from MATLAB simulation is
proposed system used it for classification and clustering loaded to WEKA and various image features are displayed in
purpose. The system has taken input in the form ARFF file the form of histograms. WEKA calculates a statistical range
and this input file is prepared for image processing. and display as min and max values for all the image features
mean, skewness, energy, contrast, Homogeneity [13] [14].
In weka, with input a sample of m classified records
(x, c(x)), a learning algorithm must provide output as a Based on the feature statistics range histograms for all
decision tree. Most algorithms follow top down approach the three types of images are constructed as shown in fig. 4,
i.e., they consider root of the tree as result (usually a test) and where blue represent document images, red represents
then, recursively, choose label of child [11], [12], [13]. caption images and green represents scene images.
VI.FUTURE WORK
In the future, the proposed system can be enhanced by
commentating and classifying text from videos specially
educating tutorials, news reports etc. More efforts can be
made to obtain accurate classification in present system as
well as in the system which have a combination of two or
more categories within a single image. Selection of low level
features can be modified and improved by using Gray Level
Run Length Matrix (GLRM).
VII. CONCLUSION
Fig 10: Classifier output showing confusion matrix As more and more textual image database is available on
the web, mining of these databases has become increasingly
important. Based on high and low level features, low, RGB
V. IMPLEMENTATION RESULTS content etc images were being classified earlier. But the main
issue is to get a classifier that gives efficient and improved
result as well as that uses minimum number of features for 13) B. Raikind, M. Lee, S. Chang, H. Yu, “Exploring text and image
classification. features to classify images in Bioscience literature,” In: Proceedings
of the BioNLP Workshop on Linking Natural Language Processing
Thus, the proposed system is very efficient in and Biology, pp. 73-80 Association for Computational Linguistics
(2006).
classifying textual images into three categories, namely:
document, scene, caption. The technique uses a minimum 14) L. Breiman, Leo, J. Friedman, C. Stone, and R. Olshen,
“Classification and regression trees” In: CRC press (1984).
number of low level features (mean, skewness, energy,
15) R. John, Smith and S. Chang, “ Transform features for texture
contrast, homogeneity) which are easy to calculate and classification and discrimination in large image databases,” In: IEEE
classify. The selected five features are enough to classify International Conference on Image Processing, pp. 407-411,IEEE
textual images in 3 defined categories. (1994).
16) Adegorite, O. Basir, M. Kamel and K. Shaban, “An approach to
Other than classification that is followed in existing mining picture objects based on textual cues,” In: Machine Learning
techniques or systems, the proposed system follows and Data Mining in Pattern Recognition, pp. 466-475 ,Springer Berlin
clustering technique (via visualization) that results in Heidelberg(2005).
improved classifying rate. The result than compared with 17) L. Tian, D. Zheng and C. Zhu, “Research on image classification
classification technique that was followed in earlier research. based on a combination of text and visual features,” In: Fuzzy
Systems and Knowledge Discovery (FSKD), Eighth International
Conference, pp. 1869-1873, IEEE (2011).
REFERENCE
1) S. Chitrakala,P. Shamini & D. Manjula, “ Multi-class Enhanced
Image Mining of Heterogeneous Textual Images Using Multiple
Image Features,” In: . IEEE International Computing
Conference,IACC, pp. 496-501, IEEE (2009).
2) S. Rasoul, Safavian & D. Landgrebe, “A Survey of decision tree
classifier methodology,” In: IEEE Transactions on Systems, Man, and
Cybernetics (1991).
3) T. Duong, J. Lim, H. Vu, & J. Chevallet, “Unsupervised Learning for
Image Classification based on Distribution of Hierarchical Feature
Tree,” In: IEEE International Conference on Research, Innovation
and Vision for the Future RIVF, IEEE International Conference on,
pp. 306-310. IEEE, (2008).
4) S. Nandgonkar, R. Jagtap, P. Anarase, B. Khadake, and A. Betale,
“Image mining of textual images using low-level image features,” In:
IEEE International Conference on Computer Science and Information
Technology (ICCSIT), 3rd IEEE International Conference on, pp.
588-592. IEEE,( 2010).
5) Y. Liu, Z. Dengsheng, L.Guojun and M. Wei-Ying, "A survey of
content-based image retrieval with high-level semantics," In: Pattern
Recognition, pp. 262-282 (2007).
6) X. Huang, T. Zhao, C. Yu, M. Xiangming, and T. Pierre, "Towards
the improvement of textual anatomy image classification using image
local features," In: Proceedings of the 2011 international ACM
workshop on Medical multimedia analysis and retrieval, pp. 25-30,
ACM (2011).
7) O. Augereau, N. Journet, A. Vialard and J. Domenger, “Improving
the classification of an industrial document image database by
combining visual and textual features,” In: Document Analysis
Systems (DAS),11th IAPR International Workshop on, pp. 314-318,
IEEE (2014).
8) J. Wang, H. Yonghao, K. Cuicui, X. Shiming, and P. Chunhong,
"Image-Text Cross-Modal Retrieval via Modality-Specific Feature
Learning." In: Proceedings of the 5th ACM on International
Conference on Multimedia Retrieval, pp. 347-354, ACM (2015).
9) Datta, Ritendra, J. Li, and J. Wang, “Content-based image retrieval-
approaches and trends of the new age,” In: Proceedings of the 7th
ACM SIGMM international workshop on Multimedia information
retrieval, pp. 253-262, ACM (2005).
10) Ojala, Timo, M. Pietikäinen and D. Harwood, “A comparative study
of texture measures with classification based on featured
distributions,” In: Pattern recognition, pp.51-59 (1996).
11) Moudani, Walid, and A. Sayed, “Efficient Image Classification using
Data Mining,” In: International Journal of Combinatorial
Optimization Problems and Informatics, p.27 (2011)
12) K. Matsuo, K. Ueda & U. Michio, “Extraction of character string
from scene image by binarizing local target area,” In: Transaction of
The Institute of Electrical Engineers of Japan, pp.232-241 (2002).