You are on page 1of 14

Soft Computing (2021) 25:4451–4464

https://doi.org/10.1007/s00500-020-05455-w (0123456789().,-volV)(0123456789().,-volV)

METHODOLOGIES AND APPLICATION

Offline handwritten Gurumukhi word recognition using eXtreme


Gradient Boosting methodology
Harmandeep Kaur1 • Munish Kumar1

Published online: 22 November 2020


 Springer-Verlag GmbH Germany, part of Springer Nature 2020

Abstract
Handwritten word recognition is undoubtedly a challenging task due to various writing styles of individuals. So, lots of
efforts are put to recognize handwritten words using efficient classifiers based on extracted features that rely on the visual
appearance of the handwritten text. Due to numerous real-time applications, handwritten word recognition is an important
research area which is seeking a lot of attention from researchers for the last 10 years. In this article, the authors have
proposed a holistic approach and eXtreme Gradient Boosting (XGBoost) technique to recognize offline handwritten
Gurumukhi words. In this direction, four state-of-the-art features like zoning, diagonal, intersection & open-end points and
peak extent features have been considered to extract discriminant features from the handwritten word digital images. The
proposed approach is evaluated on a public benchmark dataset of Gurumukhi script that comprises 40,000 samples of
handwritten words. Based on extracted features, the words are classified into one of the 100 classes based on XGBoost
technique. Effectiveness of the system is assessed based on several evaluation parameters like CPU elapsed time, accuracy,
precision, recall, F1-score and area under curve (AUC). XGBoost technique attained the best results of accuracy (91.66%),
recall (91.66%), precision (91.39%), F1-score (91.14%) and AUC (95.66%) using zoning features based on 90% data as the
training set and remaining 10% data as the testing set. The comparison of the proposed approach with the existing
approaches has also been done which reveals the significance of the XGBoost technique comparatively.

Keywords Handwritten word recognition  Feature extraction  Holistic approach  Gradient boosting  XGBoost

1 Introduction the process to recognize handwritten words (which may be


written using any natural language) by the machine. Words
Document image analysis and recognition is one of the can be written using two modes, namely online mode and
significant progressions toward making society paperless. offline mode. In online mode, words are written using a pen
Handwritten word recognition is an emerging field in the on a digital tablet where the pen tip directions are noted to
domain of document analysis and recognition, which has recognize the written word, whereas in an offline mode, the
been a subject of deep research over the past 10 years handwritten word samples are written on a sheet of paper
(Gader et al. 1996; Senior and Robinson 1998; Steinherz using pen/pencil and then the paper sheet is fed to the
et al. 1999; El-Yacoubi et al. 1999; Plamondon and Srihari scanner to get a digitized image of the document. For
2000; Arica and Yarman-Vural 2001; Koerich et al. 2003; recognizing handwritten words, two approaches, namely
Vinciarelli et al. 2004). Handwritten word recognition is segmentation-based approach and segmentation free
approach, are considered. The segmentation-based
approach is also known as analytical approach, which
Communicated by V. Loia. considers the word as a collection of individual characters.
Thus, to recognize, initially the word is divided into its
& Munish Kumar
munishcse@gmail.com individual units (characters) and then the individual char-
acters are recognized using recognition algorithms.
Harmandeep Kaur
harmandeepk08@gmail.com Sometimes, there is existence of overlapping characters in
a word which leads to an issue in its segmentation. This
1
Department of Computational Sciences, Maharaja Ranjit issue can be solved by the segmentation free approach to
Singh Punjab Technical University, Bathinda, Punjab, India

123
4452 H. Kaur, M. Kumar

word recognition, which considers the whole word as an postal letters can be written using the Gurumukhi script and
individual entity. As this approach does not consider the till now, no postal automation system exists in Gurumukhi
individual characters of a word separately, thus there is no script. Hence, this article is a motivating factor in this
issue of explicit segmentation and the whole word is rec- direction that recognizes offline handwritten Gurumukhi
ognized using recognition algorithms. This approach is also words which finds its application in postal automation. The
known as a holistic approach to word recognition. There high performance of the XGBoost technique as compared
are several applications of the handwritten word recogni- to other algorithms motivates the authors to employ it for
tion system in various areas like automation of handwritten the recognition purpose. So, the aim of this paper is to
documents, processing of bank checks, postal automation, explore the efficiency of XGBoost technique (Chen and
signature verification, writer identification, storage of his- Guestrin 2016) for the proposed work.
torical documents, document authentication, etc. The entire article has been organized as follows. Sec-
In order to have a very high predictive capability, gra- tion 2 represents the literature review related to character
dient boosting technique was proposed. But its acceptance and word recognition. The proposed approach to word
is very limited due to take large time in order to train even recognition is discussed in Sect. 3 which describes all the
the straightforward models. This is due to the fact that this phases of word recognition system, like, digitization, pre-
algorithm needs a single decision tree at a time for mini- processing, feature extraction and classification. Section 4
mizing the errors of all antecedent trees in the model. In describes the various evaluation parameters to test the
order to overcome this issue, a new technique known as efficacy of the proposed system, whereas Sect. 5 mentions
XGBoost is proposed. In XGBoost technique, independent the dataset for the proposed work. Section 6 discusses the
trees are generated and data is arranged to lessen the experimental work to evaluate the system performance.
lookup time, which leads to less training time of models Section 7 demonstrates the comparison between the pro-
and thus enhances the accuracy of classification. The key posed approach and the existing approaches in the litera-
factor which supports the accomplishment of XGBoost is ture. The advantages and disadvantages of the proposed
its extensibility in all scenarios. Due to this factor, the approach are discussed in Sect. 8. Finally, inferences are
system executes over ten times faster as compared to the drawn in Sect. 9 with some future directions.
existing famous solutions on a single machine and extends
to billion examples in distributed settings. There are
numerous systematic and algorithmic optimizations avail- 2 Literature review
able that account for the extensibility of XGBoost (Chen
and Guestrin 2016). Based on these features, XGBoost In the literature, a lot of work has been done in recognition
technique has been applied to recognize offline handwritten of various Indic and Non-Indic scripts, which has been
Gurumukhi words. Gurumukhi script is utilized to write presented in this section. For example, Kumar et al. (2014)
Punjabi language which is the official language of Punjab developed a new hierarchical approach to recognize offline
state of India. Gurumukhi script comprises 41 consonants, handwritten Gurumukhi characters. They extracted a set of
9 vowels, 2 symbols for nasal sounds, 1 symbol for con- 105 features employing four features, namely horizontally
sonant lengthening and 3 subscript characters. It follows peak extent features, vertically peak extent features, diag-
left to right direction of writing in a horizontal way. onal features and centroid features. To reduce the dimen-
The main contributions of the proposed work are as sionality of the features, various feature selection
follows. Till now, no recognized work is available in off- approaches such as Correlation-based feature selec-
line handwritten Gurumukhi word recognition system. So, tion (CFS), Principal Component Analysis (PCA) and
the authors employed the novel approach to offline hand- Consistency-Based (CON) feature selection approaches
written Gurumukhi word recognition based on XGBoost were employed. Based on dataset of 3500 character sam-
technique by considering the four state-of-the-art features, ples, they attained a recognition rate of 91.80% using PCA
namely zoning, diagonal, intersection & open-end points and linear-kernel Support Vector Machine (SVM) classi-
and peak extent features. To the best of authors’ knowl- fier. They (Kumar et al. 2014) also proposed two feature
edge, XGBoost technique has been applied for the very first extraction techniques, namely power curve fitting and
time for word recognition in Gurumukhi script. From the parabola curve fitting-based features to recognize offline
past few years, holistic approach to word recognition has handwritten Gurumukhi characters. To determine the effi-
been gaining a lot of attention due to better results as ciency of the proposed feature extraction techniques, they
compared to segmentation-based approach. So the holistic also considered existing features, namely zoning, transi-
approach to word recognition has been utilized in this tion, diagonal, intersection and open-end points, direc-
work. The proposed work has been evaluated on a public tional, gradient and chain code features. They also used the
benchmark dataset in Gurumukhi script. In Punjab state, same dataset of 3500 samples of Gurumukhi isolated

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4453

characters to perform experiments and attained recognition transformations (DCT2), discrete wavelet transformations
rates of 97.14% and 98.10% using SVM and k-NN, (DWT2), fast Fourier transformations and fan beam
respectively, by considering the power curve fitting-based transformations. Considering experiments on a dataset of
features. Patel et al. (2015) proposed holistic approach for 10,500 character samples, they attained a recognition rate
the recognition of offline handwritten English words based of 95.8% using fivefold cross-validation technique and by
on structural features which finds its application in recog- employing DCT2 features to linear-kernel SVM classifier.
nition of postal addresses. For classification, Euclidean To resolve the issue of scripts having less training dataset,
distance-based k-NN classifier was employed that achieved Bhunia et al. (2018) proposed cross-language framework to
90% accuracy on a dataset comprising 300 samples of 30 recognize words and employed zone-wise approach to map
district names of Karnataka state. Das et al. (2016) pro- characters in order to locate Indian scripts that execute
posed Harmony Search (HS) algorithm-based feature training on one script having large dataset and testing on
selection approach to recognize handwritten Bangla words other scripts having less samples comparatively. They
by reducing the dimensionality of features of the method considered lexicons of sizes 1921, 1934 and 1953 for
presented in the paper by Bhowmik et al. (2014). In the Bangla, Devanagari and Gurumukhi scripts, respectively.
mentioned paper (Bhowmik et al. 2014), to recognize Based on experiments, the global mean average precision
handwritten Bangla words, a set of 65 elliptical features rate of 66.87% (66.42) was reported for Devanagari
was considered with a recognition rate of 81.37% based on (Bangla) script where the Gurumukhi script was used as the
MLP (Multilayer Perceptron) classifier. To eliminate source script and the results revealed higher script simi-
undesired features and to select only relevant features, HS- larity between Bangla and Devanagari as compared to
based feature selection approach only considered 48 fea- Bangla and Gurumukhi script. Gupta et al. (2018) pre-
tures which reported a recognition rate of 90.29% on the sented an ensemble model to combine the output of SVM
same dataset comprising 1020 words handwritten in Bangla classifiers based on three features (two handcrafted and one
script. Based on comparison with two feature dimension- machine-generated features), namely Arnold transform-
ality reduction techniques, namely Genetic Algorithm based features, curvature-based features and Deep Convo-
(GA) and Particle Swarm Optimization (PSO), the pro- lution Neural Network (DCNN)-based features, in order to
posed Harmony Search (HS) algorithm provided better recognize offline handwritten words. To combine the
recognition results using a holistic approach. Kumar et al. decisions of three classifiers, three strategies were
(2016) proposed a new technique to extract features from employed, namely vote for majority decision, vote for
the pre-segmented offline handwritten Gurumukhi charac- strongest decision and vote for the sum of the decisions.
ters. The proposed feature extraction technique considered The proposed system achieved the best recognition rates of
the boundary extent of the character sample to extract the 95.23%, 97.16% and 95.07% on public datasets, namely
desirable features which were then reduced using PCA CENPARMI, ISIHWD and IAM200, respectively.
feature selection approach. The experiments were con- Tavoli et al. (2018) proposed a novel feature extraction
ducted on a dataset comprising 7000 Gurumukhi character technique named as SGCSL (Statistical Geometric Com-
samples using three classifiers, namely k-NN, SVM and ponents of Straight lines) to recognize offline handwritten
MLP. The best recognition rate of 93.8% was attained Arabic/Persian words. The extracted features were then fed
using RBF-kernel SVM classifier and fivefold cross-vali- to the SVM classifier for classification task. They attained
dation approach. recognition rates of 67.47%, 80.78% and 86.22% by
Assayony and Mahmoud (2017) proposed a holistic experimenting the proposed approach on three public
approach to recognize offline handwritten Arabic words datasets, namely Iran-cities, IFN/ENIT and IBN SINA
based on Gabor filters. Two types of features like statistical Arabic dataset, respectively. Arani et al. (2019) demon-
Gabor features and Gabor descriptors were considered strated a holistic approach to recognize handwritten Farsi
from the word samples which were then integrated with words based on the fusion of three HMM classifiers that
Bag-of-features framework for extracting the desired fea- were trained separately using three feature sets, namely
tures. Then, based on extracted features, the words were image gradient, black-white transitions and contour chain
recognized using SVM classifier with linear-kernel func- code features. Then, the fused output of HMMs was given
tion. The proposed recognition system achieved the best to the MLP classifier for the recognition purpose. They
average recognition rate of 86.44% by considering exper- evaluated the approach on ‘‘Iranshahr 3’’ dataset and
iments on CENPARMI public dataset comprising Arabic attained a recognition rate of 89.06%, which is observed as
handwritten checks. Kumar et al. (2017) presented various superior than independent base classifiers. In order to make
transformations-based techniques to recognize offline comparison possible between the proposed approach and
handwritten Gurumukhi characters. The various transfor- the existing approaches in Gurumukhi script, Kumar et al.
mation techniques used were discrete cosine (2019) developed seven benchmark datasets in Gurumukhi

123
4454 H. Kaur, M. Kumar

script, namely HWR-Gurmukhi_1.1, HWR-Gurmukhi_1.2, recognition rate of 95.30%, which is 1.28% superior in
HWR-Gurmukhi_1.3, HWR-Gurmukhi_2.1, HWR-Gur- comparison to the recognition rate attained through the
mukhi_2.2, HWR-Gurmukhi_2.3 and HWR-Gur- original feature set. Weldegebriel et al. (2020) proposed
mukhi_3.1. Each of HWR-Gurmukhi_1.1, HWR- hybrid Convolutional Neural Network (CNN) and
Gurmukhi_1.2, HWR-Gurmukhi_1.3 benchmark datasets XGBoost classifier to recognize handwritten Ethiopian
comprise 3500 character samples, whereas HWR-Gur- characters, where CNN was employed to extract features
mukhi_2.1, HWR-Gurmukhi_2.2, HWR-Gurmukhi_2.3 from images and to XGBoost for recognition and classifi-
comprise 5600 character samples, individually and the last cation purpose. To conduct experiments, Handwritten
benchmark dataset HWR-Gurmukhi_3.1 comprises 7000 Ethiopian Character Recognition (HECR) dataset was
character samples. They also performed experiments on the proposed which comprises a mixture of scripts, numerical
proposed datasets using k-NN, RBF-SVM, MLP, neural representations, tonal symbols, special characters, punctu-
network, decision tree and random forest classifiers by ation and combining symbols. CNN generated error rate of
considering existing features, namely zoning features, 0.4630 whereas CNN-XGBoost generated error rate of
diagonal features, intersection & open-end points features, 0.1612, thus reveals the superiority of the hybrid model as
directional features, transition features and centroid fea- compared to CNN model. An accuracy of 99.84% was
tures and reported results based on Precision rate, False attained, which is considered as superior based on com-
Acceptance Rate (FAR) and False Rejection Rate (FRR). parison with some existing approaches (Niu and Suen
These benchmark datasets are publicly available at https:// 2011; Maitra et al. 2015; Zhong et al. 2015; Ramraj et al.
sites.google.com/view/gurmukhi-benchmark/. Wu et al. 2016; Katiyar et al. 2017; Younis 2017; James et al. 2019).
(2019) proposed an approach to recognize unconstrained
offline handwritten words based on integration of position
embeds with residual networks (ResNets). Based on this 3 Proposed method
integration, generated outputs were given as an input to
bidirectional long short-term memory (BiLSTM) networks The proposed method recognizes offline handwritten
for recognition of characters. They reported results of Gurumukhi words by following a step-by-step procedure
91.97% accuracy and 1.79%-character error rate on two like going through various phases, namely digitization,
standard datasets, namely 2017 ICDAR IEHHR competi- preprocessing, feature extraction and classification. For the
tion and RIMES datasets, respectively. proposed approach, at first the offline handwritten docu-
To select only relevant and discriminant characteristics ment comprising words is scanned in order to get a digi-
from the word images, feature selection has a significant tized image of the document. After that based on
role. In this direction, Ghosh et al. (2019) proposed preprocessing operations, the handwritten words get
Memetic Algorithm (MA)-based wrapper filter selection extracted from the document. The various features like
approach to reduce the dimensionality of gradient-based zoning features, diagonal features, intersection & open-end
features and modified statistical and contour-based features points features and peak extent features are extracted from
(SCF) in order to recognize offline handwritten Bangla the word image in a holistic way in order to get the desired
words using holistic approach. Using MLP classifier, the feature database. Then, based on features, the handwritten
proposed approach was experimented on dataset compris- words are classified and recognized using the XGBoost
ing 7500 words handwritten in Bangla script and attained technique as depicted in Fig. 1. In order to test the given
93% recognition accuracy after applying feature selection word on the proposed system, at first the system is trained
approach on hybrid of the considered features and gained using the training dataset.
3.33% enhancement as compared to the rate attained
through the original feature set. They also compared MA- 3.1 Digitization
based approach with Genetic Algorithm (GA)-based fea-
ture selection approach and revealed that MA achieved The process of converting offline handwritten documents
better recognition accuracy comparatively, even though into a digital form is called digitization. The digitization
GA selected less number of features. Malakar et al. (2020) phase is the first step to process the documents by computer
proposed GA-based hierarchical feature selection method system whereby the digitized image is stored in the form of
to minimize the shape (elliptical features) and texture bits. The digitized image gets formed using scanner which
(gradient-based features) based on features extracted from scans the document to generate the bitmap image of the
the word samples, which has been reduced by approxi- document. In the proposed system, the authors have scan-
mately 28%. They evaluated the proposed method on a ned the documents at 300 dpi through the scanner.
dataset of 12,000 handwritten Bangla word samples using
the MLP classification technique and attained the

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4455

Fig. 1 Schematic diagram of


proposed offline handwritten Handwritten Test image
word recognition system document

Digitization

Training Phase Testing Phase

Pre-processing

Feature
Feature extraction
Database

Zoning, Diagonal
Intersection & open
end points and Peak
extent features

Classification based
on XGBoost

Recognized
word

3.2 Preprocessing normalization operation is applied to provide a uniform


size of 256 9 64. This size has been selected due to hor-
The preprocessing phase takes place to reduce the varia- izontal writing style of Gurumukhi script.
tions in the writing styles. In the proposed system, the
authors have considered three basic preprocessing opera- 3.2.3 Thinning
tions, namely binarization, normalization and thinning
operations, which are discussed in the following To minimize the text width from multiple pixels to a single
subsections. pixel, the thinning operation takes place. In the proposed
system, the authors have considered parallel thinning
3.2.1 Binarization algorithm proposed by Zhang and Suen (1984).

As the name implies, the binary form is generated in 3.3 Feature extraction
binarization operation. To create the binary image, the
threshold constant is positioned between higher and lower The process of extracting the desirable features of the word
values which equate to white and black pixels, image is called feature extraction. The feature extraction
respectively. phase is significant for recognizing word images with high
recognition rate because based on the extracted features,
3.2.2 Normalization the word image is classified. There are various features
available to recognize word images and various combina-
Due to distinct writers and various writing styles, the size tions of features have been proposed in the literature. In the
of the words is different. So, to provide uniformity to the proposed system, the authors have considered only indi-
words, the normalization operation takes place which vidual features such as zoning features, diagonal features,
normalizes the words to uniform size. In the proposed intersection & open-end points features and peak extent
system, after slicing the words from the binarized image of features, without taking their combinations. These features
the document, the cropping operation is applied to elimi- are discussed in the following subsections.
nate the white space around the word. Then, the

123
4456 H. Kaur, M. Kumar

3.3.1 Zoning features Intersection point has more than one pixel in its neighbors
and the open-end point includes only single pixel in its
To extract zoning features, the word image is partitioned neighborhood. For the proposed work, the authors have
into a few zones and then from each zone, the features are extracted 170 intersection & open-end point-based features
extricated. These features are computed from pattern (85 intersection features & 85 open-end point features).
characteristics of each considered zone. After getting
foreground pixels from n considered zones, these are nor- 3.3.4 Peak extent features
malized to [0–1], thus corresponds to feature a set of
n elements. Kumar et al. (2013) have proposed peak extent features for
For the proposed work, initially, the authors have par- the recognition of offline handwritten Gurumukhi charac-
titioned the word image into 4 zones and then each of the 4 ters. For the proposed work, the same features have been
zones gets partitioned into 4 zones resulting in total 16 applied to recognize offline handwritten Gurumukhi words.
zones. Each of the 16 zones gets partitioned into 4 zones For extracting these features, initially, the various zones
further resulting in total 64 zones and thus the individual have been considered for the word image. Then, these
features in form of foreground pixels are extracted from the features are extricated by considering the sum of the
respective zones. One feature is considered from the whole lengths of the peak extent that places consecutive black
word image. Thus, total 1 ? 4?16 ? 64 = 85 zoning pixels along each zone. Peak extent features are extracted
features are extracted from the word image as depicted in horizontally as well as vertically. In horizontal peak extent
Fig. 2. features, the sum of the lengths of peak extent is taken that
place consecutive black pixels horizontally in each row of a
3.3.2 Diagonal features zone as shown in Fig. 3b, whereas in vertical peak extent
features, the sum of the lengths of peak extent that place
To extract diagonal features, initially, the word image is consecutive black pixels vertically in each column of a
partitioned into zones. Then, as the name implies, the zone is considered as shown in Fig. 3c. For the proposed
features are extracted from the foreground pixels along the work, the authors have considered 170 peak extent features
diagonal of each considered zone. There are multiple (85 horizontal & 85 vertical peak extent features) from the
characteristics along the diagonal of each zone, so the word image.
average of these multiple values is considered to acquire a
single value corresponding to each zone. After partitioning 3.4 Classification
the word image into a few zones as illustrated in Fig. 2, the
authors have extricated 85 diagonal features for the pro- Based on the extracted features, the words are recognized
posed work. using the classification phase. This phase determines the
class to which the word belongs to. There are many clas-
3.3.3 Intersection and open-end points-based features sification algorithms available in the literature to classify
the characters and words of various Indic and Non-Indic
In order to calculate intersection & open-end point-based scripts. In the latest research, eXtreme Gradient Boosting
features, at first, the word image is divided into considered (XGBoost) technique is performing superior to boost the
number of zones and then the intersection & open-end performance of the recognition systems in the field of
points are calculated for each of the considered zones. pattern recognition and image processing. Thus, the
authors have employed XGBoost technique for the classi-
fication purpose.

3.4.1 eXtreme Gradient Boosting (XGBoost)

eXtreme Gradient Boosting or XGBoost is an ensemble


technique that comprises sequential decision trees. So, it
also calls sequential ensemble technique. It has been pro-
posed by Chen and Guestrin (2016). It is based on a gra-
dient boosting algorithm that boosts weak classifiers
(learners) and generates a predicted model in the form of an
ensemble of weak learners. In this algorithm, initially
similar weight is assigned to all training samples. These
Fig. 2 Zoning features of considered Gurumukhi word weights specify the probability of the record getting

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4457

Fig. 3 a Zoning of bitmap word 0 1 1 1 1 0 1 1


image. b Horizontally peak
extent features. c Vertically 1 0 1 1 1 0 1 1
peak extent features
0 0 0 1 1 0 1 0
0 1 1 0 1 1 1 0
0 1 0 1 1 0 1 0
0 0 1 1 0 1 0 0
1 1 0 1 1 1 0 0
1 1 1 1 0 0 1 0
(a)

0 4 4 4 4 0 2 2 4 0 1 2 3 5 0 5 2
1 0 3 3 3 0 2 2 3 1 0 2 3 5 0 5 2
0 0 0 2 2 0 1 0 2 0 0 0 3 5 0 5 0
0 2 2 0 3 3 3 0 3 0 2 1 0 5 1 5 0
0 1 0 2 2 0 1 0 2 0 2 0 4 5 0 5 0
0 0 2 2 0 1 0 0 2 0 0 1 4 0 2 0 0
2 2 0 3 3 3 0 0 3 2 2 0 4 1 2 0 0
4 4 4 4 0 0 1 0 4 2 2 1 4 0 0 1 0
Sum= 23 Sum=24 2 2 2 4 5 2 5 2
(b) (c)

selected by the decision tree for the training purpose. Since Training Weighted Weighted Weighted

the weights are similar, so the probability of selection of all Sample Sample Sample Sample
records is equal. After training, the model is ready to
predict. After prediction, whichever records are incorrectly
classified by model, those weights get updated and thus fed 1st weak 2nd weak 3rd weak nth weak
decision decision decision decision
to the second decision tree, this model is called weak tree tree tree tree
classifier, because this has not classified all the records classifier classifier classifier classifier
correctly. For the second decision tree, whichever records
have maximum weight, those records get selected for the
training purpose. Thus, weight updation is significant in Final classifier
XGBoost. When the weight is updated for the wrong pre-
dicted results, then that is passed to the next decision tree. Fig. 4 Demonstration of eXtreme Gradient Boosting (XGBoost)
Similarly, this process continues sequentially one after
another up to nth decision tree. After combining all the dataset comprising n samples and m features D ¼ fðxi; yi )}
weak classifiers, the new final classifier gets generated (|D| = xi e Rnm , yi e Rn ), the mathematical delineation of
which generates the final class of the record. Then, the final the ensemble technique is given as follows.
classifier classifies the test sample based on the maximum X
k
of similar predictions by the weak classifiers as described y^i ¼ hk ðxi Þ; hk 2 R; ð1Þ
in Fig. 4. k¼1

where k denotes the number of trees, h represents function


3.4.1.1 Mathematical implementation XGBoost is a tree in functional space R, R denotes the set comprising all
ensemble technique which combines multiple classifica- possible classification and regression trees.
tions and regression trees (Song et al. 2016). Consider a The objective function is delineated as follows:

123
4458 H. Kaur, M. Kumar

X
n X
k • True Negative (TN) when the observation is false and is
F1 ðhÞ ¼ ^ þ
hðyi; yÞ Xðhk Þ; ð2Þ predicted to be false.
i k¼1 • False Positive (FP) when the observation is false and is
where hðyi y^Þ represents the training loss function, Xðhk Þ predicted to be true.
denotes the regularization function. h denotes the differ-
entiable convex loss function that quantifies the deviation 4.1 CPU elapsed time
between prediction y^i and the target yi . The purpose of
XGBoost technique is to lessen F1 ðhÞ. Elapsed time is an evaluation parameter to measure the
The regularization complexity can be described as speed of the processor. To quantify the performance of the
follows: processor, it is considered as inversely proportional to the
1 XT execution time. It is a superior measure to examine the
XðhÞ ¼ cL þ k x2 ; ð3Þ processor speed because it is less dependent on other sys-
2 j¼i j
tem components. It measures in milliseconds (ms).
where c denotes the gamma parameter, L denotes the
number of leaves, k denotes L2 regularization term on 4.2 Accuracy
weights in the model, x denotes the vector score on leaves.
As XGBoost technique generates trees on the basis of Accuracy is defined as the proportion of the number of
number of labels, so the time spent on training will be correct predictions and the total number of input speci-
based on number of classes present in the dataset. mens. Accuracy can also be computed in terms of positives
and negatives in case of binary classification as shown
3.4.1.2 Merits of XGBoost technique Due to the following below.
merits of XGBoost technique, it has been employed for the TP þ TN
recognition purpose. Accuracy ¼ :
TP þ TN þ FP þ FN
• It is an efficient and simple to use algorithm which
provides high performance and accuracy than other 4.3 Precision
algorithms.
• It supports parallel processing and is faster in compar- Precision is considered as the proportion of positive iden-
ison to Gradient Boosting Machine (GBM). tification that i correct. It can be computed as the ratio of
• It has inherent L1 and L2 regularization, which resolve correct positive outcomes and the number of predicted
the overfitting problem. Due to an in-built regulariza- positive outcomes.
tion facility, it is also known as a regularized form of
TP
GBM. Precision ¼ :
TP þ FP
• It has an inherent potential to handle missing values.
• It permits users to perform cross-validation at each
successive (iteration) of the boosting process and hence 4.4 Recall
it is easy to obtain the precise optimum number of
boosting iterations in a single run. Recall is considered as the proportion of actual positives
that are identified correctly. It is computed as the propor-
tion of correct positive outcomes and number of all rele-
vant specimens identified as positive.
4 Evaluation parameters
TP
Recall ¼ :
To evaluate the performance of the proposed system, var- TP þ FN
ious evaluation parameters have been considered like CPU
elapsed time, accuracy, precision, recall, F1-score and area 4.5 F1-Score
under curve (AUC). There are following four important
terms that need to be considered. F1-Score integrates precision and recall proportionate to
• True Positive (TP) when the observation is true and is certain positive class. It is computed as the weighted
predicted to be true. average of precision and recall as shown below. It has its
• False Negative (FN) when the observation is true and is best value at 1 and worst value at 0.
predicted to be false.

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4459

Precision  Recall been proposed by Kaur and Kumar (2019). They have also
F1  Score ¼ 2  :
Precision þ Recall presented a survey on word recognition for non-Indic and
Indic scripts (Kaur and Kumar 2018). This dataset com-
prises 40,000 samples of 100 place names written by 40
4.6 Area under curve (AUC) different writers, where each writer has written each word
10 times. Table 1 illustrates the few samples of the dataset
AUC is employed for binary classification problems. It
written by 4 different writers.
indicates the probability that the classifier will rank a
randomly selected positive sample higher than a randomly
selected negative sample. AUC is calculated using ROC
6 Experimental results and discussion
(Receiver Operating Characteristic) curve by mapping
True Positive Rate (TPR) along y-axis against False Posi-
In this segment, the experiments on the proposed system
tive Rate (FPR) along x-axis (Fig. 5), which are explained
are presented by considering the mentioned evaluation
as below. AUC lies in the range of [0,1]. The higher the
parameters. For performing experiments on the considered
value of the AUC, the superior is the performance of the
dataset, the dataset has been divided into training and
model.
testing sets using three partitioning strategies as depicted in
• True Positive Rate (TPR) TPR is considered as the ratio Table 2. In the strategy a, 90% data is considered in the
of positive data points that are correctly observed as training set and remaining 10% data is considered in the
positive, with respect to all positive data points. It also testing set. The strategy b considers 80% data in the
terms as sensitivity. training set and 20% data in the testing set, whereas last
strategy c considers 70% data in the training set and
TP remaining 30% data in the testing set.
TPR ¼ : The authors have performed experiments on the pro-
ðFN þ TPÞ
posed system using four feature extraction techniques,
• False Positive Rate (FPR) FPR is considered as the namely zoning features, diagonal features, intersection &
ratio of negative data points that are wrongly observed open-end point features and peak extent features. Then, the
as positive, with respect to all negative data points. It feature wise results are reported based on six evaluation
also terms as specificity. parameters, namely CPU elapsed time, accuracy, precision,
recall, F1-score, AUC by considering the XGBoost tech-
FP nique as discussed below.
FPR ¼ :
ðFP þ TNÞ
6.1 Performance based on zoning features

Based on zoning features, the best accuracy rate, precision


rate, recall rate, F1-score and AUC of 91.66%, 91.39%,
5 Dataset for the proposed work
91.66%, 91.14% and 95.66% are achieved, respectively, by
considering 90% data in the training set and 10% data in
For the experimental work, the authors have utilized the
the testing set as shown in Table 3. By employing only
benchmark dataset of Gurumukhi script which is publicly
zoning features to XGBoost technique, the best CPU
available at https://sites.google.com/view/gurmukhi-bench
Elapsed time is 43.63 ms based on 80:20 partitioning
mark/home/word-level-gurmukhi-dataset. This dataset has
strategy. Thus, the best rates are reported using strategy
a (90:10).

6.2 Performance based on diagonal features

Using diagonal features as an input to XGBoost technique,


the authors have attained the best accuracy and recall rates
of 91.30% and an AUC of 95.47% based on 90% training
and 10% testing set, whereas the best precision and F1-
score of 91.03% and 90.88% are reported, respectively,
based on 80:20 ratio of training and testing set as depicted
in Table 4. The best CPU elapsed time reported is
Fig. 5 Area under curve (AUC) 42.77 ms using 70:30 partitioning strategy.

123
4460 H. Kaur, M. Kumar

Table 1 Handwritten
Gurumukhi word samples Script
W1 W2 W3 W4
Word

ਮਾਨਸਾ

ਮਕ
ੁ ਤਸਰ

ਮਗ
ੋ ਾ

ਸਗ
ੰ ਰਰ

ਨਾਭਾ

Table 2 Dataset partitioning


Strategy Training:testing ratio Training set (words) Testing set (words)
strategies
a 90:10 36,000 4000
b 80:20 32,000 8000
c 70:30 28,000 12,000

Table 3 System performance


Partitioning strategy Evaluation parameters
based on zoning features
CPU Elapsed Time Accuracy Precision Recall F1 Score AUC

90:10 50.09 91.66 91.39 91.66 91.14 95.66


80:20 43.63 91.34 91.21 91.34 91.04 95.50
70:30 46.33 89.93 90.01 89.93 89.84 94.76

Table 4 System performance


Partitioning strategy Evaluation parameters
based on diagonal features
CPU Elapsed Time Accuracy Precision Recall F1 Score AUC

90:10 51.60 91.30 90.95 91.30 90.73 95.47


80:20 46.86 91.18 91.03 91.18 90.88 95.41
70:30 42.77 90.00 90.06 90.00 89.92 94.80

6.3 Performance based on intersection 6.4 Performance based on peak extent features
and open-end points features
Based on peak extent features, the maximum accuracy and
By employing intersection & open-end point features to the recall rates of 86.27%, a precision rate of 86.13%, F1-score
XGBoost technique, the maximum accuracy and recall of 85.73% and an AUC of 92.85% have been attained using
rates are 88.37%, precision and AUC are 88.40% and 90:20 partitioning strategy, whereas the best CPU elapsed
93.94%, respectively, using 90:10 partitioning strategy, time of 57.23 ms is reported using 70:30 partitioning
whereas the best F1-score of 87.95% has been reported strategy as mentioned in Table 6.
based on 80:20 partitioning strategy as illustrated in
Table 5. The authors have attained the best CPU elapsed
time of 57.20 ms using 70% training and 30% testing set.

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4461

Table 5 System performance


Partitioning strategy Evaluation parameters
based on intersection and open-
end points features CPU Elapsed Time Accuracy Precision Recall F1 Score AUC

90:10 73.51 88.37 88.40 88.37 87.87 93.94


80:20 67.28 88.31 88.24 88.31 87.95 93.91
70:30 57.20 87.22 87.54 87.22 87.21 93.35

Table 6 System performance


Partitioning strategy Evaluation parameters
based on peak extent features
CPU elapsed time Accuracy Precision Recall F1 Score AUC

90:10 79.16 86.27 86.13 86.27 85.73 92.85


80:20 64.82 84.44 84.53 84.44 84.26 91.90
70:30 57.23 83.36 83.27 83.36 83.12 91.34

7 Comparison with state-of-the-art work approach, the latter approach surpassed the proposed
and analysis approach.
• Even without using the feature selection approach, the
In the literature, the authors have studied various machine proposed approach attained significant results in com-
learning approaches like HMM, k-NN, MLP, SVM, DCNN parison to Das et al. (2016) who used HS-based feature
and various feature dimensionality reduction approaches selection approach to minimize the dimensionality of
such as GA, PSO and HS, which were utilized to recognize the set of 65 elliptical features to 48 features and
handwritten words of various scripts. These state-of-the-art enhanced the Bangla handwritten word recognition rate
approaches have been compared with the proposed from 81.37 to 90.29%.
approach in terms of recognition accuracy as depicted in • Kumar et al. (2014) attained recognition rate of 91.80%
Table 7. Due to non-availability of word recognition for Gurumukhi handwritten characters using a PCA
approaches in Gurumukhi script, the proposed approach feature selection approach which is very close to the
has also been compared with character recognition recognition rate attained through the proposed approach
approaches in Gurumukhi script as presented in Table 8. presented in this paper. Thus, the proposed approach
can provide better recognition rate using feature
7.1 Comparative analysis selection approaches which remain a further area of
research.
Based on comparison with the existing approaches to word • Gupta et al. (2018) achieved the best recognition rates
recognition, the authors have drawn the following of 95.23%, 97.16% and 95.07% on three public
inferences. datasets, namely CENPARMI, ISIHWD and IAM200,
respectively. The spike in their recognition results is
• Based on a benchmark dataset of Gurumukhi script due to the amalgamation of the outcome of SVM
(Kaur and Kumar 2019), the authors have achieved the classifiers based on three features, namely Arnold
best word recognition rate of 91.66% using zoning transformation-based features, curvature-based features
features which surpasses the recognition rates achieved and DCNN-based features. The same amalgamation
through some state-of-the-art methods in other scripts approach can be applied to the proposed word recog-
(Assayony and Mahmoud 2017; Tavoli et al. 2018; nition system in the future. But without hybrid
Arani et al. 2019; Kessentini et al. 2010). approach, the proposed system provides significant
• Comparatively, the proposed approach attained recog- results comparatively when the former approach was
nition rates of 91.30%, 88.37% and 86.27% based on tested on Arnold transformation-based features and
diagonal features, intersection & open-end points curvature-based features individually.
features and peak extent features, respectively. • As elucidated in Table 8, SVM classifier is mostly used
• The proposed approach attained a higher recognition classifier in character recognition of Gurumukhi script
rate in comparison to the approach proposed by Ghosh and has provided the best result based on the power
et al. (2019) in the case where the latter approach curve fitting-based features (Kumar et al. 2014).
considered a hybrid of the features without using • It is clear from the comparative analysis that XGBoost
feature selection approach. But with feature selection technique provided a better machine learning model to

123
4462 H. Kaur, M. Kumar

Table 7 Comparison of the proposed work with the existing methodologies


Authors Dataset Feature extraction approach Classification Recognition accuracy
approach

Kessentini IFN/ENIT Density and contour-based features HMM 79.8%


et al. (2010) IRONOFF 89.8%
Patel et al. 300 English word samples of 30 Structural features k-NN 90%
(2015) district names of Karnataka
state
Das et al. 1020 Bangla words HS-based feature selection approach MLP 90.29%
(2016)
Assayony and CENPARMI Gabor filters integrated with Bag-of- SVM 86.44%
Mahmoud features
(2017)
Gupta et al. CENPARMI Arnold transform-based features, SVM 95.23%
(2018) ISIHWD curvature-based features and DCNN- 97.16%
based features
IAM200 95.07%
Tavoli et al. Iran-cities SGCSL SVM 67.47%
(2018) IFN/ENIT 80.78%
IBN SINA 86.22%
Arani et al. Iranshahr 3 Image gradient, black-white transitions HMM and 89.06%
(2019) and contour chain code features MLP
Ghosh et al. 7500 Bangla handwritten words Gradient-based features and modified MLP 89.67% (without feature
(2019) SCF; MA-based wrapper filter selection selection) 93% (with
approach feature selection)
Proposed 40,000 Gurumukhi handwritten Zoning features XGBoost 91.66%
approach words Diagonal features 91.30%
Intersection and open-end points 88.37%
Peak extent features 86.27%

recognize handwritten words of Gurumukhi script. But it eliminates all the limitations raised through touching
this technique proved good in combination with several characters, overlapping characters, cursive writing
features considered and has provided the best result in style, etc. in segmentation-based approach.
combination with zoning features using 90:10 parti- • Maximum recognition accuracy of 91.66% has been
tioning strategy where 90% data has been used to train attained, which is significant based on comparison with
the model and remaining 10% data has been used to test state-of-the-art work as delineated in Tables 7 and 8.
the proposed model. • Till now no postal automation system exists in Guru-
mukhi script. The proposed work is an endeavor in this
direction as this work recognizes handwritten place
names in Gurumukhi script.
8 Advantages and disadvantages
• This approach can be applied to other scripts by training
of proposed approach
the proposed model using a dataset of that script.
In this section, the authors have presented the advantages
and disadvantages of the proposed approach as an indicator 8.2 Disadvantages of proposed approach
for further research in the field of word recognition.
• As the training time of XGBoost depends on the num-
8.1 Advantages of proposed approach ber of classes present in the dataset. The proposed
approach has been evaluated on 100 classes, thus
• The proposed approach extracts the features from the XGBoost training time is comparatively more in this
complete word image without segmenting the word work.
image into its primitive components (characters). Thus,

123
Offline handwritten Gurumukhi word recognition using eXtreme Gradient Boosting methodology 4463

Table 8 Comparison of the proposed work with the existing Gurumukhi character recognition methodologies
Authors Dataset in Gurumukhi script Feature extraction approach Classification Recognition
(character samples) approach accuracy

Kumar et al. 7000 Horizontal peak extent, vertical peak extent, shadow, Linear-SVM 95.62%
(2013) centroid features k-NN 95.48%
MLP 94.74%
Kumar et al. 3500 Horizontal peak extent, vertical peak extent, diagonal SVM 91.80% (PCA)
(2014) and centroid features
Feature selection approaches: CFS, PCA and CON
Kumar et al. 3500 Power curve fitting and parabola curve fitting-based SVM 97.14%
(2014) features k-NN 98.10%
Kumar et al. 7000 Features based on boundary extent of the character k-NN 93.8% (RBF-
(016) sample SVM SVM)
PCA MLP
Kumar et al. 10,500 DCT2, DWT2, fast Fourier transformations and fan SVM 95.8% (DCT2)
(2017) beam transformations
Proposed 40,000 Gurumukhi handwritten Zoning features XGBoost 91.66%
approach words Diagonal features 91.30%
Intersection and open-end points 88.37%
Peak extent features 86.27%

• This approach may fail for overlapped and touched efficiency of XGBoost technique for the considered fea-
characters in a word. tures. In the future, several combinations of the features
• This approach may not work for overlapping words. can be tested as an input to XGBoost technique to test its
• The approach has been tested on limited dataset efficiency for the hybrid features. Moreover, different
comprising only 40,000 word samples. The results can feature selection techniques can be applied to reduce the
be enhanced by using a larger training dataset. dimensionality of the original feature set in order to lessen
the burden of classification and to enhance the system
performance. The proposed approach can also be applied to
other North Indian scripts such as Devanagari which are
9 Inferences and future directions
having a similar structure as that of the Gurumukhi script.
The authors proposed a holistic approach to offline hand-
written Gurumukhi word recognition based on XGBoost
technique. A few state-of-the-art features such as zoning Compliance with ethical standards
features, diagonal features, intersection & open-end point-
based features and peak extent features are given as an Conflict of interest The authors declare that they have no conflict of
interest.
input to XGBoost technique to test the efficacy of the
proposed system. The proposed approach finds its appli-
cation in postal automation due to recognition of hand- References
written place names of Gurumukhi script present in the
public benchmark dataset. Among all the features consid- Arani SAAA, Kabir E, Ebrahimpour R (2019) Handwritten Farsi
ered, the zoning features provide the best accuracy and word recognition using NN-based fusion of HMM classifiers
recall rate of 91.66%, a precision rate of 91.39%, F1-score with different types of features. Int J Image Gr 19(1):1–21
Arica N, Yarman-Vural FT (2001) An overview of character
of 91.14% and an AUC of 95.66% based on 36,000 training recognition focused on off-line handwriting. IEEE Trans Syst
and 4000 testing words handwritten in Gurumukhi script. Man Cybern Part C Appl Rev 31(2):216–233
Except zoning features, the least CPU elapsed time is Assayony MO, Mahmoud SA (2017) Integration of gabor features
reported in the case of 70% training and 30% testing set. with bag-of-features framework for Arabic handwritten word
recognition. In: Proceedings of the 9th IEEE-GCC conference
The comparison of the proposed approach with the existing and exhibition (GCCCE), pp 1–4
approaches in character and word recognition indicates the

123
4464 H. Kaur, M. Kumar

Bhowmik S, Malakar S, Sarkar R and Nasipuri M (2014) Handwritten transformations. Proc Natl Acad Sci India Sect A Phys Sci
Bangla word recognition using elliptical features. In: Proceed- 87:137–143
ings of international conference on computational intelligence Kumar M, Sharma RK, Jindal MK, Jindal SR, Singh H (2019)
and communication networks (CICN), pp 257–261 Benchmark datasets for offline handwritten Gurmukhi script
Bhunia AK, Roy PP, Mohta A, Pal U (2018) Cross-language recognition. In: Sundaram S, Harit G (eds) Document analysis
framework for word recognition and spotting of Indic Scripts. and recognition. DAR 2018. Communications in computer and
Pattern Recogn 79:12–31 information science, vol 1020, pp 143–151
Chen T, Guestrin C (2016) XGBoost: a scalable tree boosting system. Maitra DS, Bhattacharya U, Parui SK (2015) CNN based common
In: Proceedings of the 22nd ACM SIGKDD international approach to handwritten character recognition of multiple
conference on knowledge discovery and data mining, scripts. In: Proceedings of 13th international conference on
pp 785–794 document analysis and recognition (ICDAR), pp 1021–1025
Das S, Singh PK, Bhowmik S, Sarkar R, Nasipuri M (2016) A Malakar S, Ghosh M, Bhowmik S, Sarkar R, Nasipuri M (2020) A
harmony search based wrapper feature selection method for GA based hierarchical feature selection approach for handwritten
holistic Bangla word recognition. Procedia Comput Sci word recognition. Neural Comput Appl 32:2533–2552
89:395–403 Niu XX, Suen CY (2011) A novel hybrid CNN–SVM classifier for
El-Yacoubi A, Gilloux M, Sabourin R, Suen CY (1999) Uncon- recognizing handwritten digits. Pattern Recogn 45(4):1318–1325
strained handwritten word recognition using hidden markov Patel MS, Reddy SL, Naik AJ (2015) An efficient way of handwritten
models. IEEE Trans Pattern Anal Mach Intell 21(8):752–760 english word recognition. In: Proceedings of the 3rd interna-
Gader PD, Mohamed MA, Keller JM (1996) Fusion of handwritten tional conference on frontiers of intelligent computing: theory
word classifiers. Pattern Recognit Lett 17:577–584 and applications (FICTA), pp 563–571
Ghosh M, Malakar S, Bhowmik S, Sarkar R, Nasipuri M (2019) Plamondon R, Srihari SN (2000) On-line and off-line handwriting
Feature selection for handwritten word recognition using recognition: a comprehensive survey. IEEE Trans Pattern Anal
memetic algorithm. In: Mandal J, Dutta P, Mukhopadhyay S Mach Intell 22(1):68–89
(eds) Advances in intelligent computing. Studies in computa- Ramraj S, Uzir N, Sunil R, Banerjee S (2016) Experimenting
tional intelligence, vol 687, pp 103–124 XGBoost algorithm for prediction and classification of different
Gupta JD, Samanta S, Chanda B (2018) Ensemble classifier-based datasets. Int J Control Theory Appl 9:651–662
off-line handwritten word recognition system in holistic Senior AW, Robinson AJ (1998) An off-line cursive handwriting
approach. IET Image Proc 12(8):1467–1474 recognition system. IEEE Trans Pattern Anal Mach Intell
James J, Lakshmi C, Kiran U, Parthiban A (2019) An efficient offline 20(3):309–321
handwritten character recognition using CNN and XGBoost. Int Song R, Chen S, Deng B, Li L (2016) eXtreme gradient boosting for
J Innov Technol Explor Eng (IJITEE) 8(6):115–118 identifying individual users cross different digital devices.
Katiyar G, Katiyar A, Mehfukz S (2017) Off-line handwritten Springer, Berlin, pp 43–54
character recognition system using support vector machine. Am Steinherz T, Rivlin E, Intrator N (1999) Offline cursive script word
J Neural Netw Appl 3(2):22–28 recognition—a survey. Int J Doc Anal Recogn 2:90–110
Kaur H, Kumar M (2018) A comprehensive survey on word Tavoli R, Keyvanpour M, Mozaffari S (2018) Statistical geometric
recognition for non-Indic and Indic scripts. Pattern Anal Appl components of straight lines (SGCSL) feature extraction method
21(4):897–929 for offline Arabic/Persian handwritten words recognition. IET
Kaur H, Kumar M (2019) Benchmark dataset: offline handwritten Image Proc 12:1606–1616
Gurmukhi city names for postal automation. In: Sundaram S, Vinciarelli A, Bengio S, Bunke H (2004) Offline recognition of
Harit G (eds) Document analysis and recognition. DAR 2018. unconstrained handwriting texts using HMMs and statistical
Communications in computer and information science, vol 1020, models. IEEE Trans Pattern Anal Mach Intell 26(6):709–720
pp 152–159 Weldegebriel HT, Liu H, Haq AU, Bugingo E, Zhang D (2020) A
Kessentini Y, Paquet T, Hamadou AMB (2010) Off-line handwritten new hybrid convolutional neural network and eXtreme gradient
word recognition using multi-stream hidden Markov models. boosting classifier for recognizing handwritten ethiopian char-
Pattern Recogn Lett 31(1):60–70 acters. IEEE Access 8:17804–17818
Koerich AL, Sabourin R, Suen CY (2003) Large vocabulary off-line Wu X, Chen Q, You J, Xiao Y (2019) Unconstrained offline
handwriting recognition: a survey. Pattern Anal Appl handwritten word recognition by position embedding integrated
6(2):97–121 resNets model. IEEE Signal Process Lett 26(4):597–601
Kumar M, Sharma RK, Jindal MK (2013) A novel feature extraction Younis KS (2017) Arabic handwritten character recognition based on
technique for offline handwritten Gurmukhi character recogni- deep convolutional neural networks. Jordan J Comput Inf
tion. IETE J Res 59(6):687–692 Technol (JJCIT) 3(3):186–200
Kumar M, Jindal MK, Sharma RK (2014a) A novel hierarchical Zhang TY, Suen CY (1984) A fast parallel algorithm for thinning
technique for offline handwritten Gurmukhi character recogni- digital patterns. Commun ACM 27(3):236–239
tion. Natl Acad Sci Lett 37:567–572 Zhong Z, Jin L, Xie Z (2015) High performance offline handwritten
Kumar M, Jindal MK, Sharma RK (2014b) Efficient feature chinese character recognition using GoogLeNet and directional
extraction techniques for offline handwritten Gurmukhi character feature maps. In: Proceedings of international conference on
recognition. Natl Acad Sci Lett 37(4):381–391 document analysis and recognition (ICDAR), pp 846–850
Kumar M, Jindal MK, Sharma RK, Jindal SR (2016) Offline
handwritten pre-segmented character recognition of Gurmukhi Publisher’s Note Springer Nature remains neutral with regard to
script. Mach Gr Vis 25(1):45–55 jurisdictional claims in published maps and institutional affiliations.
Kumar M, Jindal MK, Sharma RK (2017) Offline handwritten
Gurmukhi character recognition: analytical study of different

123

You might also like