Professional Documents
Culture Documents
ABSTRACT
This Automated text scoring (ATS) or subjective-answer evaluation is one of the big hurdles in the technical advancement of
academics. Reading each answer meticulously and scoring them impartially becomes a monotonous task for many in the
teaching profession, especially if the answers are long. Another big challenge is comprehending the student’s handwriting.
Marking criteria may also vary largely with domain, for instance, credit is given to usage of correct grammar in some cases,
while other domains may require certain keywords to be present in student’s answers. In this paper, we have tried to approach
this problem with three perspectives- two standard linguistic approaches and a deep learning approach. The first approach
employs the presence of certain keywords as a marking criteria and also includes a handwriting recognizer that can extract text
from scanned images of the handwritten answers. The second approach uses similarity between an understudy and a benchmark
answer. This paper also proposes the use of a sequential model, which is trained on the Automated Student Assessment Prize -
Automated Essay Scoring (ASAP-AES) dataset for evaluating long answers.
Keywords :- Natural Language Processing (NLP), Cosine Similarity, Jaccard Similarity, Synonym Similarity, Bigram
Similarity, Sequential Model, LSTM, Root Mean Squared Error (RMSE)
usage of latent semantic analysis and similarity measures for 3.1 Answer Evaluation using the keyword
scoring a text. Even though standard linguistic approaches matching technique.
have given substantial performance over text scoring, they For domains that include many pre-specified words and
still need to get free of domain specificity. figures, scoring must be based on the presence of that
Impedovo et al.[6] proposes the use of optical character information in the answer. This is where the keyword
recognizer(OCR) for handwriting Extraction. And also matching criteria comes into picture. The flow process of this
discusses the problems in recognizing handwritten and printed approach is shown in figure 1.
characters.An advanced version of OCR model is proposed by
Vanni et al.[9]. It uses an artificial neural network at the
backend to get high accuracy. Nagy et al. [7] also discusses
the OCR model and its strengths and weaknesses. Pradeep et
al.[8] uses a new approach called diagonal based feature
extraction by training a neural network using many
handwritten alphabets. Shi el al.[10] discusses the scene text
recognition and checks the performance on both lexicon-free
and lexicon- based recognition. The study of Oganian et al.[15]
gives insight on the bilingual word recognition system.
The study of Lahitani et al. [17] has been a source of
motivation for our second approach. They use similarity
scores for the answer and generate the final score. Bluche et al.
[18] used LSTM-RNN to predict open bigram and performed
experiments on public databases, Rimes and IAM. Nau et al.
[19] have combined latent semantic analysis and linear
regression to predict the score of answers.
With advances in deep learning, such systems have
surpassed the past benchmarks in terms of performance. A
paper by Tai et al.[11] proposes usage of Tree-LSTM instead
for general LSTMs (linear chain structure), for semantic
representation of text. The basic thought is that any natural
language combines its words with phrases. This custom
LSTM structure outperforms in the task of predicting semantic
relatedness of two sentences. Another paper by Tang et al.[12] Figure 1: An overview of answer evaluation using keyword
proposed learning the sentiment-specific semantic matching technique
representation for the analysis of the entire document. A. Providing Inputs
Alikaniotis et al.[13] proposed usage of LSTMs to represent Before starting with the evaluation process, we need to
semantics of the understudy answer and word representation provide our system with some inputs to get accurate and fair
model that learns the impact of a word on the text’s score. marking for the students' answers. These inputs are:
Laxmi et al.[14] propose usage of ANN for comparing the a) Students Answer: The most obvious input here is the
understudy answer with a benchmark answer and keyword list. students answer which is to be evaluated based on some
The same answer is also evaluated using a NLP system for predefined rules. This can be in the form of a scanned
deducing marks over grammatical/spelling mistakes. The image or a text, as suitable.
scores predicted by the two systems are compared and a final b) Keywords List: This list consists of the keywords that are
score is calculated. expected to be present in the student answer. These
The Hewlett Foundation sponsored a contest on Kaggle in keywords are related to the domain of the question of
2012 named “Automated Student Assessment Prize”(ASAP). which answers are being evaluated.
They released a dataset, which has been used by many text c) Maximum Marks: This is the maximum marks out of
scoring systems(Shermis, [16]). which the student answer is to be marked.
d) Minimum Length: This is the minimum length of the
III. METHODOLOGY answer above which the answer will be considered a good
answer, and no marks will be deducted if a student's
Proposed system involves 3 approaches viz. keyword answer exceeds this limit. This number depends on the
matching, similarity measures and sequential model, each of maximum marks of the answer.
which are elaborated in sections 3.1, 3.2 and 3.3. e) Maximum Matching Keywords: Expecting the student to
mention all the keywords from the provided keywords list
in his/her answer is not practical, so we provide a
minimum number of keywords which is expected to Each word present in the list is matched in the ‘Wordnet’
match with the list for the answer to score good marks. database and if it is present in the database, all the words
related to that word are also added to the list.
B. Text Extraction using Google vision API. E. Keyword matching
The student answer is provided to the system in the form of After extending the keywords list, all the words in the
an image, which is the scanned image of the student’s answer preprocessed student's answer are matched against the
sheet. This answer extracting function increases the use of the extended keyword list. The numbers of matching keywords
answer evaluation system in real life. In this step, input image are counted for further calculations.
is provided to the system. Then the text written on the answer F. Calculating percentages
sheet is extracted using vision API[20]. Sample input is shown The required percentages are keyword percentage, word
below in figure 2. percentage, and sentence percentage. The weight of these
parameters are different, which is calculated by manually
evaluating answers and checking the role of these parameters
in the final score. The weights are derived from the combined
observation of Kapoor et al.[28], Mahmud et al.[29], and
Bharadia et al.[30]. The weights shown in equations 1,2,3.
…. (1)
…. (2)
.....(3)
• Dropout layer: This layer is added to prevent the model Same evaluation process as discussed for the first method is
from overfitting. During training, this layer sets random used to check the performance of our second method. Our
inputs to 0 with a pre-specified frequency rate. second system is also giving satisfactory results as shown in
• Dense layer: It is a regular fully-connected neural network table 2.
layer. This layer was added in the model to provide
learning features from all combinations. It is stacked after Sample Marks allotted in Marks allotted Squared
LSTM because many frameworks give the internal (or No. manual checking by our system Error
hidden) state h as output. The dimensionality of this output 1 2.96 3.70 0.5476
(which is the number of units) may not be equal to the 2 3.22 2.98 0.0576
dimensionality of the target that we want. Dense layer 3 4.00 3.84 0.0256
allows us to tune the dimension of output. 4 2.21 2.28 0.0049
5 3.5 3.58 0.0064
The model is feeded with uniformly-dimensioned vectors 6 4.2 4.78 0.3364
representing the essays and predicts the score for the answer. 7 3.9 3.22 0.4624
The length of intersection of the keyword set input by the 8 4.5 4.22 0.0784
examiner and the student’s answer is considered to have 9 4.5 3.79 0.5041
weightage(on a scale of 0 to 1) over the marks obtained. 10 3.7 3.45 0.0625
Consequently, we multiply this weightage to the score
predicted by the sequential model
Table 2 :Comparison between manual checking and by the
similarity measure evaluation system.
IV. EVALUATION
In result analysis of 2000 samples RSME value is
The results of our first method (answer evaluation system 456716542 for method 2.
using keywords matching) is compared with the marks The approach 3 for answer evaluation includes usage of a
allotted manually by teachers answers taken from ASAP-SAS deep learning model which was trained on ASAP-AES. The
dataset[21] . Results of 10 sample are shown in table 1, the answer evaluation model gives a quadratic(or Cohen's) kappa
second column shows the marks allotted in manual checking score of 0.96 upon being tested on 25% of the ASAP-AES
(from dataset) and the third column shows the marks allotted dataset. Cohen's Kappa (shown in equation 9) measures the
by our system. The marks obtained are close and the results of agreement between two raters who classify N items into C
our system are satisfactory. mutually exclusive categories.
……................................. (9)
Sample Marks allotted in Marks allotted Squared
No. manual checking by our system Error where, po is the relative observed agreement among raters,
1 2.5 2.9 0.16 and pe is the hypothetical probability of chance agreement.
2 3.22 3.5 0.078 Kappa score is 1 if both rates are in complete agreement.
3 4.00 3.42 0.3364 A scatter plot of the predicted scores over the actual scores
4 2.21 2.26 0.0025 in the test dataset is shown in Figure. 5, where the x-axis
5 3.5 3.18 0.1024 represents various essays(or student answers) and y-axis
6 4.2 3.4 0.64 represents the scores obtained.
7 3.5 3 0.25
8 4.5 4 0.25
9 4.5 3.9 0.36
10 3.7 4.1 0.16
.................................................(8)
In result analysis of 2000 samples RSME value is
0.252012678 for method 1. Figure 5: Scatter plot of test scores(red) vs. predicted
scores(green)
dual bachelor's degrees in economics and physics there. He [7] George Nagy, Stephen V. Rice, and Thomas A. Nartker,
moved to California in 1995 for Stanford University. But he “ Optical Character Recognition: An Illustrated Guide to
decided instead to pursue a business career. He then co- the Frontier (The Springer International Series in
founded the web software company Zip2 with his brother Engineering and Computer Science)
Kimbal. The startup was acquired by Compaq for $307 [8] J.Pradeep , E.Srinivasan and S.Himavathi, “ Diagonal
million in 1999. Musk co-founded online bank X.com that Based Feature Extraction For Handwritten Alphabets
same year. It merged with Confinity in 2000. This later Recognition System Using Neural Networks ”,
formed the company PayPal, PayPal was subsequently bought International Journal of Computer Science &
by eBay in 2002 for $1.5 billion.''' Information Technology (IJCSIT), Vol 3, No 1, Feb
Output 2011.
The system provides a score of 8.5 to the first answer, [9] B vanni, M. shyni, and R. Deepalakshmi, ”High
while a score of 2.5 to the second answer for the given accuracy optical character recognition algorithms using
keyword list, which is quite close to what we want to achieve. learning array of ANN” in Proc. 2014 IEEE International
Conference on Circuit, Power and Computing
Technologies (ICCPCT), 2014 International Conference.
VI. CONCLUSION AND FUTURE WORK [10] Baoguang Shi, Xiang Bai and Cong Yao. 2015. An End-
to-End Trainable Neural Network for Image-based
Our system successfully evaluates answers from different Sequence Recognition and Its Application to Scene Text
domains using the three mentioned approaches. As the Recognition. arXiv:1507.05717v1 [cs.CV].
approaches use different courses of action, the application of [11] Kai Sheng Tai, Richard Socher, and Christopher D.
our system is expanded. The system can be improved by using Manning. 2015. Improved semantic representations from
a more efficient technique of handwritten text extraction (for tree-structured long short-term memory networks. Feb.
bad handwritings) and employing it to the entire system. [12] Duyu Tang. 2015. Sentiment-specific representation
Another functionality that can be added is grammar and learning for document-level sentiment analysis. In
spelling check to deduce marks over grammatical mistakes, Proceedings of the Eighth ACM International
which is quite an important feature for a subject of language Conference on Web Search and Data Mining - WSDM
domain. '15. Association for Computing Machinery (ACM).
[13] Dimitrios Alikaniotis, Helen Yannakoudakis and Marek
REFERENCES Rei. 2016. Automatic Text Scoring Using Neural
Networks. Proceedings of the 54th Annual Meeting of
the Association for Computational Linguistics, pages
[1] M. D. Shermis & J. Burstein (Eds.). Page, E. B. (2003).
715–725,
Project Essay Grade: PEG. In, Automated essay scoring:
[14] V. Lakshmi and Dr. V. Ramesh. 2017. Evaluating
A cross-disciplinary perspective. Lawrence Erlbaum
Students’ Descriptive Answers Using Natural Language
Associates Publishers, pages 43-54.
Processing and Artificial Neural Networks. International
[2] Piyush Patil, Sachin Patil, Vaibhav Miniyar and Amol
Journal of Creative Research Thoughts (IJCRT1704424),
Bandal. 2018. Subjective Answer Evaluation Using
volume 5.
Machine Learning. International Journal of Pure and
[15] Y. Oganian, M. Conrad, A. Aryani, K. Spalek and H. R.
Applied Mathematics, volume 118.
Heekeren, "Activation Patterns throughout the Word
[3] Thomas K. Landauer, Darrell Laham, and Peter W. Foltz.
Processing Network of L1-dominant Bilinguals Reflect
2003. Automated scoring and annotation of essays with
Language Similarity and Language Decisions," in
the Intelligent Essay Assessor. In M.D. Shermis and J.C.
Journal of Cognitive Neuroscience, vol. 27, no. 11, pp.
Burstein, editors, Automated essay scoring: A cross-
2197-2214, Nov. 2015.
disciplinary perspective, pages 87–112.
[16] Mark D Shermis. 2015. Contrasting state-of-the-art in
[4] L.M. Rudner and Tahung Liang. 2002. Automated essay
the machine scoring of short-form constructed responses.
scoring using Bayes’ theorem. The Journal of
Educational Assessment, volume 20(1), pages 46–65.
Technology, Learning and Assessment, 1(2):3–21.
[17] A. R. Lahitani, A. E. Permanasari and N. A. Setiawan,
[5] Keisuke Sakaguchi, Michael Heilman, and Nitin
"Cosine similarity to determine similarity measure:
Madnani. 2015. Effective feature integration for
Study case in online essay assessment," 2016 4th
automated short answer scoring. Proceedings of the
International Conference on Cyber and IT Service
Tenth Workshop on Innovative Use of NLP for Building
Management, Bandung, 2016, pp. 1-6.
Educational Applications.
[18] T. Bluche, C. Kermorvant, C. Touzet and H. Glotin,
[6] S. Impedovo, L. Ottaviano and S.Occhinegro, “ Optical
"Cortical-Inspired Open-Bigram Representation for
character recognition ,'' International Journal Pattern
Handwritten Word Recognition," 2017 14th IAPR
Recognition and Artificial Intelligence, Vol. 5(1-2), pp.
International Conference on Document Analysis and
1- 24, 1991
Recognition (ICDAR), Kyoto, 2017, pp. 73-78.