Professional Documents
Culture Documents
Assessment for
Indira Gandhi Learning
National Open University
School of Education
Block
2
TECHNIQUES AND TOOLS OF
ASSESSMENT AND EVALUATION
UNIT 5
Techniques of Assessment and Evaluation 5
UNIT 6
Criteria of a Good Tool 21
UNIT 7
Tools for Assessment and Evaluation 44
UNIT 8
ICT Based Assessment and Evaluation 64
EXPERT COMMITTEE
Prof. I. K. Bansal (Chairperson) Prof. Anju Sehgal Gupta
Former Head, Department of Elementary School of Humanities,
Education, NCERT, New Delhi IGNOU, New Delhi
Prof. Shridhar Vashistha Prof. M. C. Sharma (Retd.)
Former Vice-Chancellor B.Ed. Prog. Coordinator till 31.3.17
Lal Bahadur Shastri Sanskrit School of Education
Vidhyapeeth, New Delhi IGNOU, New Delhi
Prof. N. K. Dash
Prof. Parvin Sinclair School of Education
Former Director, NCERT IGNOU, New Delhi
School of Sciences,
IGNOU, New Delhi Dr. Gaurav Singh
(Programme Co-coordinator-B.Ed.)
Prof. Aejaz Masih School of Education
Faculty of Education, IGNOU, New Delhi
Jamia Millia Islamia, New Delhi
Prof. Pratyusha Kumar Mandal
DESSH, NCERT, New Delhi
SPECIAL INVITEES (Faculty of School of Education)
Prof. D. Venkateshwarlu Dr. Bharti Dogra
Prof. Amitav Mishra Dr. Vandana Singh
Ms. Poonam Bhushan Dr. Niradhar Dey
Dr. Eisha Kannadi Dr. Elizabeth Kuruvilla
Dr. M. V. Lakshmi Reddy
PRINT PRODUCTION
Prof. Saroj Pandey Mr. S.S. Venkatachalam
Director Assistant Registrar (Publication)
SOE, IGNOU, New Delhi SOE, IGNOU, New Delhi
October, 2017
Indira Gandhi National Open University, 2017
ISBN: 978-81-266-
All rights reserved. No part of this work may be reproduced in any form, by mimeograph or any
other means, without permission in writing from the Indira Gandhi National Open University.
Further information on the Indira Gandhi National Open University courses may be obtained from
the University’s Office at Maidan Garhi, New Delhi-110068.
Printed and published on behalf of the Indira Gandhi National Open University, New Delhi, by the
Director, School of Education, IGNOU, New Delhi.
Laser Typeset : Rajshree Computers, V-166A, Bhagwati Vihar, Uttam Ngr. (Near Sec.-2, Dwarka),
N.D.59
Printed at :
BES 127 ASSESSMENT FOR LEARNING
BLOCK 1 Evaluation in Teaching-Learning Process
Unit 1 Concept and Purpose of Evaluation
Unit 2 Perspectives of Assessment
Unit 3 Approaches to Evaluation
Unit 4 Issues, Concerns and Trends in Assessment and Evaluation
ASSESSMENT AND
EVALUATION
Structure
5.1 Introduction
5.2 Objectives
5.3 Purpose of Assessment and Evaluation
5.4 Techniques of Assessment and Evaluation
5.4.1 Concept Tests
5.4.2 Self-report Techniques
5.4.3 Assignments
5.4.4 Observation Technique
5.4.5 Peer Assessment
5.4.6 Sociometric Technique
5.4.7 Portfolios
5.4.8 Project Work
5.4.9 Debate
5.4.10 School Club Activities
5.1 INTRODUCTION
In Unit 1, you have studied about the concepts of measurement, assessment
and evaluation and by now, you might be able to differentiate these terms from
one another. Assessment is an integral part and ongoing process in the educational
context. To learn effectively, students need to know as to how they are
performing. Similarly, to be an effective teacher, you must be informed about
what the student knows, feels, and can do, so that you can help him/her build
on her/his skills, knowledge, and attitudes. Therefore, you and your students
need continuous feedback on their progress and difficulties in order to plan
appropriate learning activities and to make adjustments to those already planned.
Assessment means the way by which teachers and other stakeholders involved
in students’ learning collect information systematically and then use it to know
their level of achievement in academic, behavioural and social areas. In this Unit,
you will study various techniques of evaluation.
5
Techniques and Tools of
Assessment and Evaluation 5.2 OBJECTIVES
After going through this Unit you will be able to:
describe the techniques of assessment and evaluation,
explain the significance of self-report techniques,
analyze students’ characteristics through observation,
use interview as a tool of evaluation,
recognize the use of assignment as an assessment technique,
identify the importance of peer ratings,
conduct socio-metric technique in classroom,
evaluate project work by using a rubric,
assess portfolios for understanding students’ progress, and
evaluate the techniques used for group processes.
In the above example, the teacher presented a situation and then asks some
questions related to prediction, problem-solving and clarification of concept. Such
kind of test helps you to know how much your students attain understanding
of concepts.
5.4.2 Self-report Techniques
Self-report techniques are data gathering tools where respondents provide
information about themselves. There are various self-reporting techniques such
as interview, diaries, questionnaire, reflective journal, etc. These techniques are
widely used to assess various traits or attitudes possessed by the individuals.
Though these are subjective techniques, they provide means for extracting the
hidden treasure of one’s own behaviours and patterns of personality. Let us
discuss in detail about the various self-reporting techniques such as interview,
focus group discussion, questionnaire and reflective journal.
Interview: An interview is a data collection method through which an interviewer
8
asks questions to an interviewee on a particular topic. It can be semi-structured
or structured and may use different media (e.g., phone, e-mail, and in person). Techniques of Assessment
and Evaluation
In semi-structured interview, the interviewer has an idea about the types of
questions but the order and way of asking questions may vary. A structured
interview on the other hand, is a specified set of questions that are asked in
the same order during each interview. Interviewing is considered as an
interpersonal encounter. Establishing rapport is an important element in interviewing.
Keeping impartiality to the responses given by the interviewee is important;
otherwise it leads to bias responses.
During interviewing, you may know whether the student understands a particular
concept. The student may be asked to give a personal explanation of a specific
concept and to use that concept to solve a problem. Through this technique,
besides getting information, it is also possible to observe a student’s body
language and facial expressions. For assessment purpose, interview may serve
the following functions:
to identify the areas of difficulty and gaps in understanding a concept among
students,
to understand how students are able to apply their knowledge in new
situations,
to examine how students’ understanding change with teaching-learning
process,
to obtain verbal feedback about teaching techniques, new concepts, etc.
Through the feedback received from students, you may be able to refine your
teaching along with assessing the understanding level of students.
Focus group discussion: It is a systematically planned discussion led by the
teacher who acts as a moderator to collect detailed information on a particular
topic. In focus group discussion, a moderator leads the discussion with a small
group of individuals to analyze in detail the thinking and feeling of the group
members about a topic. It is called ‘focus group’ because the moderator keeps
the individuals in the group focused on a particular topic taken for discussion.
The moderator generates group discussion through the use of open-ended
questions, and he/she acts as a facilitator of the group process. It is used for
multiple purposes as follows:
Generating new ideas and concepts,
Obtaining general background information about a topic of interest,
Generating research hypotheses that can be submitted to further research,
Diagnosing the problem in a new way,
Examining how respondents talk about the selected topic, and
Interpreting previously obtained results.
You might be thinking how an interview differs from a focus group discussion.
In interview, the interviewer has a prominent role in asking specific questions.
The depth of information received from an interview is less. In the case of focus
group discussion, the number of participants varies from four to eight members
and you will get a great deal of information.
Questionnaire: A questionnaire is a self-report data collection technique that
each participant fills out in the given form. There are open-ended and closed-
ended questionnaires. Through questionnaires, one can obtain information about
the thoughts, feelings, attitudes, beliefs, values, perceptions, personality, and 9
Techniques and Tools of behavioural intentions of participants. Open-ended questions are valuable when
Assessment and Evaluation
the researcher needs to know what people are thinking. Such questions help
the interviewer to understand participants’ inner worlds in their natural languages
and categories.
When do we use questionnaires?
Questionnaires will be used for the following purposes:
To collect feedback from a large number of students,
To know students’ opinions, attitudes, feelings and perception about a
particular issue, and
To allow each student to give anonymous feedback.
Reflective Journal: Reflective journal is a tool where students can reflect on
and write about their progress in learning. They can write their achievements
and difficulties in learning. Students can write whatever they found to be
innovative or what made them think and ponder about their learning experiences.
Journal writing is a non-threatening way to get information about students’ thinking
and feeling. The following points give you an idea about how to use and what
to be included in the journal.
When should I use it?
Daily or weekly
How could I use it?
As a self- assessment tool for students
At the beginning or end of a unit, topic or project
As a communication tool for students and teachers
Why should I use it?
It enhances higher order thinking
It leads to self-assessment
It sets goals for future learning
It gives students a sense of ownership and control on their own learning
Students’ reflective thinking can be supported by giving opportunities to reflect
on their own learning and their own context of learning. Thereby, it promotes
students’ responsibility of learning. Reflective journal permits students to practice
their writing skill in a free environment that encourages critical thinking. In this
way, reflective journals can be used as an assessment tool that provides teachers
an insight into how students value their own learning and achievement.
Benefits of peer-assessment
An educational process that installs autonomy in learners.
Empowerment of the learner in a learning environment.
Development of learner confidence in assessing/ marking peers (through
practice).
Development of learner ability to self-evaluate and reflect.
Greater understanding of what is required by tutors for assessment at
various levels.
Interactive classes for marking/feedback.
Reflection on recently completed assessments with full explanation of
the answer (improving information and understanding).
Clear, open marking systems (seeing what is required and improving
work).
Seeing standards set by peers as well as mistakes of others (and
avoiding them in the future).
Gaining an ability to ‘stand back’ from own work for assessment
purposes (an essential ability of an ‘objective’, ‘unbiased’ scientist).
Rapid way for a tutor to assess a large amount of student work and
provide specific feedback.
(Source: Langan, Mark A. & Wheater, P., 2003)
12
Techniques of Assessment
Check Your Progress 3 and Evaluation
13
Techniques and Tools of
Assessment and Evaluation A
C
D
H
E
F
B
In the figure 5.2, you notice that student ‘B’ is not selected by others. Hence,
‘B’ is considered as an isolate. Students ‘A’, ‘D’ and ‘E’ choose each other
and thus form a clique. Student ‘E’ is liked by most, thus becoming a star.
Sociogram also provides details about the relationship among the members of
the group. The sociogram, thus helps the teacher to build a supportive learning
environment. Between students ‘C’ and ‘H’, reciprocity is both sided while for
‘E’ and ‘F’ it is one-sided only. Sociometry is thus an important technique for
peer ratings.
5.4.7 Portfolios
The details of portfolio have been discussed in Unit-7 of this Block. A portfolio
is a systematically organized collection of a student’s work that covers a specific
period of time. It proved to be an excellent vehicle for gathering together student
writing and displaying the changes in the proficiency over a period of time.
Portfolio assessment represents a blend of the instructional process and the
evaluation process. You might know that portfolio is not something static but
it is an ongoing dialogue between teacher and student and, often, between the
individual student and others in the class or between the students and parents.
Components of portfolios will vary depending on the subject that is being taught
and preferences of teachers and students. Portfolios can contain the following
14 items:
Techniques of Assessment
Completed assignments and Evaluation
Journal writings
Reflections on discussions that have been held in or out of class
Photos, sketches, and other visuals
Summary statements made at different points regarding what has been
learned
Self-assessment statements
Process
Outcome/
Product
5.4.9 Debate
Debate is a kind of dialogue about a significant topic. It acts as a useful technique
to promote independent learning about a topic selected for debating. Sometimes
you are familiar with this technique and its procedure. In this technique, you
have to play the role of a moderator. The first task is to decide the topic to
be debated. You have to give an overview of the topic prior to the task being
set. Then you should set the scene and allocate the team. Generally, the ideal
team size is 10, out of which 5 present a case in favour of the topic and the
rest 5 defend the topic. During the conduct of debate you, as a moderator,
can give clear guidelines on the structure of the debate, timings and the roles
to be performed.
The next task is how to evaluate the debate. For evaluation, you can prepare
a rating scale containing the following indicators:
Relevance of the content
Presentation style
Clarity of the argument
Logical approach
Quality of the rebuttal
17
Techniques and Tools of This technique develops skills such as listening, rapid response, speaking, and
Assessment and Evaluation
debating the issues in a professional manner. It also promotes skills such as
team spirit, synthesizing data, negotiation, argumentation, and rebuttal skills.
6.1 INTRODUCTION
In Unit-5, you have studied about the techniques of evaluation such as
observation, interview, rating scale, socio-metric techniques, seminar, group
discussion, etc. From that Unit, you might have understood the techniques of
using various assessment tools for assessment and evaluation. In this Unit, we
will discuss the types of tools and the essential criteria of a good tool. During
the course of discussion you will understand that there are different criteria to
be taken into consideration while evaluating a tool or determining the worth
of a tool to be used. In this regard, broadly we will discuss the criteria of
an effective tool such as reliability, validity, usability, objectivity and norm of
a tool and how to determine the index of reliability and validity for making
it an effective tool. Again, the effective tools are used for satisfying the purpose
of its evaluation. According to this purpose, we can classify the tools into two
broad categories, that are self-made tool and standardized tool. This Unit will
also make you understand the concept of self-made and standardized tools and
also to differentiate between them.
6.2 OBJECTIVES
After going through this Unit, you should be able to:
define self-made and standardized tools,
differentiate self-made and standardized tool,
discuss the different criteria of an effective tool,
explain the relationship between validity and reliability,
describe the procedure to find out reliability and validity of a tool,
calculate the index of reliability and validity of a tool, and
explain the criteria of an effective tool. 21
Techniques and Tools of
Assessment and Evaluation 6.3 EVALUATION TOOLS : TYPES AND
DIFFERENCES
We use a variety of tools for the purpose of assessment of performances of
the students in the school. We also use different types of tools for collecting
data to conduct various projects and research studies. It is also equally used
for various administrative purposes, for example, recruitment for various jobs.
That is why we say that for assessing something, we need to use a tool. Now
the question is as to which type of tool can be used for what purpose. On
the basis of the purpose, tools are classified into two broad types namely self-
made and standardized tools. In this Section, you will learn about both types
of tools and also you will be able to differentiate them.
22
Criteria of a Good Tool
Table 6.1: Difference between self-made and standardized tool
Self or Teacher-made Tool StandardizedTool
Teacher-made tool is made by the teacher Standardized tools are prepared by the test
for the students to whom he/she teaches. specialists and the experts in the field.
Teacher-made tool is prepared for the small Standardized tools are prepared for the large
group students in a school or class. group population.
The use of teacher-made tool is verylimited. The use of standardized tool is very vast.
Norms are not developed for developing Developing norms are compulsory in
teacher-made tool. standardized tool.
Item analysis and item discrimination are Item-wise analysis is made to know the item
not calculated in teacher-made tool. difficulty and discrimination before
including the item in the final test.
The difficulty level of the item in teacher- The difficulty value of the item in
made tool depends upon the standard of standardized tools is generally average in
teaching of the teacher to the group. It may standard.
be easy, average or difficult.
Subjectivity in preparing and applying the Standardized tools are always objective in
teacher-made tool is always there. nature.
Teacher-made tool supports the principles Standardized tools support the principles
and spirit of continuous comprehensive and spirit of summative evaluation.
evaluation, specifically the formative
assessment.
The Unit tool, quarterly and half-yearly The tests prepared for the Board Exami-
examinations conducted in the schools are nation may be central or state level and the
examples of teacher-made tool. test prepared for admission, placement and
recruitment for various vocations/jobs are
also the examples of standardized tool.
Activity 1
Analyse the Table 6.1 and summarize it on the following aspects :
Purpose of both type of tools :
........................................................................................................
........................................................................................................
........................................................................................................
Process followed to prepare the tools :
........................................................................................................
........................................................................................................
........................................................................................................
Main beneficiaries of the tools and the jurisdiction of its use :
........................................................................................................
........................................................................................................
........................................................................................................
Examples of the tools :
........................................................................................................
........................................................................................................
........................................................................................................
Administer Twice
Group Single
After analysis of Table 6.2, we can say that test-retest method is one of the
easiest method to determine reliability of the test. In this method, the same test
is administered twice within a gap of time limit to minimum 15 days to maximum
six months. After getting the scores of test and re-test results, any method of
correlation either Spearman’s Rank Difference or Pearson’s Product Moment
Correlation method are used to get the coefficient of correlation (please read
Unit-16, Block-4, BES-127). As this has already mentioned in the Table, if
the coefficient of correlation (r) is = +0.5, we can say that the test is reliable.
The following methods can be employed to calculate correlation of the testing
and re-testing scores.
6 d2
ρ = 1 n(n2 - 1)
r =
r = Correlation coefficient
x = Deviations taken from mean of the distribution (X-M) [X – Individual
Score and M – Mean of the first distribution]
y = Deviations taken from mean of the distribution (Y-M) [Y – Individual Score
and M – Mean of the second distribution]
N = Total number of the candidates
σx = Standard deviation of first distribution
σy = Standard deviation of second distribution
Limitations of using the method : The following are the limitations of the
test-retest method :
As the same test is administered twice on the same group, there will be
the threat of carry over effect, it means, during the second administration,
the candidates may remember many items from the first administration.
The scoring of second administration is usually high than the first one.
Maintaining a gap of time between test and re-test is also again one of
the important aspects to determining exact value of reliability. If time gap
is very less, then carry over effect will be high and on the other side,
if time gap is very high, maturity effects of the candidates may hamper
the test results.
This method is not free from errors. Memory, carry over, practice and
maturity effects are high in this technique.
Activity 2
Prepare a small test and find out reliability of the test by using test-
retest method of reliability.
.........................................................................................................
.........................................................................................................
.........................................................................................................
.........................................................................................................
.........................................................................................................
In Table 6.3 we have stated the processes adopted in the parallel-form method
to determine the reliability of a test. In this method, two parallel tests are
developed which are equivalent in all respects, such as content, objectives, types
of items, time given, difficulty level of the item, etc. When two parallel tests
are used to get the reliability of the test, it is quite natural that the carry over,
memory, and practice effects are highly minimized. For example, when items
are different, there is very less chance to remember the earlier questions. Just
like the test-retest method, the first form of the test is administered on the group
and after a gap of some time period say, from a minimum of 15 days to a
maximum of 6 Months, the second form of the test may be used. Spearman’s
or Karl Pearson’s method can be used to get the coefficient of correlation of
the two administrations. If correlation coefficient is, r > +0.5, then the test is
said to be reliable.
Limitation of parallel form method : Parallel form method is also not
completely free from errors. There are possibilities of making errors in this
method also:
Practice and carry over effect is not totally minimized, as both the tests
are equivalent in nature in many respects except only the items are
different and a time interval of 15 days to 6 months is given for testing
28 the second form of the test. During this period, there is a chance that
Criteria of a Good Tool
the students may practice the similar content and items and hence chances
for getting better scores in second testing are generally more.
Preparing two parallel forms of the tests is also a complex task.
This method is comparatively time taking to get the reliability.
(iii) Internal consistency reliability : Internal consistency reliability indicates
the homogeneity of the test. If all the items of the test measure the same
function or trait, the test is said to be a homogeneous one and its internal
consistency reliability would be pretty high. The most common methods
of estimating internal consistency reliability are the (a) Split-half method
and (b) Rational equivalence method. Let us discuss split-half method first.
(a) Split-half method : This method is also called as ‘odd-even method’.
Let us discuss the procedures and conditions of using split-half method
for determining reliability.
Table 6.4 : Precedures of using split-half reliability method
Conditions Use
Test Single form
Administer Once
Group Single
Time interval Not necessary as the test will administer only once.
Determining reliability By employing correlation statistics as well as split-
half reliability formula.
Decision Reliable if reliability index is positive and
high. (r = +0.5).
Threat of error Error of ‘carry over effect’, ‘memory effect’,
‘practice effect’, etc. are minimized, but still
splitting the full test into two halves is also difficult
because splitting the full test into two-halves can
be made in many different ways such as odd and
even, first fifty percent and last fifty percent, and
so on.
Suggestion Is one of the highly used method as most errors
are minimized.
Table 6.4 reveals the procedures of using split-half reliability method. Most of
the errors that occur in test-retest method are minimized in split-half method
as the test is not conducted twice on the same group. In this method, the test
is to be administered only once to a single group and after getting the scores
of the students, it has to be split into two equal halves like in odd and even
items, or any other technique to divide it into two halves. The scores of the
candidates for odd items and for even items will be separated and suitable
statistical techniques are used to get the reliability of the scores of two halves.
The Spearman-Brown Prophecy formula is generally used for calculating
reliability of the full test. The formula is as follows:
29
Techniques and Tools of
Assessment and Evaluation rtt =
Just like, test-retest method, correlation of the two halves of the test (r1/2)
will be calculated by using the method of Spearman or Pearson’s correlation.
Splitting of the item can be done as shown in Table 6.4.
Table 6.5: Worksheet for the Odd-Even Reliability
Examinee Total Number of correct Number of
No. of scores on odd- correct score
correct numberred items on even-numbered
scores items
1 50 29 21
2 46 20 26
3 55 28 27
4 39 19 20
5 55 25 30
6 49 27 22
7 60 32 28
8 42 20 22
9 45 24 21
10 53 28 25
Table 6.5 represents the total number of correct scores of the examinees and
also the number of correct scores on odd numbered items and on even numbered
items. For example, the examinee at serial number 1 scored a total of 50 out
of which 29 scored from the odd items and 21 from the even items. Accordingly,
all examinee’s scores are grouped for odd and even items and made into two
groups for getting the reliability of the whole test.
Limitations of this method : The important error in this method is to split
the full test into two equal halves. This can be done in many ways like odd-
even, first 50 percent items and next 50 percent items, etc. The correlation
of two-haves of the test and the reliability of the full test will also be different
based on different ways of splitting the test items. Besides this type of errors,
the other type of errors such as ‘carry over error’, ‘maturity error’, ‘memory
and practice effect’ etc. are mostly minimized in this method.
Activity 3
With the example presented in Table 6.5, calculate reliability coefficient
of whole test by using Spearman-Brown Prophecy formula.
.........................................................................................................
.........................................................................................................
30 .........................................................................................................
Criteria of a Good Tool
(b) Rational equivalence method : This is purely a statistical process of
determining reliability of a test. The formula for calculating reliability in
rational equivalence method is developed by Kuder and Richardson. They
developed two formulae popularly known as Kuder and Richardson (KR)
formula number 20 and 21 (KR 20 and KR 21). Lee J. Cronbach has
called it Coefficient of Internal Consistency. This method of reliability
is an improvement of all the earlier methods as it minimises errors in
calculating reliability. You should also understand that rational equivalence
method is the improvement of split-half method. The assumption in Kuder
and Richardson formula is that all items have the same or equal difficulty
value, but not necessarily the same persons solve each item correctly. The
difficulty value of an item depends on knowing aspect of a person. If the
person knows the item, it is easy for him/her and if he/she does not know
the item, it is difficult for him/her. The assumption is that the test is
homogeneous in nature.
Table 6.6: Precedures of using rational equivalence method
Conditions Use
Test One
Administer Once
Group Single
Time interval No need
Determining reliability By employing KR-20 and KR-21 formullae.
Decision Reliable if reliability index is positive and
high. (r > +0.5).
Threat of error Errors are mostly minimized
Suggestion Can be used effectively in research
r = Reliability Index
σ 2 = Variance of the total test (square of the standard deviation)
n = Number of items in the test
p = Proportion of right responses
q = (1-p) proportion of wrong responses 31
Techniques and Tools of KR-21 formula is given as follows :
Assessment and Evaluation
r = —— KR-21
r = Reliability Index
σ 2 = Variance of the total test (square of the standard deviation)
n = Number of items in the test
M = Mean of the test scores
Limitations of rational equivalence Method : The following are the main
limitations of this method :
The item variation of item difficulty is not considered in calculating reliability
index.
The reliability coefficient is approximate and lower than the accurate
reliability coefficient.
It can not be used for power test or heterogeneous test.
The different KR formula yields different reliability coefficient.
Factors influencing Reliability Coefficient : Factors influencing reliability of
a test can be classified as extrinsic factors and intrinsic factors.
Extrinsic factors : The important extrinsic factors (i.e. the factors which remain
outside the test itself) influencing the reliability are :
i. Group variability : When the group of pupils being tested is homogeneous
in ability, the reliability of the test scores is likely to be lowered and vice-
versa. It is therefore, the method of randomization is used to get the groups
having a representative of all abilities.
ii. Guessing and chance errors : Guessing in test results in to increased
error variances and thus reduces reliability. For example, in two-alternative
response options there is a 50% chance of answering the items correctly
in terms of guessing. It is therefore difficult to determine the true scores
of the examinees.
iii. Environmental conditions : As far as practicable, testing environment
should be uniform like the proper facilities of light, air, seating arrangement,
audibility of the instructions, time of testing, qualities of materials used for
the test, etc.
iv. Momentary fluctuations : Momentary fluctuations may raise or lower
the reliability of the test scores. A broken pencil, momentary distraction
by the sudden sound of a train running outside, anxiety regarding non-
completion of homework, mistake in giving the answer and knowing no
way to change it are the factors which may affect the reliability of the
test scores.
Intrinsic factors : The principal intrinsic factors (i.e. those factors which lie
within the test itself) which affect the reliability are :
i. Length of the test : Reliability has a definite relation with the length of
32
the test. The more the number of items the test contains, the greater will
Criteria of a Good Tool
be its reliability and vice-versa. Therefore adequate number of items needs
to be included in the test.
ii. Homogeneity of the items : Homogeneity of items has two aspects :
item reliability and the homogeneity of traits measured from one item to
another. If the items measure different functions and the inter-correlation
of items is ‘zero’ or near to it, then the reliability is ‘zero’ or very low
or vice-versa.
iii. Difficulty value of the items : Broadly, items having indices of difficulty
at 0.5 or close to it yield higher reliability than items of extreme indices
of difficulty. It emphasizes that the items should be of the average difficulty
values. Too much difficulty and too much easy items are usually avoided
in the test.
iv. Discriminative value : When items can discriminate well between superior
and inferior, the item total correlation is high, then the reliability is also
likely to be high and vice-versa. It implies that the item has the quality
to discriminate the lower and the higher group.
v. Scorer reliability : Scorer reliability, otherwise known as reader reliability,
also affects the reliability of a test. Scorer reliability speaks of how closely
two or more scores agree in scoring the same set of responses. The
reliability is likely to be lowered if they do not agree.
(Source : Factors Influencing Reliability, ES-333, IGNOU, 2010)
6.4.2 Validity
In the preceding sub-section you have studied about the reliability. Let us now
discuss about validity. Validity tells us about the accuracy and truthfulness of
a test. The accuracy of a test can be said if and when the test measures the 33
Techniques and Tools of the purpose it was constructed. Cronbach (1970) defines validity as ‘validity
Assessment and Evaluation
is the extent to which a test measures what it purports to measure’. Further
Freeman (1965) defines validity as ‘an index of validity shows the degree to
which a test measures what it purports to measure when compared with accepted
criterion.’
From the above three definitions given we can say that validity talks not only
about consistency but also accuracy and truthfulness of the test results. This
can also be said that, validity is above reliability and it includes reliability also.
See Figure 6.1 given below :
Total 40 20 20 20 100
iv. Factors in students’ response : There are certain personal factors which
influence the students’ response to the test situation and invalidate the test
interpretation. Students’ emotional disturbance, lack of motivation, test
anxiety, etc. may affect the validity.
v. Nature of the group and the criterion : Factors such as age, sex,
ability level, educational and cultural background of the students influence
the test measures. Therefore, the nature of the validation group should
find a mention in the test manual. The nature of the criterion used is another
important consideration while evaluating validity coefficient. For example,
scores on a scientific aptitude test are likely to provide a more accurate
prediction of achievement in an ‘environmental studies’ course. Other things
being equal, the greater the similarity between the performance measured
by the test and the performance presented in the criterion, the larger the
validity coefficient.
39
Techniques and Tools of
Assessment and Evaluation Check Your Progress 3
Note : a) Write your answers in the space given below.
b) Compare your answers with those given at the end of the
unit.
6. Define validity in your own words.
.........................................................................................................
.........................................................................................................
.........................................................................................................
7. Explain predictive validity with an example.
.........................................................................................................
.........................................................................................................
.........................................................................................................
8. State any four important factors that influence validity of a test.
.........................................................................................................
.........................................................................................................
.........................................................................................................
6.4.3 Usability
Usability refers how successfully you, as a teacher, use the test in a classroom
situation. It has been observed that, many highly valid tests lack the quality
of usability. The user fails to understand or feels it difficult to use the test.
Therefore, a good test should have the quality of usability. While selecting an
evaluation tool, you should look for certain practical considerations like easy
for administration and scoring, easy for interpretation, availability of comparable
forms and cost of testing. All these considerations induce a teacher to use tools
of evaluation and such practical considerations are referred to as the ‘usability’
of a tool of evaluation. In other words, usability means the degree to which
the tool of evaluation can be successfully used by the teacher and school
administrators. So usability of a test includes comprehensibility, easy for
administration and scoring, easy for interpretation, appearance of the test,
economy and availability of the test for use.
6.4.4 Objectivity
Objectivity is another important feature for a good test. Objectivity of a test
refers to two aspects of the test, viz;
Item-objectivity, and
Scoring-objectivity.
Item-objectivity means the item is having only one right answer. Many a times,
you might have observed that a single item has two or more related answers.
That affects the validity of the test. Apart from these, ambiguous questions, lack
of proper direction, double-barreled questions, questions with double negatives,
essay type questions, etc. do not have objectivity. So much care has to be
40
exercised while framing the questions.
Criteria of a Good Tool
Scoring-objectivity refers the test paper would fetch the same score. Scoring
objectivity can be ensured by carefully maintaining the item-objectivity. In
objective-type items, it is easy to ensure scoring-objectivity whereas in subjective
item, certain precautions needs to be taken to ensure scoring-objectivity such
as carefully phrasing the essay items, making proper directions of scoring, making
the items short-answer type instead of essay type item, etc.
6.4.5 Norm
Generally norm is considered as a standard, but technically, there is difference
between the concepts of norm and standard. Norm can be defined as the average
or standard score on a particular test made by a specified population.’ Thorndike
and Hegen (1977) defines norm as ‘average performance on a particular test
made by a standardized sample.’ Determining norm is one of the important criteria
of a good test. Most standardized tests determine norm. Norm can be
characterized as follows :
It acts as a basis for interpreting test scores and minimize interpretive error
of the test.
It helps to transform the raw scores into standard scores or derived scores
and put meaning to it.
Norm suggests a level and therefore the individual departure from the level
can be evaluated in quantitative term.
Norms are necessary for the purpose of promotion, gradation, selection
and classification of examinees.
It refers to the average performance on a particular test made by
standardized sample or specified population.
The procedure of deterring the norm is a challenging task. Without determining
the norm of a test we can not say that the test is standardized. It is therefore,
in a standardized test, usually the age, level, habitation, etc. are mentioned on
which the test can be used and it also equally reflects in the interpretation of
the test scores. For interpretation of scores of educational and psychological
tests, different norms like age, grade, percentile, standard score, etc. are broadly
employed.
42
Criteria of a Good Tool
6.7 ANSWERS TO CHECK YOUR PROGRESS
1. Test prepared by the class-teacher for formative assessment of the students
to whom he/she teaches which is not standardize is called as self-made
or teacher-made test.
2. Standardized tool is prepared for the large group. The complete procedures
for test construction/standardization are followed to standardize a test. The
reliability, validity, norm, etc. are usually determined to standardize a tool.
3. Reliability refers to consistency. A test which shows a consistent result in
its frequent uses in different situations and places is called reliability of the
test.
4. Rational equivalence method is the best method for editermining reliability
of a test as statistifically most of the errors are minimised in this method.
5. (a) In case number of items in the test is less; (b) If items are not
homogeneous; (c) If memory and carry-over effect works; (d) Error in
scoring the test, and (e) In case items fail to discriminate the higher and
lower group students.
6. Validity means accuracy and truthfulness of the test and which satisfies the
very purpose of the test.
7. Predictive validity indicates the effectiveness of the test in forecasting or
predicting future outcomes in a specific area.
8. Self exercise.
9. Item-objectivity means the item is having only one right answer and it does
not carry ambiguous questions, lack of proper direction, double-barreled
questions, questions with double negatives, broad essay type questions, etc.
10. Comprehensibility, easy for administration, easy for scoring, easy for
interpretation, appearance of the test, economy and availability of the test.
43
Techniques and Tools of
Assessment and Evaluation
UNIT 7 TOOLS FOR ASSESSMENT
AND EVALUATION
Structure
7.1 Introduction
7.2 Objectives
7.3 Tests
7.3.1 Paper Pencil Test
7.3.2 Oral Test
7.3.3 Aptitude Test
7.3.4 Achievement Test
7.3.5 Diagnostic–Remedial Test
7.3.6 Intelligence Test
7.4 Tools
7.4.1 Rating Scales
7.4.2 Questionnaire
7.4.3 Inventories
7.4.4 Checklist
7.4.5 Interview Schedule
7.4.6 Observation Schedule
7.4.7 Anecdotal Records
7.4.8 Learners Portfolios and Rubrics
7.1 INTRODUCTION
In Unit-6, you have studied about the techniques of evaluation and criteria of
a good tool. In Unit-5 of this Block, you have studied various techniques, such
as self-report, assignment, projects, socio-metric techniques, peer assessment,
debates, school and club activities, cooperative and collaborative learning and
assessment, and seminars use for assessing students performance in scholastic
and co-scholastic areas. Units 5 and 6 of this Block will help you to understand
the discussion in this Unit.
In this Unit, we will come to know about some prevailing tests being used
in our schools to comprehend, evaluate and accomplish various aspects of
teaching and learning namely achievement, oral, intelligence, and aptitude of the
students.
These aspects not only help you to evaluate the learners and diagnose their
problems but also assess your own teaching too, so that it can be improved.
The most common tools that are used in the schools such as rating scale,
questionnaire, inventories, checklists, interview and observation schedules,
(Note : Some sections of this Unit have been partially reproduced from Course:
ES-333 ‘Educational Evaluation’, B.Ed. Programme, IGNOU, 2010)
44
Tools for Assessment
anecdotal reports, portfolio and rubrics are discussed in the Unit. This will help and Evaluation
you to do an effective and enjoyable teaching and learning.
7.2 OBJECTIVES
After going through this Unit, you should be able to:
discuss the concept of tests and tools,
differentiate between paper – pencil and oral tests,
explain the concept of aptitude tests, their types, uses and limitations,
discuss and administer the aptitude, achievement and diagnostic tests,
explain the concept of intelligence test, its types, uses and limitations,
explain a rating scale, its types, uses and limitations,
explain the process of developing and use of questionnaire, inventory,
checklist, interview and observation schedule,
discuss the use of anecdotal records, and
discuss the use of portfolios and rubrics.
7.3 TESTS
Tests may be of different forms, such as psychological test used to measure
mental and behavioural traits, achievement test to assess performances of
students, etc. They can be used to assess both the scholastic and co-scholastic
abilities of the students. Let us now understand the concept of the test. The
tests are those instruments by which you, as a teacher, collect information as
data through verbal and non-verbal responses of the students. A concise
definition may be : a test is an instrument or systematic procedure for measuring
a sample of behaviour (Gronlund, 1990). Further, a psychological test is defined
as a standardized, repeatable procedures used to elicit and measure samples
of human behaviour’ (Kazdin, 2000). From the above two definitions, you can
summarize the meaning of a test and the sample behaviour it measures as follows :
Adjustment and mental health, which involves detecting signs and symptoms
of psychological and neurological disorders and appraising the effectiveness
of psychological functioning.
A test can be used for two broad objectives. First, it attempts to compare
the same student on two or more than two aspects of a trait, such as
attitude and aptitude of the same student; and second, two or more than two
students may be compared on the same trait like attitude of two students (Singh,
2002). 45
Techniques and Tools of The tests which you are going to study in this Unit are classified based on
Assessment and Evaluation
the various objectives or purposes to be fulfilled which are not only related
to the learner but also useful to the teacher and teaching-learning process. Some
tests have been discussed in the subsequent sub-sections.
These paper pencil tests are easy to administer and score. They are economic
and have standard questions which provide uniform procedure for all the students.
These tests provide standard scores to the teachers.
I.Q. =
Chronological age : The exact age of the individual whose intelligence will
be measured. This can be calculated counting the age from the date of birth
of an individual. Chronological age of a child never varies.
Mental age : Mental age is calculated by comparing a child’s obtained score
from the intelligence test with the average score at his/her age level. For example,
if average score in an intelligence test for six years age group children is 70,
and a child of this age group scored 70, then it will be considered that the
mental age of that child is 7. In case, the same child scored the average score
of another intelligence test meant for the age group of eight years, then the
child’s mental age will be 8. Therefore, mental age of a child may vary from
the chronological age. In different cases it may be same as chronological age
or above or below of it.
Let us discuss it with an example :
Let the chronological age of a child is nine (CA – 9) and mental age is 10
(MA-10).
I.Q. =
=
= 111.11
To interpret the IQ score, you can refer the intelligence table as follows :
Table 7.1: IQ Reference Table
IQ score range Category
Under 70 Feeble mindedness
70-79 Border line
80-89 Dullness
90-109 Average or normal
110-119 Superior
120-140 Very superior
Over 140 Genius or almost genius
50 (Source : https://www.iqtestforfree.net/iq-scale.html, retrieved on 10.8.2017)
Tools for Assessment
As per the above example, you can say that the child having the IQ of 111.11 and Evaluation
is in superior category.
7.4 TOOLS
In the preceding section, you have studied about the various tests. In this section,
you will study about various tools including the scales, schedules and inventories.
To measure the scholastic abilities of the students, we usually use the tests such
as achievement, intelligence, aptitude, etc., but for measuring most of the co-
scholastic abilities, we use the tools such as rating scale, inventories, checklists,
schedules, questionnaires, etc. In simple language, you can understand a tool
as an instrument to measure something. In education, the word tool is used
to measure various traits of students. Now, let us discuss each tool with details. 51
Techniques and Tools of
Assessment and Evaluation
7.4.1 Rating Scales
Rating scale is one of the important tools widely used in psychology and
education. It is used for assessing attitudes of students on any situation, idea,
object, character, person or an attribute. In a rating scale, the opinions are given
in various degrees such as strongly agree, agree, and disagree; highly satisfied,
satisfied, and dissatisfied, etc. A rating scale is prepared always in odd number
points like 3-point scale, 5-point scale, 7-point scale, or 9-point scale. It is
in odd number points because, a definite middle measuring point will be possible
only when the scale is odd points. You might have observed that many persons
have the attitude to opine their opinions at the middle rating. Rating scales can
be presented in different categories. The most commonly used categories of
rating scales used in the schools are as follows :
(i) Numerical scale
(ii) Graphic scale
Let us discuss the above mentioned scales briefly.
i) Numerical scale : In this type of rating scale, numbers are assigned in
different degrees. It is therefore called numerical rating scale. In a numerical
scale a sequence of definite numbers is supplied to an observer. The assigned
numbers against the item carry certain meaning/interpretation. See the example
given below :
Question : How do you consider Sudhir’s behaviour in your class? (in this
example, Sudhir is a student)
9 Most pleasant
8 Extremely pleasant
7 Moderately pleasant
6 Mildly pleasant
5 Indifferent
4 Mildly unpleasant
3 Moderately unpleasant
2 Extremely unpleasant
1 Most unpleasant
Instead of nine-point scale you can have 3-point, 5-point or 7-point scale. In
a 5-point scale you can have the following characteristics:
5 Most pleasant
4 Moderately pleasant
3 Indifferent
2 Moderately unpleasant
1 Most unpleasant
Numerical rating scales are the easiest to construct and to apply. They are also
the simplest in terms of handling the results. However, numerical scales are often
rejected in favor of other types of scales because it is believed that they suffer
from various biases and errors.
ii) Graphic Scale : The graphic scale is a popular and most widely used
52
rating scale. In this scale, a straight line is shown vertically and horizontally
Tools for Assessment
with various clues to help the rater. The line is either segmented into units and Evaluation
or continuous. If the line is segmented then the number of units can vary
from case to case. Let us try to understand from the example given below:
Question : How effective was the teacher in the class?
1 2 3 4 5
Very Moderately Average Moderately Very
Effective Effective Ineffective Ineffective
Question : Semester system of education encourages learners to engage in their
studies regularly.
1 2 3 4 5
Strongly Agree Undecided Disagree Strongly
agree disagree
The above two examples are presented in 5-point rating scale. Scoring to this
type of items are done as per the instructions given in the test. In case it is
a positive item, the highest positive degrees of the response are given highest
points and for the negative items lower points. For example, in a positive item,
5 points can be given for ‘strongly agree’, 4 to ‘agree’, 3 to ‘undecided’, 2
to ‘disagree’, and 1 to ‘strongly disagree’ in a 5-point scale. This is vice-versa
in case of negative items.
Limitations of rating scales : The rating scales have certain limitations. Some
of them are discussed as follows :
Error of leniency: There is a constant tendency among the raters to rate
those whom they know well or in whom they are closely involved, higher
than they should. Such raters are called ‘easy rater’. Some raters become
aware of the feeling of easy rating and consequently rate individuals lower
than they should. Such raters are called ‘hard raters’. The leniency error
refers to a general and consistent tendency for a rater to rate too high
or too low for whatever reasons.
Error of central tendency: Most of the raters hesitate to rate the
individuals on the extremes of the scale, instead they follow the middle
of the scale. Hence the result gets distorted.
Halo effect: It is an error which obscures the cluster of traits within an
individual. The rater develops a general opinion about the person’s merit
and his ratings on specific traits and is greatly influenced by the impression.
It results in a false positive co-relation among the traits which are rated.
The logical error: It is due to the fact that judges are likely to give similar
ratings for traits which they feel logically related to each other.
The contrast error: It is due to a tendency of a rater to rate others
in the opposite direction (contrasting) from himself in a trait.
The proximity error: It has been seen that adjacent traits on a rating
tend to inter correlate higher than remote ones, their degree of actual
similarity being approximately equal. This error may be counteracted to
some extent by placing similar traits further apart and the different ones
close together.
(Source : ES-333, B.Ed., IGNOU, 2010) 53
Techniques and Tools of 7.4.2 Questionnaire
Assessment and Evaluation
A questionnaire is a device comprising a series of questions dealing with some
psychological, social, and educational topic(s) given to an individual or a group
of individuals, with the object of obtaining data with regard to some problems
under investigation. (Koul,1997) Questionnaire is a common tool that we, the
teachers usually apply to collect data from a situation, condition or prevailing
practice.
The questionnaire can be classified in terms of the nature of the questions
constructed to gather the data from the stakeholders. It may be closed or open
ended. Closed questionnaires are those where the respondents answer in limited
way, like responding in ‘yes’ or ‘no’; underlining the replied among the pre-
defined responses, putting the sign ‘correct’ or ‘incorrect’. Whereas in open
questionnaires the respondents are free to share, clarify and put their view.
Let us discuss certain examples of closed and open ended questions.
Examples of open-ended questions :
Describe any one of the wars fought for independence in India.
Write an essay on Quit India Movement.
Examples of closed-ended questions :
In which year Quit India Movement took place?
a) 1941 b) 1942 c) 1943 d) 1944
According to other classification, questionnaire may be Structured or Non-
Structured. A structured questionnaire contains definite, concrete and directed
questions, whereas non-structured questionnaires are often used as the interview
guide which is non-directive. It is an important instrument being used to gather
information from widely scattered sources and also used when factual information
is desired. (IGNOU, 2007)
Advantages of questionnaires : Some of the advantages of questionnaires
are enumerated as follows :
The responses are gathered in a standardized way.
Questionnaires are more objective, certainly more so than interviews.
Relatively quick to collect information.
Potential information can be collected from a large portion of a group.
Limitations of questionnaires : Although the questionnaire is one of the widely
used tools but it has also certain limitations which can be enumerated as follows :
Not easy to be used with children or illiterates.
Respondents may not agree to respond in writing.
Sometimes it is difficult to construct questions on complex and crucial topics.
Respondent interprets the questions from his/her angle of understanding.
External factors may affect the response.
7.4.3 Inventories
An inventory is constructed in the form of a questionnaire. But the inventories
are more exhaustive than questionnaire. Inventories have been mostly used for
measuring personality traits, interests, values and adjustments i.e. for assessing
self-reporting affective domain of behaviour. It consists of a series of questions
54
or statements to which the subject responds to by answering ‘Yes’ or ‘No’,
Tools for Assessment
or ‘Agree’ or ‘Disagree’. This can also be answered in some similar ways to and Evaluation
indicate preferences or to make those items that describes the subject’s typical
behaviour.
In the inventory, the statements are put in first person, For example, “I think
I am comparatively more tense than others”. In the questionnaire, there is a
question in a second person, for example “Do you think you are more tense
than other persons around you?”
The term inventory is also used in another sense. That is, an inventory of various
items is available in an office room or inventory of clothes to be given to the
washer man, etc, but in the educational evaluation, in a school and classroom
situation, inventories are used for the aforementioned concepts.
Guidelines for the administration of the inventory are as follows
The teacher should explain printed instructions very clearly to the learners.
The teachers should make it clear to the learners’ that the data will be
kept confidential.
The teacher should remove doubts, if any, regarding the manner of filling
the inventory or the questionnaire.
The teacher should take all other timely precautions for preparing a state
of mind conducive to response.
7.4.4 Checklist
According to Koul (1997), a checklist is a simple device consisting of a prepared
list of items which are prepared by the teacher to be relevant to the problem
being studied. After each item a space is provided for the observer to indicate
the presence or absence of the item by checking ‘yes’ or ‘no’, or a type of
number of items may be indicated by inserting the appropriate word or number.
The checklist is a systematic and quick way to gather data of the relevant factors
and take actions accordingly. A very simple example is being given below :
For example :
Is the school building fire proof? Yes/No
Does the school follow rain harvesting policy? Yes/No
Is the school building earthquake proof? Yes/No
It is noteworthy that the responses collected on the checklist are as a matter
of fact not any judgment. It is a good tool in gathering facts for educational
survey, checking your school library, laboratory, game facilities, school buildings,
textbooks, etc. It may also be used to check the availability of other facilities
in your school.
Checklists are also applied to classroom instructional activities such as studying
the working habits of students, supervising classroom instructions, teacher –pupil
relationships, etc.
7.4.5 Interview Schedule
Interview is a communication or conversation by which a person asks interviewer
and interviewee responds verbally in the face-to-face situation. An interview can
also be conducted through skype electronic media. It can be conducted through
telephonic conversation or through by using internet. It is a common technique
for collecting required information about an individual. You have learnt about
interview technique in Unit-5 of this Block. The interview schedule is a tool 55
Techniques and Tools of with the help of which the interview is conducted.
Assessment and Evaluation
Interview schedule can be classified according to the purpose for which it is
structured and used. If it is to resolve the research hypothesis, it is a research
interview schedule and if it is for clinical purpose, it a clinical interview schedule.
On the bases of the structure, the interview schedules are categorized as
structured or unstructured.
A structured interview schedule is one in which the procedure to be
followed is standardized and it is determined in advance of the interview.
The same types of questions are presented to the interviewee and the
wordings of the instructions to the interviewee are specified.
In the unstructured interview schedule, the series of questions are decided
in advance but the interviewer is largely free to reorganize the questions
and timing to attain the objective of the interview.
7.4.6 Observation Schedule
Observation has been the first practice of assessment that we do in our
classroom. Each observed incident, expression and reaction has useful data for
teachers, hence the observation is an effective tool for us. Observation is the
process in which one or more persons observe what is happening in a real-
life situation, and he/she classifies and records pertinent happening according
to pre-planned scheme. It is used to evaluate the overt behaviour use uniformly
either American or British English i.e., spellings of words, etc. of individuals
in both the controlled and uncontrolled situations. (Koul, 1997, p.170)
The observation schedules are the enumerations, listing of the facts or other
data that are observed under observation process. Like questionnaires, the
observation schedules are also classified as structured or unstructured. This can
also be classified as participant and non-participant observation. You have already
studied various types of observation techniques in Unit-5 of this Block. In case
the things to be observed are properly defined, the style of recording the
observed information, standardized conditions of observation and the selection
of pertinent data of observation then it is a structured observations (Kothari
& Garg, 2014, p.91) and the schedule is a structured observation schedule.
On the other hand when these are not pre-determined aspects of observation,
such observation schedules are unstructured schedules.
Merits of observation : The following are some of the merits of observation
Through observation the data are gathered from a natural setup.
The data observed are direct or first hand.
As the data are first hand or direct, while doing observation we can correlate
what is being said and what is being shown.
Does not rely upon people’s willingness or ability to provide information.
Demerits of observation : The observation has also the following demerits :
The observation if not done with planning then it will not bring out authentic
information.
Observer’s biases may affect the result.
There may come ‘Hawthorne effect’, that is when the person gets to know
that he/she is being observed then the real problems may not be shared
or shown.
Does not develop the understanding why people behave in a particular
manner.
56
Tools for Assessment
Table 7.2: Strengths and weaknesses of various tools and Evaluation
Tool Strengths Weaknesses
Tests* Economical Norms may be inappropriate
Standard Questions Standardized tests may be too
Commercial tests Strong in broad and general
Technical Qualities Standard Score may distort
Objective Tests Easy to differences
Score Standardized tests may give
Standardized tests provide false sense of validity
Uniform Procedure for all Locally developed tasks often
subjects and standards technically weak test anxiety
Restricted to subjects who can
read and write
Scales Consumes less time to Consumes time to construct
administer Items if not made valid data can
Interesting to rater be misleading
Can be used by little More than three point scale are
training difficult for some raters to
Wide range of Application answer
Question- Economical Response rate of mailed
naires* Can be anonymous questionnaire
Standard Questions & Inability to probe and clarify
Uniform Procedure Scoring open ended items
Usually easy to score Faking and social desirability
Provides time for subject to Restricted to subjects who can
think about responses read or write
Biased or ambiguous terms
Response set
Activity 1
Cite an example of an activity that you would like to conduct among
the students to whom you teach and prepare a rubric for its assessment.
Also specify the scoring procedure.
.........................................................................................................
.........................................................................................................
......................................................................................................... 61
Techniques and Tools of
Assessment and Evaluation 7.5 LET US SUM UP
In this Unit, you have studied about the tests and tools which can be used
in schools easily. We discussed the tests, their types such as paper-pencil tests,
oral tests, etc. We also discussed about commonly used tests in the school
system such as aptitude tests, achievement tests, diagnostic-remedial tests and
intelligence tests. You have also studied when to use the aptitude test and how
the achievement tests have become the most used tests in our schools.
You also studied about the different tools like rating scales, questionnaires,
inventories, checklists, interview and observation schedules, anecdotal records,
portfolios and rubrics and their uses and limitations.
The contents and discussions covered in this Unit will help you to conceptualise
and understand Unit-9, 10 and 11 presented in Block-3 of the same Course.
The details about developing achievement test, diagnostic remedial test, and
commonly used tests in schools have been elaborately discussed in Block-3
of this Course.
63
Techniques and Tools of
Assessment and Evaluation UNIT 8 ICT BASED ASSESSMENT
AND EVALUATION
Structure
8.1 Introduction
8.2 Objectives
8.3 Importance of ICT in Assessment and Evaluation
8.4 Use of ICT in Various Types of Assessment and Evaluation
8.5 Role of Teacher in Technology Enabled Assessment and Evaluation
8.6 Online and E-examination
8.7 Learners’ E-portfolio and E-rubrics
8.8 Use of ICT Tools for Preparing Tests and Analysing Results
8.9 Let Us Sum Up
8.10 References and Suggested Readings
8.11 Answers to Check Your Progress
8.1 INTRODUCTION
In Units-5 and 7 of Block-2 of this Course, we have discussed various tools
and techniques used for assessing learning of children. Many of them such as
achievement, aptitude, and intelligence tests are used in the day-to-day classroom
situations. Nevertheless, as you know, the constructivist approach in the teaching–
learning processes, facilitate teachers to apply innovative information and
communication technologies (ICTs) to assess children. For the last few decades,
ICT has become an important component of the classroom activities. Teachers
gather information via internet and other ICT resources for utilizing it in the
teaching and learning process. Children are also using computer and other
electronic devices for self-learning and also acquiring more knowledge with the
existing curriculum they study. It is under this context; you need to be oriented
and equipped in utilizing ICT for evaluating student performance. Various ICT
tools are available that help teachers to assess the performance of students.
In this Unit we will discuss various aspects such as importance of ICT in
evaluation, role of teacher in ICT based evaluation, few forms of ICT techniques
used to asses children’s progress and use of ICT in preparing and analyzing
evaluation tools.
8.2 OBJECTIVES
After going though this Unit, you should be able to:
discuss the role of ICT in assessment and evaluation,
describe the use of ICT in various types of evaluation such as online and
e-examination,
64
discuss the of use of different tools to create online/e-exams, ICT Based Assessment
and Evaluation
discuss e-portfolio and e-rubrics as types of evaluation,
describe the use of different tools to create online/e-exams, and
identify ICT tools to prepare tests and analyse results.
66
Figure 8.1: Types of Evaluation ICT Based Assessment
and Evaluation
evaluation
is conducted at the
end of an instruction
will
(Sources: https://moodle.org/.,https://www.edmodo.com/.,http://
anz.blackboard.com/.,https://classroom.google.com/h)
Activity 1
How will you use technology in formative evaluation? Suggest your
plan?
.........................................................................................................
.........................................................................................................
.........................................................................................................
......................................................................................................... 67
Techniques and Tools of
Assessment and Evaluation Check Your Progress 2
Note: (a) Write your answers in the space given below.
(b) Compare your answers with the one given at the end of
this Unit.
2. Differentiate between diagnostic, formative and summative evaluation.
.........................................................................................................
.........................................................................................................
.........................................................................................................
3. How will you use technology for summative evaluation? Suggest a plan.
.........................................................................................................
.........................................................................................................
.........................................................................................................
Coming to technology based assessment, let us see what you have done. First,
you would have listed the technologies and various applications like computer,
online testing, rubric, computer-based testing, learning management systems,
applications for developing quizzes and multiple choice questions (MCQs), etc.
From the list, you may select computer testing and applications used to develop
MCQs. In the process, you may display the below given figure and ask children
to identify the ball that would make bigger splash. For that, a MCQ was prepared
(may use Hot Potatoes, application for creating MCQ’s) as given below:
Q.1: Which ball will make bigger splash?
a) Small
b) Big
c) Both
Activity 2
Prepare a teaching module employing TPACK framework. Discuss
the factors that you have considered in selecting the technologies
while preparing the teaching module.
........................................................................................................
........................................................................................................
........................................................................................................
........................................................................................................
........................................................................................................
Source:https://www.classmarker.com/Source:https://www.onlineexambuilder.com 71
Techniques and Tools of
Assessment and Evaluation
Source:http://speedexam.net/Source: https://exambuilder.com/
Activity 3
Select any free online or e-exam creator website/software and prepare
an e-exam for the children of your class. Try them in your class
and identify the difficulties you have faced in developing and
administering it.
.......................................................................................................
.......................................................................................................
.......................................................................................................
.......................................................................................................
.......................................................................................................
(Source: http://www.informationweek.com/software/7-ways-to-create-e-
portfolios/d/d-id/1110673)
Activity 4
Select a free e- portfolio creator (even google apps) and develop
one e-portfolio suitable for children of your class. Describe the
process that you have carried out in developing e-portfolio.
........................................................................................................
........................................................................................................
........................................................................................................
77
Techniques and Tools of Types of e-portfolios: As discussed in preceding sub-sections, e-portfolios are
Assessment and Evaluation
also known as digital portfolios, web portfolios and online portfolios. Basically
the nomenclature is based on the format of portfolios. Generally, online portfolio
functions with the help of internet while other formats are internet independent.
Based on the functions of e-portfolios, they are classified into three; developmental
(or working), assessment and showcase.
Developmental (working) e-portfolios: These e-portfolios collect and
record students’ activities. For example, students are assigned a project
of two months as a learning activity. During the period of the project, they
have to submit short reports every 25 days in MS Word format. Such
a digital report will be stored in the database of e-portfolio. Thus
developmental e-portfolios record activities over a period of time. The
primary objective of recording the progress is to communicate the progress
to the students and the teachers.
Assessment e-portfolios: These e-portfolios collect and record children’s
activities for the entire duration of the programme/course. The primary
objective is to assess and judge the competency of children in specific
skills/ courses and programmes. These portfolios are submitted for assessing
and evaluating the performance in a learning activity/programme. For
example, digital formats of assignments in a single subject submitted during
his/her studies in eighth class.
Showcase e-portfolios: These portfolios exhibit exceptional works/skills
of an individual. This may include works in preparing reports, inserting
graphic/pictures in preparing write-ups, development of quality video/audio
clips and so on. These portfolios happen to judge one’s performance,
selection to a job, etc. Thus, as the name indicates, showcase e-portfolios
are images on individual persona.
Holding
Thermometer
Measuring
temperature
Recording
Reading
http://rubistar.4teachers.org
http://www.assessmentfocus.com
http://www.rcampus.com
http://www.teach-nology.com
http://www.learner.org
http://www.essaytagger.com
http://www.edudemic.com
http://edtechteacher.org
79
Techniques and Tools of
Assessment and Evaluation
(Source: http://rubistar.4teachers.org.,http://www.assessmentfocus.com/.,
80 http://www.rcampus.com/., http://www.teach-nology.com)
Importance of e-rubric : We have discussed the basic purpose of building ICT Based Assessment
and Evaluation
of e-rubric. Basically, e-rubric helps us evaluating a students’ performance. Let
us elaborate its importance in the teaching learning and evaluation process. In
a teacher’s point of view, e- rubric is a tool to set his learning objectives. While
deciding on the topic or activity to be transacted, he/she should think of the
learning objectives and skills to be developed among students. Accordingly
different criteria will be developed and recorded in the e-rubric template. Later,
the students will be evaluated based on the e-rubric and that will help you,
as a teacher to know the performance of the students. This will also help you
to reflect on your teaching process. In the case of students, e-rubric is an
instrument for self reflection. While teacher identifies the criterion for any
particular learning activity/task, the same could also be discussed with students.
Thus students are aware of the criteria based on which they are going to be
assessed. Thus as students proceed with the task, they would pay attention
on the criteria leading to development of learning skills as desired by the teacher.
Such a process provides scope for students’ self-reflection and to become aware
of their own learning.
Steps to Create E-Rubric
As we have discussed, e-rubric is an effective tool to assess children’s
performance. Generally teachers can prepare two types of e-rubric; one general
e-rubric and the other specific e-rubric. In general e- rubric, general activities
and its criterion will be included. For example, writing an assignment; in writing
assignment, the criterion includes, writing a good introduction, body (content),
conclusion, supporting examples, figures and so on. Irrespective of the subject,
the format followed in writing assignment remains the same. In specific e-rubric,
criterion for specific activities is included. For instance, performing experiment
in a chemistry laboratory. In the laboratory, each student is required to perform
certain activities; may be mixing chemical in a specific proportion, taking
measurements and so on. Such tasks are specific tasks and each student is
supposed to develop those specific skills. The following steps should be used
to develop an e-rubric.
i. Select the learning activity from your subject of teaching. It can be either
general or specific in nature
ii. List the criterion relating to the learning activity. At this stage, children may
also be encouraged to suggest criterion.
iii. Decide levels for each criterion ranging in different modes; can be 1 to
5, low to high, minimum to maximum, poor to high and so on. A five
point level would be reasonable for assessing children on varied tasks.
iv. Develop re-rubric template. The e-rubric template will show the criterion
and different levels for specific or general learning activity. You may also
leave a column for students to write their comments. The e-rubric can also
be shared with students.
v. Execute the learning activity and record scores against each criterion in
the e-rubric template.
vi. Asses and evaluate children’s performance based on the scoring in e-
rubric.
vii. If required, revise the e-rubric based on the feedback and performance
of children. 81
Techniques and Tools of
Assessment and Evaluation Activity 5
Identify a task of your choice and develop its criterion in a rubric
template. Carryout the task in your class and record the level of
children against each criterion.
........................................................................................................
........................................................................................................
........................................................................................................
........................................................................................................
........................................................................................................
https://www.classmarker.com/ http://www.fiftysneakers.com/
http://www.contentgenerator.net/
http://www.easytestmaker.com/ http://funnelbrain.com/
http://elearningbrothers.com/
http://www.ecomscotland.com/ http://web.uvic.ca/hrd/hotpot/
http://www.exambuddy.com/
https://www.igneon.com/ http://www.kubbu.com/
http://www.mygradebook.com/
http://www.cram.com/
https://www.onlineexambuilder.com/
http://www.proprofs.com/quiz -school/ https://quizslides.com/
http://www.questiontools.com/
http://www.questionwriter.com/ https://www.studyblue.com/
https://www.quia.com/web
http://www.quiz-builder.com/ http://www.btl.com/surpass/
http://ww17.quizlet.cm/
http://www.respondus.com/ http://www.schoolsonic.com/
http://www.helpteaching.com/
http://www.qarbon.com/ http://www.learnclick.com/
http://eng.smartlite.it/
https://yacapaca.com/ http://www.mobilestudy.org/
83
Techniques and Tools of
Assessment and Evaluation
Thereafter, you may insert questions and it may be distributed to your students
for answering. After attempting the test, the system itself will evaluate and test
results will be displayed. Closer screen shots are given below:
85
Techniques and Tools of
Assessment and Evaluation
is
Activity 6
89