Professional Documents
Culture Documents
Measurement
process of collecting data on attribute of interest. Measurement is an act or
process that involves the assignment of numerical values to whatever is being
tested. So it involves the quantity of something
is the term used to describe the assignment of a number to a given assessment.
The number can be a raw score or a score based on a normal distribution curve.
The process of quantifying this number is separate from using this information to
evaluate student outcomes and achievement.
In the classroom, to determine a child’s performance, you need to obtain
quantitative measures on the individual scores of the child. If the child scores 80
in Mathematics, there is no other interpretation you should give it. You cannot
say he has passed or failed.
Assessment
is a fact finding activity that describes conditions that exists at a particular time.
Assessment often involves measurement to gather data. However, it is the
domain of assessment to organize the measurement data into interpretable forms
on a number of variables.
in educational setting may describe the progress students have made towards a
given educational goal at a point in time. However, it is not concerned with the
explanation of the underlying reasons and does not proffer recommendations for
action. Although, there may be some implied judgement as to the satisfactoriness
or otherwise of the situation.
In the classroom, assessment refers to all the processes and products which are
used to describe the nature and the extent of pupils’ learning. This also takes
cognisance of the degree of correspondence of such learning with the objectives
of instruction.
Evaluation
adds the ingredient of value judgement to assessment. It is concerned with the
application of its findings and implies some judgement of the effectiveness, social
utility or desirability of a product, process or progress in terms of carefully
defined and agreed upon objectives or values.
Evaluation often includes recommendations for constructive action. Thus,
evaluation is a qualitative measure of the prevailing situation. It calls for
evidence of effectiveness, suitability, or goodness of the programme.
KINDS OF TEST
1. Objective Test
it is a test paper and pencil test where in students’ answers can be compared
and quantified to yield a numerical score. This is because it requires
convergent and specific response.
2. Subjective Test
it is a paper and pencil test which is not easily quantified as students are given
the freedom to write their answer to a question, such as an essay test. Thus,
the answer to this type of test is divergent.
3. Individual Test
it is a test administrated to one student at a time.
4. Group Test
it is one administrated to a group of students.
5. Unstandardized Test
it is one prepared by teachers for use in the classroom, with no established
norms for scoring and interpretation of results.
6. Standardized Test
it is a test prepared by an expert or specialist. This type of test samples
behavior under uniform procedures.
Cognitive Targets
Knowledge – acquisition of facts, concepts and theories
Comprehension – understanding, involves cognition or awareness of the
interrelationship
Analysis – breaking down of a concept or idea into its components and explaining
the concept as a composition of these concepts
Synthesis – opposite of analysis, entails putting together the components in order
to summarize the concept
Evaluation and Reasoning – valuing and judgment or putting the “worth” of a
concept or principle
VALIDITY
Appropriateness, correctness, meaningfulness and usefulness of the specific
conclusions that a teacher reaches regarding the teaching-learning situation.
TYPES OF VALIDITY
RELIABILITY
Is the extent to which an experiment, test, or any measuring procedure yields the
same result on repeated trials.
Equivalency reliability – is the extent to which two items measure identical concepts at
an identical level of difficulty. It is determined by relating two sets of test scores to one
another to highlight the degree of relationship or association.
Stability reliability – is the agreement of measuring instruments over time. To determine
stability, a measure or test is repeated on the same subjects at a future date.
Internal consistency – is the extent to which tests or procedures assess the same
characteristic, skill or quality. It is a measure of the precision between the observers or of
the observers or of the measuring instruments used in a study.
Interrater reliability – is the extent to which two or more individuals agree. It is a
reliability addresses the consistency of the implementation of a rating system.
FAIRNESS
The concept that assessment should be fair covers a number of aspecits
o Student knowledge and learning targets of assessment
o Opportunity to learn
o Prerequisite knowledge and skills
o Avoiding teacher stereotype
o Avoiding bias in assessment tasks and procedures
MODE OF ASSESSMENT
1. Placement Assessment
are used to “place” students into a course, course level, or academic
program
For example, an assessment may be used to determine whether a
student is ready for Algebra I or a higher-level algebra course, such as
an honors-level course
For this reason, placement assessments are administered before a
course or program begins, and the basic intent is to match students
with appropriate learning experiences that address their distinct
learning needs
2. Diagnostic Assessment
Carried out at the end of a particular learning session or section within
a wider programme in order to establish how much has been learned.
(eg. Via an oral test, short written test or practical test)
3. Formative Assessment
Carried out during a programme of instruction, its main object being
to provide the learners with feedback on how they are doing and thus
help them to learn more effectively.
It does not normally count towards a final grade, mark or award, nor
is it normally used to determine whether the learner will be allowed to
progress to a later stage of a course.
4. Summative Assessment
Carried out at the end of a programme of instruction or section thereof
in order to establish or measure what the learner has achieved.
It differs from formative assessment in that it generally does count
towards a final grade, mark or award or is used to determine whether
the learner is allowed to make progress through the course
Traditional assessments
are “tests” taken with paper and pencil that are usually true/false, matching,
or multiple choice. These assessments are easy to grade, but only test
isolated application, facts, or memorized data at lower-level thinking skills.
Traditional assessment provides little evidence of what a language learner
actually can do with the language.
In order to evaluate what a language learner can do with the language, a
student must be evaluated using various performance tasks and assessments.
c) Product-based
is a kind of assessment where in the assessor views and scores the final product made
and not on the actual performance of making that product.
It is concern on the product alone and not on the process. It is more concern to the
outcome or the performance of the learner. It also focuses on achievement of the
learner. - Product assessment focuses on evaluating the result or outcome of a
process.
d) Portfolio Assessment
is a collection of student work that can exhibit a student's efforts, progress, and
achievements in various areas of the curriculum. A portfolio assessment can be an
examination of student-selected samples of work experiences and documents related
to outcomes being assessed, and it can address and support progress toward achieving
academic goals, including student efficacy. Portfolio assessments have been used for
large-scale assessment and accountability purpose
e) the Rubrics
are effective and efficient tools which allow for objective and consistent assessment
of a range of performances, assignments, and activities. Rubrics can help clarify your
expectations and will show students how to meet them, making students accountable
for their performance in an easy-to-follow format. The feedback that students receive
through a grading rubric can help them improve their performance on revised or
subsequent work. Rubrics can help to rationalize grades when students ask about your
method of assessment. Rubrics also allow for consistency in grading for those who
team teach the same course, for TAs assigned to the task of grading, and serve as
good documentation for accreditation purposes.
ADVERTISEMENTS:
ADVERTISEMENTS:
ADVERTISEMENTS:
ADVERTISEMENTS:
ADVERTISEMENTS:
Weightage of a topic:
ADVERTISEMENTS:
ADVERTISEMENTS:
For example:
The items in the test should be so prepared that it will cover all the
instructional objectives—Knowledge, understanding, thinking skills
and match the specific learning outcomes and subject matter
content being measured. When the items are constructed on the
basis of table of specification the items became relevant.
Example:
5. The test item must be free from technical errors and irrelevant
clues:
6. Test items should be free from racial, ethnic and sexual biasness:
But the validity and reliability of the test items to a great extent
depends upon the instructions for the test. N.E. Gronlund has
suggested that the test maker should provide clear-cut direction
about;
Clear cut instruction must be supplied to the pupils about the time
allowed for whole test. It is also better to indicate the approximate
time required for answering each item, especially in case of essay
type questions. So that the test maker should carefully judge the
amount of time taking the types of items, age and ability of the
students and the nature of the learning outcomes expected. Experts
are of the opinion that it is better to allow more time than to deprive
a slower student to answer the question.
Instructions about basis for answering:
In the rating method the answers are rated on the basis of degrees of
quality and determine the credit assigned to each answer. Thus a
scoring key helps to obtain a consistent data about the pupils’
performance. So the test maker should prepare a comprehensive
scoring procedure along with the test items.
3. The teacher should not give any hints to any student who has
asked about any item.
When the pupils do not have sufficient time to answer the test or the
students are not ready to take the test at that time they guess the
correct answer, in recognition type items.
c. Whether the item is free from irrelevant clues and other defects?
2. Select 27% test papers from highest and 27% from lowest end.
3. Keep aside the other test papers as they are not required in the
item analysis.
4. Tabulate the number of pupils in the upper and lower group who
selected each alternative for each test item. This can be done on the
back of the test paper or a separate test item card may be used (Fig.
3.1)
In our example (fig. 3.1) out of 32 students from both the groups 20
students have answered the item correctly and 30 students have
tried the item.
The item difficulty is as following:
Where RU= Students from upper group who got the answer correct.
RL= Students from lower group who got the answer correct.
T/2 = half of the total number of pupils included in the item analysis.
Upper group than the lower group. It indicates that distracter ‘D’ is
not an effective distracter. ‘E’ is a distracter which is not responded
by any one. Therefore it also needs revision. Distracter ‘A’ and ‘B’
prove to be effective as it attracts more pupils from lower group.
Once the item analysis process is over we can get a list of effective
items. Now the task is to make a file of the effective items. It can be
done with item analysis cards. The items should be arranged
according to the order of difficulty. While filing the items the
objectives and the content area that it measures must be kept in
mind. This helps in the future use of the item.