Professional Documents
Culture Documents
Prepared by:
FERNANDO R. SEQUETE JR., LPT
LEARNING OUTCOMES
At the end of the chapter, you should be able to:
1. Define the basic concepts regarding item analysis;
2. Identify the steps in improving test items;
3. Solve difficulty index and discrimination index;
4. Identify the level of difficulty of an item;
5. Perform item analysis properly and correctly;
6. Identify the item to be rejected, revised, or retained; and
7. Interpret the results of item analysis.
INTRODUCTION
•The teacher normally prepares a draft of the test.
•The draft is subjected to item analysis and validation to be useful
and functional.
•Try-out phase – the draft test will be tried out to a group of
students of similar characteristics to the test takers.
•Item analysis phase – each item will be analyzed in terms of its
ability to discriminate, and also its level of difficulty.
•Item revision phase – allow the teacher to decide whether to
revise or replace an item.
INTRODUCTION
•The final draft of the test is subjected to validation if the intent is
to make use of the test as a standard test for the particular unit
or grading period
ITEM ANALYSIS: DIFFICULTY INDEX &
DISCRIMINATION INDEX
✓There are two important characteristics of an item that will be of interest
to the teacher: (1) item difficulty and (2) discrimination index.
✓Item difficulty – the number of students who are able to answer the item
correctly divided by the total number of students.
✓The item difficulty is usually expressed in percentage.
✓A high percentage indicates an easy item/question while a low
percentage indicates a difficult item.
EXAMPLES:
1. What is the item difficulty index of an item
if 25 students are unable to answer it
correctly while 75 answered it correctly?
2. What is the item difficulty index of an item
if 25 students answered the item correctly
while 75 students did not?
❑One problem with this type of difficulty index is that it may not actually
indicate that the item is difficult or easy.
❑A student who does not know the subject matter will naturally be unable to
answer the item correctly even if the question is easy.
Here,
DU = 0.60
DL = 0.20
Action to be
Index Range Interpretation
done
-1.0 – -0.50 Can discriminate but item is questionable Discard
-0.55 – 0.45 Non-discriminating Revise
0.46 – 1. Discriminating item Include
EXAMPLE:
Consider a multiple choice type of test of which the
following data were obtained:
ANSWER:
Difficulty index = # of students getting correct response/total
= 40/80 = 50%, within range of a “good item”
6−2
𝐷= = 0.40
10
Where,
𝑟𝑜𝑡 = 𝑟𝑒𝑙𝑖𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑡ℎ𝑒 𝑜𝑟𝑖𝑔𝑖𝑛𝑎𝑙 𝑡𝑒𝑠𝑡
𝑟𝑜𝑒 = 𝑟𝑒𝑙𝑖𝑎𝑏𝑖𝑙𝑖𝑡𝑦 𝑜𝑓 𝑡ℎ𝑒 𝑐𝑜𝑟𝑟𝑒𝑙𝑎𝑡𝑖𝑜𝑛 𝑜𝑓 𝑜𝑑𝑑 𝑎𝑛𝑑 𝑒𝑣𝑒𝑛 𝑖𝑡𝑒𝑚𝑠
KR-20 & KR-21 FORMULAS
The KR-20 formula is also known as the Kuder-Richardson formula.
𝑘 𝑝𝑞
𝐾𝑅20 = 1− 2
𝑘−1 𝑠
k = number of items
p = proportion of the students who got the item correctly (index of
difficulty)
q=1–p
S2 = variance of the total score
KR-20 & KR-21 FORMULAS
𝑘 𝑥 (𝑘 − 𝑥)
𝐾𝑅21 = 1 − 2
𝑘−1 𝑘𝑠
k = number of items
x = mean value
S2 = variance of the total score
THE FOLLOWING TABLE IS A STANDARD FOLLOWED ALMOST
UNIVERSALLY IN EDUCATIONAL TEST & MEASUREMENT
Reliability Interpretation
0.90 and above Excellent reliability; at the level of the best standardized tests
0.80 – 0.90 Very good for a classroom test
Good for a classroom test; in the range of most. There are probably a few times which
0.70 – 0.80
could be improved.
Somewhat low. This test needs to be supplemented by other measures (e.g., more tests) to
0.60 – 0.70
determine grades. There are probably some items which could be improved.
Suggests need for revision of tests, unless it is quite short (ten or fewer items). The test
0.50 – 0.60
definitely needs to be supplemented by other measures (e.g., more tests) for grading.
Questionable reliability. This test should not contribute heavily to the course grade, and it
0.50 and below
needs revision