Professional Documents
Culture Documents
Item Analysis: Improving Multiple Choice Tests
Item Analysis: Improving Multiple Choice Tests
Improving Multiple
Choice Tests
Crystal Ramsay
September 27, 2011
Schreyer Institute
for Teaching Excellence
http://www.schreyerinstitute.psu.edu/
This workshop is designed to help you do three
things:
To differentiate between
well-performing items and
poor-performing items
Stem
Item Difficulty
Item Discrimination
Distracter information
Test score reliability is an index of the likelihood that
scores would remain consistent over time if the same
test was administered repeatedly to the same learners.
Reliability
coefficients range
from .00 to 1.00.
Ideal score
reliabilities are >.80.
Higher reliabilities =
less measurement
error.
6 0 0 0 5+ 0 95 E 95 -0.11
Number of
alternatives for
each item
performing item % % % % % %
8 0 0 43 0 57 0 D 57 0.46
performing item % % % % % %
6 0 0 0 5+ 0 95 E 95 -0.11
What is an ‘ideal’ item discrimination statistic
depends on 3 factors.
Item Difficulty
Test heterogeneity
Item characteristics
Item difficulty
Which items on
your exam
performed ‘well’?
Item
Item Difficulty
Discrimination
Distracters Instruction
Ultimately, it’s a
judgment call that
you have to make.
What if I have a relatively
short test or I give a test
in a small class? I might
not use the testing
service for scoring. Is Yes.
there a way I can
understand how my
items worked?
Item 1 A B* C D From: Suskie, L. (2009).
Top 1/3 10 Assessing student
learning: A common
Item 2 A* B C D
Top 1/3 8 2
Bottom 1/3 7 3
Item 3 A B C* D
Top 1/3 5 1 4
Bottom 1/3 2 4 4
Item 4 A* B C D
Top 1/3 10
Bottom 1/3 9 1
Student feedback