Professional Documents
Culture Documents
-Item analysis is a method that is used in education to evaluate test item. This can ensure that questions
are in appropriate standard and measure the effectiveness of individual test item. Item analysis is
purposed to improve test items and identify unfair or biased item.
- Quantitative item analysis meant to know the quantitative statics as regards to the test items
and item concerned and is based on the following three numerical indicators which are: (Item
difficulty index. Item discrimination index. Distractor power.)
Difficulty Index
- Refers to the proportion of the of the number of students in the upper and lower groups who
answered an item correctly and ranges from 0.0 – 1.0. (Determine who are only correct and the
difficulty of the test)
- Formula in solving Index Difficulty
Ru – Number of students in the upper group who answer the items correctly.
Ri – Number of students from the lower group who answered the items correctly.
Discrimination Index
- Refers to the power of the item to discriminate the students between those who scored high
and low in the over all text. (Determine who are correct yet to identify whether the items were
effective)
- Types of discrimination Index: Positive Discrimination, Negative Discrimination, Zero
Discrimination
- Positive Discrimination- Happens when more students in the upper group got the items
correctly than those students in the lower group
- Negative Discrimination - Occurs when more students in the lower group got the items
correctly than the upper group.
- Zero discrimination -is when a number of students in the upper group and lower group who
answers the test correctly are equal.
- Formula in solving Discrimination Index
Ru – Number of students in the upper group who answer the items correctly.
Ri – Number of students from the lower group who answered the items correctly.
Distractor analysis
-term used for incorrect options in the multiple-choice types of tests while the correct answer
represents the key.
-The distractor analysis is an important component of an item, as they show a relationship between the
total test score and the distractor chosen by the student. Distractor efficiency is one such tool that tells
whether the item was well constructed or fail to perform its purpose. It allows us to examine how many
students in the upper-class group and the lower-class group selected each option on a multiple-choice
item.
- take note: Any distractor that has been selected by less than 5% of the student is considered to be a
non-functional distractor (NF-D).
- Qualitative Item Analysis According to Cohen, Swerdlik, and Smith (1992) as cited by Zurawski,
students who took the examination are asked to express verbally their experience in answering each
item in the examination.
-How to Improve the Test Items Consider the following examples in analyzing the test item and some
notes on how to improve the item based from the results of item analysis.
Factors why students failed to get the correct answer in the given question:
- The test item is a potential mis key if there are more students from the upper group who choose
the incorrect options than the key.
Guessing item
- Students from the upper group have equal spread of choices among the given alternatives. The
students from the upper group guess their answers because of the following reasons:
• the content of the test is not discussed in the class or in the test.
• The item is very difficult.
• The question is trivial.
Ambiguous item
- This happens when more students from the upper group choose equally an incorrect option
and the keyed answer.
Reliability
There are different factors that affect the reliability of a measure. The reliability of a measure can be
high or low, depending on the following factors:
-The more the items a test has, the likelihood of reliability is high. The probability of obtaining consistent
scores is high because of the large pool items.
2. Individual differences of participants
-Every participant possesses characteristics that affects their performance in a test, such as fatigue,
concentration, innate ability, perseverance, and motivation. These individual factors change over time
and affect the consistency of the answers in the test.
3. External environment
- it includes room temperature, noise level, depth of instruction, exposure to materials, and quality of
instruction, which could affect the changes in the responses of examinees in a test.
There are different ways in determining the reliability of a test. The specific kind of reliability of
reliability will depend on (1) Variable that you are measuring, (2) type of test, and (3) number of versions
of the test.
1. Test-retest You have a test and you need to Correlate the test scores from
administer it at one time to a the first and the next
group of examinees. Administer Administration. Significant and
it again at another time to the positive correlation indicates
''same group'' of examinees. that the test has temporal
There is a time interval of not stability over time. Correlation
more than 6 months between refers to a statistical procedure
the first and second where linear relationship is
administration of the tests that expected for two variables. You
measure the stable may use Pearson product
characteristics, such as the moment correlation or Pearson r
standardized aptitude test. The because the test data are usually
post test can be given with a in an interval scale (refer to a
minimum time interval of 30 statistics book for Pearson r).
minutes. The responses in the
test should more or less be the
same across the two points in
time.
1. Linear Regression
- It is demonstrated when you have two variables that are measured, such as two sets of scores
taken at two different times by the same participants. When the two scores are plotted in a
graph (with x-axis and Y-axis), they tend to form a straight line. The straight line formed from
the sets of scores can produce a linear regression. When a straight line is formed, we can tell
that there is correlation between the two sets of scores. This can be seen in the graph shown.
The correlation is shown in the graph given. The graph is called a scatterplot. Each point in the
scatterplot is correspondent with two score (one for each test.)
-when the value of the correlation coefficient is positive, it means that the higher scores in X, the higher
score in y, in the case of two spelling scores this is called a positive correlation. In the case of the two
spelling scores, a positive correlation is obtained. When the value of the correlation coefficient is
negative, it means that the higher the scores in x axis, the lower the score in the y axis, and vise versa.
This is called a negative correlation. When the same test is administered to the same group of
participants, usually a positive correlation indicates reliability or consistency of the scores.
-The strength of the correlation also indicates the strength of the reliability of the test. This indicated by
the value of the correlation coefficient. The closer the value to 1.00 or -1.00, the strongest is the
correlation.
Value Meaning
0.80-1.00 Very strong relationship
0.6-0.79 Strong relationship
0.40-0.59 Substantial/marked relationship
0.2-0.39 Weak relationship
0.00-0.19 Negligible relationship
The correlation obtained between two variables could be due to chances. In order to determine if the
correlation is free of certain errors, it is tested with significance. When a correlation is significant, it
means that the probability of the two variables being related is free of certain errors.
Example:
A Kendall’s w coefficient value of 0.38 indicates the agreement of the raters in the 5 demonstrations.
There is moderate concordance among the three raters because the value is far from 1.00.
-fin-
Item analysis:https://thejoyoflanguageassessment.wordpress.com/2012/12/19/item-analysis-3/
#:~:text=Item%20analysis%20is%20a%20method%20that%20is%20used,test%20items%20and
%20identify%20unfair%20or%20biased%20item.