Professional Documents
Culture Documents
Provide learners with the means to organize their efforts and activities toward
accomplishing the intent of instruction .1
Statistics are used in the field of education for the following purposes : 123
To study the underlying factors and events that affect the behavioural outcome of
students . 1
To spot trends in student performance using data visualizations . 2
The testing effect, also known as retrieval practice, active recall, practice testing, or
test-enhanced learning, suggests that long-term memory is increased when part
of the learning period is devoted to retrieving information from memory . The 1
APA Dictionary of Psychology defines the testing effect as “the finding that taking a
test on previously studied material leads to better retention than does restudying
that material for an equivalent amount of time” . Testing provides feedback on the
2
effectiveness of the learning and studying processes, identifies areas of strength and
weakness, and helps to promote deeper encoding for faster, more accurate recall . 3
Data. The;
Nominal or categorical: used to measure variables that can be broken down into
groups with distinct attributes, such as gender, race, or religion.
Ordinal: used to measure variables that signify the order or rank of the different
groups, such as education level, income level, or satisfaction level.
Interval: used to measure variables that have equal intervals or distances between
the values, such as temperature, IQ score, or year.
Ratio: used to measure variables that have a meaningful zero point and can be
compared using ratios, such as height, weight, or age.
Indicate how far above or below the average (the "mean") an individual score falls,
using a common scale, such as one with an "average" of 100 . 1
Multiple Response
Short Answer
Short answer items ask a test taker to synthesize, analyse, and evaluate
information, and then to present it coherently in written form.
Matching
Build List
A build list item challenges a candidate’s ability to identify and order the
steps/tasks needed to perform a process or procedure.
There are many guidelines for creating tests (see this handy guide, for
example), but this list sticks to the most important points. Little things can really
make a difference when developing a valid and reliable exam!
1. Institute Fairness
Although you want to ensure that your items are difficult enough that not
everyone gets them correct, you never want to trick your test takers! Keeping
your wording clear and making sure your questions are direct and not
ambiguous is very important. For example, asking a question such as “What is
the most important ingredient to include when baking chocolate chip cookies?”
does not set your test taker up for success. One person may argue that sugar is
the most important, while another test taker may say that the chocolate chips
are the most necessary ingredient. A better way to ask this question would be
“What is an ingredient found in chocolate chip cookies?” or “Place the following
steps in the proper order when baking chocolate chip cookies.”
o Always make sure your correct option is 100% correct, and your
incorrect options are 100% incorrect. By using partially correct or
partially incorrect options, you will confuse your candidate. Doing
this could keep a truly qualified candidate from answering the item
correctly.
o Make sure your distractors are plausible. If you’re correct response
logically answers the question being asked, but your distractors are
made up or even silly, it will be very easy for any test taker to figure
out which option is correct. Thus, your exam will not properly
discriminate between qualified and unqualified candidates.
o Try to make your options parallel to one another. Ensuring that your
options are all worded similarly and are approximately the same
length will keep one from standing out from another, helping to
remove that testwiseness effect.
objectives document the purposes for which measurement and analysis are done and specify
the kinds of actions that can be taken based on results of data analyses . Objective
2
measurement is something that is measured consistently, without any factors that can alter
the data gathered . Measurable objectives are objectives that can be tracked with the help of
3
Distorting factors for performance appraisal: The factors that distort can be described are
as follows –
Evaluator’s marks high, while others mark low. The former is referred ‘to as positive
leniency error, Leniency error: Relative to the true or actual performance an
individual exhibits, some and the latter as negative leniency error.
Halo error: The halo error of effect is a tendency to rate high or low on all factors
due to the impression of a high or low rating on some specific factor.
Similarity error: When evaluators rate other people in the same way that the
evaluators perceive themselves, they are making a similarity error.
Low appraisal motivation: The evaluator may be reluctant to give a realistic
appraisal. The tendency is more harmful than other pitfalls.
Central tendency: It is possible that regardless of whom the appraiser evaluates and
what traits are used, the pattern of evaluation remains the same.
Inappropriate substitutes for performance: It is the usual job where the definition
of performance is absolutely clear and direct measures are available for appraising the
incumbent.
Attribution theory: According to this theory, employee evaluations are directly
affected by a supervisor’s perceptions of who is believed to be in control of the
employee’s performance – the employer or the manager.
Explain four variables that should be taken into account in any good
accountability in evaluation in education
Explain four variables that should be taken into account in any good accountability in evaluation in
education
The six levels of learning in the cognitive domain according to Benjamin Bloom are :
12345
A Table of Specifications (TOS) is a tool used to ensure that a test or assessment measures
the content and thinking skills that the test intends to measure . It can provide response
12
content and construct validity evidence . The TOS can also improve the validity of teacher-
1
made tests and student learning . Additionally, the specification can provide clear
2
1. Achieve balance in the test and to identify the achievement domains being measured.
2. Ensure that the test has a fair and representative sample of questions appeared on the test.
3. Provide evidence that a test has content validity, that it covers what should be covered.
4. Improve long-term recall for students.
5. Inform instruction or curriculum.
6. Provide the opportunity to reduce test anxiety in students and help build content mastery.
In summary, a TOS is an essential tool that can help ensure that a test or assessment
measures the content and thinking skills that the test intends to measure. It can also
improve the validity of teacher-made tests and student learning. The TOS can
provide clear instructions on project intent, performance, and construction, and can
help reduce test anxiety in students and build content mastery 1.
The table of specifications (TOS) is a tool used to ensure that a test or assessment measures
the content and thinking skills that the test intends to measure 1. It is a two-way chart that
describes the topics to be covered by a test and the number of items or points that will be
associated with each topic 2. The TOS helps the examiner to construct a test that focuses on
the key areas and weights those different areas based on their importance 2. It also provides
the examiner with evidence that a test has content validity, that it covers what should be
covered 3.
In summary, the TOS is an essential tool for examiners as it ensures that the test measures the
content and thinking skills that it intends to measure, and it provides evidence that the test has
content validity 13.
The table of specifications (TOS) is a tool used to ensure that a test or assessment
measures the content and thinking skills that the test intends to measure 1. It is a
two-way chart that describes the topics to be covered by a test and the number of
items or points that will be associated with each topic 2. The TOS helps the examiner
to construct a test that focuses on the key areas and weights those different areas
based on their importance 2. It also provides the examiner with evidence that a test
has content validity, that it covers what should be covered 3.
1. Ensures content validity: The TOS helps the examiner to ensure that the test measures the
content and thinking skills that it intends to measure 13.
2. Helps to focus on key areas: The TOS helps the examiner to construct a test that focuses on
the key areas and weights those different areas based on their importance 2.
3. Provides evidence of content validity: The TOS provides the examiner with evidence that a
test has content validity, that it covers what should be covered 3.
4. Helps to avoid bias: The TOS helps the examiner to avoid bias by ensuring that the test is
not skewed towards any particular area 1.
5. Helps to ensure fairness: The TOS helps the examiner to ensure that the test is fair to all
students by covering all the topics that should be covered 1.
6. Helps to improve test quality: The TOS helps the examiner to improve the quality of the
test by ensuring that it measures the content and thinking skills that it intends to measure 1
Construct validity: This type of validity focuses on whether a test measures what it is
supposed to measure.
Content validity: This type of validity focuses on whether a test covers all aspects of what
it is supposed to measure.
Face validity: This type of validity focuses on whether a test appears to measure what it is
supposed to measure.
Criterion-related validity: This type of validity focuses on whether a test can predict future
outcomes.
Test validity: This type of validity focuses on whether a test is measuring what it is
supposed to measure.
Essay tests are a popular form of assessment that can evaluate higher cognitive
processes. They are relatively valid assessments of learning outcomes such as
summarizing, integrating ideas, and expressing oneself in one’s own way 12. However,
essay tests are harder to score reliably and are not amenable to objective scoring as
they give scope for halo-effect, inter-examiner variability, and intra-examiner
variability in scoring 123.
One of the most well-known disadvantages of essay testing is time. Instructors must
read every essay, grading the content and composition, while making a judgment on
how thorough the answer is 4. Additionally, only a small part of the material covered
in class is reinforced on the test, which can be a disadvantage for some students 4.
In summary, essay tests are useful for evaluating higher cognitive processes, but they
are harder to score reliably and can be time-consuming for instructors. Students
must study all of the material covered in class, but only a small part of it is reinforced
on the test.
Advantages:
1. Assesses higher-order thinking: Essay tests are designed to evaluate higher-order thinking
skills such as analysis, synthesis, and evaluation 1.
2. Allows for creativity: Essay tests allow students to express themselves in their own words
and demonstrate their understanding of the material in a unique way 1.
3. Evaluates writing skills: Essay tests are an effective way to evaluate a student’s writing skills,
including grammar, spelling, and punctuation 1.
4. Measures depth of knowledge: Essay tests can measure the depth of a student’s
knowledge on a particular topic 5.
5. Encourages critical thinking: Essay tests encourage students to think critically about the
material and develop their own ideas and opinions 5.
6. Provides feedback: Essay tests provide students with feedback on their writing and
understanding of the material 5.
Disadvantages:
1. Subjectivity: Essay tests are subjective and can be influenced by the grader’s personal
biases 1.
2. Time-consuming: Essay tests are time-consuming to grade, which can be a disadvantage for
instructors who have many students 4.
3. Limited scope: Essay tests only cover a small part of the material covered in class, which can
be a disadvantage for some students 4.
4. Inter-examiner variability: Essay tests are not amenable to objective scoring and can be
influenced by inter-examiner variability 23.
5. Intra-examiner variability: Essay tests can also be influenced by intra-examiner variability,
where the same grader may assign different scores to the same essay at different times 23.
6. Halo-effect: Essay tests can be influenced by the halo-effect, where the grader assigns a
higher score to an essay based on a single positive attribute 23.
Analyse the causes of low reliable coefficient for test retest method
A low reliability coefficient for the test-retest method could be caused by several
factors, including:
1. Test taker: The subject may be having a bad day, or their mood may have changed between
the two administrations of the test.
2. Test itself: The questions on the instrument may be unclear, or the test may not be
measuring what it is intended to measure.
3. Testing conditions: There may be distractions during the testing that detract the subject, or
the testing environment may be different between the two administrations of the test 13.
It is important to note that a less-than-perfect test-retest reliability causes test-retest
variability. Such variability can be caused by, for example, intra-individual variability
and inter-observer variability 4.
Explain five factors that should be put into consideration when constructing a test
When constructing a test, there are several factors to consider. Here are some of the
most important ones:
1. Appropriateness of test items: Test items should be appropriate for the learning
outcome to be measured. They should measure all types of instructional objectives
and the whole content area. The test items should be free from ambiguity and of
appropriate difficulty level 1.
2. Validity: Validity is the degree to which a test measures what it is intended to
measure. It is the most important criterion of a good examination. The test should
measure the abilities, skills, and information it is supposed to measure 2.
3. Reliability: Reliability is the consistency of test scores across different
administrations or different ratters. A reliable test should produce consistent results
over time and across different ratters 3.
4. Objectivity: Objectivity refers to the degree to which a test is free from subjective
judgment. It is important to minimize the influence of the examiner’s personal bias
on the test results 4.
5. Practicality: Practicality refers to the ease of test administration, scoring, and
interpretation. A test should be easy to administer, score, and interpret, and should
not be too time-consuming or expensive 1.
To increase the validity of a test in the classroom, you can consider the following six
ways:
1. Identify the test purpose by setting SMART goals: Clearly define the purpose of
the test by setting Specific, Measurable, Achievable, Relevant, and Time-bound
(SMART) goals. This will help you to design a test that measures what it is intended
to measure 1.
2. Measure the right skills: Ensure that the test measures the skills that you want to
assess. This can be achieved by aligning the test with the learning objectives and
course content 1.
3. Prioritize accessibility, equity, and objectivity: Ensure that the test is accessible to
all students, regardless of their background or abilities. The test should be fair and
objective, and should not disadvantage any particular group of students 1.
4. Encourage students to perform their best: Encourage students to perform their
best by providing clear instructions, and by creating a supportive and positive
learning environment. This will help to reduce test anxiety and improve student
performance 1.
5. Match the assessment difficulty to the students’ ability levels: Ensure that the
test is neither too easy nor too difficult for the students. The test should be
challenging enough to assess the students’ knowledge and skills, but not so difficult
that it discourages them 1.
6. Have scoring criteria that are available and well understood by students before
they start the assignment: Clearly communicate the scoring criteria to the students
before they start the assignment. This will help them to understand what is expected
of them, and will reduce confusion and anxiety 1.
Advantages of marking schemes and teacher made questions
Asking questions is an important tool for teachers . Effective questioning can help
1234
teachers diagnose student understanding of material, ascertain where children are with their
learning, and reveal evidence of existing knowledge, gaps in knowledge, and
misconceptions . Good questioning reinforces and revisits learning objectives/goals, shows
2
connections between previous and new learning, and gives the teacher immediate feedback
on students’ understanding, which they can then use to modify their teaching . 3
Marking schemes have several advantages. They promote fairness and consistency in
marking assessments, develop shared understandings around assessment of formative and
summative work related to, for example, an assignment or exam answers, help judge the
quality of discursive work like an argument in an essay or an open-ended text response to an
exam question, allow you to see common errors made by candidates, improve the ways you
support pupils in their development of exam technique, enable you to more easily develop
exam style questions and create model answers, enable the marking of your pupils work to
become more focused on the requirements of the specification and mark scheme, state the
correct answer or a range of possible answers, although these may not be exhaustive, indicate
how a second mark is awarded for a second point or developed idea, and give an indication of
unacceptable answers 1.
I hope this helps!
Learn more:
1. Lo.unisa.edu.au2. bing.com3. examining.jcq.org.uk4. cambridgeinternational.org5.
cambridgeinternational.org6. papers.gceguide.com7. cambridgeinternational.org+5 more
1of30
What is a marking scheme? How do I create an effective marking scheme? Can you give me some
examples of good and bad answers to exam questions?
Response stopped
New topic
New topic
0/2000
Feedback
There are four frames of references for interpreting test scores: 1 Ability-referenced
interpretation 2 Growth-referenced interpretation 3 Norm-referenced interpretation
4 Criterion-referenced interpretation. Norm-referenced test interpretation compares
the scores that the applicant receives with the test performance of a particular
reference group. Criterion-referenced test interpretation indicates the amount of skill
or knowledge the test taker possesses in a particular subject or content area 1.
The two most common methods of interpreting test scores are norm-
referenced and criterion-referenced interpretations 23. In norm-referenced test
interpretation, the scores that the applicant receives are compared with the test
performance of a particular reference group. In this case, the reference group is the
norm group. Criterion-referenced test interpretation indicates the amount of skill or
knowledge the test taker possesses in a particular subject or content area. This
method of interpretation is based on the assumption that there is a specific set of
skills or knowledge that the test taker should possess 3.
The functions of grading and reporting are to assess and communicate students'
academic progress and achievement . Grading and reporting systems can help improve
123
affective domain moving from the lowest order to the highest :12
Receiving: Refers to the student’s willingness to attend to particular phenomena or stimuli
(classroom activities, textbook, music, etc.).
Responding: Refers to active participation on the part of the student.
Valuing: Refers to the worth or value a student attaches to a particular object,
phenomenon, or behaviour.
Organizing: Refers to the student’s ability to organize values into priorities by contrasting
different values, resolving conflicts between them, and creating a unique value system.
Characterization by Value or Value-Set: Refers to the student’s ability to internalize values
and beliefs that will guide their behaviour.
Objective tests are a popular form of testing that have several advantages. They are
objective, efficient, reliable, and comprehensive 1. Objective tests are also engaging and can
be used to test a wide range of knowledge 12. However, they have some disadvantages. One
of the main disadvantages is that objective tests can only assess a limited range of
knowledge 3. Another disadvantage is that objective tests have the potential for guessing,
which can lead to inaccurate results 1. Additionally, objective tests are unable to distinguish
between students with deep and superficial understanding 1. Despite these limitations,
objective tests are widely used in modern education systems 4.
Test and assessment are two different terms that are often used interchangeably . Here 12
Testing is the process of putting something through a series of trials to determine its
value .
32
Assessment is the process of documenting knowledge, skills, attitudes and beliefs, usually
in measurable terms . 3
Testing measures the level of skill or knowledge that has been reached .
3
Portfolios and assessments are two different ways of evaluating student learning .The 1234
Portfolios give students the opportunity to direct their own learning, while assessments
assess students across a limited range of reading and writing assignments which may not
match what students do . 2
Portfolios allow students a more creative way to showcase their experiences and
academic progress, while assessments typically involve students verbally presenting their
work and arguing persuasively what they have learned in a unit or subject .3
Assessment portfolios are used to evidence what a student has learned and achieved
against the intended learning outcomes, while portfolios are simply a selection from the
working portfolio . 4
Rubrics and rating scales are both used to evaluate student work, but they differ in the
following ways : 1234
Rating scales provide three or four response selections to describe the quality or
frequency of student work, while rubrics can be summative and often associated with
skills assessment.
Checklists ask for dichotomous decisions (typically has/doesn't have or yes/no) for each
criterion, while rating scales ask for decisions across a scale that does not describe the
performance.
Rubrics demarcate different levels of performance (typically four) in relation to a standard,
while scoring guides illustrate how students can earn a certain amount of points for
specific questions and responses.
Learning objectives are the end goal of a learning experience, while learning activities are
the means by which the goals are achieved . Learning goals describe general outcomes for
1
an entire course or program, while learning objectives are more focused outcomes for
specific lessons or activities . Activity objectives help young people understand what they
2
will be learning and why, and keep you focused when developing your activity . Unlike3
learning goals, learning objectives are more specific, defined, and measurable . 4
Learning objectives and learning activities are two different concepts that are
crucial to the learning process. Learning objectives are the desired outcomes of a
learning experience, while learning activities are the means by which the objectives
are achieved 1. Here are six differences between learning objectives and learning
activities:
1. Focus: Learning objectives are focused on the desired outcomes of a learning experience,
while learning activities are focused on the means by which the objectives are achieved 1.
2. Specificity: Learning objectives are more specific and measurable than learning activities 1 .
3. Scope: Learning objectives are focused on specific lessons or activities, while learning goals
describe general outcomes for an entire course or program 12.
4. Purpose: Activity objectives help learners understand what they will be learning and why,
and they keep instructors focused when developing their activities 3.
5. Design: Learning objectives are developed before learning activities, and they guide the
design of the activities 1.
6. Assessment: Learning objectives are used to assess whether the desired outcomes have
been achieved, while learning activities are used to facilitate the achievement of the
outcomes 1.
To ensure the validity of a teacher-made test, educators can maintain reliability by:
Describe ways in which the school administration can use the results of assessment in school
Reliability
Validity
Technology
Legal context
Face validity/applicant reactions
When selecting a test format, it is important to consider the following factors 1:
1. Reliability: The consistency of the test scores over time and across different evaluators.
2. Validity: The extent to which the test measures what it is intended to measure.
3. Technology: The availability of technology to administer the test and score the results.
4. Legal context: The legal requirements and restrictions that apply to the test.
5. Face validity/applicant reactions: The extent to which the test appears to measure what it
is intended to measure and the reactions of the test-takers to the test.
6. Objectivity: The extent to which the test is free from subjective judgment.
7. Economy: The cost-effectiveness of the test.
These factors are crucial in selecting an appropriate test format that meets the needs
of the test-takers and the test administrators.
Matching test items are a type of assessment question that present learners with two
items separated into two columns and ask them to match items from the first column
to the corresponding items in the second 1. Here are four demerits of matching test
items:
1. Matching test items are not suitable for measuring learning objectives that require more
than simple recall of information 1.
2. They are difficult to construct due to the problem of selecting a common set of stimuli and
responses 1.
3. Matching test items are time-consuming for learners to answer 2.
4. They are ill-suited for gauging the learners’ higher understanding (analysis and synthesis
levels) 3.
When constructing matching test items, it is important to take the time to construct
items that are valid and reliable. Here are some guidelines for this 3:
1. Two-part directions. Your clear directions at the start of each question need two parts: 1)
how to make the match and 2) the basis for matching the response with the premise.
2. Keep questions short and straightforward. Avoid unnecessary words.
3. Do not get carried away adding additional items. Having 10-12 items between both columns
(5-6 “question - answer” pairs) is the sweet spot.
4. It is best to arrange the items in the left column according to some criterion (alphabetically,
chronologically, etc.).
5. Make sure that no items in the right column can be matched to more than one item in the
left one. However, having an item in the left column serves as the key for more than one
item in the right column is all right.
6. Avoid positioning matching test questions in such a way that the list is separated in two by a
page break. Learners should not have to go back and forth trying to match questions on one
page to answers on the other.
7. When constructing answers, try to keep them interconnected by theme and the manner of
presentation.
Provide feedback to the students and help them improve their learning 12
There are three general types of assessments in the classroom: diagnostic, formative, and
summative .The three types of evaluation are :
1 12
Diagnostic: This type of evaluation is used to identify students' strengths and weaknesses
before instruction begins.
Formative: This type of evaluation is used to monitor student progress during instruction.
Summative: This type of evaluation is used to evaluate student learning at the end of an
instructional unit by comparing it against some standard or benchmark
Difficulty index and discrimination index are two parameters that help evaluate the
quality of test items . Difficulty index measures how easy a question is by determining
12345
the proportion of students who got it right . Discrimination index measures how well a
125
question can differentiate between high and low scoring students . The standard range
1245
for difficulty index is 0.25 to 0.75, and the standard for discrimination index is 0.20 or
higher .
45
affective domain includes objectives relating to interest, attitude, and values relating to
learning the information .The psychomotor domain focuses on motor skills and actions that
3
Test-retest reliability: giving the same test more than once to the same group of people
and comparing the results.
Parallel forms reliability: giving different types of tests to the same group of people and
comparing the results.
Inter-ratter reliability: having different ratters evaluate the same test or application and
comparing their ratings.
Internal consistency reliability: measuring how well the items or components of a test or
application are related to each other.
Objective tests are a popular form of assessment that have several advantages and
disadvantages. Here are six of them:
Advantages:
1. Objectivity: Objective tests are scored using a predetermined answer key, which reduces the
potential for subjective grading and increases the objectivity of the assessment 12.
2. Efficiency: Objective tests can be graded quickly and efficiently, which makes them a useful
tool for large-scale assessments 12.
3. Reliability: Objective tests are generally reliable, meaning that they produce consistent
results over time and across different graders 12.
4. Comprehensive coverage: Objective tests can cover a wide range of topics and can assess a
large amount of material in a relatively short amount of time 12.
5. Engagement: Objective tests can be engaging for students, as they provide immediate
feedback and can help students identify areas where they need to improve 12.
6. Standardization: Objective tests are standardized, meaning that all students receive the
same questions and are assessed using the same criteria, which helps ensure fairness and
equity in the assessment process 12.
Disadvantages:
1. Potential for guessing: Objective tests can be susceptible to guessing, as students may be
able to eliminate some answer choices or make educated guesses based on the structure of
the question 13 .
2. Inability to distinguish between deep and superficial understanding : Objective tests may
not be able to distinguish between students who have a deep understanding of the material
and those who have a superficial understanding 13 .
3. Limited assessment of higher-order thinking skills: Objective tests may not be able to
assess higher-order thinking skills, such as critical thinking, creativity, and problem-solving,
as effectively as other forms of assessment, such as essays 3.
4. Limited feedback: Objective tests may not provide detailed feedback to students, which can
limit their ability to learn from their mistakes and improve their understanding of the
material 3.
5. Limited assessment of real-world skills: Objective tests may not be able to assess real-
world skills, such as communication, collaboration, and leadership, as effectively as other
forms of assessment, such as projects or presentations 3.
6. Limited assessment of student growth: Objective tests may not be able to assess student
growth over time, as they only provide a snapshot of a student’s understanding at a
particular point in time 3.
It is important to note that objective tests are not the only form of assessment, and
other forms of assessment, such as essays, may be better suited to evaluate certain
skills or knowledge 3. Overall, objective tests can be a useful tool for evaluating
student learning, but it is important to consider their limitations and use them in
conjunction with other forms of assessment to gain a more complete picture of
student understanding.
When written well, instructional objectives will assist instructors in identifying course
content, help structure lectures, and allow them to select activities and assessments
that are relevant and meaningful for learning 3. They illustrate the skills and
applicable knowledge students will master 2.
1. Constancy of the speed of working: If the speed of working is not consistent, it can
lead to unreliable results.
2. Variability in accuracy: If the accuracy of the test-taker varies, it can lead to
unreliable results.
3. Incentive or effort: If the test-taker is not motivated or does not put in enough
effort, it can lead to unreliable results.
4. The obtrusion of competing ideas: If the test-taker is distracted by competing
ideas, it can lead to unreliable results.
5. Illness: If the test-taker is ill, it can lead to unreliable results.
6. Worry: If the test-taker is worried, it can lead to unreliable results.
7. Excitement: If the test-taker is excited, it can lead to unreliable results.
8. Accidents during examination: If accidents occur during the examination, it can
lead to unreliable results.
9. Cheating: If the test-taker cheats, it can lead to unreliable results.
Other factors that affect reliability include the length of the test, moderate item
difficulty, objectivity, heterogeneity of the students’ group, and limited time.
Assessment and evaluation are different ways of collecting and using information about
student learning 12345
. Assessment is about providing feedback and improving learning,
while evaluation is about making judgments and measuring outcomes . Assessment
125
teaching and learning strategies, while evaluation is used to guide strategic planning and
resource allocation . Assessment is based on individual performance or achievement, while
3
Assessment and evaluation are two distinct methods of collecting and utilizing
information about student learning. Assessment is the process of providing
feedback and improving learning, while evaluation is the process of making
judgments and measuring outcomes. Here are six differences between assessment
and evaluation:
together in a way not expected on the basis of chance alone . Correlation can be causal or
124
not, and can be used for various data sets . Correlation is usually measured by the degree to
45
Correlation is a statistical term that refers to the degree or measure of how two or more
variables move in coordination or in the same or opposite direction 1. It is the state or
relation of being correlated, which means varying, being associated, or occurring together in
a way not expected by chance alone 1. Correlation is a mutual relationship or connection
between two or more things, and it can be established by the process of establishing a
relationship or connection between two or more things 1. Correlation is useful because it can
indicate a predictive relationship that can be exploited in practice 2. However, the presence of
a correlation is not sufficient to infer the presence of a causal relationship 2. There are several
correlation coefficients, often denoted or, measuring the degree of correlation 2. The most
common of these is the Pearson correlation coefficient, which is sensitive only to a linear
relationship between two variables 2. Other correlation coefficients, such as Spearman’s rank
correlation, have been developed to be more robust than Pearson’s, that is, more sensitive to
nonlinear relationships 2.