Professional Documents
Culture Documents
Year and Section: BEEd II-A and II-B, BSEd II-A, II-B and II-C
Semester: Second Semester
School Year: 2020 – 2021
Guidelines to Students
1. Read the learning materials given to you or posted in your group chat (GC).
2. Work on the exercises/activities indicated at the end of the unit or chapter. Use long size bond
paper for your answer sheet.
3. Answers to the exercises/activities will be sent back through e-mail, messenger or google
classroom. These will serve as your output in this subject.
e-mail address: exalmonje@gmail.com
mariaandaradar@wvsu.edu.ph
4. For those who can’t submit their output through email or messenger, you placed it inside an
envelop properly sealed and labeled (with your name, name of your subject teacher, and name of
the subject) and submit to security guard on duty.
5. If you have queries/questions regarding the topic, just send message through messenger (GC or
private messenger) or call me in this cp number: 09306481618 -----Asso. Prof. Exaltacion Monje
09199950207------Dr. Marian Argie Daradar
6. Queries/questions regarding the topic will be entertained during your class schedule.
1. explained and discussed the nature of assessment and its relevance to learners, teachers, parents
and stakeholders;
2. compared and contrast measurement, assessment, testing and evaluation;
3. explained the various roles/purposes of assessment to learners, teachers, parents, and other
stakeholders;
4. showed appreciation on the role of assessment in their teaching learning process.
Measurement- deals with quantitative description of data; refers to the process of quantifying an
individual’s achievement, personality and attitude by means of appropriate measuring instruments.
- deals with the systematic procedure of determining the quantity or extent of all the measurable
dimensions in the educative process
Types of measurement
The issue that that test alone cannot measure the total development of an individual, makes it
imperative that the term “measurement” be changed to “assessment”.
The overall goal of assessment is to improve students’ learning and provide students, teachers and parents
with reliable information regarding students’ progress and extent of attainment of the expected learning
outcomes.
Assessment- a process of gathering and organizing quantitative or qualitative data into an interpretable
form to have basis for judgment or decision making. It is a prerequisite to evaluation. It provides the
information which enables evaluation to take place.
2
- the ongoing process of gathering evidences of students’ performance over a period of time to determine
learning and mastery of skills.
Traditional Assessment– refers to the use of paper and pencil objective test
Alternative Assessment- refers to the use of methods other than pen-and –paper objective test which
includes performance tests, projects, portfolios, journals, and the like.
Authentic Assessment- an assessment method that stimulate true- to -life situations. This could be
objective test that reflect real-life situations or alternative methods that are parallel to what we experience in
real life.
Test - an instrument designed to measure any characteristic, quality, ability, knowledge or skill. It comprised
of items in the area it is designed to measure.
1. Assessment FOR Learning – this includes three types of assessment done before and during
instruction. These are placement, formative, and diagnostic.
a) Placement –this is done prior to or before instruction. Its purpose is to assess the needs of the
learners to have basis in planning for a relevant instruction. Teachers use this assessment to know
what their students are bringing into the learning situation. The results of this assessment place
students in specific learning groups to facilitate teaching and learning. Examples: Aptitude tests,
NCEE, NCAE, NEAT, pre-tests
b) Formative – this is done during instruction. This assessment is where teachers continuously
monitor the students’ level of attainment of the learning objectives. The results of this assessment
are communicated clearly and promptly to the students for them to know their strengths and
weaknesses and the progress of their learning.
-guides the teacher in his day to day teaching activity; the results are the basis of teachers’
decision to proceed to the next lesson or re-teach the lesson. Examples: Quizzes, daily recitation,
long test
c) Diagnostic – this is done during instruction and used to determine students’ recurring and
persistent difficulties. It searches for the underlying causes of students’ learning problems that do
not respond to first aid treatment. It helps formulate a plan for detailed remedial instruction.
-It is the basis of teachers on what to do next in the teaching-learning process.
- determine the gaps in learning or learning processes, hopefully to be able to bridge these gaps.
Example: Diagnostic test on the topic “sentence construction”, Diagnostic test in “performing the
four fundamental operations, addition to division”.
2. Assessment OF Learning – this is done after instruction. This is usually referred to as the summative
assessment. It is used to certify what students’ know and can do and the level of their proficiency or
competency. Its results reveal whether or not instructions have successfully achieved the curriculum
outcomes. The information from assessment of learning is usually expressed as marks or letter grades.
The results of which are communicated to the students, parents, and other stakeholders for decision
making. Example: summative tests, periodical tests, midterm/finals
3. Assessment AS Learning – this is done for teachers to understand and perform well their role of
assessing FOR and OF learning. It requires teachers to undergo training on how to assess learning and
be equipped with the competencies needed in performing their work as assessors.
3
4. using assessment results when making decisions about individual students, planning teaching,
developing curriculum, and school improvement.
5. developing valid pupil grading procedures which use pupil assessments.
6. communicating assessment results to students, parents, other lay audiences and other
educators.
7. Recognizing unethical, illegal, and otherwise inappropriate assessment methods and uses of
assessment information.
ACTIVITY #1
1. Daily quizzes
2. Periodical test
3. National scholastic aptitude test
4. IQ test
5. District wide test
6. Licensure Examination for Teachers
7. Attitude test
8. Pre-test
9. College Admission Test
10. Unit test
B. Discussion
Benjamin Bloom identified three domains of learning: These are cognitive, affective and psychomotor.
Affective – includes the manner in which we deal with things emotionally, such as feelings, values,
appreciation, enthusiasms, motivations and attitudes.
Psychomotor – includes physical movement, coordination, and use of the motor-skill areas.
Learning targets should be clearly stated, specific and center on what is truly important.
Learning targets need to be stated in behavioral terms or terms which denote something which can be
observed through the behavior of the students. Thus the objective “to understand the concept of
buoyancy” is not stated in a behavioral term. It is not clear how one measures “understanding”. On the
other hand, if we restate the target as “to determine the volume of water displaced by a given object
submerged”, then we can easily measure the extent to which the students understand “buoyancy”
Below is Bloom’s revised taxonomy of educational objectives/targets (Cognitive):
4
Figure 1.
Figure 2.
Below is a matrix consisting of cognitive level of skills and the corresponding verbs to be used in formulating
objectives/learning targets.
5
computes; constructs; contributes; controls;
demonstrates; solve; discovers; establishes; extends;
implements; instructs; participates; prepares;
produces; relates; transfers; uses; utilizes; operates;
sketches;
Analysis/Analyze Analyzes; breaks down; categorizes; compares;
correlates; differentiates; discriminates; distinguishes;
focuses; illustrates; infers; outlines; separates;
subdivides; organizes; examines
Evaluation/Evaluate Appraises; compares and contrasts; concludes;
criticizes; decides; interprets; judges; justifies; values;
supports; argues; weighs;
Synthesis/Create adapts; anticipates; collaborates;
formulates; devises; generates; hypothesizes;
integrates; modifies; negotiates; rearranges;
reconstructs; validates; revises; invents; develops;
incorporates; combines; creates; compiles; designs;
assembles; investigates; reorganizes
Behavior
6
d. Observation– this can be formal or informal. A tally sheet is used by teachers to record the
frequency of student behaviors, activities or remarks.
e. Self Reports – a self-checklist is used in this method. A self-checklist is a list of several
characteristics or activities presented to the subject of a study. The individuals are asked to study
the list and then to place a mark opposite the characteristics when they possess the given
characteristics. This helps diagnose or to appraise the performance of students from the point of
view of the students themselves.
Principle 3.Balance
A balance assessment sets targets in all domains of learning (cognitive, affective and psychomotor) or
domains of intelligence (verbal-linguistic, logical-mathematical, bodily-kinesthetic, visual-spatial,
musical-rhythmic, intrapersonal-social, intrapersonal-introspection, physical world-natural, existential-
spiritual).
A balanced assessment makes use of both traditional and alternative assessments.
Principle 4. Validity
1. Face Validity–refers to the outward appearance of the test. It is done by examining the physical
appearance of the instrument to make it readable and understandable. It is the lowest form of test
validity.
2. Content Validity –refers to the content and format of the instrument. This answers the following
questions: How appropriate is the content? How comprehensive? Does the instrument logically get
the intended variable or factor? How adequately does the sample of items or questions represent
the content to be assessed? It is done through a careful and critical examination of the objectives
of assessment to reflect the curricular objectives.
3. Criterion-Related Validity – the test item is judged against a specific criterion; the degree to
which the item measures a criterion. It is established statistically such that a set of scores revealed
by measuring instrument is correlated with the scores obtained in another external predictor or
measure. It has two purposes: concurrent and predictive.
a) Concurrent validity – describes the present status of the individual by correlating the set
of scores obtained from two measures given at a close interval. This can be established by
correlating the scores in the newly constructed test with the scores in another test with
already known validity. Ex. If the school wanted to determine whether the entrance test
constructed is valid or not, the scores in the newly constructed test can be correlated in the
scores in the previous entrance test which is already known to be valid.
b) Predictive validity – describes the future performance of an individual by correlating the
sets of scores obtained from two measures given at a longer time interval. The assessment
instrument must gather data that correlates with students ’previous performance in order to
predict future performance. Ex. Scores of students in an entrance examination can be
correlated with their grades in academic subjects at the end of a semester to determine
whether the entrance test constructed possess predictive validity.
4. Construct Validity – is established statistically by comparing psychological traits or factors that
theoretically influence scores in a test. Ex. The newly constructed measuring instrument possess
construct validity, if when administered to two groups of individuals (fast learners and slow
learners), the fast learners are expected to get higher scores than the slow learners. If the slow
learners get higher scores than the fast learners, then the instrument has a poor construct validity.
Principle 5. Reliability
This refers to the consistency of scores obtained by the same person when retested using the same or
equivalent instrument. It is also a term that is synonymous with dependability or stability.
Reliability can be estimated in several ways as shown in the matrix below.
7
B) with close time
interval between forms
3. Test-retest with Measure of Stability Give parallel forms of Pearson’s r
Equivalent Forms and Equivalence tests (Form A and Form
B)with increased time
interval between forms.
4. Split Half Measure of Internal Give a test once to Pearson’s r and
Consistency obtain scores for Spearman Brown
equivalent halves of the Formula
test..ex. odd and even -
numbered items
5. Kuder-Richardson Measure of Internal Give the test once then Kuder-Richardson
Consistency correlate the proportion Formula 20 and 21
/percentage of the
students passing and
not passing a given
item.
Principle 6. Fairness
A fair assessment provides all students with an equal opportunity to demonstrate achievement. The
key to fairness are as follows:
When assessing learning, the information obtained should be worth the resources and time required to
obtain it. The factors to consider are as follows.
Teacher Familiarity with the Method. The teacher should know the strengths and weaknesses of
the method and how to use it.
Time required. Time includes construction and use of the instrument and the interpretation of results.
Other things being equal, it is desirable to use the shortest assessment time that provides valid and
reliable results.
Complexity of the Administration. Directions and procedures for administrations are clear and that
little time and effort is needed.
Ease of Scoring. Use scoring procedures appropriate to the method and purpose. The easier the
procedure, the more reliable assessment is.
Ease of Interpretation. Interpretation is easier if there is a plan on how to use the results prior to
assessment.
Cost. Other things being equal, the less expense used to gather data/information, the better.
Principle 8. Continuity
Assessment takes place all phases of instruction. It could be done before, during, and after instruction.
Principle 9. Authenticity
8
Principle 11. Positive Consequences
Assessment should have positive a positive consequence to students; that is, it should motivate them
to learn.
Assessment should have positive consequence to teachers; that is, it should help them improve the
effectiveness of their instruction.
Teachers should free the students from harmful consequences of misuse or overuse of various
assessment procedures such as embarrassing and violating students’ right to confidentiality.
Teachers should be guided by laws and policies that affect their classroom assessment.
Administrators and teachers should understand that it is inappropriate to use standardized student
achievement to measure teaching effectiveness.
ACTIVITY #2
A. Classify the cognitive objectives below in terms of revised Bloom’s Taxonomy (remember,
understand, apply, analyze, evaluate, and create)
B. Discussion
Types of Tests
1. True-False
2. Multiple choice
3. Matching type
4. Filling the blanks
5. Essay
Development of objective tests requires careful planning and expertise in terms o f actual test construction.
9
Planning a Test and Construction of Table of Specifications (TOS)
Identifying Test Objectives. An objective test, if it is to be comprehensive, must cover the various levels
of Bloom’s taxonomy (remember, understand, apply, analyze, evaluate, create). Each objective consists of a
statement of what is to be achieved and, preferably, by how many percent of the students.
Deciding on the type of objective test to be prepared. The test objectives dictate the kind of
objective test that will be designed and constructed by the teacher. For instance, for the first four (4) levels,
we may want to construct a multiple choice type of test while for the two (2) higher levels, we may opt to
give an essay or a modified essay test.
Preparing a table of specification (TOS). A table of specifications or TOS is a test map that guides the
teacher in constructing a test. The TOS ensures that there is a balance between items that test lower level
thinking skills and those which test higher order thinking skills (or alternately, a balance between easy and
difficult items in the test.
Understand
Apply
Analyze
Evaluate
Create
(Hours)
Time Spent
No. of Items
Process
Content/Topic
TOTAL 31 21 4 7 2 5 26 70
The table above presents a sample/prototype of a table of specification (TOS) with columns for
content/topic, cognitive skills (there are six), time spent (in hours or minutes), number of items in each topic.
The other element that is not obviously noticed is the item placement. As you can see in the table, item
numbers 61, 62, 63 and 64 are placed in a content “Demand and Supply”; the cognitive skill is “application”.
This means that these items should be taken from the said content, and constructed such as they measure
how the examinee applies the concept in real life situation.
10
How to Compute for the Number of Items in Each Topic
1. Use the total number of items in the test and the time spent/allotment for each topic
2. Refer to the TOS above for an example:
There are 22 items for the content “Introduction to Economics”. It is determined by dividing time
spent for this subtopic (8) by the total number of time spent for the whole lesson, then multiply the
result by the total number of items in the test.
8
x 70 = 21.53 or 22 number of items to be constructed in the
26
subtopic “Introduction to Economics”
9
x 70 = 24.23 or 24 number of items to be constructed in the
26
subtopic “ Demand and Supply”.
After developing a table of specification, the teacher should make a decision as to type of item formats
that will be constructed. The choice of the item format should be based on clearly stated objectives. If the
item requires the enumeration of important facts, then the test item format should be enumeration. There
are guidelines for constructing different test item formats. As soon as the table of specification is developed,
the teacher is ready for test construction.
Constructing the draft test items. The actual construction of the test items follows the TOS. As a general
rule, it is advised that the actual number of items to be constructed in the draft double the desired number of
items. The subsequent test try-out and item analysis will most likely eliminate many of the constructed items
in the draft (either they are too difficult, too easy or non-discriminatory), hence it will be necessary to
construct more items than will actually be included in the final test form.
Try-out and validation. The test draft is tried out to a group of pupils or students. The purpose of this try
out is to determine the a) item characteristics through item analysis, and b) characteristics of the test itself,
validity, reliability, practicality, etc.
Binomial-choice tests are tests that have only two (2) options such as true or false, right or wrong, good or
better and soon. A modified true-false test can offset the effect of guessing by requiring the students to
explain their answer and to disregard a correct answer.
Ex. The Philippines gained its independence in 1898 and therefore celebrated its centennial year in
2000.
Obviously, the answer is false because 100 years from 1898 is not 2000 but 1998.
2. Avoid using the words “always”, “never”, “often” and other adverbs that tend to be either always true
or always false.
3. Avoid long sentences as these tend to be “true”. Keep sentences short.
4. Avoid trick statements with some minor misleading word or spelling anomaly, misplaced phrases etc. a
wise student who does not know the subject matter may detect this strategy and thus get the answer
correctly.
Ex. The principle of our school is Mr. Albert P. Pamadero.
The principal’s name may actually be correct but since the word is misspelled and the entire sentence
takes a different meaning, the answer would be false! This is an example of a tricky but utterly useless
item.
5. Avoid quoting verbatim from reference materials or textbooks. This practice sends wrong signal to the
students that it is necessary to memorize the textbook word for word and thus, acquisition of higher
level thinking skills are not given due importance.
6. Avoid specific determiners or give-away qualifiers. Students quickly learn that strongly worded
statements are more likely to be false than true, for example statements with “never”, “no”, or
“always”. Moderately worded statements are more likely to be true than false. Statements with “many”,
“often”, “sometimes”, “generally”, “frequently”, or “some” should be avoided.
7. With true or false questions, avoid grossly disproportionate number of either true or false statements
or even patterns in the occurrence of true and false statements.
11
Multiple Choice Test
Each item in a multiple choice test consists of two parts: a) the stem and b) the options. In the set of
options, there is a correct or best option while all the others are considered distracters. The distracters are
chosen in such a way that they are attractive to those who not know the answer or are guessing but at the
same time, have no appeal to those who actually know the answer.
Ex. Which of the following will not cause inflation in the Philippine economy? (Poor)
Which of the following will cause inflation in the Philippine economy? (Better)
5. Each item stem should be as short as possible; otherwise you risk testing for reading and
comprehension skills.
6. Distracters should be equally plausible and attractive.
Ex. The Filipino author who wrote the story “may Day’s Eve”, was______. (Poor)
a) Jose Garcia Villa b) Nick Joaquin c) Genoveva Edrosa d) Robert Frost e) Edgar Allan Poe
The distracters are not all Filipino authors, only the first three, the last two can be essentially
disregarded by the students since these are not Filipino authors. If the distracters had all been Filipino
authors, the value of the item would be greatly increased.
7. All multiple choice options should be grammatically consistent with the stem.
8. Avoid stems that reveal the answer to another item.
9. Avoid alternatives (options) that are synonymous with others or those that, include or overlap others.
Ex. what causes ice to transform from solid to liquid state? (Poor)
a) Change in temperature
b) Changes in pressure
c) Change in the chemical composition
d) Change in heat levels
The options a and d are essentially the same. Thus a student who spots these identical choices would
right away narrow down the field of choices to a, b, c. The last distracter would play no significant role
in increasing the value of the item.
10. Avoid use of unnecessary words or phrases, which are not relevant to the problem at hand.
11. Avoid presenting sequenced items in the same order as in the text.
12. Use the “None of the above” option only when the keyed answer is totally correct. When choice of the
“best” response is intended, “none of the above “is not appropriate.
13. The main stem of the test item may be constructed in question form, completion form (filling
the blank), or direction form.
12
Example of a direction for a multiple choice type of test
Direction: Read the statements carefully and choose the correct answer from the given alternatives. Write
the letter of your choice on the space/blank provided before each item number.
1. Include only homogeneous material in each matching exercise. Do not mix such dissimilar items as
persons and places in a single exercise.
2. Put items on the left and number them, put options on the right and designate them by letters.
3. Place all the items and options for a matching exercise on a single page, if possible.
4. Limit a matching exercise to not more than 10-15 items.
Example:
Direction: Match the items in column A with the items in column B. Write the letter of your choice on the
space or blank provided before each item number.
Col. A col. B
Example:
Direction: Write G if the item on the left is greater than the item on the right; L if the item on the left
is less than the item on the right; E if the item on the left is equal to the item on the right and D if the
relationship cannot be determined.
Col. A Col. B
The data sufficiency test above can, if properly constructed, test higher order thinking skills.
This type of test can also test lower order thinking skills. Like the multiple choice test, the items in this kind
of test consist of a stem and a blank where the students would write the correct answer. Supply type tests
depend heavily on the way the stems are constructed. These tests allow for one and only one answer. Provide
blanks that are of equal length.
1. Avoid indefinite statements. A pupil has the right to know the type of response desired.
Ex. Jose Rizal was born in ________. (poor)
Jose Rizal was born in the year ________. (Better)
2. Avoid over mutilated statements. If too many key words are left out, it is impossible to know what
meaning was intended.
3. Avoid lifting statements directly from the text. This tests nothing more than rote memory.
4. Make the blanks of uniform length. If the blanks vary in length the pupil has a clue to the length of the
correct answer.
5. Try to choose statements in which there is only one correct response for the blank.
6. The required response should be a single word or brief phrase.
13
Example:
Direction: Read the statement carefully. Fill in the blank with the word that will make the statement correct.
Write your answer on the blank provided before each item number.
Essays
Essays, classified as non-objective tests, allow for the assessment of higher order thinking skills. Such
tests require students to organize their thoughts on a subject matter in coherent sentences in order to inform
an audience. In essay tests, students are requested to write one or more paragraphs on a specified topic.
Essay questions can be used to measure attainment of a variety of objectives. Below is the list of types of
abilities that can be measure by essay items:
ACTIVITY # 3
A. Write T if the statement is correct and F if otherwise. Write your answer on your answer sheet.
1. Test blue prints are necessary to ensure direction in the preparation of a test.
2. Learning targets or objectives can be measured by any test format.
3. Tricky questions are necessary to test pupils’ understanding.
4. Care must be taken in rewording faulty items.
5. A multiple choice type of test can measure various level of cognitive skill.
6. Constructing a draft test is the first step in planning a test.
B. Discussion
C. Prepare a table of specification (TOS) for a 10 or 15- item test for the topic that you have selected
(refer to your answer in activity #2, letter C).
14
D. Construct a 20-item test based on the TOS that you have prepared.
E. Construct 5-item test on the topic which you have selected (refer to your answer in activity #2,
letter C) using each type of test below. Observe the principles in constructing them.
1. Matching Type
2. True-False
3. Multiple Choice
4. Completion/Supply Type
Test assembling – means recording test items; review test items; arrange test items
Items should be arranged according to types
Items should be arranged in order of difficulty
Items of the same subject matter content maybe grouped together
How to assemble the test
Guidelines:
Item Analysis – a statistical method used to determine the quality of a test by looking at each
individual item or question and determining if they are sound or effective.
- It helps identify individual item or question that are not good question and whether or not they
should be discarded, kept, or revised.
- It is used for selecting and rejecting the item on the basis their difficulty value and discriminating
power.
Item analysis procedure provides the following information:
1. The difficulty of the item
2. The discriminating power of the item
3. The effectiveness of each alternatives/options
Benefits derived from item analysis
1. It provides useful information for class discussion of the test.
2. It provides data which helps students improve their learning.
3. It provides insights and skills that lead to the preparation of better tests in the future.
15
Qualitative Item Analysis - a non-numerical method for analyzing test item not employing
students’ responses but considering test objective, content validity and technical item quality.
– a process in which the teacher or expert carefully proofreads the test before it is
administered to check if there are typographical errors, to avoid grammatical clues that may lead
to giving away the correct answer , and to ensure that the level of reading materials is appropriate.
Quantitative Item Analysis – a numerical method for analyzing test item employing
students’ responses to the alternatives or options.
-evaluating items in terms of statistical properties such as item difficulty and item
discrimination.
Difficulty index – is a proportion of students who answered the item correctly. It indicates
whether the item is easy, of average difficulty or difficult.
Steps:
1. Rank the scores of the students (arrange from highest to lowest or vice versa).
2. Compute 27% of the total number of students who took the test to determine the number
of students that belong to the upper/lower 27%.
Ex. Number of students who took the test (N) = 60
27% of 30 = 0.27 x 30 = 8.1 or 8
This means that there will be 8 students from the top and another 8 from the
bottom so only 16 test papers out of 60 will be analyzed.
3. Separate the test papers of upper 27% from the lower 27%.
4. Determine (tally) the number of students from the upper 27% who got the item right. Label
it (U). The same procedure will be applied to the lower 27%. Label it (L)
5. Compute for the difficulty and discrimination indices using the formula below:
𝑈+𝐿
Difficulty Index (DfI) = 𝑇
, where T = total no. of test paper to be analyzed ( upper 27% + lower27%)
For U and L (refer to step 4 for the meaning)
𝑼−𝑳
Discrimination Index (DsI) = 𝟏
𝑻
𝟐
6. Interpret the results using the scale below.
16
𝑼−𝑳 𝟏𝟒−𝟐 𝟏𝟐
DsI = 𝟏 = = = 0.75
𝑻 𝟏𝟔 𝟏𝟔
𝟐
Establishing Validity
After performing the item analysis and revising the items which need revision, the next step is
to validate the instrument/test. The purpose of validation is to determine the characteristics of the
whole test itself namely, the validity and reliability of the test. Validation is the process of collecting
and analyzing evidence to support the meaningfulness and usefulness of the test.
There are three main types of evidence that may be collected: content- related evidence of
validity, criterion- related evidence and construct –related evidence of validity.
Content- related evidence of validity – refers to the content and format of the instrument/test.
How appropriate is the content? How comprehensive? Does it logically get at the intended variable?
How adequately does the sample of items or questions represent the content to be assessed?
Criterion-related evidence of validity – refers to the relationship between scores obtained using
The instrument and scores obtained using one or more other tests (often called the criterion). How
Strong is this relationship? How well do such scores estimate present or future performance of a
certain type?
Construct-related evidence of validity – refers to the nature of the psychological construct or
characteristic being measured by the test. How well does the measure of the construct explain
differences in the behavior of the individuals or their performance on a certain task.
To determine content validity the teacher writes out the objectives of the test based on the table of
specifications and then gives these together with the test to at least two (2) experts along with the description
of the intended test
takers. The experts look at the objectives, read over the items in the test and placed a check mark in front of
each question or item that they feel does not measure one or more objectives. They also place a check mark
in front of each objective not assessed by any item in the test. These will be followed by some revisions until
the experts approve the test.
In order to obtain evidence of criterion-related validity, the teacher usually compares scores on the test
in question with the scores on some other independent test criterion which presumably has already high
validity. For example, if a test is designed to measure mathematics ability of the students and it correlates
highly with a standardized mathematics achievement test (external criterion), then we say we have a high
criterion-related evidence of validity. This type of criterion-related validity is known as the concurrent
validity. Another type of criterion-related validity is predictive validity wherein the test scores in the
instrument are correlated with scores on a later performance (criterion measure) of the students. For example,
the mathematics ability test constructed by the teacher may be correlated with their later performance in the
Division wide math achievement test.
Estimating Reliability
Reliability – refers to the consistency of the scores obtained – how consistent they are for each
individual from one administration of an instrument to another and from one set of items to another.
Validity and reliability are related concepts. If an instrument is unreliable, it cannot yield valid
outcomes. As reliability improves, validity may improve (or it may not). However, if an instrument is shown
scientifically to be valid then it is almost certain that it is also reliable.
1. Test-Retest – Give a test twice to the same learners with any time interval between tests
from several minutes to several years.
2. Equivalent Forms – Give parallel forms of test with close interval between forms.
3. Test-Retest with parallel forms – Give parallel forms of tests with increased time
interval between forms.
17
4. Split Half – Give a test once to obtain scores for equivalent halves of the test (the odd and
even numbered items.)
5. Kuder –Richardson – Give the test once then correlate the proportion/percentage of the
students passing and not passing a given item.
The following table is a standard followed almost universally in educational test and measurement
ACTIVITY #4
Objectives:
Statistics plays a very important role in assessment. Statistical techniques allow us to describe the
performance of students and make proper scientific inferences about their performance.
18
–the tabular arrangement of data by classes or by categories together with their corresponding class
frequencies.
Total N= 20 100
In the table above, 20 – 29 and 30 – 39 are examples of class intervals. End numbers like 20 and 29 are
class limits, 20 is the lower
Limitand29 is the upper limit.19.5 - 29.5 and 29.5- 39.5 are called class boundaries, 19.5 is the lower
boundary and 29.5 is the upper boundary.
Class boundaries are more precise expressions of the class limits by at least 0.5 of their values.
Class width or class size is the difference between the upper and the lower class boundaries.
Class mark/Midpoint – the midpoint between the upper and lower class limit.
Steps:
1. Compute for the range. Highest value in the set ( HV) – lowest value (LV)
2. Determine the desired number of class intervals or categories. The ideal number of class intervals is
from 10 and15.
3. Determine the class interval size by dividing the range by the desired number of class intervals. The
class size is preferably an odd number.
4.Organize the class intervals either in descending or ascending order starting from a lower limit which, is
a) divisible by the class size or b) start with the lowest value in the set.
Example:
Below are the scores of 50 students in Assessment 1. Construct a grouped frequency distribution.
18 29 42 57 61 67 37 49 53 47
24 34 45 58 63 70 39 51 54 48
28 36 46 60 66 77 40 52 56 49
19 31 44 58 62 68 38 50 54 48
27 36 46 59 64 74 39 51 55 48
a) Range = 77-18 = 59
b) The ideal number of class intervals is 10-15. Divide the range by any number from 10-15 to find
the class size in letter c.
c) Class interval size (i) = 59/10 = 5.9 or 5 (the nearest odd number)
d) Organize the class intervals either in descending or ascending order. The lower limit of the lowest
class interval can be the value a)divisible by the class size or b) it can be the lowest value
in the set.
19
Below is the Frequency Distribution of the scores of 50 Students in Assessment 1
a) The lower limit of the lowest interval starts with the number divisible by the class interval
size( i). Refer to step no. 4-a above.
15-19 14.5-19.5 II 2 17 2 50
20-24 19.5-24.5 I 1 22 3 48
25-29 24.5-29.5 III 3 27 6 47
30-34 29.5-34.5 II 2 32 8 44
35-39 34.5-39.5 IIII-I 6 37 14 42
40-44 39.5-44.5 III 3 42 17 36
45-49 44.5-49.5 IIII-IIII 9 47 26 33
50-54 49.5-54.5 IIII-II 7 52 33 24
55-59 55.5-59.5 IIII-I 6 57 39 17
60-64 59.5-64.5 IIII 5 62 44 11
65-69 64.5-69.5 III 3 67 47 6
70-74 69.5-74.5 II 2 72 49 3
75-79 74.5-79.5 I 1 77 50 1
N= 50
b) The lower limit of the lowest class interval starts with the lowest value in the set (sample class
intervals and the corresponding class boundaries are shown). Refer to step no. 4-b above.
18-22 17.5-22.5
23-27 22.5-27.5
28-32 27.5-32.5
33-37 32.5-37.5
38-42 37.5-42.5
43-47 42.5-47.5
48-52 47.5-52.5
53-57 52.5-57.5
58-62 57.5-62.5
63-67 62.5-67.5
68-72 67.5-72.5
73-78 72.5-78.5
ACTIVITY # 5
A. Given below are the scores of 44 students in arithmetic test. Construct a grouped frequency distribution.
Use step number 4-b to organize your data.
8 3 18 13 23 8 19 13 6
18 18 12 8 17 11 3 16 5
12 17 4 22 6 8 13 8 3
22 30 19 26 8 8 12 21 15
8 7 18 11 10 22 9 21
Study the frequency distribution table that you have constructed and answer the following questions:
20
Measures of Central Tendency
- are measures indicating the center of the set of data.
- they are typical representative values of the set of data.
1. Arithmetic mean or the average refers to the sum of all values/observations divided by the
number of values/observations.
Formulas:
∑𝐱 ∑ 𝑓𝑀
𝐱̅ = , where x = observation ̅=
𝑥 , where f = frequency of the class interval,
𝐍 𝑁
N = no. of cases; M=midpoint of the interval
3. Median is a single value which divides the set of data into two equal parts such that half (50%) of
the values fall below it and half (50%) fall above it.
N
cf
x LB 2
~ i , where LB = lower class boundary of the median class; cf = cumulative frequency of
f
the class interval below the median class; f = frequency of the median class; N = number of values in the
sample; i = class interval size
4. Mode – the value which occurs most frequently in the given set of data; the value with the highest frequency
Rough estimate of the mode can be represented by the midpoint of the modal class (class interval
with thehighest frequency).
f f
xˆ LB mo 1 i , where LB = lower class boundary of the modal class; fmo= frequency of the
2
f mo
f 1
f 2
modal class; f1 = frequency of the class interval below the median class; f2 = frequency of the class interval
above the median class
Examples:
1. Find the mean, median, and mode of the scores of 9 students. The scores are:
53, 45, 59, 48, 54, 48, 51, 58, and 55.
Median= 53→this means that 53 is the middlemost score. It divides the set of data into two parts such that
50% lie above 53 and another 50% lie below it.
Mode = 48 since 48 has the highest frequency or 48 is the most occurring score in the set
Mean= Σx= 6 + 10 + 8 + 6 + 8 + 9 + 7 + 5 = 59 = 7. 38
21
N 8 8
Ranked scores: 5, 6, 6, 7, 8, 8, 9, 10
N ( N 2) 8 (8 2)
th th th th
2 2 2 2 4th 5th 78 15
Median = = = = = = 7.5
2 2 2 2 2
Mode = 6 and 8 are most occurring score or the score with the highest frequency
3. Solve for the mean, median and mode of the scores of 50 students in assessment 1.
∑ 𝑓𝑀
C.I. f M fM <cf 𝑥̅ =
𝑁
N
15-19 2 17 34 2 𝑥̅ = 2410
20-24 1 22 22 3 50
25-29 3 27 81 6 𝑥̅ = 48.2 Mean score
30-34 2 32 64 8
35-39 6 37 222 14
40-44 3 42 126 17
Median 45-49 9 47 423 26
classs 50-54 7 52 364 33
interval 55-59 6 57 342 39
60-64 5 62 310 41
65-69 3 67 201 47
70-74 2 72 144 49
75-79 1 77 77 50
N = 50 ΣfM = 2,410
To compute for the median of grouped data, first determine the median class by computing N/2,
then
Under column <cf, select <cf greater than or equal to N/2.
The median class is the class interval corresponding to <cf greater than or equal to N/2.
N
cf
x LB 2
~ i f = 9, cf= 17, LB = 44.5 , N/2 = 50/2 = 25
f
~ 25 17
x 44.5 5
9
~ 8
x 44.5 5
9
Mode
f mo f 1
xˆ LB i xˆ 44.5 6 5
2 Given: LB = 44.5, fmo=9, f1=3, f2= 7, i=5
8
f mo f 1 f 2
93
xˆ 44.5 5 xˆ 48.25
2(9) 3 7
ACTIVITY #6
B. The following are the scores of seven students in English 101: 7, 12, 17, 7, 15, 5 and 12
22
C. Compute for the mean, median and mode of the data set below.
40-44 3
35- 39 4
30-34 6
25-29 12
20-24 10
15-19 6
10-14 4
N = 45
Measures of Dispersion/Variability
– Describe the extent of scattering or spread of individual observations relative to the average or mean.
In this topic, all the measures of variability were introduced but we will give emphasis on the application of
the most stable and reliable measure which is the standard deviation.
∑(𝑥−𝑥̅ )2
𝑠2 =
𝑛−1
Standard deviation – is the positive square root of the distributions’ variance. It is also known as the
root- mean-square deviation. The variance and the standard deviation are the most stable and
reliable measure of variability.
∑(𝑥−𝑥̅ )2 ∑ 𝑓𝑀 2
S=√ (sample standard deviation) 𝑠=√ − 𝑥̅ 2
𝑛−1 𝑁
∑(𝑥−𝑥̅ )2
𝜎=√ (population standard deviation
𝑛
Examples:
The scores of two groups of students in a 10-item test in Math are given below. Solve for the range,
mean deviation, standard deviation, variance and quartile deviation of each group.
Male: 2, 1, 5, 6, 10, 6
Female: 4, 5, 5, 6, 6, 4, 5
Solution: First construct a table as indicated below. Compute for the value of the mean by dividing 30 (sum
of all the values) by 6 (number of values) or use the formula,
∑𝐱
𝐱̅ = 𝐍
2 -3 9
1 -4 16 N=6
5 0 0
6 1 1
10 5 25
6 1 1
Σ│x-x│= 14 Σ(x-x)2 = 52
30
Mean(𝐱
̅)= 6
=5
s
x x 2
variance ( s2 ) = 10.4
n 1
23
52
s
6 1
52
s
5
s 10.4
s = 3.22
4 -1 1
5 0 0
5 0 0
6 1 1
6 1 1
4 -1 1
5 0 0
Σ│x-𝐱
̅│=4 ̅)2 =4
Σ(x-𝐱
35
𝐱̅= 7 = 5 (mean)
s
x x 2
The standard deviation (s) of male students is 3.22 .This indicates that the scores of male deviate from
the mean at about 3.22 or approximately 3 points.
The standard deviation of female students is 0.82. This indicates that female scores deviate from the
mean at about 0.82 or approximately 1 point.
The standard deviation of males is greater than the standard deviation of the females. This indicates
that the scores of male are more variable than the scores of female.
ACTIVITY #7
Math scores: 11, 10, 12, 10, 6, 3, 10, 7, 17, 14 Physics scores: 10, 8, 9, 12, 10, 7, 10, 9, 14, 11
Normal Distribution
- One of the continuous distributions used in the entire theory of Statistics. The graph is a bell-
shaped curve that extends indefinitely in both directions. This indicates that most of the values are
on the average, few are above and below the mean and very few are extremely high and extremely
low values. Examples of data that approximate a normal distribution: weights/ heights of a large
number of individuals, IQ (Intelligence Quotients) and scholastic achievements of students.
Pierre Laplace, Abraham de Moivre and Karl Friedrich Gauss were the first to investigate the
mathematical properties of this kind of continuous distribution curve and its theoretical basis.
-
24
Properties/Characteristics of Normal Distribution
*The area under the normal curve may represent several things like the probability of an event, the
percentile rank of a score or the percentage distribution of the whole population. P (a ≤ x ≤ 𝑏) =
area under the normal curve between x =a and x = b. This is read as “the probability that x is between a
and b is equal to the area between a and b.
We simply obtain areas under a normal curve by performing a simple change of scale called a
standard score or the z-score or simply the z-value.
Standard score (z) –a kind of transformed score that relates a raw score to the mean and standard
deviation of a distribution. It tells us how many standard deviations the corresponding value lies above or
below the mean. The z- values o z-scores range: ….-3, -2, -1, 0 1, 2, 3……,The z-value of 2 (z = 2)
means that a raw score is two standard deviations above the mean. The z-score of -2 (z = -2) indicates
that the raw score is two standard deviation below the mean.
𝒙−𝐱̅
z=
𝒔 → (if based on a sample), where x = raw score; 𝐱= sample mean;
̅
s = standard deviation
𝒙−𝛍
z= → (if based on population), where𝝁 = population mean;
𝝈
𝝈 = population standard deviation
Examples:
1). On a final examination in Biostatistics the mean is 76 and standard deviation is 10.
a) Determine the standard score of a student when receiving the grade of 90.
𝑥−𝐱̅ 90−76 14
z= = = = 1.4
𝑠 10 10
25
x = zs + 𝐱̅
x = 0.5(10) + 76
x = 5 + 76
x = 81
Example 1.
Refer to the figure at the left. The area between z =-1 and
z = 1is approximately 0.6826 or 68.3%. The area
between z= -2 and z = 2 is approximately 0.9544 or
95.4%. Just still refer to the figure for the area between
other two z-scores.
Example 2.
The shaded area in figure 1 below represent z = 0.56. Please secure a copy (have a photocopy) of the
table of areas under the normal curve like the one indicated in figure 2 below. This table can help you
find the area, given the z-values or the z-value corresponding to the given area. Refer to the table,
what is the area when z = 0.56? The answer is 0.2123. It is the intersection of the row through z
=0.5 and the column headed by 0.06. Try to search for a copy of table of areas under the normal
Curve for your reference.
Fig.1
Fig.2
Example 3.
26
Example.4 (figure at the left)
To find the area in example 4 above, use the area below z = 0 which is 0.5000 or 0.5. Subtract the area
z = -0.99 from the area below z = 0.
ACTIVITY #8
A. Convert the following raw scores (x) to a z-score. Show your computation/solution.
Given: 𝐱̅= 30; s = 5; Find z corresponding to the raw scores below.
1) x = 37 2) x = 44 3) x =19
1) z= 2 2) z = -1.5 3) z = 2.5
C. Find the area under the normal curve as indicated below. Draw the normal curve and shade the
area indicated.
1. Between z = -2 and z = 2.5
2. Below z = -1.5
3. Above z = -2.5
*To determine the percentage of cases, proportion or probability in a normal distribution within
given limits.
* To find limits in any normal distribution which include a given percentage of cases.
Problem:
1. In a reading ability test, with a sample of 100 cases, the mean score is 40 and the standard
deviation is 4.
2.
Assuming normality,
a) What is the percentage of the cases that fall between the mean and a score of 46?
b) What is the probability that a score picked at random will be above a score 46?
c) What is the probability that a score will be below 46?
d) How many cases fall between scores of 42 and 48?
To answer problem letter D number 1(a-d), first convert the raw scores 42, 46, and 48 to a z-score.
Use the formula in conversion.
27
1. Perfect correlation (perfect positive or perfect negative) –the points in the scatter diagram
form a straight line either rising or falling to the right.
2. Some degree of correlation ( positive or direct and negative or indirect) – the points
form a band either rising or falling to the right As one variable increases the other also increases
as in the case of positive correlation or as one variable increases the other decreases as in the
case of negative correlation
3. No correlation (zero correlation) –indicates no relationship between two variables. The points
are all scattered in the coordinate plane
Scatter Plot or scatter diagram – the graph showing the relationship between two variables
3. Perfect Negative 4-a. Negative or Inverse Correlation 4-b. Negative or inverse Correlation
5. Zero Correlation
The qualitative interpretation of the degree of linear relationship existing is shown in the following
range of values or scale:
r-value Interpretation
Formula in computing for the value of Pearson Product Moment Correlation Coefficient
or Pearson’s r in short.
𝑵∑𝒙𝒚−∑𝒙∑𝒚
r=
√[𝑵∑𝒙𝟐−(∑𝒙)𝟐][𝑵∑𝒚𝟐−(∑𝒚)𝟐
Example:
Given:
x 3 5 6 8 9 11
y 2 3 4 6 5 8
Solution:
28
1. Draw the scatter diagram (use 1 whole sheet of graphing paper).
12
11
10
9
8 *
7
6 *
5 *
4 *
3 *
2 *
1
0 1 2 3 4 5 6 7 8 9 10 11
0
Solution:
x y xy x2 y2
3 2 6 9 4
5 3 15 25 9
6 4 24 36 16 (∑x)2 =(42)2= 1,764
8 6 48 64 36 (∑y)2 = (28)2 = 784
9 5 45 81 25
11 8 88 121 64
∑x = 42 ∑y=28 ∑xy=226 ∑x2=336 ∑y2=154
𝑵∑𝒙𝒚−∑𝒙∑𝒚
r=
√[𝑵∑𝒙𝟐−(∑𝒙)𝟐][𝑵∑𝒚𝟐−(∑𝒚)𝟐
6(226)– 42(28)
r=
√[6(336)− 1.764][6(154)− 784]
1,356 – 1,176
r=
√[2,016− 1.764][924− 784]
180
r=
√[252][140]
180 180
r= = = 0.96→ answer
√35,280 187.83
ACTIVITY #9
A. Given:
x 1 3 4 6 8 9 11 14
y 1 2 4 4 5 7 8 9
B. The following data show the results of a study indicating the number of accidents per
hundred thousand vehicle kilometers.
For letter B, answer the same set of questions in A shown above (1-4)
UNIT VI. GRADING SYSTEM
29
Objectives: At the end of the topic the students must have:
Norm-referenced grading refers to a grading system where in student’s grade is placed in a relation
to the performance of a group. In this system a grade of 80 means that the student performed better than
or same as 80% of the class or group.
Example 1: Consider the following two sets of scores in an English class for two sections A and B of ten
students.
A=(30,40,50,55,60,65,70,75,80,85)
In section A the student who get a raw score of 75 would get a grade 0f 80% while in section B the grade
of 80% would correspond to a raw score of 90.If the same test was used for the two sections ,it would be “unfair”
system of grading. A wise student would opt to enrol in class A since it is easier to get a higher grades in that
class than in the other class(class B).
Example 2: A teacher may establish a grading policy whereby the top 15 percent of students will receive a mark
of excellent or outstanding which is a class of 100 enrolled students will be
15 persons. Grading Policy:
The underlying assumption in norm-referenced grading is that students have abilities (as reflected in
their raw scores) that obey the n0rmal distribution. The objective is to find out the best performers in this
group.
Example 3:I n a class of 100 students ,the mean score in a test is 70 with a standard deviation of 5. Construct
a norm-referenced grading table that would have seven-grade scales and such that students ‘scoring between
plus or minus one standard deviation from the mean receives an average grade.
Solution: The following intervals of raw scores to grade equivalents are computed:
Below 55 Fail 1%
55-60 Marginal 4%
30
Above 85 Excellent 1%
Criterion-Referenced grading system is based on a fixed criterion measure. There is a fixed target and
the student must achieve that target in order to obtain a passing grade in a course regardless of how the
other students in the class perform. The scale does not change regardless of the quality, or lack therefore, of
the student. For example, in a class of 100 students using the table below, no one might get a grade of
excellent if no one scores 98 above or 85 above depending on the criterion used.
There is no fixed percentage of students who are expected to get the various grades in the criterion-
referenced grading system.
Criterion-referenced systems are often used in situations where the teachers are agreed on the
meaning of a “standard of performance” in a subject but the quality of the student is unknown or uneven;
where the work involves student driving factor such as needing to systematically reduce a pool of eligible
students.
Note that in criterion-referenced grading system, students can help a fellow student in a group work
without necessarily worrying about lowering his grade in that course. This is because the criterion-referenced
grading system does not require the mean (of the class) as basis for distributing grades among the students.
6.3 Four Questions in Grading
Marinila D. Svinicki (2007) of the center for Teaching Effectiveness of the University of Texas at
Austin poses four intriguing questions relative to a grading.
The grading system an instructor selects, reflects his or her educational philosophy. There are no
right or wrong systems, only systems which accomplish different objectives. The following are questions
which an instructor may want to answer when choosing what will go into a student’s grade.
1. Should grades reflect absolute achievement level or achievement relatives to others in the same class?
2. Should Grades reflect achievement only or non academic components such as attitude, speed and
diligence?
Should grades report status achieved or amount of growth?
3. How can several grades on diverse skills combine to give a single mark?
Pass –Fail Systems. Other colleges and universities, faculties, schools and institutions use pass-fail
grading systems in the Philippines, especially when the student work to be evaluated is highly subjective (as in
the fine arts and music) there are no generally accepted standard gradations (as with independent studies),or
the critical requirement is meeting a single satisfactory standard(as in some professional examinations and
practicum.
Non-Graded Evaluations. While not practiced in Philippine schools, and institutions, non-graded
evaluation do not assign numeric or letter grades as a matter of policy. This practice is usually based on a
belief that grades introduce an inappropriate and distracting element of competition into the learning process,
or that they are not as meaningful as measures of intellectual growth and development as are carefully
crafted faculty evaluation. Many faculty, schools and institution that follow a no-grade policy will, if requested,
produce grades or convert their student evaluation into formulae acceptable to authorities who require
traditional measures of performance.
31
Test standardization is a process by which teacher or researcher-made tests are
validated and item analyzed. After a thorough process of validation, the test characteristics are established.
These characteristics include: test validity, test reliability, test difficulty level and other characteristics as
previously discussed. Each standardized test uses its own mathematical scoring system derived by the
publisher and administrator, and these do not bear any relationship to academic grading systems.
Standardized test are psychometric instrument whose scoring systems are developed by norming the test
using national samples of test takers, centering the scoring formula to assure that the likely score distribution
describes a normal curve when graphed, and then using the resulting scoring system uniformly in a manner
resembling a criterion-referenced approach. If you are interested in understanding and interpreting the scoring
system of a specific standardized test, refer to the policies of the test’s producers.
In the Philippines, there are two types of grading systems used: the averaging and the commutative
grading systems. In the averaging system, the grade of a student on a particular grading period equals the
average of the grades obtained in the prior grading periods and the current grading period. In the cumulative
grading system, the grade of a student in a grading period equals his current grading period grade which is
assumed to have the cumulative effects of the previous grading periods. In which grading system, would
there be more fluctuations observed in the students’ grades? How do these systems relate with either norm
or criterion-referenced grading?
ACTIVITY #10
1. Define a norm-referenced grading. What are some of the issues that confront a
80, 81, 82, 83, 84, 82, 80, 81, 79, 77, 88, 83,
89, 90, 91, 90, 78, 79, 82, 91, 92, 90, 88, 85,
88, 87, 85, 88, 83, 82, 80, 79, 77, 76, 77, 78,
Using a norm-referenced grading with a seven-point scale, determine the scores that would get a
failing mark. What is general impression on this?
c. Define a criterion-referenced grading. What are some of the issues that confront
a teacher using a criterion-referenced grading system? Discuss.
For these guidelines, the Department will use a floor grade considered as the lowest possible
grade that will appear in a learner’s report card.
Learners from Grades 1 to 12 are graded on Written Work, Performance Tasks, and Quarterly
Assessment every quarter. These three are given specific percentage weights that vary according to the
nature of the learning area.
32
For Kindergarten
For Grades 1 to 12
In a grading period, there is one Quarterly Assessment but there should be instances for students to
produce Written Work and to demonstrate what they know and can do through Performance Tasks. There is
no required number of Written Work and Performance Tasks, but these must be spread out over the quarter
and used to assess learners’ skills after each unit has been taught.
How to Compute for Final Grades and General Average in DepEd K to 12 Grading System The
following are the steps in computing for the Final Grades.
This results in the total score for each component, namely Written Work, Performance Tasks, and
Quarterly Assessment.
Raw scores from each component have to be converted to a Percentage Score. This is to ensure that
values are parallel to each other.
Step 2: The sum for each component is converted to the Percentage Score.
To compute the Percentage Score (PS), divide the raw score by the highest possible score then
multiply the quotient by 100%. This is shown below:
Step 3: Percentage Scores are then converted to Weighted Scores to show the importance of
each component in promoting learning in the different subjects.
To do this, the Percentage Score is multiplied by the weight of the component found in Table 4 for
Grades 1 to 10 and Table 5 for Senior High School. The product is known as the Weighted Score (WS).
33
The grading system for Senior High School (SHS) follows a different set of weights for each
component. Table 5 presents the weights for the core and track subjects.
Step 4: The sum of the Weighted Scores in each component is the Initial Grade. This Initial
Grade will be transmuted using the given transmutation table to get the Quarterly Grade (QG).
Step 5: The Quarterly Grade for each learning area is written in the report card of the student.
For a better understanding of how to record the summative assessments, Table 6 presents a sample class
record showing three learners for the first quarter of Grade 4 English. On the basis of this class record, Table
7 presents a step-by-step process on how to compute for the Quarterly Grade
34
Table 6. Sample Class Record for English Grade 4 (First Quarter)
35
Table 7. Steps for Computing Grades
36
For MAPEH, individual grades are given to each area, namely, Music, Arts, Physical Education,
and Health. The quarterly grade for MAPEH is the average of the quarterly grades in the four
areas.
For Kindergarten
There are no numerical grades in Kindergarten. Descriptions of the learners’ progress in the
various learning areas are represented using checklists and student portfolios. These are
presented to the parents at the end of each quarter for discussion. Additional guidelines on the
Kindergarten program will be issued.
The General Average is computed by dividing the sum of all final grades by the
total number of learning areas. Each learning area has equal weight.
37
The Final Grade in each learning area and the General Average are reported as
whole numbers. Table 8 shows an example of the Final Grades of the different learning
areas and General Average of a Grade 4 student.
The summary of learner progress is shown quarterly to parents and guardians through
a parent-teacher conference, in which the report card is discussed. The grading scale, with its
corresponding descriptors, are in Table 10. Remarks are given at the end of the grade level.
38
Table 10. Descriptors, Grading Scale, and Remarks
Using the sample class record in Table 6, LEARNER A received an Initial Grade of 84.86 in
English for the First Quarter, which, when transmuted to a grade of 90, is equivalent to
Outstanding. LEARNER B received a transmuted grade of 88, which is equivalent to Very
Satisfactory. LEARNER C received a grade of 71, which means that the learner Did Not Meet
Expectations in the First Quarter of Grade 4 English.
When a learner’s raw scores are consistently below expectations in Written Work and
Performance Tasks, the learner’s parents or guardians must be informed not later than the fifth week
of that quarter. This will enable them to help and guide their child to improve and prepare for the
Quarterly Assessment. A learner who receives a grade below 75 in any subject in a quarter must be
given intervention through remediation and extra lessons from the teacher/s of that subject.
How are learners promoted or retained at the end of the school year?
This section provides the bases for promoting a learner to the next grade level or for
retaining a learner in the same grade level. These decisions must be applied based on evidence and
judiciously.
A Final Grade of 75 or higher in all learning areas allows the student to be promoted to the
next grade level. Table 11 specifies the guidelines to be followed for learner promotion and
retention.
For Grades 1-10, a learner who Did Not Meet Expectations in at most two learning areas must
take remedial classes. Remedial classes are conducted after the Final Grades have been computed.
The learner must pass the remedial classes to be promoted to the next grade level. However,
teachers should ensure that learners receive remediation when they earn raw scores which are
consistently below expectations in Written Work and Performance Tasks by the fifth week of any
quarter. This will prevent a student from failing in any learning area at the end of the year.
For Grade 11-12, learners who fail a unit/set of competencies must be immediately given
remedial classes. They should pass the summative assessments during remediation to avoid a failing
grade in a learning area/subject. This will prevent students from having back subjects in Senior High
School (SHS). However, if the learner still fails remedial classes, s/he must retake the subject/s
failed during the summer or as a back subject. Guidance teachers/career advocates must provide
support to the SHS student for his/her choices in SHStracks.
Summative Assessments are also given during remedial classes. These are recorded,
computed, weighted, and transmuted in the same way as the Quarterly Grade. The equivalent of
the Final Grade for remedial classes is the Remedial Class Mark (RCM). The Final Grade at the end
of the school year and the Remedial Class Mark are averaged. This results in the
Recomputed Final Grade. If the Recomputed Final Grade is 75 or higher, the student is promoted to the next
grade level. However, students will be retained in the grade level if their Recomputed Final Grade is below
75.
The teacher of the remedial class issues the Certificate of Recomputed Final Grade, which is
noted by the school principal. This is submitted to the division office and must be attached to both Form
137 and School Form Number 5. Figure 1 below shows a sample certificate.
39
Figure 1. Sample Certificate of Recomputed Final Grade
The learner can enrol in the next grade level for Grades 1-10 and in the next semester for Grades 11-
12 upon presentation of the Certificate of Recomputed Final Grade. This certificate can be verified in the
division offices as needed.
ACTIVITY #11
Given the raw scores of Grade 3 learner in Mathematics, solve for the quarterly grade using the
steps below:
1 20 25
2 18 20
3 22 25
4 23 25
5 21 25
Total
1 12 15
2 18 20
3 16 20
4 18 20
5 14 15
Total
45 50
Total
40
References:
41