Professional Documents
Culture Documents
LANGUAGE ASSESSMENT
A. INTRODUCTION
1. Teaching and Testing
2. Backwash
WHAT IS A BACKWASH?
Harmful backwash, as the name suggests, are the negative effects of testing on
the teaching and learning activities.
Examples:
• If a test is regarded as important, if the stakes are high, preparation for it can
dominate the teaching and learning activities
• Test content and testing are not appropriate to measure the intended learning
outcomes
Examples:
• It helps to determine whether to revise the curriculum and teaching styles for
the betterment of teaching and learning activities
1|P age
3. The Needs for Tests
▪ One conclusion drawn from understanding why tests are so mistrusted by
language teachers and how this mistrust is often justified is that this might be
better off without language test.
▪ Teaching is, after all the primary activity; if testing comes in conflict with it, then
it is testing that should go, especially when it has been admitted that so much
testing provides inaccurate information.
▪ Teaching systems need dependable measures of language ability to provide
information about the achievement of groups of learners, without which it is
difficult to see how rational educational decisions can be made.
▪ Even without considering the possibility of bias, we have to recognize the need
for a common yardstick, which test provide, in order to make meaningful
comparisons.
4. Reasons for Testing
a. Finding out about progress
• The type of test to be given will depend very much on the purpose in
testing. One should always one’s self about the real purpose of the test to be
given to the students.
• One major reason is to find out how well the students have mastered the
language areas and skills which have been taught. These tests look back at
what students have achieved and are called progress tests, the most
important kinds of tests for a teacher.
• Progress tests should produce a cluster of high marks for it is expected. But
if most students fail, something must have been wrong with the teaching,
the syllabus or the materials.
• It also acts as a safeguard against hurrying on to complete a syllabus or
textbook regardless of what the students are actually achieving--or failing to
achieve.
• A teacher has to avoid over-testing although one should try to give progress
tests regularly
• The best progress test is one which students do not recognize as a test but
see as simply an enjoyable and meaningful activity.
b. Encouraging students
• This is one important function of a teacher-made test, to encourage
students.
2|P age
• In learning a foreign language, and it is often very difficult indeed for us to
judge our own progress.
• A classroom test can help to show students the progress which they are
undoubtedly making. It can serve to show them each set of goals which they
have reached on their way to fluency.
c. Finding out about learning difficulties
• In teaching, sometimes we concentrate on following the syllabus and ignore
the needs of our students leading to failure.
• . A good diagnostic test helps us to check our students’ progress for specific
weaknesses and problems they may have encountered. One must be
systematic when designing the test and must select areas where we think
there are likely to be problems or weaknesses.
• Usually a diagnostic test forms part of another type of test especially a
classroom progress test. As such, it is useful to regard diagnostic testing as
an ongoing part of the teaching and testing process.
• When marking a diagnostic test, one should try to identify and group
together a student’s marks on particular areas of language
• Diagnostic tests of all kinds are essential if we wish to evaluate our teaching.
One can also evaluate the syllabus, the course book and the materials used.
• Whatever the reason, a classroom test can enable teachers to locate
difficulties and to plan appropriate remedial teaching.
d. Finding out about achievement
• Achievement test is also like a progress test but it is usually designed to
cover a longer period of learning than a progress test.
• Achievement tests should attempt to cover as much of the syllabus, the
contents of the test will not reflect all that has been learned.
• A test of achievement measures a student’s mastery of what should have
been taught.
• It is concerned with covering a sample which accurately represents the
contents of a syllabus or a course book.
• In setting an achievement it is necessary to base your test on a syllabus
instead of simply what you taught to maintain a certain constant standard.
• In order to resolve problems regarding topics to cover, it is necessary to
attain cooperation with colleagues for it will accommodate different
3|P age
perspectives and summarize fields to cover.
e. Placing Students
• A placement test enables us to sort students into groups according to their
language ability at the beginning of a course.
• Questions measuring general language ability can form a useful part of a
placement test.
• The most important part of the test should consist of questions directly
concerned with the specific language skills.
• A placement test should try to spread out the student’s scores as much as
possible. In this way, it’s possible to divide students into several groups
according to their various ability levels.
Selecting Students
• The purpose of this test is to compare the performances of all the candidates
and select only the best. Often refer to a selection test as being norm-
referenced.
• A good selection test will usually spread out students’ scores over most of
the scale that one is using (0%-100 %).
• A norm-referenced test is used to show how a student’s performance
compares with the performance of other students in the same group.
• Selection test are rarely set by the class teacher. They are usually set by
outside examining bodies that have washback effect (refers to the way an
exam or test influences teaching and learning in the classroom.) Teachers
will gear closely to the exam if it is good then, it is useful. If it's bad, then it
has a damaging effect on teaching
f. Finding out about proficiency
• Proficiency tests are used to measure how suitable candidates will be
performing certain task or following a specific course.
• This test has different parts which candidates can choose to do according to
their different purposes.
• In designing a proficiency test, one should pay careful attention to those
language areas and skills which the candidate will need.
• The main concern is to find out only the degree of success of someone
rather than comparing the abilities of the various candidates.
4|P age
• A criterion-referenced test is used to find out whether a student can perform
a particular task or not.
2. Achievement Tests
5|P age
Two Kinds of Achievement Tests
Examples:
• A course may aim to develop a reading ability in German, but the test may limit itself
to the vocabulary that the students are known to have met.
• A course intended to prepare students for a university study in English, but the
syllabus (and so the course and the test) may not include listening (with note taking)
to English delivered in lecture style on topics of the kind that the students will have
to deal with at a university.
Test results will fail to show what the students have achieved in terms of course
objectives. The alternative approach is to base the test content directly on the objectives
of the course.
6|P age
This in turn puts pressure on those responsible for the syllabus and for the selection of
books and materials to ensure that these are consistent with the course objectives.
Progress Achievement Tests are tests that are intended to measure the progress
that the students are making.
3. Diagnostic Assessment
• It identifies weaknesses, strengths, and problems of students’ learning.
• Diagnostic assessment can be the teacher’s basis of planning of what to do next
in the teaching and learning process.
• The teacher will be able to design classroom activities that address their actual
learning needs if he knows students’ strengths and weaknesses.
7|P age
• They are intended primarily to ascertain what learning still needs to take place.
At the level of broad language skills this is reasonably straightforward. We can
be fairly confident of our ability to create tests that will tell us that someone is
particularly weak in, say, speaking opposed to reading in a language.
• We may be able to go further, and analyze samples of a person’s performance in
writing or speaking in order to create profiles of the student’s ability with respect
to such categories as ‘grammatical accuracy’ or ‘linguistic appropriacy’. Indeed
raters of writing and oral test performance should provide feedback to the test
takers as a matter of course.
The lack of good diagnostic tests is unfortunate. They could be extremely useful for
individualized instruction or self-instruction. Learners would be shown where gaps exist in their
command of the language, and could be directed to sources of information, exemplification and
practice.
Well-written computer programs will ensure that the learner spends no more time than
is absolutely necessary to obtain the desired information, and without the need for a test
administrator. Whether or not they become generally available will depend on the willingness of
individuals to write them and of publishers to distribute them.
Placement Tests
8|P age
teaching programme. One possible exception is placement tests designed for use
by language schools, where the similarity of popular text books used in them
means that the schools’ teaching programme also tend to resemble each other.
• The placement tests that are most successful are those constructed for particular
situations. They depend on the identification of the key features at different
levels of teaching in the institution. They are tailor-made rather than bought off
the peg. This usually means that they have been produced ‘in house’. The work
that goes into their construction is rewarded by the saving in time and effort
through accurate placement.
Examples:
The tasks, and the texts that are used, should be authentic as possible.
Every effort is made to make them as realistic as possible.
Indirect Testing
Indirect testing attempts to measure the abilities that underlie the skills in
which we are interested.
Examples:
9|P age
• Developing an indirect measure of writing ability. It contains items
of the following kind where the student has to identify which of
the following underlined elements s erroneous or inappropriate in
formal standard English.
• Testing pronunciation ability by a paper and pencil test in which
student has to identify pairs of words which rhyme each other.
▪ The main appeal of indirect testing is that it seems to offer the possibility of testing a
representative sample of finite number of abilities which underlie a potentially indefinite
large number of manifestations of them.
▪ The main problem with indirect tests is that the relationship between performance on
them and performance of the skills in which we are usually more interested tends to be
rather weak in strength and uncertain in nature.
▪ Some tests are referred to as semi-direct.
Example:
Speaking tests where students respond to a tape-recorded stimuli, with their own
responses being recorded and later scored. These test are semi- direct in the sense that,
although not direct, they stimulate direct testing.
10 | P a g e
3. Norm-Referenced vs. Criterion Referenced Testing
Norm-Referenced Testing
Criterion-Referenced Testing
11 | P a g e
2. In Criterion Referenced Test, the quality of achievement of a student is determined by
the distance of his score from the criterion established.
3. In criterion referenced tests, the variability is irrelevant.
4. Criterion referenced test are used to determine the level or skill or knowledge of
individuals if they are capable or qualified to apply such skill or knowledge.
5. In criterion referenced tests, too easy or too difficult items are not removed, rather they
should be included if they truly reflect being measured.
6. In criterion referenced tests, an individual’s scores simply above or below standard or
criterion.
7. In criterion referenced tests, a pupil is not supposed to tackle a higher learning tasks if
has not passed the standard set for preceding learning task.
12 | P a g e
5. Computer adaptive testing
Computerized adaptive testing (CAT, sometimes called computer-adaptive
testing) are designed to adjust their level of difficulty—based on the
responses provided—to match the knowledge and ability of a test taker.
In most paper and pencil tests, the candidate is presented with all the items,
usually in ascending order of difficulty, and is required to respond to as many
of them as possible. This is not the most economical way of collecting
information on someone’s ability. People of high ability (in relation to the test
as a whole) will spend time responding to items that are very easy for them –
all, or nearly all, of which they will get correct. We would have been able to
predict their performance on these items from their correct response to more
difficult items. Similarly, we could predict the performance of people of low
ability on difficult items, simply by seeing their consistently incorrect
response to easy items. There is no real need for strong candidates to
attempt easy items, and no need for weak candidates to attempt difficult
items.
13 | P a g e