Professional Documents
Culture Documents
3. “the reporting process used by the teacher based on any response or performance by the
learner, which gives the teacher, the learner themselves and other learners in the group,
information about that individual’s learning, progress or achievement”
(Englewood Clifs, NJ. Prentice hall. In Bachman, L. 1990. Fundamental considerations in
language testing. Oxford. Oxford University press)
A Assessment
B Evaluation
C Testing
1 2 3
3. Approaches to testing. Fill in the table with usual test formats (1-8).
1
marking
Scientific Audio-lingual structuralism Mastery of Objective,
(1950s-1960s) method separate language rise of
elements(gramma testing
r, experts
vocabulary,
phonology), no
context, separate
skills
Integrative Cognitive, Language as a Meaning and Objective
and communicative complex system communication, and
communicativ approach and a means of language in subjective,
e communication context, several using
(1980s-1990s) things tested at a marking
time scales,
specially
trained
teachers
and
markers
1. multiple choice
2. oral interview
3. writing essays
4. translation
5. cloze test
6. ‘real-life’ tasks
7. grammar analysis
8. dictation
4. Look at these comments from teachers and decide why they are giving a test in each
case. What do you think their comment tells us about what they believe about testing?
● Right! As you can’t be quiet and won’t do what I tell you, we’ll have a test!
Reason for testing:
Comments:
● They never do any homework, so I give them lots of short tests instead.
Reason for testing:
Comments:
● I haven’t got enough grades in the register for class 10B, so I’ll give them a test next
lesson.
Reason for testing:
Comments:
● What do you mean I didn’t tell you we were going to have e test? It shouldn’t
matter. You should be prepared anyway – it’s on the material we’ve just done.
Reason for testing:
2
Comments:
● Now we’ll find out what you don’t know. I’ve got a nice, tricky, little test for you!
Reason for testing:
Comments:
6. Criteria for evaluating a test. Match test characteristics (A-D) to their definitions (1-4).
2. The extent to which a test is consistent in its results (produces the same or the similar
results under the same conditions and with the same performance by students).
4. The extent to which a test measures what it is supposed to measure and nothing else (e.g.
writing ability, not background knowledge + writing ability).
A Practicality
B Reliability
C Validity
D Washback effect
1 2 3 4
3
7. Match these examples of assessment with their washback effects.
1. A progress test which concentrates on a) Good washback. Students will feel they
one out of four chapters covered in have been tested fairly.
the textbook.
2. An end of year test which b) Good washback. The information will help
concentrates on grammar and you plan your course to suit the needs of
vocabulary, even though you have your students.
done lots of speaking and listening.
3. A diagnostic test which has shown c) Bad washback. Students will feel that luck
that your class is very weak on was more important than hard work.
speaking skills.
4. An end of year test on reading, d) Bad washback. Students will feel that
listening and speaking covering a wide classwork has been a waste of time.
range of material.
8. Practicality. Types of resources in the Test Development Process. List the resources
under each category.
1. Human resources
●
●
●
●
2. Material resources
●
●
●
3. Time
●
●
●
9. The process of examination paper production. Based on the processes given below
order the following stages.
1 2 3 4 5 6 7 8 9 10
1. Writing test specifications. A test’s specifications provide the official statement about
what the test tests and how it tests it. This is the beginning and a very important
beginning to the whole testing process.
2. Item writing and moderation. Issues such as who is qualified to write test items, how
does one start writing a test item, which methods are most suitable for testing
particular abilities need to be considered. What principles and guidelines should we
follow when writing test items? Who decides whether a test item is a good or not? Who
edits and moderates the items?
3. Pretesting and analysis. Testers need to find out how difficult an item is and tests what it
is intended to test. This has to be done by piloting and by trials.
4. The training of examiners and administrators. Examiners need to become familiar with
the marking systems that they are expected to use and learn to apply them consistently.
Training needs to be ongoing especially with regards to assessing speaking and writing.
5. Monitoring examiner reliability. It is important that a candidate’s score does not depend
upon who is marking the test, nor upon the consistency of an individual marker.
Procedures can be designed to improve examiner reliability.
6. Reporting scores and setting pass marks. Decisions need to be made whether some are
more important than others. Testers have to decide whether candidates have
performed well enough to have passed the test or not.
7. Validation. This basically means to make sure that the test does indeed test what it is
supposed to test. Is the test valid?
8. Writing post-test reports. Information about the test results can be important to the
testing body organizing the test, for teachers who prepare students for the test and for
those who administer the test.
9. Developing and improving tests. Monitoring and revising tests.
1 2 3 4 5 6
11. Match the definitions below with a correct term from the box. There are more terms
than you need.
5
informal assessment validity discrete point test summative assessment indirect
test
format rubric practicality formative assessment item washback effect
evaluation reliability integrative test
12. Match the definitions below with a correct term from the box. There are more terms
that you need.