Professional Documents
Culture Documents
net/publication/308749113
CITATIONS READS
5 14,132
1 author:
Sahbi Hidri
Université de Tunis
29 PUBLICATIONS 180 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Measuring the construct of writing in a static vs. dynamic assessment exam among ESP learners View project
All content following this page was uploaded by Sahbi Hidri on 01 April 2021.
Suggested Readings
Coombe, C., Davidson, P., & Lloyd, D. (Eds.). (2005). Proceedings of the 7th and 8th Current
Trends in English Language Testing (CTELT) Conferences, Vol. 4. Dubai, UAE: TESOL Arabia
Publications.
Gu, L., & So, Y. (2015). Voices from stakeholders: What makes an academic English test
international? Journal of English for Academic Purposes, 18, 9–24.
Milanovic, E. (2002). Common European Framework of Reference for Languages: Learning,
teaching, assessment: Language examining and test development. Strasbourg, France: Council
of Europe.
Framing the Issue
Language testing trends are dictated by the views of language and language learn-
ing. Teaching methods and/or approaches, such as the grammar-translation
method (GTM), the audio-lingual method (ALM), pragmatic expectancy grammar
(PEG), total physical response (TPR), communicative language teaching (CLT)
require different learning skills, strategies, methodologies, teaching styles and phi-
losophies all of which have a deep impact on testing, leading to many testing tech-
niques (Spolsky, 1985). There have been major disparities between discrete-point
(1960s) vs. integrative testing (1970s). In order to grasp the unremitting develop-
ments in teaching and testing, an investigation of the relationship between lan-
guage testing and applied linguistics (AL), and whether the latter has impacted
language testing paradigms, has an overriding consideration. AL tries to find
remedies to language problems by probing into the learning aspects on the one
hand and knowing about the language on the other and it is at this level that
Making the Case
Pedagogical Implications
Example 1
Test takers hear:
Mary got dizzy and hit her head on the wall.
Then they read:
Mary got dizzy and hit/hid her head on the wall.
Testing paradigms
Discrete-point Integrative
The guessing effect for this item would be 50%. This task can be criticized for
being inauthentic, since it is very rare for test takers to be exposed to similar situ-
ations. The second testing technique is paraphrase and recognition. For instance, test
takers hear an utterance then they are instructed to select one option from a range
of options, preferably four, that best reflects the main idea of the sentence or
utterance.
Example 2
Test takers hear:
Three men broke into the house
They read
The purpose of the task is to check whether test takers are capable of discerning
the meaning of “broke into.” The four options should focus on the meaning of one/
two items of the original sentence. It is a paraphrase in a sense that the test takers
should be equipped with the ability to select the most paraphrased item that best
reflects the word(s) in the main sentence. The third testing technique is called
response evaluation. Test takers hear a question then select the most appropriate
option from a four-option test item. The tasks are administered in short decontex-
tualized sentences or utterances.
Example 3
Test takers hear
How often do you go to the cinema?
They read
a. Yes. I do.
b. Once a month.
c. No. I don’t.
d. Around 200 dollars.
The problem with dictation, unlike communicative dictation, is that test takers are
instructed to provide the exact spelling of the correct word; thus being tested on a very
narrow range of skills. Therefore, it would be safer if they are provided with longer
pauses to allow them to insert the right word. Test developers should avoid deleting
words from the very first sentences of the input. They should allow some time for test
takers to get used to the text and build some background knowledge on it. Reading or
playing longer dictation passages in one shot may impose a very heavy memory bur-
den on test takers. Rather, it should be done in meaningful units. Sometimes, there are
problems with a deleted word, as it may turn out to be a high frequency word that is
overused in the passage, such as words 12 and 13 in example 4. It remains a matter of
common sense then for the test designer to decide on what to delete.
Cloze tests are often administered in reading tests where every nth word (e.g.,
fifth, seventh, or tenth word) is deleted and test takers are asked to complete the
blanks with the missing word. Cloze tests measure a wider range of skills of
language ability than dictation.
Example 5
I have always kept wondering about cloning and its side effects on people. I have read a
lot about the reasons behind cloning and I still cannot come to grips with the real reasons
that push people to (1) ………………….. to cloning: why do people (2) …………………..
to have a replica of (3) …………………..? Why do they pay all (4) …………………..
money to be cloned or (5) ………………….. clone other people? Do they (6)
………………….. want to live forever? And (7) ………………….. do scientists invest
a lot (8) ………………….. such an enterprise? Is cloning (9) …………………..
accepted by all nations and (10) …………………..?
Apart from constituting low percentages of face validity, cloze tests depend on the
nature of deleted items, especially when they are functional words that could be
easily discerned without even being exposed to the listening or reading input. To
achieve a higher reliability index, cloze tests should consist of as many deleted
words as possible. For Heaton (1988), the number of blanks for advanced-level test
takers should be from 40 to 50. Test takers are encouraged to provide either the most
acceptable or exact word and they should not be penalized for spelling mistakes;
however, they should be penalized for grammar mistakes if the provided word
distorts meaning. No matter how discrete and/or integrative tests might be, it is
salient to take into account both approaches to test design. As invaluable classroom
testing and teaching techniques and even though they seem different, both testing
paradigms cannot circumvent the pitfalls of measuring a very narrow range of skills.
SEE ALSO: Behaviorist Learning Theory; Cloze Tests; Mentalist Learning Theory;
Reliability; Validity
References
Cai, H. (2012). Partial dictation as a measure of EFL listening proficiency: Evidence from
confirmatory factor analysis. Language Testing, 30(2), 177–99.
Suggested Readings
Brown, H. D. (2004). Language assessment: Principles and classroom practices. New York, NY:
Pearson.
Davies, A. (2003). Three heresies of language testing research. Language Testing, 20(4), 355–68.