Professional Documents
Culture Documents
net/publication/350755164
CITATION READS
1 4,128
1 author:
Ciler Hatipoglu
Middle East Technical University
53 PUBLICATIONS 265 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
Catch a Tiger by the Toe: Modal Hedges in EFL Argumentative Paragraphs View project
Metadiscourse across Genres: Mapping out interactions in spoken and written discourses March 2017, CYPRUS View project
All content following this page was uploaded by Ciler Hatipoglu on 09 April 2021.
Çiler Hatipoğlu
Middle East Technical University, TURKEY
ciler@metu.edu.tr
https://orcid.org/0000-0002-7171-1673
C H A P T E R VI
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 121
Pre-reading Questions
7. Can ‘speaking’ be divided into types? If “YES”, list the types of speak-
ing that you know.
(a) Compare your answers with your classmates and make a list of all
the types of speaking mentioned in class.
(b) In pairs, think of criteria that you could use to place these types of
speaking in different groups.
122 LANGUAGE ASSESSMENT AND TEST PREPARATION
Pre-reading Questions
10. Form groups and try to define the following terms without using any
extra sources:
(a) speaking vs writing
(b) testing vs assessment of speaking
(c) validity of a speaking test
(d) reliability of a speaking test
(e) practicality
(f ) rubrics
As a class, compare the definitions written by the different groups.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 123
Introduction
It is easy to take spoken language for granted – for most of us it is perhaps our most
natural form of expression. However, when learning a foreign language, it is the spoken
language that is often found the most challenging, and often frustrating, perhaps
precisely because speech is so natural for us (Bradshaw, 2020, p. 212).
Speaking is one of the most natural and one of the most crucial language skills in
our lives. It is the most needed skill in our everyday interactions, and the way we
speak reveals our identities and views of the world (Hatipoğlu, 2017b). Speaking is
also the “most complex and demanding of all human mental operations” (Taylor,
2011, p. 70). In today’s globalised world and with the technological advances and
economic globalisation, differently from the previous centuries when speaking in
the foreign language was either ignored or deemed with secondary importance,
the ability to speak in a foreign language is a “highly coveted skill and a source
of cultural capital” (Isaacs, 2016, p. 131). Speaking skills are also “often considered
the most important part of an EFL course, and yet the difficulties in testing oral
skills frequently lead teachers into using inadequate oral tests or even not testing
speaking skills at all” (Knight, 1992, p. 294). The difficulty related to the assessment
of speaking, according to testing experts, stems from the following distinctive
characteristics of the skill:
the most. If to all these, issues related to assessing speaking in a valid, reliable
and practical manner are added, then it would not be difficult to understand why
many of the well-known international language tests (e.g., TOEFL) and editions
of prominent books discussing “current issues” in assessing second language
proficiency (e.g., Lowe & Stansfield, 1988) until very recently did not have sections
focusing on speaking.
of comprehension since “speakers react to each other, take turns or overlap each
other, ask for clarification, fill in gaps in their interlocutor’s utterances and, finally,
produce the text of their joint speech” (Hatipoğlu, 2017b, p. 121). Because of all
those reasons, it is difficult to separate the speaking skill from listening skill.
Achieving the objectives listed by Hughes (2003) becomes more difficult with the
increase in the level of freedom of the exam tasks. With more open-ended tasks,
test takers have the liberty to respond with a wider variety of words and structures
not anticipated by the test writers. To avoid problems that might damage both the
reliability and validity of the exam, test writers should prepare detailed analytical
rubrics where every type and piece of information is allocated an individual score.
Speaking rubrics should be as detailed as possible and, depending on their aims of
the exam, should allocate points for pronunciation, fluency, grammar, vocabulary,
discoursal elements, and level of pragmatic appropriacy.
Number of test-takers
If the tasks are well designed, it is possible to assess the listening, reading, and
writing skills of a large number of students in a relatively short period of time in a
valid and reliable manner. However, due to the speaking skill’s real-time interactive
nature, assessing it becomes notoriously difficult when the group taking the test
is big, and the time allocated for the assessment process is limited. To be done
successfully, enough time, meticulous planning, and the involvement of a large
number of well-trained testers are required. Therefore, test administrators should
126 LANGUAGE ASSESSMENT AND TEST PREPARATION
plan and schedule their speaking exams accordingly. With the more significant
emphasis on speaking in the new educational system, nowadays, the pressure on
assessment experts to find quick, reliable and valid ways to test speaking is bigger
than ever before.
Theoretical Background
In this part of the Chapter, the developments in the last two centuries or so related
to how speaking in the foreign language were conceptualised and how those
changes have been reflected in the construction of speaking tests over the years
are reviewed.
Up until the end of the 19th C, when the predominant method of language teaching
was the Grammar-Translation Method (GMT), there was “little consideration
given to the need for a speaking test of any type” (Bradshaw, 2020, p. 214).
Originally, the aim of the proponents of GMT was to adapt the scholastic study of
the teaching of Latin and Greek “to the circumstances and requirements of school
students” (Howatt, 1984, p. 131). To carve out a place for GMT in the teaching of
modern languages, classes focused on grammar, reading and translation. This
training enabled learners to access the Classical cultures’ literature in their original
form and translate it, while speaking was defined mostly as reading aloud. During
that period, foreign language tests included mainly translation as a device for
assessing grammar and vocabulary (Hatipoğlu, 2017a; Hatipoğlu & Erçetin, 2016).
This situation started to change towards the end of the 19th C when the Reform
Movement in Germany began (around 1882). The new movement took as its
focus the recently introduced science of phonetics (i.e., in 1886, the International
Phonetics Association was founded) and argued that teaching foreign languages
should “begin with the spoken language” (Sweet, 1899, p. 49). This movement
led to the development of a new pedagogical approach rooted in the spoken
language – the Direct Method (DM). DM supporters suggested more of an
immersion approach to language learning. According to them, rather than
learning L2, it should be acquired, and among all skills, “the acquisition of the
spoken language was seen as the main purpose” (Bradshaw, 2020, p. 2015). This
change in perspective started to be slowly reflected in the structure and content of
foreign language proficiency exams as well. The Certificate of Proficiency in English
(CPE) (now known as Cambridge English Proficiency), which was first administered
in 1913 and lasted for 12 hours, included both Written (11 hours) and Oral (1 hour)
sections (See Table 1).
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 127
Table 1
1913 CPE Examination Parts (Adapted from Weir, 2013, p. 3)
Main Parts Sub-sections Time (hours)
The Written Section of the exam had five sub-sections and lasted for 11 hours,
while the Oral Section comprised two sub-parts: Dictation (0.5 hours) and Reading
and Conversation (0.5 hours) and was just an hour-long (Weir, 2013). As can be
seen from the content of the exam and the weights of each of the sub-sections,
at that time, the effect of the GTM was still strongly felt. The Written section of
the exam (including English Phonetics) focused on form but, for the first time,
“attention was clearly paid to active language as well” (Weir, 2013, p. 3).
The next development in the field of linguistics that affected how foreign languages
were taught and assessed was the birth of ‘structural linguistics’ (i.e., an approach
aiming to uncover the rules that underlie language and govern how it functions)
in the 1920s. Structural linguistics had a long-lasting (from the 1920s well into the
1970s) but varied effects on the teaching and assessing foreign languages on both
sides of the Atlantic. In the USA, where the leading structuralists were Bloomfield
(1926, 1933) and Fries (1945), the primary focus of foreign language classes and
exams was the testing of isolated language structures and groups of words using
objectively scored items (e.g., multiple-choice items, True/False questions). It is not
surprising then that the Test of English as a Foreign Language (TOEFL), which
was first administered in 1964, consisted of only reading, listening, structure and
vocabulary sections for the following 20 years. Even when a speaking section was
added to TOEFL, it was optional for quite a long period of time. Factors such as
reliability, practicality and lack of experts (i.e., variables that are still valid today)
were cited as the main reasons for the non-adoption of a speaking section by
TOEFL writers. In Britain, on the other hand, where the leading linguist was Palmer
(1921a, 1921b), the structuralist principles and content were “usually combined
with the direct method with its emphasis on the spoken language” (Weir, 2013,
p. 3). Word lists and grammatical structures were tested, but exam writers tried
to locate them in relevant and interesting situations to make their meaning clear
leading to the introduction of the “situational approach” in the UK.
128 LANGUAGE ASSESSMENT AND TEST PREPARATION
The gradual shift away from structuralism and towards the Communicative
Approaches to Language Teaching and Testing started at the beginning
of the 1980s and flourished during the 1990s. The communicative approach
underlined the importance of meaningful tasks and authentic materials, made
use of the rapidly developing field of sociolinguistics and gave birth to the
notional-functional syllabus. Because of these, differently from the previous
periods, a considerable amount of teaching and learning of foreign languages
began to be done orally. Experts, language teachers, and students became
interested in language as a means of communication rather than its structures
per se. Consequently, developing learners’ speaking proficiency started to be held
high among most foreign language programs’ objectives. These changed views
“forced language testers out of their narrow conception of language ability as
an isolated trait, and required them to take into consideration the discoursal and
sociolinguistic aspects of language use, as well as the context in which it takes
place” (Bachman, 2000, p. 3). Finally speaking has come to the fore and exam
tasks began to be constructed around this skill. Nowadays, communicative test
designers try to ensure that their exams involve ‘performance’ (i.e., test takers are
engaged in the act of communication) and are real-life like/authentic (i.e., the
social roles assumed by the test takers during the exam are similar to the ones
they are likely to undertake in real-world settings and the test takers know and
understand the communicative purpose of the task) (Fulcher, 1999; Harding, 2014;
McNamara, 2000). Therefore, now the aim of the speaking sections of exams such
as TOEFL and IELTS is to “measure test takers’ ability to actually communicate
in English about what they have read and heard” (Bridgeman et al., 2011, p. 91-
92). That is, they try to approximate as closely as possible the tasks students are
expected to perform in their English medium institutions.
Procedure:
The test taker who is given a pair of pictures of contrasting words hears: Listen and choose the correct
picture
The test taker listens and chooses a picture.
Example 2: Imitative Assessment Task Examples: PRODUCTION (adapted from Hatipoğlu, 2017b)
Procedure:
The test taker hears: Repeat after me
Test taker repeats the stimulus
(i) (ia) permit /pəˈmɪt/ [pause], Contrasting stress patterns leading to meaning and/or word
permit /ˈpɜː.mɪt/ [pause] category change
Note 1:
(ib) address /ˈæd.res/ [pause], - permit /pəˈmɪt/ (Verb): to allow something
address /əˈdres/ [pause] - permit /ˈpɜː.mɪt/ (Noun): an official document that allows
you to do something or go somewhere
Note 2:
- address /ˈæd.res/ (Noun): a place where someone lives
- address /əˈdres/ (Verb): to speak or write to someone
(iii) (iiia) Are they busy? Such exercises are useful in focusing students’ attention on
(iiib) Where did you go the existing intonational patterns in the target language (e.g.,
yesterday? Yes/No questions, Wh-questions, tag questions) as well as on
(iiic) They are students, aren’t the differences between the first and target languages of the
they? students.
One of the biggest problems related to testing speaking is its scoring since various
factors related to the administration of the exam (e.g., a noisy environment), the
examiner (e.g., his/her background, (lack of ) knowledge of other foreign languages,
acquaintance with the test taker’s accent) and/or the examinee (e.g., being shy/
introvert, having other problems related to speaking) can affect the evaluation
process. To make the assessment as reliable as possible, the scoring criteria should
be clearly defined and explained in detail while keeping in mind the goals of the
task. Brown and Abeywickrama (2019, p. 161) state that with imitative tasks, it
would be more practical to adopt a two- or three-point system for each response.
Scoring scale for repetition tasks
2 Acceptable pronunciation
When the stretch of language that test-takers are asked to repeat gets longer, the
likelihood of making mistakes increases, and the evaluation of the oral production
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 133
becomes a difficult task. With longer utterances, it gets more difficult for the
examiners to isolate and describe the patterns that they are interested in. To
avoid such validity problems (e.g. instead of focusing on the intended pattern,
the examiner’s evaluation is affected by other more prominent mistakes related
to the pronunciation of the repeated phrase/sentence), Brown and Abeywickrama
(2019) advice testers to “score only the criterion of the task” (p. 161). That is, in
instances where the focus is on meaning and word category changing stress
patterns (e.g., I am content with the content of this blog), test takers should be
awarded points for the correct stress patterns in the focus words, regardless of
any other problems in the utterance. If the focus is on stress patterns in the target
language, the examiners should refrain from penalising students for mistakes
related to pronunciation or intonation.
Read-aloud Tasks
The read-aloud assessment tasks are the most controlled and the least
interactional INATs. As the name suggests, the test takers, depending on their age
and proficiency level, are asked to read out loud texts with various lengths (e.g.,
from a single sentence up to 150 words long texts). These ‘diagnostic passages’
(Prator & Robinett, 1985, p. ix), are constructed according to the objectives of
the test. That is, they include words and types of structures that enable testers
to elicit information related to the phonological (e.g., problematic consonants
and consonant clusters; vowels, diphthongs and triphthongs; difficult or distinct
stress and intonation patterns), structural (e.g., regular interrogative sentences vs
question tags; irregular nouns and verbs) and lexical (e.g., homophones, synonyms,
antonyms) knowledge of test takers’ in thetarget language (TL)
134 LANGUAGE ASSESSMENT AND TEST PREPARATION
Until recently, general ELT methodology literature did not recommend the read-
aloud practice as it was not seen as ‘genuinely communicative’ (e.g., Broughton
et al., 1980). Assessment experts also argued that it was not a good technique
to test reading in a foreign language (Heaton, 1990; Hughes, 2003). However, a
significant number of studies conducted in the last two decades revealed the
benefits of using such assessment tasks and started calling for the reappraisal
of the technique (Gibson, 2008; Huang, 2010; Seo, 2014; Stroh, 2012; Supraba et
al., 2020; Weir & Wu, 2006). They have also shown where, why and how language
teachers, learners and testers can benefit from this practice. Below is a summary of
the most frequently mentioned benefits/advantages of using read-aloud activities
as speaking assessment tasks:
Versatile
Read-aloud tasks are versatile (Gibson, 2008). With well designed ‘diagnostic texts’,
read-aloud tasks can be used to assess the language abilities of learners with
different ages (young learners, adults) and proficiency levels (A1, A2, B1 etc.).
Reliable
Read-aloud tasks are easy to administer and score as all of the test takers’
production is controlled (Brown, 2004; Weir & Wu, 2006). Since all of the students
are asked to read the same text, test developers will not have to worry about the
parallel-test reliability either (Weir & Wu, 2006), and they will have a firm ground to
compare students’ performances to each other.
accelerate learning, Sadler, 1998, p. 77). After introducing new sounds, stress patterns,
structures or lexical groups, students might be asked to read aloud assigned texts
and record themselves in their own time. These recordings can be submitted to
the teacher for analysis and/or could also be employed as self-assessment tools
(i.e., students listen to themselves and evaluate their performance). Such exercises
can develop students’ pronunciation, intonation, chunking of a text into sense
groups and comprehension as well as their memorisation of new vocabulary items.
According to Prator and Robinett (1985), “reading aloud is clearly one of the most
effective mechanisms for learning to monitor one’s own pronunciation” (p. xxvi).
By hearing and analysing their own speech, students will gain self-confidence,
and this eventually will lead to “autonomous learning and may help some anxious
students to feel more able to speak” (Gibson, 2008, p. 29). It can also help students
to move between different levels of comprehension of a text as with more practice
in reading and speaking they become increasingly engrossed in the meaning of
what they are reading and saying.
The read aloud activities in the tests (especially the ones prepared for progress and final
achievement tests) should match the curriculum goals. When selecting or preparing
‘diagnostic paragraphs’, preference should be given to texts written in simple modern
language (almost) free from unusual private names (e.g., person, town, place) and
dialectal peculiarities. The font of the text should be big enough, and there should
be enough space between the lines.Numbering every sentence will help testers take
notes and follow test takers’ progress (see Example 3).
Procedure:
The test taker hears: This is the passage that you will be asked to read aloud. First, you will have 2
minutes to look over the text and read it silently. Then, you will be asked to read the passage aloud
with attention to pronunciation, intonation, and flow of delivery. The reading should be done at
normal speed, and should sound as much like natural speech/conversation as possible. Do you
have any questions?
Test takers go over the passage and when the time is up, they start reading the diagnostic
passage out loud.
(1) When a student from another country comes to study in the United States, he has to find
out for himself the answers to many questions, and he has many problems to think about. (2)
Where should he live? (3) Would it be better if he looked for a private room off campus or if he
stayed in a dormitory? (4) Should he spend all of his time just studying? (5) Shouldn’t he try to
take advantage of the many social and cultural activities which are offered? (6) At first, it is not
easy for him to be casual in dress, informal in manner, and confident in speech.
Depending on the test-takers’ age, proficiency level, and exam goals, the evaluation
of their performance on the read-aloud activities can be done using a very detailed
evaluation rubric (see Example 4) or a more general one (see Example 5).
136 LANGUAGE ASSESSMENT AND TEST PREPARATION
Example 4: Checklist of problems for a Read-aloud test task (adapted from Prator &
Robinett, 1985, p. xii-xiv)
EVALUATION/
CRITERIA POINTS TO CHECK
NOTES
VI. General
Comments
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 137
Example 5: Read-aloud test task -Rating Scale (adapted from Weir & Wu, 2006, p. 196)
EVALUATION/
RATING INTERPRETATION
NOTES
When all of the students in the tested group are given the same words, phrases or
sentences to translate, it is possible to both limit the output and to compare the
test takers among each other against the same criteria.
(ii) An integrative assessment tool
To be able to translate a phrase or a sentence from L1 into L2 or vice versa,
students need to mobilise a diverse set of interdisciplinary skills (e.g., problem-
solving, creative and critical thinking) and types of knowledge (e.g., metalinguistic
competence, grammar, vocabulary, pronunciation skills) (Laviosa, 2014; Presas,
2000). What is more, according to Sun and Cheng (2013), and House (2015),
translation is unique in scaffolding students’ learning via their first language.
(iii) A practice that creates multilingual identities and facilitates communication
across cultures (House, 2015; Laviosa, 2014).
A good translation is fair, balanced, and grounded. It is not only a linguistic act,
where forms from one language are just converted into the other. It is also an
act of communication across cultures (House, 2015, p. 3). Eugene Nida (1964),
one of the leading figures of translation theory, argues that translation always
involves L1, L2 and the cultures of the two languages since language and
culture cannot be neatly separated. According to him, translation is one of the
major means of constructing representations of other cultures since languages
are culturally embedded. Languages serve to express and shape cultural reality,
and the “meanings of linguistic units can only be understood when considered
together with the cultural contexts in which they arise, and in which they are used.
In translation, therefore, not only two languages, but also two cultures invariably
come into contact” (House, 2015, p. 4). When viewed from that perspective,
translation is a form of intercultural communication.
Similarly to the read-aloud tasks, the translation assessment tasks are non-
interactional and allow for different levels of control of the test taker’s output.
In translation assessment tasks, instead of a picture or a written stimulus, test-
takers, depending on their age and proficiency level, are given a word, phrase, or
sentence in their native language and are asked to translate it into the TL. The test
takers could be asked to translate the stimulus instantly or might be allowed 1-2
minutes of thinking and organisation time before orally producing a translation
of somewhat longer texts. Example 6 shows instances where the test-takers are
asked to translate words (Part A), phrases (Part B) and sentences (Part C).
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 139
Part B
Examiner: lezzetli bir muz/ iki ağaç sandalye / koyu kırmızı bir elbise
/ kırılan omuzum / kızgın olan taraftarlar
Tes taker (expected answer): a delicious banana / two wooden chairs/ a dark red dress
/ my broken shoulder / fans who are angry
Part C
Examiner: Çocuklar bu oyunu seviyor. / Bu hoteldeki yemekler
lezzetli değil. / Bu bilet kaç para? / Partiye gelecek misin?
Test taker (expected answer): Kids/Children love this game. / Food in this hotel is not
tasty. / How much is this ticket? / Will you come to the
party?
The quality of the translation done by test takers can be based on a number of
different criteria. Example 7 shows a Rubric where a more holistic approach
to speaking is adopted. Pronunciation, word and sentence stress, intonation,
vocabulary usage and grammar are evaluated together.
140 LANGUAGE ASSESSMENT AND TEST PREPARATION
Example 7: Read-aloud test task -Rating Scale (adapted from Angelelli, 2009, p. 33):
EVALUATION/
RATING INTERPRETATION
NOTES
(a) Versatile
They can be utilised with students of any age (e.g., young learners and adults)
and proficiency level (e.g., students with limited speaking skills and students
approaching intermediate level).
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 141
(b) Reliable
As all students are asked to repeat the same prompts, there is a stable criterion
against which all of the students are compared (i.e., there is a consistency with
which the assessment measures what it aims to measure, Popham, 2018).
(c) Practical
In schools with language labs or big classrooms with the required equipment, big
groups of test-takers can be tested using DRT.
With DRT, the level of difficulty and authenticity, and the degree of creativity
required from the test-takers can be manipulated depending on the goals of
the exam (see Example 8). Examples 8a-8c are all more artificial, but they have
different levels of difficulty as they go from the one requiring a simple repetition
of a part of the original sentence (8a) to prompts that get longer (8c) and require
one (8c) or more transformations (8b). Examples 8d-8g are more authentic as they
are related to the test takers’ lives (8d), imitate message giving, which is frequently
encountered in real life (8e, 8f ), and require the analysis of the contexts within
which the conversation is taking place (8e, 8g).
142 LANGUAGE ASSESSMENT AND TEST PREPARATION
8a. Tell me they finished their They finished their *Artificial and very basic
homework. homework. *No transformation is required;
simple repetition of the
last part of the examiner’s
prompt.
8b. Tell me that you aren’t taller than I am not taller than *Artificial
Mary. Mary. *Modification required:
Change of the subject (‘you’ to
‘I’) and contraction (‘you aren’t’
to I am not)
8c. Tell me that you want to be a part I want to be a part *Artificial if not related to the
of the team that will visit Italy for of the team that will test taker’s real situation.
ground-breaking research next visit Italy for ground- *More difficult as the students
winter. breaking research will have to remember a long
next winter. prompt.
*Modification required:
Change of subject (‘you’ to ‘I’)
8e. There are three people in the (i) Mrs Taylor says *More authentic as in real
exam room: the test-taker, the that she can visit you life, we are required to give
examiner (Mrs Taylor) and the on Wednesday. messages to other people.
interlocutor (John): (ii) Hey John/ *Modification required:
Examiner: Tell him I can visit him Pardon me, John. Change of subject and
on Wednesday. Mrs Taylor says that object (as in i)
she can visit you on *If the aim is to assess the
Wednesday. pragmatic development
of the students (i.e., giving
socially more appropriate
messages), answer (ii) is
expected.
*Answer (ii) requires a more
detailed analysis of the
context and more creativity
on the part of the test taker.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 143
8f. There are three people in the (i) Where is the *Authentic as in real life, we
exam room: the test-taker, the chemistry lab? are required to request
examiner (Mr Wilson) and the (ii) Pardon me, something as speakers/
interlocutor (Linda): Linda. Where is the representatives of a group.
Examiner: Ask her where the chemistry lab? *Modification required:
chemistry lab is. (iii) Good morning. Change of subject and
Could you tell me object (as in ii)
where the chemistry *If the aim is to assess the
lab is, please? pragmatic development
of the students (i.e., giving
socially more appropriate
messages), answers (ii) or (iii)
can be expected depending
on the level of formality of
the interaction.
*Answers (ii) and (iii) require
analysis of the situation,
initiative, and creativity on
the test taker part.
8g. It is 5:30 in the afternoon. You (i) Sorry. I didn’t see *Authentic
are in a hurry on your way you there. *As in real life, test takers have
home from the department (ii) I’m really sorry. more information about
store because you are expecting Are you alright? the context in which the
important guests. While you are interaction is taking place.
getting out of the department *This exercise demands more
store, you bump into a well- critical and creative thinking
dressed elderly lady. It is your as well as more initiative on
fault, and the lady seems really the part of the test takers.
upset. What would you say *To be successful in such
in that situation? (Hatipoğlu, tasks, students would need
2009). practice with commutative
activities in class and
information about the target
culture
about the appropriate utterances to fill in the gaps in the exchange. Then, the test
administrator, an interlocutor, or the tape reads aloud/produces the non-deleted
lines in the dialogue while the test-taker responds.
Depending on the level of proficiency of the test takers (lower vs higher) and on
how much control the test writers want to have on the output produced by the test
takers two types of DCT can be created. In Type 1 (i.e., the guided DCT), students
are informed of what is expected from them. The expected answer is summarised
in brackets (see Example 9), while in Type 2, students see the prompts/questions
but have more freedom to construct the responses the mselves (see Example 10).
Example 9: Guided Dialogue completion task (Type 1): Focus: Speech Acts
Instructions
First, read the dialogue carefully and think about appropriate answers. You have X minutes
to prepare.
Now, listen to the statements carefully and respond with the appropriate lines.
At the local shop
Shop assistant: Hi there. Can I help you?
(You want to buy a present for your friend’s birthday)
You:
Shop assistant: Oh, we have many of those. What colour and what size are you looking for?
(You think that he/she likes darker colours and he/she is S or M size.)
You:
Shop assistant: Oh, ok. What do you think about this T-shirt and this shirt?
(Tell the shop assistant that you like the T-shirt and ask its price)
You:
Interviewer: What have you learned about the city where you are going?
Candidate:
(i) Reliability
Examiners read the same prompts, and all test takers are assessed against identical
criteria. Since test takers can see the dialogues in written format, they have
relatively more time to process and anticipate the exam prompts. In turn, this
removes or, in some instances, at least decreases the “ambiguity created by aural
misunderstanding” (Brown, 2004, p. 151).
(ii) Focus students’ attention on certain aspects of the target spoken language
When created well, DCT are very useful assessment tools in contexts where English
is taught as a foreign language and grammar, and reading are the foci of education
(Heaton, 1990). Such tasks help teachers focus learners’ attention on specific
aspects of the target spoken language (e.g., irregular nouns/verbs; collocations;
speech acts).
146 LANGUAGE ASSESSMENT AND TEST PREPARATION
(iii) The examiners can (moderately) control the output of the test takers
By creating dialogues focusing on specific characteristics of the target language,
(and by providing guides related to the expected answers), test writers limit the
acceptable responses (e.g., in Example 10, by asking Question 2 in the past tense,
test writers limit the possible answers for the test-takers).
Expected answer: sum /sʌm/ vs. thumb /θʌm/ Expected answer: cap /kæp/ vs. cup /kʌp/
Instructions
The test taker sees the pictures showing different actions.
Test taker hears [test administrator points to each picture in succession]:
Situation 1: If the assessed students have a lower level of proficiency, they can say: What is the
monkey doing in 1/here?
Situation 2: When the test is constructed for students with a higher level of proficiency, and
the aim is to elicit longer stretches of language, the questions might be: Tell me
about picture 1. What do you see? What is happening here?
Test taker: [Gives the answer required by the question.]
Scoring: The task is achieved if the candidates can produce comprehensible and grammatically
correct utterances fulfilling the requirements of the task.
(1) pulling (a rope) (1) This is a picture of a cute monkey that looks
worried. The monkey is pulling a rope, and it
looks strained.
(2) throwing a ball (2) This is Chimp. He is a young monkey that likes
throwing balls. In the picture, we see Chimp
doing his favourite activity. He is throwing
his favourite ball. Chimp is very excited about
what he is doing. We can see that from his
facial expression.
Diagrams and Figures coming from authentic research papers can also be used in
speaking tests to elicit limited oral output. In Example 13, a Figure from a study
(Hatipoğlu et al., 2020) comparing the use of crowdsourcing platforms for learning
foreign languages in Turkey, Poland, The Republic of North Macedonia and, Bosnia
and Herzegovina is given. By exposing students to such authentic materials, test-
writers can create experiences of real contexts in the target language. This gives
language learners self-confidence (i.e., they know that they are able to understand,
analyse and discuss authentic materials) and motivates them to continue learning
the foreign language, which, in the end, leads to the desire to continue learning
autonomously (Umirova, 2020).
Example 13: Picture-cued assessment tasks: Describing information given in a diagram/
figure
Instructions
The test takers are given the diagram to study for a few minutes. Then, they are asked to
answer the questions they hear/read.
Test administrator: You have one minute to study Figure 1. The given numbers represent
percentages. [When the time is up] Look at Figure 1 and answer the questions that you
hear.
Figure 1: Which languages have you learned while using crowdsourcing platforms? (from
Hatipoğlu et al., 2020, p. 86)
More open-ended responses can also be elicited when picture-cued tasks are
used. Students can be asked to describe a picture or the people in a picture (i.e.,
assessment of descriptive adjectives, physical appearance, prepositions, colours).
Test takers can be asked to talk about the relationships between the people in
the picture and/or their feelings as well (see Example 14). To complete such tasks,
students will have to use a wide variety of grammatical and semantic units.
Example 14: Picture-cued assessment tasks: Describing people, relationships and feelings
Instructions
The test takers are given a picture and told that they have 2 minutes to study it. When the time
is up, they are asked to describe in detail the place and the people they see in the picture.
Then they are instructed to look at the picture again and guess the relationship between the
people in the picture or what they might be discussing. Students can be given 2-3 minutes to
complete the task.
Test administrator:
(a) Look at the picture and describe the place and the people you see in the picture in
detail.
(b) What do you think the people in the picture are discussing? Why do you think so?
(c) What could be the relationship between them? Why do you think so?
4. identical with the ones used in class 4. printed with high quality on quality
(i.e., The use of the same pictures may paper (e.g., problems related to the
lead to just parroting of information quality of the pictures may lead to
students remember from class.) invalid exam results)
The writing, ordering and combining of the questions in the ‘Question and Answer’
tasks require careful planning and precise implementation as the exam time is
limited, and every question should serve a specific purpose. Depending on the
aims of the exam, the questions should elicit as detailed and accurate information
as possible about the phonetic, grammatical, semantic, discourse, pragmatic
knowledge of the students in the foreign language.
Paraphrasing
Paraphrasing is “restating a passage text using different words or sometimes even
rearranging the sentence structure” (Chen et at., 2015, p. 22). It is a vital skill that
we frequently employ at work, at school, or in our everyday lives. We paraphrase
because we cannot remember the exact words (i.e., lexical gaps), want to simplify
the topic, and make it more comprehensible (e.g., when we have to teach a
grammatical structure to young learners), need to summarize a long story when
we do not have much time for lengthy explanations, or even try to make something
more acceptable for the audience. It is also a critical skill for language learners to
be successful in their academic lives and survive in everyday interactions. That is
why paraphrasing is a skill that should be taught and closely assessed in foreign
language classes. Research shows, however, that this is not the case and that many
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 155
foreign language learners struggle when faced with the task (Frodesen, 2002; Sun,
2009) mainly due to insufficient lexical knowledge and lack of training (McInnis,
2009; Milićević & Tsedryk, 2011).
To be able to use this technique (i.e., paraphrasing), test developers have two
options. They can combine speaking with either listening or reading (Madsen,
1983). Combining speaking with listening brings more freedom for the test
developers as those tasks can be used with almost all groups of students, even
with illiterate ones (e.g., very young and young learners). The technique is simple.
The teacher/tester reads a sentence/ story/ joke etc., and asks the student to tell it
in their own words. Example 18 is a paraphrasing task appropriate for intermediate
level students where students hear a separate sentence each time, and they are
asked to paraphrase it. To control the output and check whether the students can
paraphrase the given statements in a wanted manner, test developers can provide
the beginning of the expected utterance.
When speaking is combined with reading (i.e., when the students are literate and
older), teachers can give students a story to read silently in one room and then, in
a separate room, they can be asked to orally paraphrase the story individually to
the teacher (see Example 19). To reduce the burden on students’ memory, teachers
might keep the stories short or might prepare drawings reminding students of the
main points in the story (see Example 20).
156 LANGUAGE ASSESSMENT AND TEST PREPARATION
Instruction
Read the story on this handout carefully. When you finish reading, you will be asked to tell
me the story in your own words.
I found a frog
By T. Albert
Even though I have grandchildren of my now, it seems like it was only yesterday when I
returned home from school to find a frog on my bed. My mother chuckled when I yelled out.
“I found a frog on my bed.”
Now, she knew that I would eventually find one, but she let me discover a wonder of Nature
that many people miss. I am glad she did.
You see, a little earlier that spring, when I was 6 years old; I saw some little black fish in a pond.
Since I did not have any pets, I went home and asked my mother if I could have one. She
agreed. She gave me a bowl and told me to catch a few. Off I went. There were so many that
they were easy to catch. I filled the bowl and ran home. When my mom looked at the fish, she
said with a big smile,
“Tadpoles. Wow! You are in for a surprise.”
A week or so later, when I got up, I was amazed. There were more changes. My fish did not
have tails, their legs were bigger, and they did not look like the little black fish I had caught
earlier in the spring.
“Mom,” I yelled with excitement. “Come here, my fish are really different.”
She came to my room, looked, smiled and told me that a surprise was very close.
That day, when I returned home from school, is when I yelled out,
“I found a frog on my bed.”
“Surprise,” yelled mon. “You watched a miracle right before your eyes. A fish changed into a
frog. Now you had better catch and take him and the other almost frogs back to the pond. I
do not need 50 more surprises tomorrow morning.”
Off I went…
Example 20: Pictures used to remind test-takers of the main parts of the story.
Picture 1 Picture 2
Picture 3 Picture 4
(Source: https://cdn.shopify.com/s/files/1/2081/8163/files/022-I-FOUND-A-FROG-Free-
Childrens-Book-By-Monkey-Pen.pdf?v=1589890638)
usually difficult for language learners to master (Hatipoğlu, 2009, 2012, 2016a,
2017c). For such speech to sound natural and more authentic, the inclusion
of proverbs, colloquial language, slang, or ellipses is expected (Can Daşkın &
Hatipoğlu, 2019c; Hatipoğlu & Can Daşkın, 2020). The aims of the transactional
interactions, on the other hand, are to inform our interlocutors and to ‘get what
we need to get’. Conversations, where somebody is informed about the opening
hours of the supermarket, where the conference is going to be and what is the
abstract submission deadline, who was appointed as the new project manager
and what are the team members’ new responsibilities, which topics are included in
the midterm exam and what is the weight of each of the topics etc., are examples
of transactional interactions. In such interactions, speakers usually say not what
they want to but “what they have to” (Harley, 1993, p. 22).
Assessment tasks such as oral interviews, role-plays, games, and discussions
are classified as interactive. In this Chapter, we are going to discuss the most
common and the most widely used one of these tasks, i.e., the oral interview.
Oral Interviews
Oral interviews can include description tasks, narrative tasks, instruction
tasks, comparing and contrasting tasks, explaining and predicting tasks, and
decision tasks. Among these, the chapter focuses on the most frequently used
ones: The description and narrative interviews.
This task is authentic since every test taker will provide a different description,
and the examiner will ask genuine questions if more information is needed. These
types of exercises are particularly suitable if the aim is to know how well test takers
can describe something they know. What is more, to avoid the negative effects
of students comparing notes after the exam, every test-taker can be asked to
describe a different object (e.g., something that s/he mentions in her/his talk) as
long as there is ‘comparable parallelism’ between the tasks given to all students
(e.g., all students are asked to provide comparable descriptions of people, places,
animals).
The main criterion for scoring in such tasks is whether or not the listener can
picture what is being described by the test-taker.
takers to talk about events that happened in their lives. Such tasks are authentic
and relevant to the students and might encourage some students to talk more.
However, “while stories about ‘what happened to me’ are very common in real
life, they usually belong to social chatting, which is difficult to replicate in a test
situation” (Luoma, 2004, p. 144). That is, the stories that test-takers tell can be
very different from each other (e.g., content, place, time frame), and this might
make the comparison between them very difficult. Some personal stories might
require sharing sad, anxiety-causing events or revealing embarrassing details,
which might dishearten some of the test-takers and make introvert, shy students
go even quieter. Conversely, without any preparation, students might not be able
to think of a story interesting enough to tell, which again might prevent them
from exhibiting their real/best skills in speaking. Because of all these, picture-
based sequences are more common in speaking exams. Sequences provide
structure and a common ground for comparison and evaluation of the test takers.
However, while selecting them, test designers have to be really careful as for a
picture sequence to work well, it should both generate enough talk and give test
takers an opportunity to show what they know concerning all criteria deemed
important in the exam (e.g., various grammatical structures, vocabulary, register,
cohesive devices etc.). They should also enable students to “show their control of
the essential features of narratives: setting the scene, identifying the characters
and referring to them consistently, identifying the main events, and telling them in
a coherent sequence” (Luoma, 2004, p. 144). To make sure that picture sequences
will elicit the variety of language targeted in the exam, they should be tried out
with similar groups of students or even among the teachers who were not involved
in the selection of the pictures and the preparation of the exam.
Example 23 is a describe and rearrange exercise, which can be used both as a pair
and group task. Here, each student is given a different set of pictures in mixed
order, and the students have to describe, discuss and put them in the correct order
as a pair/group.
Step 1: Students are given 1 minute to look at their own pictures and to prepare to
describe the pictures without showing them to the other group members.
Step 2: Each student describes his/her pictures, while others are asked to take
(mental) notes.
Step 3: Students compare notes and decide on the appropriate sequence.
Step 4: Each student puts down the picture(s) they hold in the agreed order.
Step 5: Students are given 1 minute to comment and to change the order of the
pictures if needed.
Step 6: Students tell the story following the final agreed-upon order. Each student
talks about his/her pictures.
162 LANGUAGE ASSESSMENT AND TEST PREPARATION
Picture 3 Picture 4
Picture 5 Picture 6
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 163
Oral Presentations
Oral presentations are a part of our academic, professional and everyday lives.
Young children, students, teachers, researchers, and other professionals are
frequently required to speak in public about their favourite toys and games,
courses, final projects, new marketing strategies, the books they read recently.
Since oral presentations can be about more general topics as well as about more
specific ones requiring expertise, they can be used both in general and in English
for specific purposes exams. What is more, oral presentations can be used with
learners of different ages and proficiency levels. While selecting topics and working
on evaluation criteria, however, test designers have to keep in mind a number of
rules, some of which are listed in Example 25.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 165
Example 25: Oral Presentations: Topic selection rules and topics (Adapted from
https://www.myspeechclass.com/good-2-minute-speech-topics-for-
students.html)
Age Group Rules for topic selection Example Topics
Young learners Topics selected for young learners * My favourite (stuffed) animal
should be * My best friend
* My favourite food for
* simple and fun breakfast/ lunch/ dinner
* omething they are familiar with * My favourite colour(s)
* something that they can see and * Something I love to do for fun
experience in their environment * How to build with Lego
* something that comforts them and * How to eat an ice cream/
they are happy to talk about apple/ pizza
* A time I was brave
Topics selected for young learners * The best day of my life
should NOT be * The smartest cartoon
* too difficult character(s)
* something that requires too much * Why does it rain?
preparation * Why is the sky blue?
* something that they are
uncomfortable about
Depending on the length and the type of content of the oral presentation, students
can be given from only 2-3 minutes to a number of days to prepare. In English for
specific purposes exams and with more advanced learners, test-takers might be
asked to include references, prepare notes and PowerPoint presentation so that
real-life context is replicated as much as possible.
166 LANGUAGE ASSESSMENT AND TEST PREPARATION
Practice Activities
Picture-cued Items
Ask your students to draw pictures of their favourite food/ place/ cartoon character.
Collect those and prepare questions related to each of those pictures. Make sure
that the questions are appropriate for the level of your students, and they cannot
be answered with one-word responses.
Give sample answers for each of the questions and prepare scoring rubrics.
Reading-aloud Task
Revise the topics and structures you covered in class and would like to include in
your first Midterm exam. Based on your analysis, prepare a ‘diagnostic passages’ for
a reading-aloud task on the exam.
Write the instructions for this task and list the criteria that will be used for grading
students’ performance in the exam.
Paraphrasing
Keeping in mind your students’ age, level of proficiency, interest and the material
covered in class, select/write two little stories for your students to paraphrase. The
first story is to be read aloud to the students, while they will have a chance to read
the second one for themselves.
Prepare rubrics for both of those tasks.
Types of Speaking
Think about the five types of speaking tasks discussed in this Chapter (i.e., imitative,
intensive, responsive, interactive, and extensive). First, in pairs and then as a class,
prepare lists for the groups of students and the types of exams where each of
them can be used.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 167
References
Angelelli, C. V. (2009). Using a rubric to Bridgeman, B., Powers, D., Stone, E.
assess translation ability: Defining & Mollaun, P. (2011). TOEFL iBT
the construct. In C. V. Angelelli, speaking test scores as indicators
& H. E. Jacobson (Eds.), Testing of oral communicative language
and assessment in Translation and proficiency. Language Testing, 29(1),
Interpreting studies: A call for dialogue 91–108.
between research and practice (pp. 13- Brown, H. D. (2004). Language
47). John Benjamins. Assessment: Principles and Classroom
Ayas, A., Ayaydin, A., Öncü, E., Kaymakçi, Practices. Pearson Education.
S., Börkan, B., Hatipoğlu, Ç., Durukan, Brown, H. D., & Abeywickrama, P. (2019).
E., Karataş, F. Ö., Çetinkaya, F. Ç., Aslan Language assessment principles and
Tutak, F., Ateşkan, A., Orçan, F. & Dinç classroom practices (Third Edition).
Altun, Z. (2020). Okul ve sınıf tabanlı Pearson.
değerlendirmeye dayalı öğretmen Broughton, G., Brumfit, C., Flavell, R.,
kapasitesinin güçlendirilmesi: Yabanci Hill, P., & Pincas, A. (1980). Teaching
dil olarak İngilizce dersi öğretmen English as a foreign language (Second
rehber kitapçiği. Ankara: Milli Eğitim edition). Routledge and Kegan Paul.
Bakanlığı. https://odsgm.meb.
Buck, G. (1992). Translation as a language
gov.tr/www/okul-ve-sinif-tabanli-
testing procedure: Does it work?
degerlendirmeye-dayali-ogretmen-
Language Testing, 9(2), 123-148.
kapasitesinin-guclendirilmesi-
calismasi/icerik/554 Can Daşkın, N., & Hatipoğlu, Ç. (2019a).
Reference to a past learning
Bachman, L. F. (2000). Modem language
event emerging as a practice of
testing at the turn of the century:
informal formative assessment in
Assuring that what we count counts.
L2 classroom interaction. Language
Language Testing, 17(l), 1-42.
Testing, 36(4), 524-551.
Bachman, L. F. & Palmer, A. S. (1996).
Can Daşkın, N., & Hatipoğlu, Ç. (2019b).
Language testing in practice:
Reference to a past learning event in
Designing and developing useful
teacher turns in an L2 instructional
language tests. Oxford University
setting. Journal of Pragmatics, 142,
Press.
16-30.
Bloomfield, L. (1926). A set of postulates
Can Daşkın, N., & Hatipoğlu, Ç. (2019c).
for the science of language.
A proverb learned is a proverb
Language, 2, 153–164.
earned: Proverb instruction in EFL
Bloomfield, L. (1933). Language. Henry classrooms. Eurasian Journal of
Holt. Applied Linguistics, 5(1), 57–88.
Bradshaw, D. (2020). Assessing Candlin, C. N. (1987). Towards task-
speaking: Then and now. In D. based language learning. In C. N.
Gonzalez-Alvarez & E. Ramma- Candlin and D. F. Murphy (Eds.),
Martinez (Eds.), Languages and Language Learning Tasks (pp. 5-22).
the internationalisation of higher Prentice Hall International: Lancaster
education (pp. 212-228). Cambridge Practical Paper in English Language
Scholars Publishing. Education, Vol. 7.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 169
Chen, M. H., Huang, S. T., Chang, J. S., & Hatipoğlu, Ç. (2012). Apologies and gender
Liou, H. C. (2015) Developing a corpus- ina Turkish and British English. Dilbilim
based paraphrase tool to improve Araştırmaları, 1, 55-79.
EFL learners’ writing skills. Computer Hatipoğlu, Ç. (2013). First Stage in the
Assisted Language Learning, 28(1), Construction of METU Turkish
22-40. English Exam Corpus (METU TEEC).
Chun, D. (2002). Discourse intonation Boğaziçi University Journal of
in L2: From theory to practice. John Education, 30(1), 5-23.
Benjamins Publishing Company. Hatipoğlu, Ç. (2016a). Explicit Apologies
Council of Europe. (2001). Common in L2 Turkish. In Ayşe Gürel (Ed.), L2
European Framework of Referencefor Acquisition of Turkish (pp. 221-248).
Languages: Learning, teaching, John Benjamins.
assessment. Cambridge: Cambridge Hatipoğlu, Ç. (2016b). The impact of
University. the University Entrance Exam
Fries, C. C. (1945). Teaching and Learning on EFL education in Turkey: Pre-
English as a Foreign Language. service English language teachers’
University of Michigan Press. perspective. Procedia-Social and
Frodesen, J. (2002). Developing Behavioral Sciences, 232, 136-144.
paraphrasing skills: A pre- Hatipoğlu, Ç. (2017a). History of
paraphrasing mini-lesson. Retrieved Language Teacher Training and
from http://www.ucop.edu/dws/ English Language Testing and
lounge/dws_ml_pre_paraphrasing. Evaluation (ELTE) Education in
pdf Turkey. In Y. Bayyurt, & N. Sifakis
Fulcher, G. (1999). Assessment in English (Eds.), English Language Education
for academic purposes: Putting Policies and Practices in the
content validity in its place. Applied Mediterranean Countries and Beyond
linguistics, 20(2), 221-236. (pp. 227-257). Peter Lang.
Fulcher, G. (2014). Testing second Hatipoğlu, Ç. (2017b). Assessing speaking
language speaking. Routledge. skills. In E. Solak (Ed.), Assessment
in language teaching (pp. 118-148).
Gibson, S. (2008). Reading aloud: a useful
Pelikan.
learning tool? ELT journal, 62(1), 29-
36. Hatipoğlu, Ç. (2017c). Status and
apologising in L2: A problem? In
Harding, L. (2014). Communicative
S. Nalan Büyükkantarcıoğlu, Işıl
Language Testing: Current issues
Özyıldırım and Emine Yarar (Eds.),
and future research. Language
45. Yıl Yazıları (pp. 195-213). Ankara:
Assessment Quarterly, 11(2), 186-197.
Hacettepe University Publications
Hartley, J. (1993). Writing, thinking (ISBN: 978-975-491-455-9).
and computers. British Journal of
Hatipoğlu, Ç., & Can Daşkın, N. (2020). A
Educational Technology, 24 (1), 22-31.
proverb in need is a proverb indeed:
Hatipoğlu, Ç. (2009). Culture, Gender and Examination of the proverbs in the
Politeness: Apologies in Turkish and coursebooks used in high schools
British English. VDM Verlag Dr. Muller in Turkey. South African Journal of
Aktiengesellschaft & Co. KG. Education, 40(1), 1-15.
170 LANGUAGE ASSESSMENT AND TEST PREPARATION
Tajeddin, Z., Alemi, M., & Yasaei, H. (2018). Underhill, N. (1987). Testing spoken
Classroom assessment literacy for language: A handbook of oral testing
speaking: Exploring novice and techniques. Cambridge University
experienced English language Press.
teachers’ knowledge and practice. Valette, R. M. (1997). Modem language
Iranian Journal of Language Teaching testing. Harcourt Brace Jovanovich.
Research, 6(3), 57-77.
Weir, C. (1988). Communicative Language
Taylor, L. (2011). Studies in language Testing. Exeter University Press.
testing, 30. Examining speaking:
Weir, C. (1993). Understanding and
Research and practice in assessing
Developing Language Tests. Prentice
second language speaking.
Hall International.
Cambridge University Press.
Weir, C. (2013). Measured constructs:
Thornbury, S. (2003). How to Teach
A history of Cambridge English
Grammar. World Affairs Press.
language examinations 1913-2012.
Tsagari, D., & Floros, G. (Eds.). (2013). Cambridge English Research Notes, 51,
Translation in language teaching and 2-10.
assessment. Cambridge Scholars
Weir, C., & Wu, J. R. W. (2006). Establishing
Publishing.
test form and individual task
Umirova, D. (2020). Authenticity and comparability: A case study of a
authentic materials: History and semi-direct speaking test. Language
present. European Journal of Research Testing, 23(2), 167-197.
and Reflection in Educational Sciences,
8(10), 129-133.
CHAPTER VI: Testing and Assessment of Speaking Skills, Test Task Types and Sample Test Items 173