Professional Documents
Culture Documents
4.1 INTRODUCTION
Language teachers, as Amrani (2011: 270) notes, are constantly engaged
in evaluating materials:
Materials evaluation is nothing new to teachers. They do it all the time
from wandering casually around a bookshop choosing a new book or
borrowing a new idea gleaned from a conference presentation or
colleague to surfing the internet discussion and resource sites to
gather together new materials.
To this list of informal evaluation activities, we could add the well-
known ‘flick test’, which involves no more than a rapid glance through the
pages of a coursebook, perhaps taking in the topics, illustrations and the
main language points covered in a particular book. Tomlinson notes wryly,
‘publishers are well aware of this procedure and sometimes place attractive
illustrations in the top right-hand corner of the righthand page in order to
influence the flicker in a positive way’ (2013a: 31).
Such activities are, however, basically intuitive. Where this intuition is
professionally informed, these activities may be of some value to the
person concerned, but the evaluation remains, to use Tomlinson’s (2003b:
17) words, ‘ad hoc and impressionistic’: it is not based on an explicit set of
teaching and learning principles articulated before the evaluation. In this
chapter we are mainly concerned with the principled evaluation to which
Tomlinson refers, and we will follow his definition (2003b: 15): ‘Materials
evaluation is a procedure that involves measuring the value (or potential
value) of a set of learning materials. It involves making judgments about the
e#ect of the materials on the people using them.’ Tomlinson (2003b)
chooses his words carefully in his definition: words such as ‘procedure’,
‘measuring’, ‘value’, ‘judgment’ and ‘effect’ combine to suggest a process
which is more systematic, more considered and more complex than
informal, intuitive processes such as browsing the book stalls at an ELT
conference or carrying out the ‘flick test’.
4.2 THE NEED FOR PRINCIPLED EVALUATION
A question we need to ask at the outset, then, is: ‘(Why) is
systematic and principled materials evaluation necessary?’ The first reason
is that, as we established in the introduction to this volume, materials play
an important role in the language learning process. Indeed, in some
contexts, materials will constitute the main exposure which learners have to
the TL. As Littlejohn (2011) argues, the added extras that many publishers
now offer (e.g. workbooks, CDs, interactive whiteboard (IWB) resources)
potentially mean that materials will occupy an even more prominent role in
class than previously. We can add to this that the materials teachers use
may even, in practice, constitute teacher education in the hands of
inexperienced and/or untrained teachers who rely heavily on materials for
guidance on how to conduct their classes. The second reason that
systematic and principled evaluation is necessary is that materials
potentially represent a significant professional, "financial and/ or political
investment (Sheldon 1988). This is particularly the case when national or
regional governments are providing English language coursebooks for the
state school system. To these two strong practical reasons for taking
materials evaluation seriously we can add a third, perhaps more academic
reason: materials evaluation can be a powerful professional development
activity for those who engage in it. Seen in this way, materials evaluation is
not an additional component to materials development, but an integral part
of it. The argument that materials evaluation is developmental rests on the
idea that drawing up evaluation criteria and making judgments based on
those criteria reveals much about the assumptions, knowledge and values
of the evaluators themselves. It can also broaden the horizons of the
evaluators: McGrath (2013: 107), for example, argues that when teachers
and supervisors are involved in evaluation, this can prompt ‘reflection on
aspects of the materials that they might otherwise not have thought about
and [offer] an outlet for feelings that might otherwise not have been
expressed’. In similar vein, Al Khaldi (2011) reported that many teachers
who took part in his coursebook evaluation exercise in Jordan commented
that they found it an enlightening experience, as the evaluation criteria they
were given and the process of applying those criteria helped them to see
more clearly what they wanted from the materials they used.
While we have defined materials evaluation above, we can perhaps
delineate it more "firmly by also considering what it is not. We need to note,
then, that Tomlinson (2003b) contrasts evaluation with analysis: whereas
evaluation is concerned with judging the effects of the materials in a given
context, analysis is concerned with what is in the materials, e.g.
The first question can be answered with a number; the second can be
answered yes/ no. We need to acknowledge, however, that evaluation can
sometimes be implicit in apparently simple, objective materials analysis
questions. In the second point above, for example, we could argue that the
very fact of including a question about authenticity shows that the analyst
thinks authenticity is important. We might go further and suggest that
definitions of authenticity might vary from analyst to analyst. Analysis, as
we have defined it here, might, for example, be useful in the ‘weeding out’
process, i.e. eliminating materials at the selection stage which are clearly
unsuitable in terms of straightforward analysis criteria such as level, target
age range; number of units, exam focus and so on. The materials which
survive the initial analysis can then be subjected to more detailed
evaluation.
Task 4.1
Decide on a target learner group, e.g. an International English Language
Testing System (IELTS) preparation class. Formulate four or five analysis
questions you would ask about potential materials for that group, such as:
Pre-use evaluation
Pre-use evaluation is more common than both whilst-use and post-use
evaluation for two reasons: it is usually designed to inform the important
decision as to which materials to adopt, and it is easier to carry out than
whilst-use or post-use evaluation, which, as we shall see, may both require
more time-consuming evaluation methods such as classroom observation
or tests. As Tomlinson (2003b) laments, pre-use evaluation is often carried
out impressionistically, though there are, as we shall also see, clear
alternatives, e.g. a framework of evaluation criteria. Pre-use evaluations
may cover largely the same ground as whilst-use criteria, but they will be
worded differently, e.g.:
Pre-use: How far are the materials likely to motivate the learners? 5 4 3 2 1
Whilst-use: How far do the learners find the materials motivating? 5 4 3 2 1
Post-use: How far did the learners find the materials motivating? 5 4 3 2 1
Alternative methods of pre-use evaluation are noted by McGrath
(2002), who refers to the possibility of trialling/piloting the materials,
consulting colleagues who have already used the materials (in other
contexts), and observing classes where the materials are already in use.
Whilst-use evaluation
Tomlinson (2003b: 24) notes the usefulness of whilst-use evaluation and
points to a number of potential foci which may lend themselves particularly
to it:
• clarity of instructions, layout
• comprehensibility, credibility, achievability of tasks
• achievement of performance objectives
• potential for localisation
• practicality, teachability, flexibility
• appeal, motivating power, impact
• effectiveness in facilitating short-term learning
• local (related to the context): e.g. are the materials culturally acceptable in
the context?
• media-specific (e.g. audio or computer): e.g. is the sound quality of the
audio materials good?
• content-specific (e.g. exam or English for specific purposes (ESP)): e.g.
do the materials replicate the types of real-world tasks the target group will
need to do?
• age-specific: e.g. are the visuals likely to appeal to children? Rubdy
(2003) argues that evaluation criteria can be generated from three key
notions: psychological validity; pedagogic validity and process validity.
These terms are described thus:
• psychological validity: learners’ needs, goals and pedagogical
requirements
• pedagogical validity: teachers’ skills, abilities, theories and beliefs
• process and content validity: the thinking underlying the materials writer’s
presentation of the content and approach to teaching and learning
respectively
Rubdy (2003) provides a further gloss to clarify these terms. Psychological
validity encompasses such notions as independence and autonomy; self-
development and creativity. Pedagogical validity refers to notions such as
guidance; choice and re0ection. A particularly interesting aspect of Rubdy’s
(2003) framework is the emphasis on teacher development: ‘An important
criterion for evaluating materials would . . . relate to the extent to which they
engage the teacher’s constantly evolving critical standpoint and facilitate
the expanding and refining of the teacher’s schemata in the process’ (ibid.:
50). Process and content validity seems to be rather a broad-brush
category covering aspects from methodology and content to layout and
graphics. It is, however, a framework for generating criteria rather than a
framework of set criteria.
It is evident from these three examples (Grant 1987; Rubdy 2003;
Tomlinson 2003b) that criteria frameworks can vary both in focus and in
scope, but McGrath (2002: 43) notes the following areas common to most
of them:
• design: includes both layout of material on the page and overall clarity of
organization
• language content: coverage of linguistic items and language skills
• subject matter: topics
• practical considerations: e.g. availability, durability and price
Operationalizing criteria
Whichever framework we design, adapt or adopt, we need to consider how
we actually generate the specific criteria for each category and
operationalize the framework. McGrath (2002: 40–1) suggests the following
procedure: