You are on page 1of 3

Okay, so this segment here on response

order effects fits nicely into the attitude question unit, but it
could also go somewhere else. These response order effects might appear
in factual questions, though maybe not as often because
people might have a stronger representation of what the
answer should be. Let's take this question as one example. "Which one of these
qualities is the most
desirable for a child to have?" And here comes a long list of qualities
and some of them you might have thought about
before, others not. This is all formulated in the male
version, but of course it should, in this question, relate to
both males and females. Good manners, tries hard to succeed, is
honest, neat and clean, good sense and sound judgment,
self-control, all those characteristics of a child. And what studies have shown is
that if you
have a long list, like this, people, if they self-administer these
questionnaires, where you look at this list, tend to answer more often "yes" to
items in the upper segment of this list. Then, experiments where the order of the
list was changed, so half of the respondent would get the list in
the order printed here and the other half would get them in the opposite order,
version B, all items
flipped. And in a study by Jon Krosnick
and Duane Alwin in 1987, they observed this primacy effect, more items being
picked on the top, in both forward and backward
orders. So, they concluded it's not due to the
items. They had these 13 choices about child characteristics that I just showed you
and the task for the respondent was to choose
three most important. And among those most
important they found that this primacy effect was
actually only present for respondents with low
sophistication, as they called it, so lower level of
education and lower level on scores on vocabulary
tests has pronounced this kind of effect. So, they argued that this group most
likely answered in a satisficing way, which means they responded as soon as an
acceptable option was encountered. They also argued that response order
effects of this kind are a function of mode, because under a visual representation,
respondents might become cluttered after the first few
options and the first few are likely to be
endorsed, showing this primacy effect. But if you hear it in a telephone survey
and it's an auditory presentation, then the earlier options are overwritten
by later ones and so the last few are more likely to be
endorsed, which would lead to a recency effect. And actually
that has more to do with who administers the option rather
than the actual mode. Don't get confused here, the respondent could request, of
course,
from the interviewer or the computer that there's a visual presentation in
addition. So it's not mode within the sense of what
the survey mode in general is, but whether its
visual or auditory. In a meta analysis on response order
effects that Holbrook, Krosnick, Moore, and
Tourangeau did in 2007, they examined over 500 experiments in
about 150 Gallup polls that involved dichotomous scales and
looked at response order effects. And 19.2 showed a significant recency
effect and only 1.8 showed a significant primacy
effect. Mind you, these are telephone polls. So, they had an average shift of 2.2%,
computed in this way, you know, in the
percentage of picking the option when second minus the percentage of
option when it's picked first. They saw as largest predictor of this
effect, question difficulty, so the average lengths of words, number
of words per sentence, and the number of
sentences. The response option length was also a
significant effect and so what the position within the
questionnaire. Now, just as a quick quiz question here in between, what is the
primacy and what is
recency? That's right, primacy would mean, options
that come first are more likely to be endorsed and rencency effect is, options that
come
last are more likely to be endorsed. Now, continuing with that meta-analysis,
question type is another thing that they
looked at more closely and they grouped or
categorized questions in to three types, seemingly open ended
questions, delayed processing questions, and
seemingly yes/no questions. The seemingly open ended questions would
be something like this: "How do you feel about President Bill
Clinton? Is he trustworthy or dangerous?" And, you know, you pause after Bill
Clinton, that of course doesn't have an answer category
there yet, and so "Is he trustworthy or dangerous?" is a closed
ended answer category but is something that sounds more like a
seemingly open ended question. The delayed processing question would be
like this: "Which of the following describes your view
about President Clinton?" So you know something is coming up and
now you get it, "Trustworthy or dangerous?" And the seemingly yes/no question would
be, "Do you think that President Bill Clinton is
trustworthy or dangerous?" But the "do you think" sort of implies that
there will be a yes/no but in this case the yes/no is
converted into "trustworthy or dangerous". And they found an effective question
type
here. The seemingly open ended questions were
least prone to the recency effect and the
delayed processing questions were most prone to recency effects. So
that's something to keep in mind when you consider
formulating your questions. In general, education had the largest
effect, for low education groups compared to higher
education groups. Now, let's look actually, how respondents
answer these questions. I brought with me a few video clips to
show to you. And the first one is a respondent who is
considering all answer options and chooses the best answer, so this is like
the perfect respondent that we see here. This is an eye tracking study done at the
University of Maryland by Mirta Galesic and Roger Tourangeau and colleagues,
meanwhile published but we
were, you know, allowed to use those videos and
you can see here, while I'm talking, that the respondent nicely reads the question,
thinks about the
answer. Now he reread the question, goes back up, goes back down, back and forth
reading the
question, still thinking about the answer options,
and then finally deciding on one and, "Whoo, off to the next
screen." Now, not everybody has that same response
pattern. So here is another example for a response
pattern. You have a respondent, she reads the question slowly, carefully uses the
mouse as a
pointer, and then clicks on the first option that
seems good, but then keeps reading and reconsidering. It's like, you know, you have
listen to phone options and you hover your
finger over a particular number and then you might
pick another one. So here, the first option was changed to the third and then the
respondent still
reads all the way down until she's done and realizes
there's no better option than the one she already
picked. And then, finally, we are seeing this
satisficing type behavior with this last respondent that we filmed here. And again,
the question is read nicely. You see the first two, three, four, five
answer categories are read again. And the bubble gets bigger, that means there's
time spent on this particular
players reading. And so far, the last four options have not
been read. And now, here one is picked and off we go
to the next screen. So, you do see that this is a situation
where the respondent did not read all the options and just picked
something in the upper half. So in summary, the response order effects
are common. They can be large. And Krosnick and Miller actually found
them on election ballots, which is of concern if you don't randomize
the order. The direction depends on the order of
processing, so the mode of presentation, auditory versus
visual, we talked about that, the pace with which the items are read, the type of
the item, whether it's a scale
or an ordered list of things. And it's magnitude depends on the respondent's side
on processing capacity,
age, education, interest in the topic, familiarity with the topic, item difficulty,
fatigue, motivation in general of the respondent,
and some psychological concepts, like need for
cognition, need for closure, those people who show these personality
traits, would be more likely to go through the entire
list, and fuzziness of the respondents, thinking
still about prior items and prior thoughts that might occur
in the questionnaire. What can you do to mitigate this effect as
a questionnaire designer? Well, ideally you would randomize the
order of these items. It's not always possible in particular,
if there's an inherent structure in this list. Then it might be
odd to randomize them but then you can hope that this inherent structure will also
mitigate some of
these effects, because people do expect a certain set of
items to appear later on. For example, if you have in factual
questions a list of, let's say, the study subjects, you know, and they're
alphabetical and
"Computer science", you know, you'll find that in
third place but if you study, you know, "War history",
then you will look at the very bottom of the
list. But, you know, some things are
ordered historically and, you know, it doesn't necessarily need to
be alphabetical, so that can help. But you can always at least run this list
forward and backward and randomize your respondents to one of
the two, if you can't randomize the entire order of
those questionnaires. That is one way, how you can mitigate this
effect, when you design your questionnaire.

You might also like