Professional Documents
Culture Documents
141
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
142 Randomization and Pseudo-Randomization
1
The seminal works combining the two approaches in an RCM context are that of Imbens
and Angrist (1994) and Angrist et al. (1996).
2
AIR (1996, pp. 444–445) summarize this view: “Standard IV procedures rely on judgments
regarding the correlation between functional-form-specific disturbances and instruments.
In contrast, our approach [incorporating RCM and an experimental framework] forces the
researcher to consider the effect of exposing units to specific treatments. If it is not possible
(or not plausible) to envision the alternative treatments underlying these assumptions, the
use of these techniques may well be inappropriate.”
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.2 The Ideal Instrumental Variable 143
Condition 3 (No Missing Data): We can perfectly observe the choices made
by those affected by M. That is, define P OBS as the choices observed by the
researcher and P ACT as the actual choices made by the units of study. When
there are no missing data, then P OBS = P ACT for all units.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
144 Randomization and Pseudo-Randomization
as the actual choices made by the units of study. When there are no missing
data, then P OBS = P ACT for all units.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.3 When Assignment of Treatment Is Not Independent of Potential Choices 145
and all follow directions during the experiment), random assignment comes
as close as possible to an ideal IV. Unfortunately, only a subset of usually
simple laboratory experiments are likely to be conducted in this fashion. A
number of laboratory experiments and almost all field experiments are not
ideal IVs because they violate one or more of these conditions.
Furthermore, if a researcher is working with observational data in which
treatment is not randomly assigned by the researcher, violations of these
conditions are likely as well. Later we discuss how such violations occur
in random assignment and in natural manipulations that researchers use
as IVs. We explore the various methods, either in the design of random
assignment or after the experiment in statistical analysis, that researchers
can use to deal with these problems.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
146 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.3 When Assignment of Treatment Is Not Independent of Potential Choices 147
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
148 Randomization and Pseudo-Randomization
the individual voting choices had been affected by the random assignment
mechanism that determined which voters were informed or not, then the
treatment assignment mechanism affects their potential choices and the
first condition for an ideal IV is violated.
In some laboratory experiments, of course, payoffs are manipulated as
subjects may be randomly assigned to particular roles and the role would
imply a different payoff matrix (Dasgupta and Williams randomly assign
roles as well). In Dasgupta and Williams’s study, however, these two random
assignments were independent, which is required if the causal effect of
information in this case is to be accurately measured. If the experiment does
not maintain that independence in the two random assignment mechanisms
(i.e., is poorly designed), then assignment to treatment is correlated with
potential outcomes and is no longer an ideal IV. Obviously, maintaining
independence of manipulations is also important in field experiments that
evaluate more than one treatment effect or a complexity of treatment effects.
If the assignments are correlated, then independence of the manipulations
can be violated.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.3 When Assignment of Treatment Is Not Independent of Potential Choices 149
the researcher has not randomized within these educational categories, ran-
domization is more likely to select college-educated voters and, as a result,
we no longer have independence between the manipulation and poten-
tial outcomes. We have independence within, but not across, educational
categories.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
150 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.3 When Assignment of Treatment Is Not Independent of Potential Choices 151
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
152 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.3 When Assignment of Treatment Is Not Independent of Potential Choices 153
3
We thank Macartan Humphreys for the example.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
154 Randomization and Pseudo-Randomization
P0 P1 W Pr(T = 1) T P P1 − P0 (P1 − P0 ) |T = 1
0 0 0 0.25 1 0 0 0
0 0 0 0.25 0 0 0
0 0 0 0.25 0 0 0
0 0 0 0.25 0 0 0
0 0 1 0.25 1 0 0 0
0 0 1 0.25 0 0 0
0 0 1 0.25 0 0 0
0 0 1 0.25 0 0 0
0 1 2 0.5 1 1 1 1
0 1 2 0.5 1 1 1 1
0 1 2 0.5 0 0 1
0 1 2 0.5 0 0 1
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 155
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
156 Randomization and Pseudo-Randomization
Definition 5.2 (Pretreated Subject): Subject who has experienced the desired
manipulation prior to his or her random assignment.
Definition 5.5 (Defier): Subject who chooses the opposite of the manipulation
assigned to him or her.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 157
But if that is not the purpose of the experiment, then these inferences
mean that the data gathered from the later periods may be suspect and not
useful.
A similar problem can occur when researchers wish to investigate com-
munication during an experiment. For example, suppose that researchers
are interested in how communication might affect the likelihood that sub-
jects are willing to cooperate in the aforementioned game. But the unin-
formed subjects might learn more from the communication – learning
about the information other subjects have about the benefits of coopera-
tion – which could interfere with the design of the experiment to manipulate
which subjects are informed. Again, this may be the purpose of the experi-
ment (to measure how information is transferred through communication),
but if it is not, then the researcher needs to be concerned about the possible
effects of the communication on information levels.
In other laboratory experiments, noncompliance can also occur when
subjects are asked to return for a subsequent session. In such a situation, a
subject is participating in a sequential experiment. Consider Example 5.1.
In the experiments discussed, Druckman and Nelson (2003) contacted sub-
jects in a survey 10 days after an experiment to determine whether framing
effects observed in the original experiment diminish over time, and Chong
and Druckman (2009) had subjects participate in two sequential sessions
three weeks apart. In both cases some subjects failed to comply in the second
round.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
158 Randomization and Pseudo-Randomization
Election Study sample.” In the follow-up second stage of the experiment, only
70 subjects participated.
Experiment 2 [first reported by Chong and Druckman (2007) and also ana-
lyzed in Chong and Druckman (2009)] recruited a combination of 869 students
at a large public university and nonstudents in the area. They do not report the
exact numbers of each type of subject, but report the following about the subjects
(footnote 7): “Overall, aside from the disproportionate number of students, the
samples were fairly diverse, with liberals, whites, and politically knowledgeable
individuals being slightly over-represented (relative to the area’s population).
We checked and confirmed that adults and nonadults did not significantly dif-
fer from one another in terms of the experimental causal dynamics presented
later.” The follow-up experiment involved 749 of the original participants.
Subject Compensation: Subjects were paid an unspecified amount in cash
for their participation. In experiment 1, subjects who responded to the post-
experiment survey were entered in a lottery for an unspecified payment. In
experiment 2, subjects were paid an extra $5 for taking part in the second
round and participants, if they responded, were entered in a lottery where they
could win an additional $100.
Environment: The experiments took place at university political psychology
laboratories. Experiment 1 took place during the period in which the U.S. Senate
was considering new legislation on campaign finance reform, the McCain–
Feingold bill, with sessions beginning about a week after the Senate introduced
the bill and ending before the Senate debate began. The experimenters used
articles on the McCain–Feingold bill that were made to resemble as closely as
possible articles from the New York Times’ Web site. The researchers drew from
recent similar reports, copied an actual article from the site, and then replaced
the original text with their text. Finally, the experimenters reported that there
was “a flurry of media coverage” that preceded the experiment and did not
begin again until the experiment was over. Experiment 2 used fake editorials
about a hypothetical urban growth management proposal for the city in which
the experiment took place and subjects were told that the editorials were from
a major local newspaper.
Procedures: We report on the two experiments highlighted by Chong and
Druckman (2009) in which subjects were called back for a post-election survey
or second experiment. Chong and Druckman reported on a third experiment,
which we do not discuss.
Experiment 1: Subjects were randomly assigned to one of seven manipu-
lations. In six of the manipulations, subjects received an article to read about
the McCain–Feingold bill. The articles were either framed to emphasize the
“free-speech” arguments against campaign finance reform or to emphasize the
“special-interest” arguments in favor of campaign finance reform. “Specifically,
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 159
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
160 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 161
Druckman (2009), in a reanalysis of the data, found that the effects depended
on the type of information processing used by the subjects; that is, subjects
who used a more memory-based method of processing new information were
less affected by the frames, but the effect that did occur persisted longer. In
experiment 2 for the first session, Chong and Druckman (2007, 2009) found
that framing effects depend more heavily on the strengths of the frames than
on their frequency of dissemination and that competition alters but does not
eliminate the influence of framing. In evaluating the two sessions together, they
find again that strong frames play a larger role in affecting opinions and that
subjects who used a more memory-based method of processing new information
responded more strongly to recent frames.
Comments: Chong and Druckman (2007, 2009) presented a nonformal
theory of how framing works over time and derived a number of hypotheses
that form the basis of the experimental tests. To increase compliance of subjects
to show up for the second treatment in experiment 2, Chong and Druckman
sent reminders every three days, with up to a total of three reminders (if
necessary).
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
162 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 163
data using the same candidates. Spezio et al., by using naturally occurring
candidates, compared the choices of the subjects between candidates with
who actually won election. Spezia et al. avoided candidates who had national
prominence or had participated in a California election (the experiment was
conducted in California). The researchers also collected familiarity ratings
from all of the participants in an attempt to verify that none of the subjects
had been pretreated.
In the field, avoiding pretreatment of subjects can be more difficult.
Gerber, Kaplan, and Bergan attempted to deal with pre-treatment by
excluding subjects that already subscribed to the newspapers and they also
attempted to control for subjects’ existing knowledge from other news
sources by measuring and randomly assigning subjects within strata by
their use of sources. Gaines and Kuklinski (2008) suggest that researchers
deal with pretreatment by explicitly considering the effects of treatment as
the effect of an additional treatment given prior to treatment. In either case,
the recommendation is for researchers to attempt to measure when subjects
have been pretreated in conducting the experiment and to consider the data
from these subjects separately from those who have not been pretreated.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
164 Randomization and Pseudo-Randomization
play the games more than once to facilitate possible learning, as explained in
Chapter 6.
Randomization has its limits in reducing such cross-effects that can limit
the ability of experimentalists to manipulate subjects independently. For
instance, suppose that two subjects are playing the ultimatum game repeat-
edly (see Section 3.3.3), but each period they are randomizing between being
the proposer and the responder. The experimenter wants to manipulate the
size of the amount of money the proposer is dividing and in some periods
the proposer has a larger pie to divide. Even if the experiment is set up so that
subjects always play a new subject with no possibility of contamination from
previous play as in Dal Bo’s experiment, it may be the case that subjects per-
ceive the game as a larger supergame and choose to always divide the pie in
half regardless of the size of the pie or who is the proposer. Thus, if the game
is to be repeated, the experimenter may want to not randomize the roles
and to always have the same subjects serve as proposer and the same ones as
receiver, and assign pairings as in Dal Bo’s experiment (always ensuring new
matches with no cross-effects). Of course, in this case, subjects assigned to
be proposers will likely earn more than those assigned to be receivers (based
on previous experiments), which can lead to some inequities in how much
subjects earn during an experiment. As a result, an experimenter may have
to pay more on average to all subjects as a consequence so that subjects still
earn enough to be motivated to participate in the experiment even when
their earnings are low relative to other subjects.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 165
usually give subjects a short quiz after going over the instructions and set
up the quiz so that subjects are not allowed to participate until they have
answered all questions correctly, as a method to reduce noncompliance of
this sort. We discuss methods of motivating subjects in laboratory experi-
ments further in Chapter 10.
In terms of sequential experiments in the laboratory as in Example 5.1,
researchers can use various motivational techniques and financial incen-
tives to minimize drop-off. In Druckman and Nelson’s study, the drop-off
is more than half but in Chong and Druckman’s study, nearly 85% com-
plied. Chong and Druckman had informed subjects at the beginning of the
experiment that there would be a second session three weeks later and they
sent reminders to subjects every three days, up to a total of three reminders.
In an innovative experiment, presented in Example 5.2, Casari et al.
(2007) report on a sequential experiment in which they evaluated the effects
of different incentive mechanisms on possible selection bias and compared
the design method to traditional post-experimental statistical analysis to
control for selection bias. In the experiment, subjects participated in a series
of auctions in which subjects bid on the value of unknown objects. Before
each auction, each subject received a private signal about the value of the
object. If a subject made the highest bid, he or she was forced to pay for the
object. Earlier experimental research had demonstrated that subjects often
overbid such that the winner ends up paying more for the object than it is
worth in terms of payoffs. Why? Most theorize that subjects overbid because
they fail to account for the implications of making a winning bid. That is,
suppose as a subject you receive a signal that the object has a high value.
If you just bid your signal, and everyone else does, in the event that you
win the auction, you are likely to have received a signal that is inaccurate –
an extreme one, much higher than the average signal and higher than the
other subjects’ signals. It is likely that the object is worth less than your
signal in this case. Ignoring these implications and overbidding by bidding
one’s signal has been labeled the “winner’s curse” result.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
166 Randomization and Pseudo-Randomization
et al. collected demographic and ability data on the subjects. They compared
the sample on these values with the target population. They report that “[m]en
comprise 59.7 percent of the sample, with the breakdown by major being
30.2 percent economics/business majors, 23.4 percent engineering and science
majors, and 46.4 percent all other majors. There are more men in the sam-
ple than in the university population, as well as a much larger percentage
of economics and business majors than the university population (30.2 per-
cent versus 12.3 percent). . . . Some 20.2 percent of the sample are in the top
5 percent (of the national average) with respect to composite SAT/ACT scores
(versus 4.9 percent for the university), with less than 8.9 percent scoring below
the median (versus 20.9 for the university), and 13.3 percent not having any
SAT/ACT scores (versus 21.1 percent for the university). The last group are pri-
marily transfers from regional campuses, as these students are not required to
take these tests when transferring to the main campus. If their SAT/ACT scores
were available, they are likely to be lower, because a number of these regional
campus transfers were ineligible to enter the main campus when they origi-
nally applied to college. Thus, our sample includes students with significantly
higher ability than the university population as measured by the percentage of
students scoring in the top 5 percent on composite SAT/ACT scores and below
median SAT/ACT scores.”
Subject Compensation: Subjects received cash payments based on their
choices as described in the procedures that follow. Subjects also received cash
show-up fees, which varied according to manipulation as described later.
Environment: The experiment was conducted in a computer laboratory in
an environment similar to that discussed in Example 2.6.
Procedures: In each experimental session, subjects were given a starting
capital balance which varied as described later. In each period, two auctions
were conducted simultaneously with six bidders each. Assignments to each
market varied randomly between periods. The subjects submitted simultaneous
bids for the item and if they won the auction paid their bid and received
the item. The value of the item, x0 , in terms of payoffs was the same for
all bidders but unknown to them. The value was chosen randomly from a
uniform distribution with support [$50, $950]. Each subject was told prior
to bidding a signal that was drawn independently for each subject from a
uniform distribution with support [x0 − $15, x0 + $15] . All this information
was conveyed to subjects in the instructions. At the end of each auction, all bids
were posted from highest to lowest, along with the corresponding signal values
(bidder identification numbers were suppressed) and the value of x0 . Profits
(or losses) were calculated for the high bidder and reported to all bidders.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 167
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
168 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 169
In the Field. In the field, researchers can try to avoid problems of noncompli-
ance that occur when collaborating with nonacademics through education,
extensive training, and close monitoring. When researchers are aware of
potential cross-effects within social or professional groups, researchers can
attempt to choose subjects who are unlikely to have interactions with each
other or use a multilayered randomization strategy as advocated by Sinclair
(2009). To see how this works, consider an example in which there are
just two levels at which individuals interact and cross-effects might occur:
household and neighborhood. Assume that the researcher is conducting
one manipulation and choosing other subjects to serve as a baseline (not
manipulated). First, the researcher randomly assigns neighborhoods to be
either manipulated or serve as baselines, then the researcher repeats the
random assignment by household within the nonbaseline or manipulated
neighborhoods, and finally randomizes again by individual within the non-
baseline or manipulated households. Only the subjects who were chosen
for manipulation in the final randomization are actually manipulated, but
in this fashion the researcher has created baselines at each level where these
individuals interact.
Sinclair contends that doing so allows for the identification of both the
spillover effects and the true treatment effect while removing the poten-
tial bias associated with the spillover. To use the procedure, a researcher
must assume that there exists a level at which no subjects interact, that
each lower-level group is a subset of a next higher-level group (that the
groups are nested), and that there is no intersection between groups at
the same level of hierarchy. Sinclair discusses some strategies for conduct-
ing such a randomization. For example, she points out that a researcher
can avoid the household-level randomization issue by selecting only one
subject per household, thereby reducing some of the complications of the
design.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
170 Randomization and Pseudo-Randomization
groups such that lower-level groups are nested. Randomization occurs within
groups by layer.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 171
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
172 Randomization and Pseudo-Randomization
4
Angrist and Kreuger (2001) point out that the bias in the IV case with observational data
is related to the difference between the number of instruments and endogenous variables.
If the number of instruments is equal to the number of endogenous variables, the bias is
zero. They discuss the use of a number of approaches to control for the bias when this is
not the case.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 173
ITT = E (δ M ), (5.5)
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
174 Randomization and Pseudo-Randomization
coverage on voting. They measured ITT and ITT(W) using OLS and a con-
trol function, as explained in Section 4.2. Thus, they included those in the
group who canceled the subscription or perhaps did not receive the paper
regularly through delivery problems as “manipulated” subjects and could
rely on the random assignment of intention to treat to justify their assump-
tion of ignorability of treatment assignment in their case. Furthermore,
they estimated that 8% of the subjects who were given free subscriptions
already subscribed to one of the newspapers (more often the Post), either
weekly or the Sunday edition only. Finally, 55.8% of the individuals in the
sample did not answer the post-election survey and thus they had a high
nonresponse rate, which can be a problem in measuring ITT and ITT(W)
as we observed earlier. We discuss the implications of nonresponse for mea-
suring treatment effects and Gerber, Karlan, and Bergan’s efforts to control
for problems nonresponse causes more fully later.
Gerber, Karlan, and Bergan found that the manipulation did not have
a significant effect on turnout but they did find that the manipulation
appeared to increase the probability that subjects favored Democrats even
for those assigned to the Washington Times. However, they speculate that the
reason was that even though the Times may have presented a conservative
biased version of reality, the period of the study (October 2005) was a par-
ticularly bad month for President Bush, in which his overall approval rating
fell approximately 4 percentage points nationwide. Thus, they contend that
exposure to the news media increased voters’ awareness of the problems
that Bush faced and increased their probability of favoring Democrats even
when the media source was arguably biased in favor of Bush.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 175
Axiom 5.2 u1 = u0 .
5
Wooldridge (2002, pp. 621–636) presents an excellent review of this literature and the
different estimation approaches that have been developed.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
176 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 177
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
178 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 179
test for redundancy suggested by Evans and Schwab (1995) – testing for
overidentification in the linear two-stage OLS model. To do so he creates
multiple instruments by using each PCD as an instrument as well as whether
a PCD resident was a user of services that were particularly decentralized
(child and elderly care, elementary schools). He finds significant evidence of
overidentification.
Particularly noteworthy is the care Lassen takes to consider alternative
estimating strategies, which he compares with his IV-probit estimation. As
noted earlier, there are concerns about variations in efficiency and consis-
tency in these procedures, and in some cases two-stage OLS may be more
appropriate even with an endogenous dummy variable. He estimates (1) a
single-equation probit of the effect of information on voter choices assum-
ing that information choices are exogenous, (2) a two-stage linear regression
equation, (3) a full bivariate probit which assumes that the error terms are
jointly normal, (4) inclusion of the alternative instruments described earlier,
and (5) nearest-neighbor propensity score matching on a larger sample that
includes respondents who refused to answer questions about their income.
Too often researchers focus on only one estimation strategy to establish
causality. Given the conditional nature of causal estimates and the implicit
and untested assumptions that most of these procedures make, use of a
variety of procedures can help mitigate concerns about the robustness of
the analysis.
Lassen finds that the differences in specification do not alter the conclu-
sions; that is, he finds that the effect of being informed significantly increases
the probability of voting by around 20 percentage points, although this
effect is smaller in the estimation that includes respondents who refused to
state their income, which is likely a consequence of the fact that these are
respondents who are also less likely to vote.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
180 Randomization and Pseudo-Randomization
6
As we have done previously, it is useful to define the causal effects first and then discuss
the assumptions underlying estimating causal effects. See, for example, the presentation of
ATE, ATT, and MTE in Chapter 3. This separates the questions of defining causal measures
from estimation issues.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.4 When an IV or Assignment Is Not a Perfect Substitute for Treatment 181
Estimation of LATE. Imbens and Angrist (1994) and AIR present general
assumptions that allow for the estimation of LATE.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
182 Randomization and Pseudo-Randomization
(P 1 − P 0 )
LATE∗ = , (5.11)
(T 1 − T 0 )
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.5 Missing Data 183
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
184 Randomization and Pseudo-Randomization
to biased estimates of causality (see Rubin, 1976; Little and Rubin, 1987;
Schafer, 1997).
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.5 Missing Data 185
8
FR’s presentation differs in that they make an assumption called compound exclusion
restriction for never-takers which is the equivalent of Axioms 5.12 and 5.13 combined.
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
186 Randomization and Pseudo-Randomization
Mealli et al. (2004) suggested that the response exclusion restriction for
never-takers (Axiom 5.13) of FR is unreasonable because never-takers who
are assigned to a treatment may alter their response probability. That is, their
refusal to follow through with treatment may increase their unwillingness
to respond to a post-treatment survey or instrument measuring outcomes
because they must admit basically that they did not comply. This could be
the case in a political science field experiment where subjects provided with
information about an election who did not read or use the information
may be less willing to answer questions about the election afterward than
they would be if they had not been forced to make the choice to ignore the
information. Mealli et al. propose the following substitute.
exp(α0 + α1 W)
Pr(S = 1|W, α) = π s = , (5.12a)
1 + exp(α0 + α1 W)
exp(β MS0 + β MS1 W)
Pr(R = 1|W, M, S; β) = π M,
R
S = , (5.12b)
1 + exp(β MS0 + β MS1 W)
exp(γ MS0 + γ MS1 W)
Pr(P = 1|W, M, S; γ ) = f M, S (1) = .
1 + exp(γ MS0 + γ MS1 W)
(5.12c)
Note that they assume that the slope coefficients in the outcome distribu-
tion for compliers are equal: γ011 = γ111 . Given the assumptions of latent
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.5 Missing Data 187
L (θ|M, W, T, R, P )
= π S π11
R
f 11 (P ) π S 1 − π11
R
Paraphrasing Mealli et al. (p. 216), the first two factors in the likelihood
represent the contribution of the compliers assigned to treatment, including
both respondents and nonrespondents. The second two factors represent the
contribution for never-takers assigned to the treatment, including respon-
dents and nonrespondents. The last two factors represent the contribution
to the likelihood function for those assigned to no treatment. This includes
both compliers and never-takers, and the likelihood contributions therefore
consist of averages over the distribution of compliance types. MAR is the
assumption π00 R
= π01R
in estimating this equation. FR assumes π00 R
= π10R
Incorporating the missing data structure from earlier, Mealli et al. were
able to obtain maximum likelihood estimates; that is, the complete data
structure in Equation (5.12d) allows them to estimate the missing val-
ues. In this section we have examined noncompliance and nonresponse
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
188 Randomization and Pseudo-Randomization
as binary cases, but it is possible to estimate causal effects for when these
variables are not binary as well as the treatment itself. For an example of
estimation of causal effects in a more complex situation, see Barnard et al.
(2002).
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.5 Missing Data 189
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
190 Randomization and Pseudo-Randomization
Is the fact that the subjects studied by Gerber, Karlan, and Bergan and
Horiuchi, Imai, and Taniguchi are not representative a problem with their
randomization procedures? Most political scientists would say yes and stop
there. However, such a summary confuses the purpose of randomization
in typical surveys with the purpose of randomization in experiments. That
is, in traditional surveys the purpose is to measure public opinion, and
simple randomization across the population is well known to be the best
way to construct such measures. The spectacular failures of pollsters to
predict the 1948 presidential election using designs that were not simple
randomizations led researchers who work in public opinion to prize surveys
based on how closely they matched fully randomized designs.9 However,
in experiments like those of Horiuchi et al. and Gerber et al., whether the
subject pool is a random draw of a different population matters in drawing
inferences about the application of the results beyond the subject pool, not
about the robustness of the causal inferences drawn as applied to the subject
pool or target population (see Definition 2.2).
For drawing causal inferences within the population subject to the exper-
iment, different standards of randomization are important. As Horiuchi
et al. noted, many researchers implicitly assume that simple randomization
of treatment is the best procedure to use within the experiment, much as it
is in measuring public opinion in surveys. Yet such a conclusion, as Hori-
uchi et al. argued, would be incorrect. Horiuchi et al. know from previous
research that particular variables such as an individual’s gender and vote
intention are strong predictors of whether they vote in Japanese elections.
Horiuchi et al. first surveyed voters about these basic variables and grouped
9
The history of randomization in public opinion research is reviewed by Morton (2006,
chapter 10).
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.5 Missing Data 191
them into strata. They then randomized treatments within the blocks using
a complete randomization procedure (that is, they randomized treatments
across all subjects at once rather than assigning them subject-by-subject). By
doing so, the resulting estimates of causal effects are more efficient than those
if subjects were assigned treatments individually.10 Their use of control,
mixed with random assignment, also shows how these two aspects of exper-
imental procedure work together and the importance of control in experi-
mental design. Again, their use of control is not the standard narrow view of
most political scientists, where control is simply a comparison of different
treatments or a treatment and a baseline, but a broader perspective where
control of observables allows for more efficient estimation of causal effects.
10
See Cox and Reid (2000).
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
192 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.7 Chapter Summary 193
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
194 Randomization and Pseudo-Randomization
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005
5.8 Adding in Formal Theory 195
Downloaded from https:/www.cambridge.org/core. UCL, Institute of Education, on 07 Apr 2017 at 17:39:11, subject to the Cambridge Core
terms of use, available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9780511762888.005