You are on page 1of 15

JOURNAL OF THE EXPERIMENTAL ANALYSIS OF BEHAVIOR 2013, 9999, 1–15 NUMBER 9999 (XXX)

EFFECTS OF REINFORCER MAGNITUDE ON REINFORCED BEHAVIORAL VARIABILITY


ADAM H. DOUGHTY, KENNETH G. GIORNO, AND HANNAH L. MILLER
COLLEGE OF CHARLESTON

Eight pigeons were exposed to a two-component multiple schedule. In each component, four-peck
sequences across left and right keys were reinforced according to a variability threshold contingency. In one
condition, only infrequently occurring response sequences were reinforced in each component, thereby
generating highly variable sequences. In a separate condition, when the variability threshold contingency
was lenient in each component, sequences were much less variable. In each condition, reinforcer
magnitude was manipulated across components, and the larger reinforcer magnitude produced less
variability than the smaller reinforcer magnitude. These results suggest that larger reinforcers hinder the
reinforcement of behavioral variability. The results are interpretable in terms of the larger reinforcer
inducing a greater level of behavioral repetition, particularly as the time to reinforcement was approached.
This effect may have implications for reinforcing behavioral variability in humans.
Key words: behavioral variability, reinforcer magnitude, induction, pigeon, key peck

Bonem and Crossman (1988) comprehen- Variation in behavior usually is reduced as an


sively reviewed the literature involving the organism approaches, in time or space, a
effects of reinforcer magnitude on operant reinforcing stimulus (e.g., Akins, Domjan, &
behavior. They described many inconsistencies Gutierrez, 1994; Gharib, Gade, & Roberts,
in the literature and remarked, “…varying the 2004), even when behavioral variability is
magnitude of a reinforcer is a very complex reinforced (e.g., Cherot, Jones, & Neuringer,
operation” (p. 360). They also suggested that 1996). Cherot et al. exposed two groups of rats
this literature may benefit from more theoreti- to a fixed-ratio (FR) 4 contingency in which four
cally driven research. One approach to under- “successful” four-press sequences resulted in
standing the behavioral effects of reinforcer reinforcement. For one (repeat) group, a four-
magnitude is to posit that larger reinforcers may press sequence satisfied the FR requirement
not always strengthen or select behavior more only if it repeated one of the sequences from the
effectively than smaller reinforcers (e.g., Bizo, previous three trials, whereas for the other
Kettle, & Killeen, 2001; Doughty & Richards, (vary) group, a four-press sequence satisfied the
2002). The present research tested the view that FR requirement only if it was different than each
larger reinforcers sometimes exert their effects of the last three sequences (i.e., Lag 3). Not
by inducing greater behavioral repetition than surprisingly, sequence variability was consider-
smaller reinforcers. Improving our understand- ably higher in the vary group. Importantly,
ing of the effects of reinforcer magnitude on however, sequence variability decreased for
behavioral variability is critical for several each group as the rats progressed through the
reasons, including identifying better the benefi- FR 4 contingency (e.g., sequence variability was
cial and detrimental effects of reinforcers on lowest in the fourth segment of the FR 4
problem solving and creativity (e.g., Ariely, contingency). This decreased variability yielded
2010; Neuringer, 2003). more efficient responding (i.e., response se-
quences per reinforcer) for the repeat group
but lower efficiency for the vary group. One
interpretation of this finding is that behavioral
Portions of this research were supported by a SURF repetition was induced as the time to reinforce-
(Summer Undergraduate Research with Faculty) grant as
well as a Faculty Research and Development grant, both ment decreased, and the operant contingency
through the College of Charleston. The authors thank (i.e., vary or repeat) determined whether this
several students for their contributions to this research, induced repetition was consistent with the
particularly Nina Deese. reinforced behavioral unit (cf. Baum, 2012).
Address correspondence to Adam Doughty, Department
of Psychology, College of Charleston, 57 Coming Street,
As time to reinforcement approaches, larger
Charleston, SC 29424 (doughtya@cofc.edu). reinforcers may generate a greater level of
doi: 10.1002/jeab.50 induced behavioral repetition than smaller

1
2 DOUGHTY et al.

reinforcers. In this context, one can compare It is easier to evaluate the position that larger
Reed and Wright (1988) with Doughty reinforcers induce greater behavioral repetition
and Richards (2002). Reed and Wright assessed by using a dependent measure other than
the effects of reinforcer magnitude (i.e., num- response rate. Stahlman and Blaisdell (2011)
ber of food pellets) on rats exposed to a chained assessed the effects of reinforcer magnitude on
differential-reinforcement-of-low-rate (DRL) behavioral spatiotemporal variation. Pigeons
variable-ratio (VR) schedule. Response rates were exposed to a discrete-trial procedure in
under the DRL and VR schedules decreased and which a random-ratio (RR) 5 schedule was
increased, respectively, with increases in rein- presented after a 30-s intertrial interval (ITI).
forcer magnitude. The results are consistent There were five trial types, each signaled by a
with the notion that a larger reinforcer better distinct stimulus on a touchscreen. Four trial
strengthens or selects the reinforced behavioral types involved each combination of two levels of
unit. Doughty and Richards examined the reinforcer probability (12.5 and 4.4%) and
effects of reinforcer magnitude on the behavior reinforcer magnitude (25.2 and 2.8 s of hopper
of rats and pigeons under simple DRL schedules access). The final trial type involved a 100%
(i.e., water amount to rats and duration of food reinforcer probability and a 2.8-s reinforcer
access to pigeons). The larger reinforcer magnitude, but the data from these trials were
generated higher response rates, which reduced omitted from analysis. Stahlman and Blaisdell
considerably the number of obtained rein- measured both spatial variation (i.e., where on
forcers (see also Kirshenbaum, Brown, Hughes, the touchscreen the pecks occurred) and
& Doughty, 2008). These results are inconsistent temporal variation (i.e., the time between
with the notion that a larger reinforcer always pecks). The larger reinforcer resulted in greater
strengthens or selects the reinforced behavioral behavioral repetition in both space and time,
unit more effectively. A potentially important and this effect was larger at the lower reinforcer
difference across the studies is that the uncon- probability. The results support the conclusion
ditional reinforcer immediately followed DRL- that larger reinforcers can induce behavioral
schedule completion in Doughty and Richards, repetition to a greater extent than smaller
while in Reed and Wright, DRL-schedule reinforcers. However, this behavioral repetition
completion immediately produced the stimulus was not inconsistent with occurrence of the
conditions correlated with the VR schedule. reinforced behavioral unit. A stricter test would
Doughty and Richards (2002) interpreted involve the reduction of behavioral variation by
their findings within a framework that considers larger reinforcers under a reinforcement con-
the induced effects of reinforcers (e.g., Killeen, tingency requiring behavioral variability.
1979, 1985; Segal, 1972; Staddon & Simmelhag, The purpose of the present experiment was to
1971). Under a simple DRL schedule, a determine if larger reinforcers reduce behav-
response immediately produces a reinforcer ioral variability under a reinforcement contin-
only after a sufficiently long period of time gency that requires behavioral variation.
without the response. As such, this passage Pigeons were exposed to a two-component
of time can induce responding (e.g., Killeen, multiple schedule. In each component, four-
1979; Lewis & Dougherty, 1992), with more peck sequences across left and right keys were
induced behavior in situations involving reinforced according to a relative-frequency
larger reinforcers (e.g., Balsam, Brownstein, & threshold contingency (e.g., Denney &
Shull, 1978; Doughty & Richards, 2002) or Neuringer, 1998; Doughty & Lattal, 2001).
greater reinforcer-deprivation levels (Lewis & Under this contingency, a response sequence
Dougherty, 1992). Greater induced behavioral results in a reinforcer only if its relative
repetition under a DRL schedule can take the frequency is less than a threshold value.
form of a significantly larger number of Threshold values were 0.05 and 0.30 in separate
precriterion responses. Although Reed and conditions such that higher and lower levels of
Wright interpreted their increased VR-schedule behavioral variability were reinforced, respec-
response rates with larger reinforcers through tively. In each condition, reinforcer magnitude
an enhanced response-strengthening process, differed across components (i.e., 2-s or 6-s access
an account emphasizing greater induced repe- to grain). It was predicted that behavioral
tition by larger reinforcers is consistent with variability would be higher under the 0.05
their results. contingency than the 0.30 contingency (e.g.,
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 3

Grunow & Neuringer, 2002). Most importantly, back panel provided general illumination. The
it was predicted that the larger reinforcer would keylights and houselight were darkened during
generate lower levels of behavioral variability reinforcement. Reinforcement was access to
than the smaller reinforcer. Such a result would mixed grain delivered in a food hopper located
extend the findings obtained by Cherot et al. behind a rectangular opening (approximately
(1996) and Stahlman and Blaisdell (2011) and 6 cm high by 5 cm wide) centered on the bottom
support the claim that larger reinforcers can of the front wall. A white light also was present
induce greater behavioral repetition, even when during food availability. Programming and data
that repetition interferes with occurrence of the recording were controlled by a computer in an
reinforced behavioral unit. adjacent room using Med-PC IV software (Med-
Associates).
Method
Subjects Procedure
Subjects were eight male pigeons obtained Pretraining. Due to their more extensive
from the Palmetto Pigeon Plant (Sumter, SC). experimental histories, Pigeons 2196, 3776,
Five were White Carneau pigeons, approximate- 6067, 6117, and 9925 required no pretraining.
ly 3 yrs of age, with experimental histories of Pigeons 0136, 2728, and 9222 received one
reinforced behavioral variability (2196, 3776, session of magazine training, one session of key-
6067, 6117, and 9925). The other three pigeons peck shaping, and approximately five sessions of
(0136, 2728, and 9222) were an “in-house increasing FR-schedule requirements across the
breed” (bred between White Carneau and Silver two side keys. Specifically, one of the side keys
King; labeled, Hubble), approximately 2 yrs of randomly appeared red or green and operated
age, with experimental histories unrelated to until an FR schedule was completed. The FR-
reinforced behavioral variability. The pigeons schedule values increased until each of the four
were maintained at 70% (0136, 2196, and 9222), options (i.e., left green, right green, left red,
73% (2728 and 6067), or 80% (3776, 6117, and right red) were completed under an FR 20
9925) of their free-feeding body weights. The schedule.
weights lower than 80% were necessary to Following pretraining, all pigeons were ex-
maintain responding in prior experiments, so posed, for approximately 15 sessions, to proce-
there were no attempts to increase these weights dures that were nearly identical to the
in this study. Each pigeon was housed individu- procedures used in Condition 1 of the experi-
ally, and water and health grit were available ment proper (see below). Each session began
continuously in each home cage. If necessary, with a 2-min period in which the key lights and
each pigeon was fed mixed grain at least 1 hr houselight were darkened. After this presession
after its session to maintain its target weight. blackout, one of the two multiple-schedule
Lights were on in the room from 6:00 p.m. to components randomly started. The start of the
6:00 a.m., and sessions occurred Monday component was signaled by the onset of the
through Friday. houselight and the two side keys. The side keys
were red in one component and green in the
other component. Regardless of key color, a
Apparatus relative-frequency threshold contingency was in
Three identical three-key operant condition- effect such that a four-peck sequence was
ing chambers were used. Each chamber had a eligible for food delivery only if its weighted
work area approximately 30 cm high by 25 cm relative frequency was less than a threshold
wide by 30 cm long and was enclosed in a sound- value (see below). At the start of the component,
attenuating box. White noise was presented a pigeon could peck the left or right key. This
through a stereo near the chambers. Response peck darkened both key lights for a 0.5-s
keys (2.5-cm diameter) were located on the resetting delay; if a peck occurred to a dark
front wall and could be illuminated red or key, the 0.5-s delay restarted. After this delay,
green. Each of the two side keys was approxi- both key lights turned on, and the pigeon could
mately 3 cm from the side wall and approxi- peck either key again. This process continued
mately 6 cm from the middle key. A white until four pecks had occurred. Immediately
houselight located in the upper portion of the after the fourth peck, the key lights and
4 DOUGHTY et al.

houselight were terminated and either food response sequence that satisfied the threshold
delivery or a timeout occurred (i.e., the house- contingency resulted in food delivery (see
light remained on across the four-peck Denney & Neuringer, 1998; Doughty &
sequence). Lattal, 2001). The change allowed the pigeons
Because there were two response options and to receive more frequent exposure to the
four required pecks, there were 16 possible different reinforcer magnitudes. Second, ses-
response sequences. The computer determined sion duration was reduced to 10 multiple-
after the fourth peck which of these sequences schedule components (compared to the previ-
occurred on a trial. The relative frequency of ous 20 components). This change was made to
that sequence was calculated by dividing the minimize potential satiation effects, given the
number of times that sequence had occurred by duration of food access in the component with
the total number of sequences emitted (sepa- the larger reinforcer as well as the relatively high
rate contingencies operated across components obtained-reinforcement rates (see Table 1 be-
such that the occurrence of sequences in one low). Third, the duration of food delivery in the
component did not affect their calculation in component with the smaller reinforcer was
the other component). The relative frequencies increased from 2 to 3 s for Pigeons 2196 and
at the start of each session were taken from the 9222. This change was made to maintain
end of the preceding session. If the relative responding more effectively for these two
frequency was greater than the threshold value, pigeons in this component.
a timeout occurred. If the relative frequency was Condition 1 continued for a minimum of 50
less than or equal to the threshold value, the sessions and until responding stabilized across
sequence was eligible for food delivery. During nine consecutive sessions. These numbers were
this pretraining period, food only was delivered selected because the session duration was
following a “food-eligible” sequence if a variable- relatively brief (i.e., approximately 30 min).
interval (VI) 20-s schedule also had elapsed. If Responding was judged to be stable, via visual
the VI 20-s schedule had not elapsed, a timeout inspection, when there was minimal session-to-
occurred. Following food delivery, each of the session variability, and no trend, in either
16 relative frequencies in the component was multiple-schedule component for U-values
multiplied by a weighting coefficient of 0.95 (see and response rates. U-value (Miller &
Denney & Neuringer, 1998). Frick, 1949; Page & Neuringer, 1985) is a
During this pretraining period, the threshold common measure of behavioral variability and
value was 0.05 for four pigeons (2728, 3776, was calculated according to the formula: U ¼ - S
6067, and 6117), and 0.30 for the other four [(Rfi  log2 (Rfi)]/log2 (16)], where Rfi is the
pigeons (0136, 2196, 9222, and 9925). The same unweighted relative frequency of each of the 16
threshold value operated in both multiple- response sequences. A U-value of 1 implies
schedule components. Across components, complete uncertainty, or variability, and a U-
different reinforcer magnitudes were arranged. value of 0 implies complete certainty, or
For Pigeons 0136, 2728, 6117, and 9925, the repetition. Thus, higher U-values indicate
duration of food delivery and timeout was 2 s in higher levels of between-sequence variability
the component with the red keys and 6 s in the (e.g., a U-value of 1 indicates that each of the 16
component with the green keys, whereas these sequences occurred the same number of times).
assignments were reversed for Pigeons 2196, Response rate was calculated by dividing the
3776, 6067, and 9222. Component duration was number of sequences in a component by the
2 min with the restriction that a component total amount of time spent in that component,
ended only after a food delivery or timeout. A excluding time spent during reinforcer delivery
30-s intercomponent interval, in which all lights or timeout.
remained off, separated multiple-schedule com- Condition 2. Condition 2 was identical to
ponents. The small and large reinforcer magni- Condition 1 with two exceptions. First, the
tude components strictly alternated until each threshold value for each pigeon was reversed
had occurred 10 times across the session. (i.e., 0.05 to 0.30, or 0.30 to 0.05). Second, the
Condition 1. The procedures in Condition 1 key colors correlated with the different rein-
were identical to the aforementioned proce- forcer magnitudes were reversed (i.e., the
dures with the following three exceptions. First, pigeons previously exposed to red and green
the VI 20-s schedule was eliminated. Every keys correlated with the small and large
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 5

Table 1
For each pigeon in each condition, shown are the total number of sessions, response rate, reinforcement
rate, and proportion of reinforced sequences in each multiple-schedule component (i.e., Small and Large).
The means for these measures were calculated for the final stable sessions of each condition (with the
standard deviations in parentheses).

Responses/Min Reinforcers/Min Proportion of Reinforced Sequences


Condition Pigeon Sessions Small Large Small Large Small Large

0.05 0136 86 11.78 (0.65) 11.84 (0.52) 2.97 (0.79) 2.48 (0.63) 0.25 (0.07) 0.21 (0.06)
2196 78 8.73 (0.45) 9.27 (0.50) 2.66 (0.29) 2.82 (0.63) 0.31 (0.03) 0.31 (0.07)
2728 68 10.53 (0.39) 11.86 (0.57) 2.49 (0.39) 2.00 (0.43) 0.24 (0.04) 0.17 (0.04)
3776 59 10.66 (0.45) 9.92 (0.37) 2.92 (0.76) 2.41 (0.60) 0.27 (0.07) 0.25 (0.07)
6067 74 6.03 (1.16) 6.35 (0.71) 1.64 (0.61) 1.35 (0.33) 0.27 (0.09) 0.22 (0.06)
6117 127 7.90 (0.31) 6.82 (0.31) 2.38 (0.29) 2.14 (0.52) 0.30 (0.04) 0.31 (0.07)
9222 86 11.08 (0.48) 11.68 (0.55) 3.78 (0.31) 3.32 (0.75) 0.34 (0.04) 0.30 (0.07)
9925 87 13.22 (0.45) 13.63 (0.58) 3.26 (0.44) 3.39 (0.61) 0.25 (0.04) 0.25 (0.05)

0.30 0136 71 12.14 (0.40) 11.99 (0.50) 6.21 (1.26) 9.07 (0.78) 0.51 (0.07) 0.76 (0.06)
2196 103 7.22 (0.82) 7.19 (0.35) 4.81 (0.77) 5.43 (0.66) 0.67 (0.12) 0.76 (0.09)
2728 71 12.78 (0.70) 12.09 (0.81) 10.57 (1.12) 10.10 (1.17) 0.82 (0.11) 0.85 (0.05)
3776 134 10.24 (0.18) 10.69 (0.58) 8.32 (0.74) 9.38 (0.68) 0.81 (0.06) 0.88 (0.08)
6067 94 10.66 (0.59) 12.05 (0.59) 8.87 (0.57) 9.94 (0.93) 0.83 (0.03) 0.83 (0.09)
6117 60 8.77 (0.90) 8.76 (0.50) 7.85 (1.16) 7.53 (0.68) 0.90 (0.10) 0.87 (0.10)
9222 126 10.15 (0.41) 10.18 (0.91) 9.56 (0.64) 8.37 (0.86) 0.95 (0.07) 0.83 (0.11)
9925 98 11.37 (0.55) 12.17 (0.53) 9.10 (0.60) 8.80 (0.61) 0.80 (0.05) 0.72 (0.06)

reinforcers, respectively, were exposed to the ment rate was higher with the smaller reinforcer
red and green keys correlated with the large and magnitude under the 0.05 contingency for six of
small reinforcers and vice versa). Condition 2 eight pigeons, whereas this result occurred
continued for a minimum of 50 sessions and under the 0.30 contingency for only four of
until responding was judged to be stable over the pigeons. However, neither the main effect of
nine sessions by visual inspection. This condi- magnitude nor the interaction was significant
tion was terminated accidentally after eight (p’s >.19). The main effect of contingency was
consecutive stable sessions for Pigeon 9222. significant, with reinforcement rates higher
under the 0.30 contingency, F(1, 7) ¼ 81.58,
Results p <.001. Visual and statistical inspections of the
proportion of reinforced sequences were similar
Table 1 shows, for each pigeon, the number of to the reinforcement-rate assessments. Neither
sessions in each condition as well as mean the main effect of magnitude nor the interac-
response rate, mean reinforcement rate, and tion was significant (p’s >.27), but the main
mean proportion of reinforced sequences, and effect of contingency was significant with a
their respective standard deviations, during the higher proportion of reinforced sequences
final stable sessions of each condition in each under the 0.30 contingency, F(1, 7) ¼ 308.66,
component. The proportion of reinforced p <.001.
sequences was calculated by dividing the num- Figure 1 shows the U-values in each compo-
ber of reinforced sequences in a component by nent in the final stable sessions under the 0.05
the total number of sequences that occurred in contingency for each pigeon. The left and right
that component. Visual inspection of response columns of graphs display the U-values for the
rates of individual pigeons indicated that these pigeons that were exposed to the 0.05 contin-
rates were not affected systematically by contin- gency in the first and second conditions,
gency or reinforcer magnitude. A two-way respectively. Figure 2 shows the U-values in
repeated-measures ANOVA (contingency, rein- each component in the final stable sessions
forcer magnitude) confirmed the absence of under the 0.30 contingency for each pigeon.
significant differences (both main effects and The left and right columns of graphs display the
interaction p’s >.24). Visual inspection of rein- U-values for the pigeons that were exposed to
forcement rates indicated that mean reinforce- the 0.30 contingency in the second and first
6 DOUGHTY et al.

Fig. 1. Displayed are U-values for each pigeon in each multiple-schedule component during the final stable sessions of the
0.05 contingency. The left and right columns display the U-values for pigeons that were exposed to the 0.05 contingency in
the first and second conditions, respectively. Reinforcer magnitude is differentiated by the circles and triangles.
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 7

Fig. 2. Displayed are U-values for each pigeon in each multiple-schedule component during the final stable sessions of the
0.30 contingency. The left and right columns display the U-values for pigeons that were exposed to the 0.30 contingency in
the second and first conditions, respectively. Reinforcer magnitude is differentiated by the circles and triangles.
8 DOUGHTY et al.

conditions, respectively. The 0.05 contingency are those sequences ending in three or four
generated higher U-values than the 0.30 contin- consecutive pecks to the same key (i.e., RLLL,
gency, F(1, 7) ¼ 76.06, p <.001, and the smaller LRRR, LLLL, and RRRR). Thus, the relatively
reinforcer magnitude produced higher U-val- low levels of between-sequence variability under
ues than the larger reinforcer magnitude, F(1, the lenient 0.30 contingency resulted from the
7) ¼ 19.09, p ¼.003. The interaction was not dominance of sequences that culminated with
significant (p ¼.20). In terms of observing the within-sequence repetition.
differential effects of reinforcer magnitude in Figures 5 and 6 display the relative frequen-
individual pigeons, the difference in U-values cies under the 0.05 contingency as a function of
can be seen most clearly under the 0.05 reinforcer magnitude for the pigeons that were
contingency in Pigeons 2728, 0136, 6067, exposed to this contingency in the first and
9222, and 6117. Under the 0.30 contingency, second conditions, respectively. The distribu-
it can be seen most clearly in Pigeons 3776, 9222, tions displayed in these two figures deviate
and 6117. considerably from the distributions under the
The results presented thus far indicate 0.30 contingency. As Figure 1 showed, the
that both the threshold contingency and effects of reinforcer magnitude on U-values
reinforcer magnitude impacted the degree of under the 0.05 contingency were observed most
sequence variation. To understand better how clearly for Pigeons 2728, 0136, 6067, 9222, and
these variables impacted behavioral variability, 6117. Accordingly, the distributions for these
Figures 3 to 6 display obtained relative-frequency pigeons with the larger reinforcer were more
distributions as a function of reinforcer magni- uneven than the distributions obtained with the
tude and threshold contingency. Specifically, smaller reinforcer. The source of this uneven-
the relative frequency of each response se- ness was that the larger reinforcer reduced the
quence was calculated for the final stable likelihood of particular sequences to very low
sessions of each condition in each component levels. Specifically, across components for these
(i.e., the number of times each sequence pigeons, there were a total of 15 sequences with
occurred in a component divided by the total a relative frequency lower than .025. Fourteen of
number of sequences that occurred in the these 15 sequences occurred with the larger
component [without the weighting coeffi- reinforcer, and 11 of those 14 sequences ended
cient]). In each figure, the relative frequencies in a changeover (i.e., ended in LR or RL). Thus,
are displayed according to a measure of within- although the 0.05 contingency maintained high
sequence variation. The leftmost eight sequen- levels of between-sequence variation with each
ces culminate with a changeover from one key to reinforcer magnitude, the larger reinforcer
the other (i.e., LRLR, RLRL, RLLR, LRRL, produced less behavioral variation by reducing
RRLR, LLRL, RRRL, and LLLR). The next four the likelihood of sequences that culminated in
sequences culminate with two consecutive within-sequence variation.
responses to the same key (i.e., RRLL, LLRR,
LRLL, and RLRR). Finally, the next two Discussion
sequences culminate with three consecutive
responses to the same key (i.e., RLLL and The present experiment produced three
LRRR), and the last two sequences do not primary results. First, consistent with prior
contain a changeover (i.e., LLLL and RRRR). findings (e.g., Grunow & Neuringer, 2002),
In other words, the leftmost sequences culmi- the 0.05 threshold contingency produced more
nate with within-sequence variation and the behavioral variability than the 0.30 threshold
rightmost sequences culminate with different contingency (e.g., higher U-values). Second,
degrees of within-sequence repetition. differences in reinforcer magnitude also im-
Figures 3 and 4 display these relative pacted behavioral variability, with the larger
response-sequence frequencies under the 0.30 reinforcer producing less variability (e.g., lower
contingency as a function of reinforcer magni- U-values). Third, the primary source of this
tude for the pigeons that were exposed to this reduced behavioral variability by the larger
contingency in the first and second conditions, reinforcer was lower within-sequence variability
respectively. Across the small and large rein- (indexed by changes in the relative-frequency
forcer magnitudes (columns), for all 16 cases, at distributions). Thus, the present findings ex-
least three of the four most frequent sequences tend Cherot et al. (1996) and Stahlman and
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 9

Fig. 3. Displayed are the relative response-sequence frequencies in each multiple-schedule component during the final
stable sessions of the 0.30 contingency for the pigeons that were exposed to the 0.30 contingency in the first condition.
10 DOUGHTY et al.

Fig. 4. Displayed are the relative response-sequence frequencies in each multiple-schedule component during the final
stable sessions of the 0.30 contingency for the pigeons that were exposed to the 0.30 contingency in the second condition.
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 11

Fig. 5. Displayed are the relative response-sequence frequencies in each multiple-schedule component during the final
stable sessions of the 0.05 contingency for the pigeons that were exposed to the 0.05 contingency in the first condition.
12 DOUGHTY et al.

Fig. 6. Displayed are the relative response-sequence frequencies in each multiple-schedule component during the final
stable sessions of the 0.05 contingency for the pigeons that were exposed to the 0.05 contingency in the second condition.
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 13

Blaisdell (2011) and show that larger reinforcers response-strengthening, or selection, effects.
can interfere with the reinforcement of behav- Future research can assess whether larger
ioral variability. This effect primarily was the reinforcers always produce lower levels of
result of the larger reinforcer producing greater reinforced behavioral variability, or whether
within-sequence repetition, particularly at the particular contingencies can minimize their
end of the sequence. The source of this result detrimental impact.
can be interpreted as a function of the larger It is important to underscore the different
reinforcer inducing greater behavioral repeti- variability levels engendered by the threshold
tion than the smaller reinforcer. contingencies in the present experiment. The
The present findings extend the literature higher U-values observed under the 0.05
involving the effects of induced behavioral contingency replicates past research (e.g.,
repetition (e.g., Cherot et al., 1996; Stahlman Grunow & Neuringer, 2002) and demonstrates
& Blaisdell, 2011). Stahlman and Blaisdell the powerful role of reinforcement contingen-
reported greater behavioral spatiotemporal cies in generating behavioral variation (e.g.,
repetition in pigeons with larger reinforcers, Neuringer, 2002; Neuringer & Jensen, 2013).
however they did not differentially reinforce Stated differently, although the larger reinforc-
behavioral variation. In the present experiment, er resulted in lower levels of behavioral variabil-
a larger reinforcer resulted in lower levels of ity, relatively high levels of behavioral variability
behavioral variability under a contingency were established under the 0.05 contingency
requiring behavioral variation. These detrimen- with each reinforcer magnitude. These latter
tal effects of the larger reinforcer primarily results are integral to consider in discussions
resulted from it reducing the probability of about the potential negative effects of promot-
sequences culminating in variation (i.e., se- ing creativity and problem solving in people
quences ending in LR or RL). Cherot et al. using reinforcement-based approaches (e.g.,
observed induced behavioral repetition as time Ariely, 2010; Neuringer, 2003). That is, even if
to reinforcement was approached under a the reinforcer-magnitude effects observed in
variability reinforcement contingency. The the present study can be extended to human
present results extend those findings by demon- behavior, the necessity of using reinforcement
strating that this induced behavioral repetition to maintain high overall variability levels cannot
is greater with larger reinforcers. Further, the be ignored.
results reported by Reed and Wright (1988) and Procedural limitations in the present work
Doughty and Richards (2002) can be reconciled should be noted. First, the absolute value of the
if their respective VR-schedule and DRL-sched- larger reinforcer was not studied beyond 6-s
ule response-rate changes were due, at least in access to grain so as to minimize potential
part, to greater behavioral repetition induced by satiation effects. Larger reinforcers may have
larger reinforcers. exerted even greater repetition-inducing ef-
The operant contingency is one factor that fects. Second, an enhanced impact of reinforcer
can determine whether greater induced behav- magnitude may be observed with a greater
ioral repetition by larger reinforcers is observed. difference between the two magnitudes (i.e.,
For example, Doughty, Galuska, Dawson, and this difference was 4 s for six pigeons, and only
Brierley (2012) obtained enhanced lever-press 3 s for the other two pigeons). Third, the effects
acquisition in rats under unsignaled delayed of the larger reinforcer may have been reduced,
reinforcement (i.e., a tandem FR 1 differential- to some degree, by the manipulation of timeout
reinforcement-of-other-behavior [DRO] 30-s duration concurrent with reinforcer duration
schedule) with a larger reinforcer (i.e., six, (i.e., timeout duration in each component
rather than one, food pellets). Induced lever matched the duration of food access in that
pressing by the larger reinforcer near the end of component). We manipulated timeout duration
the DRO requirement was eliminated, or at least concurrent with reinforcer duration to assess
reduced considerably, by the relatively long the negative effects of larger reinforcers under
response–reinforcer delay created by the tan- stricter conditions (i.e., conditions less likely
dem FR 1 DRO 30-s schedule (i.e., compare the to demonstrate an effect). Neuringer (1991)
results to Doughty & Richards, 2002). As such, reported higher levels of response variation after
this relatively strong operant contingency al- a longer time since the preceding response and
lowed the larger reinforcer to exert its greater following timeout rather than reinforcement.
14 DOUGHTY et al.

Based on these results, the longer timeout Lee, 2004; Hopkinson & Neuringer, 2003;
duration in the present experiment should Lee, McComas, & Jawor, 2002; Miller &
have increased behavioral variation more than Neuringer, 2000; Napolitano, Smith, Zarcone,
the briefer timeout duration, thus reducing the Goodkin, & McAdam, 2010). Translation is
repetition induced by the larger reinforcer. needed to understand fully how the present
Finally, our statistical analyses probably were results bear on that work; however, our findings
limited by the relatively small number of pigeons tentatively suggest that it may be easier
used in this research. to reinforce behavioral variability in people
There were no consistent response-rate differ- using smaller rather than larger reinforcer
ences in the present experiment as a function of magnitudes.
threshold value or reinforcer magnitude. The
absence of response-rate differences as a
References
function of threshold value may seem surprising
given the considerable discrepancy in reinforce- Akins, C. K., Domjan, M., & Gutierrez, G. (1994). Topogra-
ment rate across conditions. Grunow and phy of sexually conditioned behavior in male Japanese
Quail (Coturnix japonica) depends on the CS-US
Neuringer (2002; Experiment 1) measured interval. Journal of Experimental Psychology: Animal Behav-
both U-value and response rate as a function ior Processes, 20, 199–209.
of threshold value (i.e., 0.037, 0.055, 0.074, and Ariely, D. (2010). The upside of irrationality: The unexpected
0.37, across groups) and reinforcement sched- benefits of defying logic at work and at home. New York, NY:
HarperCollins Publishers.
ule (i.e., FR 1 [as studied in the present Balsam, P. D., Brownstein, A. J., & Shull, R. L. (1978). Effect
experiment], VI 1 min, and VI 5 min, across of varying the duration of grain presentation on
conditions). They also observed minimal re- automaintenance. Journal of the Experimental Analysis of
sponse-rate differences in their FR 1 condition, Behavior, 29, 27–36.
despite considerable U-value differences across Baum, W. M. (2012). Rethinking reinforcement: Allocation,
induction, and contingency. Journal of the Experimental
groups (e.g., the mean response-rate difference Analysis of Behavior, 97, 101–124.
across the 0.055 and 0.37 groups was two Bizo, L. A., Kettle, L. C., & Killeen, P. R. (2001). Rats don’t
sequences per min). The absence of a differen- always respond faster for more food: The paradoxical
tial effect of reinforcer magnitude on response incentive effect. Animal Learning and Behavior, 29, 66–78.
Bonem, M., & Crossman, E. K. (1988). Elucidating the effects
rate in the present experiment also is not of reinforcement magnitude. Psychological Bulletin, 104,
without precedent (e.g., Bonem & Crossman, 348–362.
1988). At this point, it only can be concluded Cammilleri, A. P., & Hanley, G. P. (2005). Use of a lag
that reinforcer magnitude is more likely to affect differential reinforcement contingency to increase
response-sequence variation than response-se- varied selections of classroom activities. Journal of Applied
Behavior Analysis, 38, 111–115.
quence rate, at least under the conditions Cherot, C., Jones, A., & Neuringer, A. (1996). Reinforced
investigated here. variability decreases with approach to reinforcers.
The present findings may be useful to consider Journal of Experimental Psychology: Animal Behavior Process-
in the context of both reinforcement theory and es, 22, 497–508.
Denney, J., & Neuringer, A. (1998). Behavioral variability is
application. Behavior-analytic theorists increas- controlled by discriminative stimuli. Animal Learning &
ingly have questioned the typical response- Behavior, 26, 154–162.
strengthening account of reinforcement (e.g., Doughty, A. H., Galuska, C. M., Dawson, A. E., & Brierley,
Baum, 2012; Shahan, 2010), including by placing K. P. (2012). Effects of reinforcer magnitude on
a greater emphasis on the induced effects of response acquisition with unsignaled delayed reinforce-
ment. Behavioural Processes, 90, 287–290.
reinforcers (e.g., Baum, 2012). A conceptualiza- Doughty, A. H., & Lattal, K. A. (2001). Resistance to change
tion of the reinforcement process that includes of operant variation and repetition. Journal of the
induction effects may benefit from considering Experimental Analysis of Behavior, 76, 195–215.
the present results as well as the general relation Doughty, A. H., & Richards, J. B. (2002). Effects of reinforcer
magnitude on responding under differential-reinforce-
between behavioral variation and schedule- ment-of-low-rate schedules of rats and pigeons. Journal of
induced effects (e.g., Cherot et al., 1996; Lee, the Experimental Analysis of Behavior, 78, 17–30.
Sturmey, & Fields, 2007; Neuringer & Jensen, Gharib, A., Gade, C., & Roberts, S. (2004). Control of
2013). Practically speaking, the reinforcement of variation by reward probability. Journal of Experimental
behavioral variability in people with and without Psychology: Animal Behavior Processes, 30, 271–282.
Grunow, A., & Neuringer, A. (2002). Learning to vary and
autism has received increased interest in applied varying to learn. Psychonomic Bulletin & Review, 9, 250–258.
behavior analysis recently (e.g., Cammilleri & Harding, J. W., Wacker, D. P., Berg, W. K., Rick, G., & Lee,
Hanley, 2005; Harding, Wacker, Berg, Rick, & J. F. (2004). Promoting response variability and stimulus
REINFORCER MAGNITUDE AND REINFORCED VARIABILITY 15

generalization in martial arts training. Journal of Applied Neuringer, A. (1991). Operant variability and repetition as
Behavior Analysis, 37, 185–196. functions of interresponse time. Journal of Experimental
Hopkinson, J., & Neuringer, A. (2003). Modifying behavioral Psychology: Animal Behavior Processes, 17, 3–12.
variability in moderately depressed students. Behavior Neuringer, A. (2002). Operant variability: Evidence,
Modification, 27, 251–264. functions, and theory. Psychonomic Bulletin and Review,
Killeen, P. R. (1979). Arousal: Its genesis, modulation, and 9, 672–705.
extinction. In M. D. Zeiler, & P. Harzem (Eds.), Advances Neuringer, A. (2003). Reinforced variability and creativity.
in analysis of behaviour, Vol. 1 Reinforcement and the In K. A. Lattal, & P. N. Chase (Eds.), Behavior Theory and
organization of behaviour (pp. 31–78). Chichester, Philosophy (pp. 323–338). New York, NY: Kluwer
England: Wiley. Academic/Plenum.
Killeen, P. R. (1985). Incentive theory: IV. Magnitude of Neuringer, A., & Jensen, G. (2013). Operant variability.
reward. Journal of the Experimental Analysis of Behavior, 43, In G. J. Madden, (Editor-in-Chief) W. V. Dube,
407–417. T. D. Hackenberg, G. P. Hanley, & K. A. Lattal (Associate
Kirshenbaum, A. P., Brown, S. J., Hughes, D. M., & Doughty, Editors). APA handbook of behavior analysis: Volume 1
A. H. (2008). Differential reinforcement-of-low-rate- (pp. 513–546). Washington, DC: American Psychologi-
schedule performance and nicotine administration: A cal Association.
systematic investigation of dose, dose-regimen, and Page, S., & Neuringer, A. (1985). Variability is an operant.
schedule requirement. Behavioural Pharmacology, 19, Journal of Experimental Psychology: Animal Behavior Process-
683–697. es, 11, 429–452.
Lee, R., McComas, J. J., & Jawor, J. (2002). The effects of Reed, P., & Wright, J. E. (1988). Effects of food reinforce-
differential and lag reinforcement schedules on varied ment on free-operant response rates. Journal of the
verbal responding by individuals with autism. Journal of Experimental Analysis of Behavior, 49, 75–85.
Applied Behavior Analysis, 35, 391–402. Segal, E. F. (1972). Induction and the provenance of
Lee, R., Sturmey, P., & Fields, L. (2007). Schedule-induced operants. In R. M. Gilbert, & J. R. Millenson (Eds.),
and operant mechanisms that influence response Reinforcement: Behavioral analyses (pp. 1–34). New York:
variability: A review and implications for future inves- Academic Press.
tigations. The Psychological Record, 57, 429–455. Shahan, T. A. (2010). Conditioned reinforcement and
Lewis, P., & Dougherty, D. M. (1992). Pigeon performance response strength. Journal of the Experimental Analysis of
on a variable-interval omission schedule at different Behavior, 93, 269–289.
levels of food deprivation. Behavioural Processes, 27, Staddon, J. E. R., & Simmelhag, V. L. (1971). The
27–36. “superstition” experiment: A reexamination of its
Miller, G. A., & Frick, F. C. (1949). Statistical behavioristics implications for the principles of adaptive behavior.
and sequences of responses. Psychological Review, 56, Psychological Review, 78, 3–43.
311–324. Stahlman, W. D., & Blaisdell, A. P. (2011). The modulation
Miller, N., & Neuringer, A. (2000). Reinforcing variability in of operant variation by the probability, magnitude, and
adolescents with autism. Journal of Applied Behavior delay of reinforcement. Learning and Motivation, 42,
Analysis, 33, 151–165. 221–236.
Napolitano, D. A., Smith, T., Zarcone, J. R., Goodkin, K., &
McAdam, D. B. (2010). Increasing response diversity in
children with autism. Journal of Applied Behavior Analysis, Received: November 20, 2012
43, 265–271. Final Acceptance: August 28, 2013

You might also like