Professional Documents
Culture Documents
Module 10 Operant & Cognitive Approaches
Module 10 Operant & Cognitive Approaches
MODULE
& Cognitive
Approaches
Operant conditioning, also called instrumental conditioning, is a kind During operant conditioning, Bart watching others.
of learning in which an animal or human performs some behavior, and the performed observable behaviors that
following consequence (reward or punishment) increases or decreases the were influenced by observable consequences. In comparison, Tony
chance that an animal or human will again perform that same behavior. learned how to skateboard through observation and imitation,
For example, if Bart performed a particular behavior, such as which involved unobservable mental processes and is called cogni-
picking up a teddy bear, the consequence that followed—getting a tive learning. We’ll discuss cognitive learning later in this module.
rewarding apple—increased the chance that Bart would again pick
up the teddy bear. Because of what Bart learned through operant
conditioning, he starred in 20 movies and became the highest paid What’s Coming
animal actor, making about $10,000 a day (Brennan, 1997). That’s a In the first half of this module, we’ll discuss the history and proce-
salary that most of us would be very happy to bear! dure of operant conditioning, how operant conditioning differs
Operant conditioning seems rather straightforward. You per- from classical conditioning, how consequences or reinforcers work,
form an action or operate on your environment, such as studying and other examples of operant conditioning. In the second half of
hard. The consequence of your studying, such as how well you do this module, we’ll explain the history of cognitive learning, the
on exams, increases or decreases the likelihood that you will per- theory behind observational learning, and the role of insight
form the same behavior—studying hard—in the future. learning.
Besides learning by having your behaviors rewarded or pun- We’ll begin with an important study that involved a cat, a
ished, you can also learn in a very different way. puzzle box, and a smelly fish.
INTRODUCTION 213
A. Operant Conditioning
Background: Thorndike and Skinner
We told you how a trainer used operant conditioning to teach Bart to perform 45 different behaviors on cue.
How did Operant conditioning has now been applied to many different settings, such as training animals to perform, train-
Bart become a
ing children to use the potty, stopping retarded children from injuring themselves, and helping autistic children
movie star?
learn social behaviors. However, the discovery of operant behavior involved two different researchers who worked
in two different laboratories on two different kinds of problems. So that you can appreciate the thinking that led to
operant conditioning, we’ll visit the laboratories of the two important researchers—E. L. Thorndike and B. F. Skinner.
Figure/Text Credit: Graph on right based on data from Behavior of Organisms, by B. F. Skinner, 1938. Appleton-Century-Crofts.
record its escape time. After Thorndike graphs the data of his picking up the teddy bear is that he receives an apple, which is a
(graph below), we see a gradual lessening in the time needed desirable effect. This desirable effect modifies his response by increasing
to escape. Notice that on the first trial the cat needed over the chances that Bart will repeat the same response.
240 seconds to hit the escape latch, but by the last trial, the By measuring or recording operant responses, Skinner can analyze
cat hits the escape latch in less than 60 seconds. animals’ ongoing behaviors during learning. He calls this kind of learn-
ing operant conditioning, which focuses on how consequences (rewards
Law of effect: In escaping the puzzle or punishments) affect behaviors.
box, the cat’s successful responses
Time to escape, in seconds
240 are strengthened and this results in A simple example of operant conditioning occurs when a rat in an
quicker escape times. experimental box accidentally presses a bar. If the bar press is followed
180
by food, this consequence increases the chance that the rat will press
120 the bar again. As the rat presses the bar more times, more food follows,
60
which in turn increases the chances that the rat will continue to press
the bar (indicated by the rise of the red line in the figure below).
0
Trials of a cat to escape
from puzzle box
Bar pressing Food
Cumulative
response
The law of effect states that behaviors followed by positive Using his newly developed procedure of operant conditioning, B. F.
consequences are strengthened, while behaviors followed by neg- Skinner spent the next 50 years exploring and analyzing learning in rats,
ative consequences are weakened. pigeons, schoolchildren, and adults.
Thorndike’s (1898) findings were significant because they The 1920s and 1930s gave learning a mighty jolt with the discovery
suggested that the law of effect was a basic law of learning of two general principles—Pavlov’s classical conditioning and Skinner’s
and provided an objective procedure to study it. Thorndike’s operant conditioning. For the first time, psychologists had two methods
emphasis on studying the consequences of goal-directed to analyze learning processes in an objective way.
behavior was further developed and expanded by B. F. Now we’ll examine Skinner’s ingenious procedure for operant condi-
Skinner. tioning in more detail.
Figure/Text Credit: Adapted from “Treating Chronic Food Refusal in Young Children: Home-Based Parent Training,” by M. A. Werle, T. B. Murphy & K. S. Budd, 1993,
be applied to teach toilet training. et al., 2002). Researchers taught parents how to
1. Target behavior. The target behavior or goal is for use the principles of operant conditioning to
Sheryl to urinate in the toilet. overcome food refusal in their young children.
2. Preparation. Before training 1. Target behavior. The target behavior or goal was
begins, put all of Sheryl’s toys for the child to taste, chew, and eat a food (usually fruits
away so that she will not be dis- or vegetables) that she or he has persistently refused to eat.
tracted. Then give her a large glass 2. Preparation. Researchers first showed mothers how to shape and rein-
of apple juice, so that she will force target behaviors. Next, each mother shaped the target behavior in her
have to urinate soon. child in the home setting.
3. Reinforcers. Select reinforc- 3. Reinforcers. Each time the child performed or emitted a target behav-
ers, which can be candy, verbal ior, the mother immediately reinforced the child with a positive reinforcer,
praise, or a hug. Each time Sheryl such as praise, attention, or a smile.
performs or emits a desired 4. Shaping. The shaping procedure consisted of having the child notice
Photo Credits: top, © Monkey Business Images Ltd./Photolibrary; left, © Peter Southwick/Stock, Boston
behavior, you immediately rein- the food and let it be placed in his or her mouth, letting the child taste the
4 steps in force it. The reinforcer increases food, and, finally, having the child chew and swallow the food.
toilet training the likelihood that the behavior The graph below explains and shows the success of using operant con-
will be repeated. ditioning to overcome food refusal in young children (Werle et al., 1993).
4. Shaping. Just as Skinner used the shaping proce-
dure in conditioning a rat to press a bar, you can use
5 Shaping
a similar shaping procedure in conditioning Sheryl
4
to use the toilet. Each time Sheryl performs a behav-
food eaten
3
Pieces of
1 Goal. The goal of operant conditioning 1 Goal. The goal of classical conditioning is to create a new
is to increase or decrease the rate of some response to a neutral stimulus. In Bart’s case, he will
response, which usually involves shaping. make a new response, salivation, to the sound of
In Bart’s case, the goal was to increase his a horn, which is the neutral stimulus because it
rate of holding a teddy bear. does not usually cause Bart to salivate.
2 Voluntary response. Bart’s behavior of holding a teddy 2 Involuntary response. Salivation is an example of a physiological
bear is a voluntary response because he can perform it at will. reflex. Physiological reflexes (salivation, eye blink) are triggered or
Bart must first perform a voluntary response before getting elicited by some stimulus and therefore called involuntary responses.
a reward.
3 Elicited response. As Bart eats an
3 Emitted response. Bart voluntarily performs or emits apple, it will trigger the involuntary physi-
some response, which Skinner called the operant response ological reflex of salivation. Thus, eating
(holding teddy bear). Skinner used the term emit to indicate the apple, which is called the uncondi-
that the organism acts or operates on the environment. In tioned stimulus, triggers or elicits an
most cases, animals and humans are shaped to emit the involuntary ref lex response, salivation,
desired responses. which is called the unconditioned response.
headache and thus increases the chances of
% !# your taking an aspirin in the future. Don’t
# be confused by the fact that both positive
# ! $ and negative reinforcers increase the fre-
Sessions &" quency of the responses they follow.
Besides positive and negative reinforc-
$ ers, there are also primary reinforcers, such
In this study, reinforcement and punishment proved an effective combination to as food, and secondary reinforcers, such as
treat a potentially dangerous problem. Next, we’ll discuss two kinds of reinforcements. money and coupons.
218 MODULE 10 OPER ANT & COGNITIVE APPROACHES
Reinforcers Punishment
A student might repeat a behav- Although placing a child in time-out and giving a
How are ior, such as study, because the Are there child an electric shock seem very different, they
chocolate and consequence is food, a primary different kinds are both consequences that can decrease the
a coupon alike? reinforcer, or be quiet on the of punishment? occurrence of certain behaviors.
school bus because the conse- There are two kinds of punish-
quence is a coupon, a secondary reinforcer. ment or consequences—positive and negative—that
decrease the occurrence of behaviors. Both positive
Primary Reinforcers
and negative punishment function as “stop signs”; they
If you made yourself study for 2
stop or decrease the occurrence of a behavior.
hours before rewarding yourself
with chocolate, you would be using Positive Punishment
a primary reinforcer. There is a school in Massachusetts that uses a controversial method of pun-
A primary reinforcer is a stimulus, Food: ishment to discourage students from engaging in dangerous behaviors.
such as food, water, or sex, that is primary reinforcer Many of the children at this school have severe behavior problems, such as
innately satisfying and requires no hitting strangers in the face, biting others or themselves, and eye gouging.
learning on the part of the subject to become pleasurable. The school requires students to wear back-
Primary reinforcers, such as eating, drinking, or hav- packs, which contain an electric device
ing sex, are unlearned and innately pleasurable. Brain that allows staff members to apply a
scans (p. 70) showed that these activities, including eating shock to electrodes attached to the stu-
chocolate, activate the brain’s built-in or inherited reward/ dent’s body whenever t he student
pleasure center (Belin & Rauscent, 2006). Although brain engages in dangerous or prohibited
scans were discovered after Skinner, they have proven behaviors (L. Kaufman, 2007). The goal
him right: Primary reinforcers are innately satisfying is to decrease the occurrence of certain
and require no training because they automatically acti- behaviors, and the method used is an
vate the brain’s built-in reward/pleasure center. In our example of positive punishment.
example, chocolate is a primary reinforcer for studying. Positive punishment refers to presenting an aversive (unpleasant) stimulus
However, many behaviors are aided or maintained by after a response. The aversive stimulus decreases the chances that the response
secondary reinforcers. will recur.
The aversive stimulus in the above example is electric shock, which decreases
Secondary Reinforcers
the chances that dangerous and self-destructive behaviors will occur.
A school teacher used a secondary reinforcer when she
There’s a second type of punishment, called negative punishment, that
gave each child a coupon good for fun prizes if he or she
can also be used to decrease children’s undesirable behaviors.
ate fruits and vegetables during lunch. A coupon is an
example of a secondary reinforcer. Negative Punishment
A secondary reinforcer is any stimulus that has acquired its One kind of undesirable behavior in young children is the persistent refusal
reinforcing power through experience; secondary reinforcers are of parental requests, a problem called noncompliance.
learned, such as by being paired with primary reinforcers or Noncompliance refers to a child refusing to follow directions, carry out a
Photo Credits: top right, © Ron Nickel/Photolibrary; left, © Simon Belcher/Alamy
Figure/Text Credit: Reinforcement diagram from Psychology: Themes and Variations, by Wayne Weiten, 2nd ed., figure 6.13. Copyright © 1992 by Wadsworth, Inc.
B. F. Skinner (1904–1990)
of reinforcement. called the cumulative record.
cumulative record shows you an animal’s ongoing responses and reinforcement schedule, which is very effective in maintaining
reinforcements across time. behavior over the long run. My dog keeps shaking my hand on
We’ll first look at how two general schedules of reinforcement— command because some of the time she gets a treat.
continuous and partial reinforcement—can greatly affect ongoing We’ll discuss the four common schedules of partial rein-
behavior. forcement and show how differently they affect behavior.
Fixed-interval schedule means that a rein- 2 involves emitting sound waves (too high for
forcer occurs following the first response that humans to hear) and analyzing the waves
occurs after a fixed interval of time. that are reflected back from objects. The dol-
A fixed-interval schedule has slow responding at first, but as phin’s echolocation is so sensitive that it can
the time for the reinforcer nears, responses increase. detect a quarter-sized metal disk 100 feet
away, no matter how murky the water is.
Variable-Ratio Schedule
2. After locating the mine, the dolphin
If a slot machine pays off after an average of returns to the surface and is trained to touch
25 pulls, the gambler is on a variable-ratio a rubber disk at the front of the boat. The
schedule. trainer rewards the dolphin with a fish for
Variable-ratio schedule means that a rein- this and many other behaviors.
forcer is delivered after an average number of
correct responses has occurred.
3
3. The trainer places on the dolphin’s nose
The variable-ratio schedule produces a hollow cone attached to a plastic cylinder
a high rate of responding because the person (gambler) doesn’t that rides on the dolphin’s back. Again, the
know which response will finally produce the payoff. dolphin dives and echolocates the mine, but
(Variable-Interval) © SuperStock; top right, © AP Images/Brien Aho/US Navy
=
of extinction.
the presence of unreinforced stimuli.
In operant conditioning, extinction
In classical conditioning, discrimination is
refers to the reduction in an operant
the tendency for some stimuli but not others to
response when it is no longer followed by
elicit a conditioned response.
the reinforcer.
One problem with Bart was that he In classical con-
would repeatedly pick up and hold the ditioning, extinction
Generalization: Bart transfers his teddy bear to receive an apple. To control refers to the reduc-
response from teddy bear to live cub. this problem, the trainer used a cue—rais- tion in a response
ing his arms in the air—to signal that only when the condi-
Although Bart was relatively tame, then would Bart receive an apple for his tioned stimulus is
his trainer took no chances of Bart’s behavior. This is an example of a discrimi- no longer followed
wilder nature coming out and kill- native stimulus. by the unconditioned
ing the bear cub. For this reason, the A discriminative stimulus is a cue that a
stimulus. Extinction: Bart
trainer started by conducting the initial behavior will be reinforced. stops behaviors if
After under- reinforcers stop.
conditioning with a stuffed teddy bear.
going extinction, Spontaneous
Only after Bart had learned to pick up recovery: After
Photo Credits: top and bottom center, © George Frey; left, © Stephen Kraseman/DRK Photo
Bart may show
and hold the teddy bear on cue did the extinction, Bart’s
spontaneous behavior returns.
trainer substitute a live bear cub. As the
recovery.
trainer had predicted, Bart transferred
In operant conditioning, spontaneous
his holding the teddy bear to holding
recovery refers to a temporary recovery in
the live bear cub, a phenomenon called
generalization. the rate of responding.
In operant conditioning, generalization In classical conditioning, spontaneous
means that an animal or person emits the recovery refers to the temporary occur-
same response to similar stimuli. rence of the conditioned response in the
Discrimination: Bart obeys signals from presence of the conditioned stimulus.
In classical conditioning, generalization his trainer but not from a stranger.
is the tendency for a stimulus similar to the Remember that all four phenom-
original conditioned stimulus to elicit a If you pay close attention to an animal ena—generalization, discrimination,
r e s p o n s e s imila r t o t h e c o n di t i o n e d trainer, you’ll notice that discriminative extinction, and spontaneous recov-
response. stimuli, such as a hand signal or whistle, ery—occur in both operant and classi-
A common and sometimes embar- are used to signal the animal that the next cal conditioning.
rassing example of generalization occurs behavior will be reinforced. One distinctive characteristic of
when a young child generalizes the word Young children learn to discriminate operant conditioning is that it usu-
“Daddy” to other males who appear sim- between stimuli when their parents rein- ally has an observable response and
ilar to the child’s real father. As quickly force their saying “Daddy” in the presence an observable reinforcer. Next, we
as possible, embarrassed parents teach of their real fathers but do not reinforce turn to cognitive learning, which may
their child to discriminate between the their children when they call strangers have neither observable response nor
real father and other adult males. “Daddy.” observable reinforcer.
ogy was splitting between those who were food present. Then, with food present in humans learn through
studying feelings and cognitive processes the maze’s food box, he would test the rat observing things. For
and those who were studying observable to see which path it took. The rat learned example, Bandura says
behaviors, such as very quickly to take the shortest path. that a child can learn to
a ni ma ls u nder Next, Tolman blocked the shortest path hate spiders simply by
controlled condi- to the food box. The first time the rat observing the behav-
t ions (f ig ure at encountered the blocked shortest path, it iors of someone who
right). In a sharp selected the next shortest path to the shows a great fear of spiders. This is an
criticism of cogni- food box. According to Tolman (1948), example of social cognitive learning.
tive learning, Skin- the rat selected the next shortest path Social cognitive learning results from
ner said, “As far as because it had developed a cognitive map watching, imitating, and modeling and does not
I’m concerned, cognitive science is the of the maze. require the observer to perform any observable
Photo Credits: (rat) © PhotoDisc, Inc.; top, © Evan Agostini/Getty Images
creationism [downfall] of psychology” A cognitive map is a Food behavior or receive any observable reward.
(Vargas, 1991, p. 1). mental representation in Just as Tolman found that learning
Skinner’s severe criticism of studying the brain of the layout occurred while rats were exploring, Ban-
cognitive processes caused many in the of an environment and dura found that humans learned while
audience to gasp and only a few to applaud its features. observing and that much (most) of human
(Vargas, 1991). Tolman showed learning takes place through observation.
Start
In the 1950s and 1960s, Skinner had that rats, in addition Observational learning, which involves
advocated that psychology’s goal should to forming a cognitive map, learned the numerous cognitive processes, is a 180-
be to study primarily observable behav- layout of a maze without being reinforced, degree change from Skinner’s position,
iors rather than cognitive processes. a position very different from Skinner’s. which had emphasized observable, non-
However, psychologists gradually dis- Tolman’s position is making a comeback cognitive behaviors.
covered that cognitive processes played a as psychologists currently study a variety Following the death of Skinner in 1990,
major role in human and animal activi- of cognitive processes in animals (Redish, the study of cognitive processes has bal-
ties and that such activities could not be 2007). Tolman’s study of cognitive pro- looned in popularity and usefulness. We’ll
understood or explained from observ- cesses in animals laid the groundwork introduce you to cognitive learning by
able behaviors alone. Today, the study for the study of cognitive processes in describing one of Bandura’s best-known
of cognitive processes is a major goal of humans, which is best shown by the cur- studies, which involved a doll and a consid-
psychology (Bandura, 2001a). rent theory of Albert Bandura (2001a). erable amount of kicking and screaming.
E. COGNITIVE LEARNING 223
E. Cognitive Learning
Observational Learning
Perhaps a dozen experiments in psychology have become classics because they were the first to demonstrate some
Do children very important principles. One such classic experiment demonstrated the conditioning of emotional responses in
learn by “Little Albert” (p. 204). Another classic is Albert Bandura (1965) and his colleagues’ demonstration that children
watching? learned aggressive behaviors by watching an adult’s aggressive behaviors. Learning through watching is called
observational learning, which is a form of cognitive learning.
Bobo Doll Experiment Learning Versus Performance
One reason this Bobo doll study is Is it possible that people learn by observing but not
Why did a classic experiment is it challenged Do you necessarily perform what they have learned? To
children kick the idea that learning occurred learn but not answer this question, Bandura and colleagues asked
the Bobo doll? through classical or operant con- show it? children to watch a movie in which someone hit and
ditioning. Children learned to kicked a Bobo doll. However, after hitting and kicking
perform aggressive responses simply from watching. the doll, the person in the film was punished by being criticized and
Procedure. In one part of the room, preschool children spanked. Next, each child was left alone in a room filled with toys, including
were involved in their own art projects. In another part a Bobo doll.
of the room, an adult got up and, for the next 10 minutes, As the experimenters watched each child through a one-way mirror, they
kicked, hit, and yelled (“Hit him! Kick him!”) at a large, found that more boys than girls imitated the model and performed aggres-
inflated Bobo doll. Some children watched the model’s sive behaviors on Bobo. But not all the children imitated the model’s aggressive
aggressive behaviors, while other children did not. behaviors. Next, each child who had not imitated the model’s aggressive behav-
Each child was later iors on Bobo was offered a reward (a sticker or some fruit juice) to imitate the
subjected to a frustrat- model’s behavior. With the promise of a reward, all of the children imitated the
ing situation and then model’s aggressive behaviors. We’ll examine in more detail the girls’ imitated
placed in a room with aggressive behaviors, which were similar to the boys’ but more dramatic.
toys, including the Bobo As the graph below shows, girls imitated an average of 0.5 aggressive
doll. Without the child’s behaviors after watch-
Average Number of Aggressive Responses
k nowledge, research- ing a film of a model
ers observed the child’s who performed aggres- 0.5 Watched punished model
behaviors. sive behaviors on Bobo
Rewarded for imitating 3.0
Results. Children who and then was punished
had observed the model’s for being aggressive.
aggressive attacks on the In other words, after observing a model being punished for
After watching, children
imitated adults kicking doll. Bobo doll also kicked, hit, aggressive behaviors, girls imitated almost none of the mod-
and yelled (“Hit him! Kick el’s aggressive behaviors. However, when the same girls were
him!”) at the doll. Through observational learning alone, promised a reward for imitating the model’s aggressive
these children learned the model’s aggressive behaviors behaviors, these girls imitated an average of 3.0 aggressive
and were now performing them. In comparison, children behaviors (Bandura, 1965).
who hadn’t observed the model’s behaviors didn’t hit or So what does this experiment show? It shows that the girls
kick the Bobo doll after they had been mildly frustrated. had actually learned the model’s aggressive behaviors through observation
Conclusion. Bandura’s point is that children learned but that some did not perform these behaviors until they were rewarded for
to perform specific aggressive behaviors not by practic- doing so (Bandura, 1965). This is an example of the learning–performance
ing or being reinforced but simply by watching a live distinction.
model perform behaviors. Observational learning is The learning–performance distinction means learning may occur but may not
called modeling because it involves watching a model always be measured by, or immediately evident in, performance.
and imitating the behavior. The learning–performance distinction may
Photo Credit: (bobo doll) © Craig McClain
It turns out that observational learning may have a be demonstrated by young children, often to the
biological foundation. Recent research on mirror neu- embarrassment of their parents. For instance, a
rons (see p. 6) shows that they become activated when we young child may overhear a “dirty” word but not
observe others perform behaviors, which suggests we are repeat the word until out in public. Repeating a
biologically programmed to learn through observation “dirty” word shows that the child had learned
(Iacoboni, 2008a; Lepage & Théoret, 2007). the word through observation but waited until
Another finding of the Bobo doll study is that a Child imitates adult’s speech. later to imitate the parent and actually say
child may learn by observing but then not perform (perform) the “dirty” word.
the behavior. This is called the learning–performance Based on the Bobo doll study and others, Bandura developed a theory of
distinction. cognitive learning that we’ll examine next.
224 MODULE 10 OPER ANT & COGNITIVE APPROACHES
Bandura’s Social Cognitive Theory
The idea that humans gather informa- Unlike operant and classical conditioning, this theory says
Would you tion about their environments and the that it is not necessary to perform any observable behav-
hold this behaviors of others through observation iors or receive any external rewards to learn.
spider? is a key part of Bandura’s (2001a) social Bandura believes that four processes—attention,
cognitive theory of learning. memory, imitation, and motivation—operate dur-
Social cognitive theory emphasizes the importance of obser- ing social cognitive learning. We’ll explain how
vation, imitation, and self-reward in the development and learning these processes operate in decreasing fear of spiders
of social skills, personal interactions, and many other behaviors. and snakes.
increase in the occurrence of some response learning that develops through watching
because it is followed either by the removal of an and imitation and that does not require
unpleasant stimulus or by avoiding the stimulus is the observer to perform any observable
called (b) . behavior or receive a reinforcer is called
6. A stimulus, such as food, water, or sex, that (a) learning. Bandura
is innately satisfying and requires no learning believes that humans gather much infor-
to become pleasurable is a (a) . mation from their (b) through social cognitive
A stimulus, such as grades or praise, that has learning.
acquired its reinforcing power through experience and learning 13. Bandura’s theory of social cognitive learning
is a (b) . involves four mental processes. The observer
7. The various ways that reinforcers occur after a behavior has must pay (a) to what the
been emitted are referred to as (a) of reinforce- model says or does. The observer must then code
ment. For example, if each and every target behavior is reinforced, the information and be able to retrieve it from
it is called a (b) schedule of (b) for use at a later time. The
reinforcement. If behaviors are not rein- observer must be able to use the coded informa-
forced each time they occur, it is called tion to guide his or her (c) in performing and
a (c) schedule of imitating the model’s behavior. Finally, the observer must be
reinforcement. (d) to perform the behavior, which involves
some reason, reinforcement, or incentive.
8. When an organism emits the same response to similar stimuli,
it is called (a) . 14. In Köhler’s study of problem solving in chimps, he identified a
When a response is emitted in the mental process marked by the sudden occurrence of
presence of a stimulus that is rein- a solution, which he termed (a) .
forced and not in the presence of This phenomenon is another example of
unreinforced stimuli, it is called (b) learning.
Answers: 1. operant conditioning; 2. (a) emits, (b) reinforcement or reinforcer; 3. shaping; 4. (a) reinforcement, (b) punishment; 5. (a) positive
reinforcer, (b) negative reinforcement; 6. (a) primary reinforcer, (b) secondary inforcer; 7. (a) schedules, (b) continuous, (c) partial; 8. (a) general-
ization, (b) discrimination, (c) extinction, (d) spontaneous recovery; 9. cognitive learning; 10. cognitive map; 11. learning–performance;
12. (a) social cognitive, (b) environments; 13. (a) attention, (b) memory, (c) motor control, (d) motivated; 14. (a) insight, (b) cognitive
Imprinting
Soon after they hatch and without Sensitive period. Unlike classical conditioning, operant condition-
Why do young any apparent learning, baby chicks ing, and cognitive learning, which occur throughout an animal’s life,
chicks follow follow their mother hen. This fol- imprinting occurs best during the first few hours after hatching. This
their mother? lowing behavior was not explained brief time period is called the critical, or sensitive, period.
by any of the principles of learning The critical, or sensitive, period refers to a relatively brief time during which
identified by Pavlov (classical conditioning), Thorndike learning is most likely to occur.
Photo Credits: top, © Mitsuaki Iwago/Minden Pictures; left, © PhotoDisc, Inc.; right, © Zoological Society of San Diego
(trial-and-error learning), or Skinner (operant condition- Normally, the first object that newly hatched ducks see is their parent,
ing). The baby chick’s seemingly unlearned behavior of fol- upon whom they imprint. Thus, imprinting is a way for newly hatched
lowing its mother was a different kind of learning that was animals to establish social attachments to members of their species.
first identified by ethologists. Although newly hatched birds will imprint on almost any moving
Ethologists are behavioral biologists who observe and study object that they first see, including a human, a colored ball, or a glove,
animal behavior in the animal’s natural environment or under rela- they imprint more strongly on moving objects that look or sound like
tively naturalistic conditions. their parent. Only birds that can walk immediately after hatching show
For example, an Austrian ethologist, Konrad imprinting, which promotes their survival (Bateson, 1991).
Lorenz (1952), studied chicks, goslings, and Irreversible. Unlike classical conditioning, operant conditioning,
ducks, which can all move around and cognitive learning, whose effects are usually reversible, imprinting
minutes after hatching. is essentially irreversible. Most likely, imprinting evolved to be irrevers-
He discovered that ible so that a young duck would not imprint on its mother one week and
these baby ani- then imprint on a sly fox the next week.
mals followed In a program to prevent the California con-
the first mov- dor from becoming extinct, condor chicks are
ing object they Baby ducks automatically follow hatched at the San Diego Zoo and raised by
saw, which was first moving object. humans. Because imprinting occurs very early
usually their mother. and is irreversible, special precautions are taken
This following behavior is an example of imprinting. so that the condor will not imprint on humans.
Imprinting refers to inherited tendencies or responses that are For example, the young condor chick
displayed by newborn animals when they encounter certain stimuli shown on the right is being fed by a puppet
in their environment. that resembles an adult condor’s head rather
Imprinting is an unlearned behavior that is based on bio- than a human’s hand. The “puppet mother”
logical factors and that has great survival value: It increases helps the young condor imprint on real condor “Puppet mother”
the chances that newly hatched birds will remain with and characteristics. When this condor grows up and feeds baby condor.
follow their parent instead of wandering off into the waiting is introduced into the wild, it will establish social
jaws of predators. Besides being unlearned, Lorenz noted relationships with, and attempt to mate with, its own species.
two other major differences between imprinting and other Another example of how biological factors increase the ease and
kinds of learning. speed of learning is evident in something called prepared learning.
Kamil & Cheng, 2001). among sounds that are essential for learning speech (Buonomano & Merzenich,
One reason nutcrackers have such phenomenal 1995). In fact, infants all around the world make similar babbling sounds, which
memories is that the areas of their brains involved indicates the influence and importance of
in memory are larger than the same areas in birds biological factors in learning to speak. Broca’s area
that do not store food. Specifically, the hippocam- Conclusion. The learning princi- is prewired to
combine sounds
pus, which is involved in transferring short-term ples of Pavlov, Thorndike, and Skin- into words.
memories into long-term memories, is larger in ner do not, by themselves, explain Wernicke’s area is
nutcrackers than in nonstoring birds (S. D. Healy how rats form cognitive maps, how prewired to combine
words into sentences.
et al., 2005). Thus, the nutcracker is biologically infants easily produce and discrim-
prepared for surviving win- inate among human speech sounds,
ters by having a larger how nutcrackers remember thousands
hippoca mpus (r ig ht of hiding places for stored food, why mon- Two brain areas are
figure), which helps it keys spend time making and rolling snowballs, prewired for speaking.
better remember the Hippocampus and why newborn chicks follow the first moving object.
locations of thousands All these examples indicate how innate biological factors, such as differences
of food stores. in size or organization of brain structures, play a major role in preparing and
Just as some birds are biologically prepared to helping animals and humans learn certain kinds of behaviors that are useful for
remember the locations of critical hidden stores, maintenance and survival.
humans are biologically prepared to make sounds Next, we’ll return to operant conditioning and discuss how its learning prin-
and speak at least one of 6,800 different languages. ciples helped parents manage childhood problems.
F. B I O L O G I C A L F A C T O R S 229
G. Research Focus: Viewing Aggression
Relational and Physical Aggression in the Media
Earlier in this module, we discussed aggression, such as verbal bullying,
Can watching observational learning, which is when an ostracizing peers, and spreading
meanness make individual learns specific behaviors sim- rumors, is also frequently portrayed.
you aggressive? ply by watching someone engage in the Is it possible that viewing relational
behaviors. For instance, in the Bobo doll aggression, such as meanness, can
study, we saw how children can learn to perform specific aggressive increase people’s relational aggres-
behaviors simply by watching a live model perform these behaviors. sion behaviors or even their physical
Researchers have long been interested in whether watching aggression behaviors? Researchers
aggression or violence in the media (television, movies) makes recently explored this very question
people more aggressive. Numerous studies have found that watch- (S. M. Coyne et al., 2008).
ing physical aggression in the media does in fact increase physical
aggression in viewers. However, physical aggression is not the only Observing physical aggression in the media increases
physically aggressive behaviors in viewers.
type of violence portrayed in the media. For example, relational
Photo Credits: top, © Edouard Berne/Getty Images; bottom left, © Radius Images/Photolibrary; bottom right, © SW Productions/Getty Images
Procedure. The women were randomly to their opponent, these women administered a louder noise than women who watched a
assigned to one of three groups, each to clip with no aggression (example of physical aggression). On an important questionnaire
watch a different video. One video featured to be used supposedly to help decide whether the rude researcher should be hired, these
physical aggression (a knife fight from Kill women rated the rude researcher lower than the women who did not watch aggression
Bill), another featured relational aggression (example of relational aggression) (S. M. Coyne et al., 2008; Toppo, 2008).
(a collection of scenes from Mean Girls), and
a third featured no aggression (a scene from Physical Aggression Relational Aggression
What Lies Beneath showing people engaging
in a ritual to communicate with spirits).
After watching the videos, all par-
ticipants completed a questionnaire about
their past aggressive behaviors. As the
women left the room to go home, another
researcher asked if they would participate
in a second study (it’s actually part of the
original study). The women were told this
second study would examine their reac-
tion times in completing puzzle tasks. Once
participants said yes and began working
on the puzzles, the researcher intentionally
behaved in a very rude manner by telling
them to hurry, pacing with a stopwatch, and
sighing loudly. When the participants later Participants who viewed either physical or relational aggression were more likely to subsequently
engage in physical and relational aggression than participants who didn’t view any aggression.
expressed low confidence about their per-
formance, the researcher blamed them for
ruining her study. She was doing a great job Conclusion. Viewing either physical or relational aggression in the media leads to an
of acting rude! increased likelihood of people subsequently exhibiting physical and relational aggression.
Participants were then asked to engage Also, watching one form of aggression can influence the occurrence of other forms of
in a competitive reaction time test (push- aggression. For example, observing relational aggression can make it more likely for
ing buttons really fast) and to choose from someone to subsequently engage in physical aggression. Parents, teachers, media watch-
ten noise levels to be administered to their dogs, and politicians should focus not only on the effects of viewing physical aggression
opponent each time they lose the button- in the media but also on the effects of viewing relational aggression.
pushing contest. At the very end, partici- Learning to behave in aggressive ways by watching aggression in the media is one
pants completed evaluation forms for the example of how learning principles can be applied to human behavior. Another example
rude experimenter. involves learning how to play the violin.
ested model for the child, play violin games that are learning involves
understand (Bandura would say code). Because a 3- to attention, memory,
fun for the child, avoid games or lessons that involve
4-year-old child does not have fully developed verbal imitation, and
competition, and never push the child beyond the level motivation.
skills or memory, little time is spent giving verbal
that he or she is capable of reaching (Slone, 1985).
instructions. Instead, the young
child is given violin information Conclusion. Parents and teachers who have used the Suzuki method
through games and exercises. For report great success (Lamb, 1990). As you can judge, the basic principles of the
example, children are taught how Suzuki method for teaching violin are quite similar to Bandura’s four mental
to hold the violin, use the bow, and processes for social cognitive learning. Both Suzuki and Bandura recognized
press the strings by first playing the importance of observational learning and how much information children
games with their hands. Children can learn from watching and imitating models. Suzuki’s successful method of
are taught how to read music (notes) teaching violin to young children provides support for Bandura’s four mental
only when they are older and have processes that he believes are involved in social cognitive learning.
gained some technical skill at playing Next, we’ll discuss how operant learning principles were used to develop
the violin. a method for teaching autistic children.
H . CU LT U R A L DI V E R S I T Y: E A S T M E E T S W E S T 231
I. Application: Behavior Modification
Definitions
In this module, we have discussed how communicating; very few activities and interests; and long periods of time
What is operant conditioning principles are used to spent repeating the same behaviors or motor patterns, or following rituals
behavior mod? decrease food refusal in young children, that interfere with more normal functioning. Symptoms range from mild
to toilet train children (p. 216), to to severe and usually appear when a child is about 2 to 3 years old
prevent eating paint chips (p. 218), and to stop children from (American Psychiatric Association, 2000). (We discussed the
engaging in dangerous behaviors (p. 219). These are all symptoms and causes of autism more thoroughly in
examples of behavior modification (Miltenberger, 2007). Module 1.)
Behavior modification is a treatment or therapy that chang- Parents, doctors, and policy makers all agree that the
es or modifies problems or undesirable behaviors by using best method to address the above symptoms, especially
principles of learning based on operant conditioning, classical deficits in forming relationships and communicat-
conditioning, and social cognitive learning. ing, in autistic children is to enroll them in intensive
For over 35 years, psychologist and researcher Ivar behavioral treatment (between 20 and 40 hours a week)
Behavior mod is used
Lovaas of the University of California at Los Angeles to treat autism. as early as possible. There is no evidence that any treat-
has used behavior modification or, more colloquially, ment leads to a complete recovery, but without treat-
behavior mod to treat autism. ment, many autistics will remain socially unresponsive (Carey,
Autism is marked by poor development in social relationships, such 2004; Kabot et al., 2003). We’ll discuss the behavior mod treatment
as not wanting to be touched, not making eye contact, and hiding to developed by Ivar Lovaas, who combined principles of operant
avoid people (see drawing); great difficulty developing language and conditioning and social cognitive learning (Lovaas & Buch, 1997).
be used? dren, and 94% of 3- and 4-year-olds have been child should not do, while positive reinforcers have the advantage
spanked during the past year (O’Callaghan, of encouraging the child to engage in desirable behaviors.
2006). Since spanking may immediately stop undesirable behaviors,
Time-Out: Negative Punishment
it has been and remains a popular form of discipline. But spanking is
Another form of discipline is time-out, which was dis-
surrounded with controversy because it is associated with numerous
cussed earlier (see p. 219). Time-out is an example of
negative side effects and may be a less desirable form of punishment
negative punishment because it involves the removal
(discipline) than the time-out procedure (Benjet & Kazdin, 2003).
of a reinforcing stimulus (desirable reward) so that
We’ll discuss the pros and cons of each.
some undesirable response will not recur. For
Spanking: Positive Punishment example, after misbehaving, a child is given a
Since spanking involves the presentation of time-out period (stays in a corner without
an aversive stimulus (pain), it is an example games, books, or toys). Time-out is most
of positive punishment. Researchers dis- effective when used consistently and com-
agree on the use of spanking. Some argue bined with teaching the child alternative
that all spanking is bad because it is associ- desired behaviors using positive reinforc-
ated with numerous short- and long-term ers (J. Taylor & Miller, 1997). Time-out has fewer
Effects of undesirable side
punishment depend negative side effects, such as the child imi- Compared to spanking, time-out has effects than spanking.
on its usage. tating and modeling aggressive behavior fewer undesirable side effects; it does not
and developing other conduct problems provide a model of aggression and does not elicit severe negative
(Gershoff, 2002). Others agree that severe spanking administered emotional reactions. Thus, when it is necessary to discipline a
by harsh parents is always bad but mild to moderate spanking used child, care should be taken in choosing between positive punish-
as one form of discipline by loving parents does not necessarily ment (spanking) and negative punishment (time-out). Although
have undesirable or negative side effects (Baumrind et al., 2002). both kinds of punishment stop or suppress undesirable behaviors,
Some of the undesirable effects of positive punishment can be spanking has more negative side effects than time-out, and time-out
reduced if it is given immediately after the behavior, if it is just has been shown to be effective in eliminating undesirable behav-
severe enough to be effective, if it is delivered consistently, if the iors. Both kinds of punishment are best used in combination with
child is told the reason for the punishment, and, perhaps most positive reinforcers so the child also learns to perform desirable
important, if punishment is used in combination with positively behaviors (Nichols, 2004).
Figure/Text Credit: (#1 graph) Adapted from “Treating Chronic Food Refusal in Young Children: Home-Based Parent Training,” by M. A. Werle, T. B. Murphy &
3. Skinner used the term operant (a) to describe
something that can be modified by its consequences. Operant 10. The stimuli of food, water, and sex, which are innately satisfy-
responses provide one way to separate ongoing behaviors into ing and require no learning to become pleasurable, are called
units that can be observed and measured. Skinner believed (a) . The stimuli of praise, money, and good
that Pavlov’s conditioning, which involves physiological grades have acquired their reinforcing properties through experi-
Photo Credits: (#8) © Simon Belcher/Alamy; (#11) © Michael P. Gadomski/Photo Researchers, Inc.; (#13) © Stephen Kraseman/DRK Photo
(b) , was not very useful in understanding ence; these stimuli are called (b) .
other forms of ongoing behaviors.
4. Suppose you wished to operantly condition your dog, Bingo, C. Schedules of Reinforcement
to sit up. The procedure would be as follows. You would give 11. A program or rule that determines
Bingo a treat, which is called a (a) , after he emits how and when the occurrence of a response
a desired behavior. Because it is unlikely that Bingo will initially will be followed by a reinforcer is called a
sit up, you will use a procedure called (b) , .
which is a process of reinforcing those behaviors that lead up to or
approximate the final desired behavior—sitting up. Immediately 12. If you received reinforcement every time you performed a good
after Bingo emitted a desired behavior, you would give him a deed, you would be on a (a) schedule. This sched-
(c) . ule is often used at the beginning of operant conditioning because it
results in a rapid rate of learning. If your good deeds were not rein-
5. Any behavior that increases in frequency because of an forced every time, you would be on a (b) schedule.
accidental pairing of a reinforcer and that behavior is called This schedule is more effective in maintaining the target behavior in
a behavior. the long run. There are four kinds of partial reinforcement schedules.
6. The essence of operant conditioning can be summed up
as follows: Consequences or reinforcers are contingent on D. Other Conditioning Concepts
Photo Credits: (#14) © Jon Lowenstein/Aurora Photos; (#19) © Zoological Society of San Diego; (#22) © SW Productions/Getty Images; (#23) © Hiroji Kubota/
Köhler used the term to describe a mental information to (c) the model’s
process marked by the sudden occurrence of a solution. actions, and having (d) to
17. Köhler’s study of insightful problem solving, Bandura’s theory perform the behavior.
of observational learning, and Tolman’s idea of cognitive maps
represent three kinds of learning.
I. Application: Behavior Modification
24. Using principles of operant conditioning
F. Biological Factors to change human behavior is referred to
18. Innate tendencies or predispositions that may either facilitate as (a) . Using these same
or inhibit learning are referred to as . principles to help individuals learn to control
(increase or decrease) some physiological
19. The innate tendency of newborn response, such as muscle activity or temper-
birds to follow the first moving object ature, is called (b) .
that they encounter soon after birth is
called (a) . This kind 25. If an aversive stimulus is presented immediately after a par-
of learning occurs best during a critical ticular response, the response will be suppressed; this procedure is
or sensitive period and is essentially called (a) . If a reinforcing stimulus is removed
(b) . One function of immediately after a particular response, the response will be sup-
imprinting is to form social attachments pressed; this procedure is called (b) . A poorly
between members of a species. chosen form of punishment, such as spanking, may have undesir-
able side effects, such as developing (c) problems
20. The innate tendency of animals to and serving as a model for future (d) behaviors.
recognize, attend to, and store certain
cues over others and associate some combinations of conditioned
Answers: 1. operant conditioning; 2. law of effect; 3. (a) response,
and unconditioned stimuli is referred to as . An (b) reflexes; 4. (a) reinforcer, (b) shaping, (c) reinforcer; 5. superstitious;
example of this tendency is observed in Clark’s nutcrackers, which 6. behavior; 7. (a) reflex, (b) unconditioned stimulus, (c) behavior,
are preprogrammed to bury and remember thousands of hidden (d) emitted, (e) unconditioned response, (f) reinforcer; 8. (a) reinforcer,
stores of food. (b) punishment; 9. (a) positive reinforcer, (b) negative reinforcer,
(c) increase, (d) decreases; 10. (a) primary reinforcers, (b) secondary rein-
forcers; 11. schedule of reinforcement; 12. (a) continuous reinforcement,
G. Research Focus: Viewing Aggression (b) partial reinforcement; 13. (a) generalization, (b) discrimination,
(c) extinction, (d) spontaneous recovery; 14. (a) social cognitive learning,
21. Verbal bullying, ostracizing peers, and spreading rumors are
(b) cognitive map; 15. learning–performance; 16. insight; 17. cognitive;
all examples of . 18. biological factors; 19. (a) imprinting, (b) irreversible; 20. preparedness,
or prepared learning; 21. relational aggression; 22. more likely;
23. (a) attention, (b) memory, (c) imitate, (d) motivation; 24. (a) behav-
Magnum Photos
ior modification, (b) biofeedback; 25. (a) positive punishment, (b) negative
punishment, (c) conduct, (d) aggressive
1 How do we know
what a reinforcer
is for a whale?
will likely repeat the behavior. Posi-
tive reinforcers for whales may
the whistle and reinforces the whale.
After this is done several times, the
include food, toys, back scratches, target is moved farther away and the
b ei ng spraye d w ith a hos e, or trainer waits for the whale to touch
3 Why is it impor-
tant to reinforce
must use a signal (whistle) to pro-
vide the whale with immediate feed-
stage, but the tr uth is training
whales to perform is very challeng-
Photo Credit: © AP Images/Amy Sancetta
6
the whale immediate- back that it has correctly performed ing work. Trainers need to be com- Can whales learn
ly after it performs a the desired behavior. A whale learns mitted, patient, and friendly to earn to do tricks by
desired behavior? just watching other
the meaning of the whistle by the the whale’s trust, and even then it
trainer whistling before giving the can take months or even years to whales perform?
4 In this training
example, what is
the neutral stimulus?
whale food. Over the course of sev-
eral trials, the whale comes to asso-
teach a whale a complex set of
behaviors. (Adapted from Animal ANS W ERS
ciate the whistle with receiving food training at SeaWorld, 2002; Aqua TO CRITI CAL
Unconditioned stimu-
and the whale performs the behav- facts: Training marine mammals, TH I NKI NG
lus? Conditioned
stimulus? ior to get the reinforcement. 2006) QUEST I ONS