Professional Documents
Culture Documents
CONNECTIONISM THEORY
Edward Thorndike's Connectionism theory gave us the original S-R framework of
behavioral psychology. More than a hundred years ago he wrote a text book entitled,
Educational Psychology. He was the first one to use this term. He explained that
learning is the result of associations forming between stimuli (S) and responses (R).
Such associations or habits" become strengthened or weakened by the nature and
frequency of the S-R pairings. The model for S-R theory was trial and error learning in
which certain responses came to be repeated more than others because of rewards.
The main principle of connectionism (like all behavioral theory) was that
learning could be adequately explained without considering any unobservable internal
states.
Thorndike's theory on connectionism, states that learning has taken place when.
a strong connection or bond between stimulus and response is formed. He came up with
three primary laws:
1. Law of Effect. The law of effect states that a connection between a stimulus
and response is strengthened when the consequence is positive (reward) and the
connection between the stimulus and the response is weakened when the consequence
is negative. Thorndike later on, revised this "law" when he found that negative rewards
(punishment) do not necessarily weaken bonds, and that some seemingly pleasurable
consequences do not necessarily motivate performance.
2. Law of Exercise. This tells us that the more an S-R (stimulus response) bond is
practiced the stronger it will become. "Practice makes perfect" seem to be associated
with this. However, like the law of effect, the law of exercise also had to be revised
when Thorndike found that practice without feedback does not necessarily enhance
performance.
3. Law of Readiness. This states that the more readiness the learner has to
respond to the stimulus, the stronger will be the bond between them. When a person is
ready to respond to a stimulus and is not made to respond, it becomes annoying to the
person. For example, if the teacher says, "Okay we will now watch the movie (stimulus)
you've been waiting for." And suddenly the power goes off. The students will feel
frustrated because they were ready to respond to the stimulus but were prevented from
doing so. Likewise, if the person is not at all ready to respond to stimuli and is asked to
respond, that also becomes annoying. For instance, the teacher calls a student to stand
up and recite, and then the teacher asks the question and expects the student to
respond right away when he is still not ready. This will be annoying to the student. That
is why teachers should remember to say the question first, and wait for a few seconds
before calling on anyone to answer.
John B. Watson was the first American psychologist to work with Pavlov's
ideas. He too was initially involved in animal studies, then later became involved
in human behavior research. He considered that humans are born with a few
reflexes and the emotional reactions of love and rage. All other behavior is
learned through stimulus-response John Watson associations through
conditioning. He believed in the power 1878 1958 of conditioning so much that
he said that if he is given a dozen healthy infants he can make them into
anything you want them to be, basically through making stimulus-response
connections through conditioning.
Skinner's work differs from that of the three behaviorists before him in
that he studied operant behavior (voluntary behaviors used in operating on the
environment). Thus, his theory came to be known as Operant Conditioning.
Reinforcement is the key element in Skinner's S-R theory. A reinforcer is anything that
strengthens the desired response. There is a positive reinforcer and a negative
reinforcer.
A positive reinforcer is any stimulus that is given or added to increase the
response. An example of positive reinforcement is when a teacher promises extra
time in the play area to children who behave well during the lesson. Another is a
mother who promises a new cell phone for her son who gets good grades. Still,
other examples include verbal praises, star stamps and stickers.
Skinner also looked into extinction or non-reinforcement: Responses that are not
reinforced are not likely to be repeated. For example, ignoring a student's misbehavior
may extinguish that behavior.
Shaping of Behavior. An animal on a cage may take a very long time to figure out
that pressing a lever will produce food. To accomplish such behavior, successive
approximations of the behavior re rewarded until the animal learns the
association between the lever and the food reward. To begin shaping, the animal
may be rewarded for simply turning in the direction of the lever, then for moving
toward the lever, for brushing against the lever, and finally for pressing the lever.
Behavioral chaining comes about when a series of steps are needed to be
learned. The animal would master each step in sequence until the entire
sequence is learned. This can be applied to a child being taught to tie a shoelace.
The child can be given reinforcement (rewards) until the entire process of tying
the shoelace is learned.
Reinforcement Schedules. Once the desired behavioral response is accomplished,
reinforcement does not have to be 100%; in fact, it can be maintained more
successfully through what Skinner referred to as partial reinforcement schedules.
Partial reinforcement schedules include interval schedules and ratio schedules.
Fixed Interval Schedules. The target response is reinforced after a fixed amount
of time has passed since the last reinforcement. Example, the bird in a cage is
given food (reinforcer) every 110 minutes, regardless of how many times it
presses the bar.
Variable Interval Schedules. This is similar to fixed interval schedules but the
amount of time that must pass between reinforcement varies. Example, the bird
may receive food (reinforcer) different intervals, not every ten minutes.
Fixed Ratio Schedules. A fixed number of correct responses must occur before
reinforcement may recur. Example, the bird will be given food (reinforcer)
everytime it presses the bar 5 times.
Variable Ratio Schedules. The number of correct repetitions of the correct
response for reinforcement varies. Example, the bird is given food (reinforcer)
after it presses the bar 3 times, then after 10 times, then after 4 times. So the
bird will not be able to predict how many times it needs to press the bar before it
gets food again.
Variable interval and especially, variable ratio schedules produce steadier and
more persistent rates of response because the learners cannot predict when the
reinforcement will come although they know that they will eventually succeed. An
example of this is why people continue to buy lotto tickets even when an almost
negligible percentage of people actually win. While it is true that very rarely there
is a big winner, but once in a while somebody hits the jackpot (reinforcement).
People cannot predict when the jackpot can be gotten (variable interval) so they
continue to buy tickets (repetition of response).