Date post: | 09-Apr-2018 |
Category: |
Documents |
Upload: | aarthiashok |
View: | 223 times |
Download: | 0 times |
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 1/15
Connectionism (E. Thorndike)
Overview:
The learning theory of Thorndike represents the original S-R framework of behavioral
psychology: Learning is the result of associations forming between stimuli and responses.Such associations or "habits" become strengthened or weakened by the nature and frequency
of the S-R pairings. The paradigm for S-R theory was trial and error learning in which certain
responses come to dominate others due to rewards. The hallmark of connectionism (like all
behavioral theory) was that learning could be adequately explained without refering to any
unobservable internal states.
Thorndike's theory consists of three primary laws: (1) law of effect - responses to a situationwhich are followed by a rewarding state of affairs will be strengthened and become habitual
responses to that situation, (2) law of readiness - a series of responses can be chained together to satisfy some goal which will result in annoyance if blocked, and (3) law of exercise -
connections become strengthened with practice and weakened when practice is discontinued.
A corollary of the law of effect was that responses that reduce the likelihood of achieving arewarding state (i.e., punishments, failures) will decrease in strength.
The theory suggests that transfer of learning depends upon the presence of identical elements
in the original and new learning situations; i.e., transfer is always specific, never general. In
later versions of the theory, the concept of "belongingness" was introduced; connections are
more readily established if the person perceives that stimuli or responses go together (c.f.
Gestalt principles). Another concept introduced was "polarity" which specifies that
connections occur more easily in the direction in which they were originally formed than the
opposite. Thorndike also introduced the "spread of effect" idea, i.e., rewards affect not only
the connection that produced them but temporally adjacent connections as well.
Scope/Application:
Connectionism was meant to be a general theory of learning for animals and humans.
Thorndike was especially interested in the application of his theory to education including
mathematics (Thorndike, 1922), spelling and reading (Thorndike, 1921), measurement of
intelligence (Thorndike et al., 1927) and adult learning (Thorndike at al., 1928).
Example:
The classic example of Thorndike's S-R theory was a cat learning to escape from a "puzzle
box" by pressing a lever inside the box. After much trial and error behavior, the cat learns to
associate pressing the lever (S) with opening the door (R). This S-R connection is established
because it results in a satisfying state of affairs (escape from the box). The law of exercise
specifies that the connection was established because the S-R pairing occurred many times
(the law of effect) and was rewarded (law of effect) as well as forming a single sequence (law
of readiness).
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 3/15
Contents
[hide]
y 1 Types
y 2 Procedure variations
o 2.1 Classical discrimination/reversal conditioning o 2.2 Classical ISI discrimination conditioning
o 2.3 Latent inhibition conditioning
o 2.4 Conditioned inhibition conditioning
o 2.5 Blocking
y 3 Applications
o 3.1 Little Albert
o 3.2 Behavioral therapies
y 4 Theories of classical conditioning
y 5 In popular culture
y 6 See also
y 7 References
y 8 Further reading y 9 External links
[edit ] Types
Diagram representing forward conditioning. The time interval increases from left to right.
y Forward conditioning: During forward conditioning the onset of the CS precedes the onset
of the US. Two common forms of forward conditioning are delay and trace conditioning.
y Delay conditioning: In delay conditioning the CS is presented and is overlapped by the
presentation of the US.
y Trace conditioning: During trace conditioning the CS and US do not overlap. Instead, the CS
is presented, a period of time is allowed to elapse during which no stimuli are presented,
and then the US is presented. The stimulus free period is called thetrace interval . It may also
be called the conditioning interval .
y Simultaneous conditioning: During simultaneous conditioning, the CS and US are presentedand terminated at the same time.
y Backward conditioning: Backward conditioning occurs when a conditioned stimulus
immediately follows an unconditioned stimulus. Unlike traditional conditioning models, in
which the conditioned stimulus precedes the unconditioned stimulus, the conditioned
response tends to be inhibitory. This is because the conditioned stimulus serves as a signal
that the unconditioned stimulus has ended, rather than a reliable method of predicting the
future occurrence of the unconditioned stimulus.
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 4/15
y Tempo
¡ l conditionin
¢
: The£
S¤ s
¥ ¦ ese
§
̈
e©
̈
¦ e
¦ y-
̈
¤ e
© ¤ §
̈
e¦ v
s
§ © cq
¤ s
¤
̈
¤ §
¤ s
© e
¥ e
§ © e
§
̈
¥ § c
¦ ¦ ec
¨ ̈
¤ ¤ §
̈
he¤ §
̈
e¦ v
e
̈
!
ee§ £
S¥ ¦
ese§
̈
̈
¤ § s
" The
c #
¦ § ©
¦ c §
̈
e$
̈
c § se ¦ ve s ̈
he S ¤ §
̈
his e$
¥ e "
y Unpai% ed conditionin
&
: The CS and US are not presented together. Us'
ally they are
presented as independent trials that are separated by a variable( or pse'
do-random,
interval. This procedure is used to study non-associative behavioral responses, such as
sensitization.y CS-alone e
)
tinction: The CS is presented in the absence of the US. This procedure is usually
done after the CR has been acquired through "Forward conditioning" training. Eventually,
the CR frequency is reduced to pre-training levels.
[edi0
] Pr1
cedure variat i1
2 3
In addition to the simple procedures described above, some classical conditioning studies are
designed to tap into more complex learning processes. Some common variations are
discussed below.
[edit ] Classical discrimi4
at i5 4
/reversal c5 4
dit i5 4
i4
g
In this procedure, two CSs and one US are typically used. The CSs may be the same modality
(such as lights of different intensity), or they may be different modalities (such as auditory
CS and visual CS). In this procedure, one of the CSs is designated CS+ and its presentation is
always followed by the US. The other CS is designated CSí and its presentation is never followed by the US. After a number of trials, the organism learns to d i6 7 8 i9 i@ A t B CS+ trials
and CSí trials such that CRs are only observed on CS+ trials.
During RB vB 8 6 A l T 8 A i@ i@ g , the CS+ and CSí are reversed and subjects learn to suppressresponding to the previous CS+ and show CRs to the previous CSí.
[edit ] Classical ISI discrimiC
at iD C
cD C
dit iD C
iC
g
This is a discrimination procedure in which two different CSs are used to signal two different
interstimulus intervals. For example, a dim light may be presented 30 seconds before a US,while a very bright light is presented 2 minutes before the US. Using this technique,
organisms can learn to perform CRs that are appropriately timed for the two distinct CSs.
[edit ] Latent inhibit iD
n cD
ndit iD
ning
In this procedure, a CS is presented several times before paired CS-US training commences.
The pre-exposure of the subject to the CS before paired training slows the rate of CR
acquisition relative to organisms that are not CS pre-exposed. Also see Latent inhibition for
applications.
[edit ] Condit ioned inhibit ion condit ioning
Three phases of conditioning are typically used:
Phase 1:
A CS (CS+) is paired with a US until asymptotic CR levels are reached.
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 5/15
Phase 2:
CS+E
US trials are continued, but interspersed with trials on which the CS+ in compound with
a second CS, but not with the US (i.e., CS+E
CS trials). Typically, organisms show CRs on
CS+E
US trials, but suppress responding on CS+E
CS trials.
Phase 3:
In this retention test, the previous CS is paired with the US. If conditioned inhibition has
occurred, the rate of acquisition to the previous CS should be impaired relative to
organisms that did not eF
perience Phase 2.
[edit ] Blocking
MaG
n aH
I
G
P
Q
R : B
Q
oP k
G
ngR ff
R P I
This form of classical conditioning involves two phases.
Phase 1:
A CS (CS1) is paired with a US.
Phase 2:
A compound CS (CS1+CS2) is paired with a US.
Test:
A separate test for each CS (CS1 and CS2) is performed. The blocS ing eff ect is observed in a
lacS
of conditioned response to CS2, suggesting that the f irst phase of training blocS ed the
acquisition of the second CS.
[edit ] Applicat ions
[edit ] Little Albert
MaT
n aU
V
T
W
X
Y : L
T
V V
X
Y A
X
bY
U
V Y xp
Y
U
T
mY n
V
John B. Watson, founder of behaviorism, demonstrated classical conditioning empirically
through experimentation using the Little Albert experiment in which a child ("Albert") was
presented with a white rat (CS). After a control period in which the child reacted normally to
the presence of the rat, the experimentors paired the presence of the rat with a loud, jarring
noise caused by clanging two pipes together behind the child's head (US). As the trials progressed, the child began showing signs of distress at the sight of the rat, even when
unaccompanied by the frightening noise. Furthermore, the child demonstrated generalizationof stimulus associations, and showed distress when presented with any white, furry object²
even such things as a rabbit, dog, a fur coat, and a Santa Claus mask with hair.
[edit ] Behavioral therapies
Ma ̀
n aa
b
̀
c
d
e : B
e hav
̀
oa
b h
e
a
apy
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 6/15
In human psychology, implications for therapies and treatments using classical conditioningdiffer from operant conditioning. Therapies associated with classical conditioning are
aversion therapy, flooding and systematic desensitization.
Classical conditioning is short-term, usually requiring less time with therapists and less effort
from patients, unlike humanistic therapies.[cit
f ti
g h h i i ded ]
The therapies mentioned are designed
to cause either aversive feelings toward something, or to reduce unwanted fear and aversion.
[edit ] Theories of classical condit ioning
There are two competing theories of how classical conditioning works. The first, stimulus-
response theory, suggests that an association to the unconditioned stimulus is made with the
conditioned stimulus within the brain, but without involving conscious thought. The second,
stimulus-stimulus theory involves cognitive activity, in which the conditioned stimulus is
associated to the concept of the unconditioned stimulus, a subtle but important distinction.
Stimulus-response theory, referred to as S-R theory, is a theoretical model of behavioral
psychology that suggests humans and other animals can learn to associate a new stimulus, the
conditioned stimulus (CS), with a pre-existing stimulus, the unconditioned stimulus (US),
and can think, feel or respond to the CS as if it were actually the US.
The opposing theory, put forward by cognitive behaviorists, is stimulus-stimulus theory (S-S
theory). S-S theory is a theoretical model of classical conditioning that suggests a cognitive
component is required to understand classical conditioning and that S-R theory is an
inadequate model. It proposes that a cognitive component is at play. S-R theory suggests that
an animal can learn to associate a conditioned stimulus (CS) such as a bell, with the
impending arrival of food termed the unconditioned stimulus, resulting in an observable
behavior such as salivation. S-S theory suggests that instead the animal salivates to the bell
because it is associated with the concept of food, which is a very fine but important
distinction.
To test this theory, psychologist Robert Rescorla undertook the following experiment.[2]
Rats
learned to associate a loud noise as the unconditioned stimulus, and a light as the conditioned
stimulus. The response of the rats was to freeze and cease movement. What would happen
then if the rats were habituated to the US? S-R theory would suggest that the rats wouldcontinue to respond to the CS, but if S-S theory is correct, they would be habituated to the
concept of a loud sound (danger), and so would not freeze to the CS. The experimental resultssuggest that S-S was correct, as the rats no longer froze when exposed to the signal light.
[3]
His theory still continues and is applied in everyday life.[1]
[edit ] In popular cult ure
One of the earliest literary references to classical conditioning can be found in the comic
novel T p
e Lif e and Opiniq ns o f T r i st r am Shandy, Gentl eman (1759) by Laurence Sterne. Thenarrator Tristram Shandy explains
[s
]how his mother was conditioned by his father's habit of
winding up a clock before having sex with his wife:
My father, [...], was, I believe, one of the most regular men in every thing he did [...] [H]e had made
it a rule for many years of his lif e,on the f irst Sunday-night of every month throughout the whole
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 7/15
year,as certain as ever the Sunday-night came,to wind up a large house-clock, which we had
standing on the back-stairs head, with his own hands:And being somewhere between fifty and
sixty years of age at the time I have been speaking of,he had likewise gradually brought some
other little family concernments to the same period, in order, as he would often say to my uncle
Toby, to get them all out of the way at one time, and be no more plagued and pestered with them
the rest of the month.[...]
[F]rom an unhappy association of ideas, which have no connection in nature, it so fell out at length, that my poor mother could never hear the said clock wound up, ²but the thoughts of
some other things unavoidably popped into her head²& vice versa: ²Which strange
combination of ideas, the sagacious Locke, who cer tainly understood the nature of these
things better than most men, aff irms to have produced more wry actions than all other sources
of pre judice whatsoever.
Another example is in the dystopian novel, A Clockwork Orant e in which the f ilm's anti hero and protagonist, Alex, is given a solution to cause severe nausea, and is forced to watch
violent acts. This renders him unable to perform any violent acts without inducing similar
nausea.
The song "Dinner Bell" by the group "They Might Be Giants" talks about "salivating dogs".
In the song "Cenoir Studies 02" by rapper "Cani bus" (Germaine Williams), he says
"Sometimes I wonder who's listening, the auditory Pavlovian conditioning's so sickening".
Operant conditioning
From Wikipedia, the free encyclopedia
Jump to: navigation, search
"O perant " redirects here. For the deu
inition ou
the word " operant ", see Wi v
tionar w
x operant .
It has been suggested thatMutual operant conditioning be merged into this article or
section. (Discussy
Operant conditioning is the use of consequences to modify the occurrence and form of
behavior. Operant conditioning is distinguished fromclassical conditioning (also called
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 8/15
respondent conditioning) in that operant conditioning deals with the modification of "voluntary behavior" or operant behavior. O perant behavior "operates" on the environment
and is maintained by its consequences, while classical conditioning deals with theconditioning of respondent behaviors which are elicited by antecedent conditions. Behaviors
conditioned via a classical conditioning procedure are not maintained by consequences.[1]
Contents
[hide]
y 1 Reinforcement, punishment, and e
tinction
o 1.1 Four conte
ts of operant conditioning
y 2 Thorndike s law of eff ect
y 3 Biological correlates of operant conditioning
y 4 Factors that alter the eff ectiveness of consequences
y 5 Operant variability
y 6 Avoidance learning
o 6.1 Discriminated avoidance learning
o 6.2 Free-operant avoidance learning y 7 Two-process theory of avoidance
y 8 Verbal Behavior
y 9 Four term contingency
y 10 Operant hoarding
y 11 An alternative to the law of eff ect
y 12 See also
y 13 Ref erences
y 14 External links
[edit ] Reinforcement, punishment, and ext inct ion
Reinforcement and punishment, the core tools of operant conditioning, are either positive
(delivered following a response), or negative (withdrawn following a response). This creates
a total of four basic consequences, with the addition of a fifth procedure known as extinction
(i.e. no change in consequences following a response).
It's important to note that actors are not spoken of as being reinforced, punished, or
extinguished; it is the actions that are reinforced, punished, or extinguished. Additionally,reinforcement, punishment, and extinction are not terms whose use is restricted to the
laboratory. Naturally occurring consequences can also be said to reinforce, punish, or extinguish behavior and are not always delivered by people.
y Reinforcement is a consequence that causes a behavior to occur with greater frequency.
y Punishment is a consequence that causes a behavior to occur with less frequency.
y Extinction is the lack of any consequence following a behavior. When a behavior is
inconsequential, producing neither favorable nor unfavorable consequences, it will occur
with less frequency. When a previously reinforced behavior is no longer reinforced with
either positive or negative reinforcement, it leads to a decline in the response.
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 9/15
[edit ] Four contexts of operant condit ioning
Here the terms positive and ne g ative are not used in their popular sense, but rather: positive
refers to addition, and ne g ative refers to subtraction.
What is added or subtracted may be either reinforcement or punishment. Hence positive
puni shment is sometimes a confusing term, as it denotes the "addition" of a stimulus or increase in the intensity of a stimulus that is aversive (such as spanking or an electric shock).
The four procedures are:
1. Positive reinforcement (Reinforcement) occurs when a behavior (response) is followed by a
stimulus that is rewarding, increasing the frequency of that behavior. In the Skinner box
experiment, a stimulus such as food or sugar solution can be delivered when the rat engages
in a target behavior, such as pressing a lever.
2. Negative reinforcement (Escape) occurs when a behavior (response) is followed by the
removal of an aversive stimulus, thereby increasing that behavior's frequency. In the Skinner
box experiment, negative reinforcement can be a loud noise continuously sounding inside
the rat's cage until it engages in the target behavior, such as pressing a lever, upon which the
loud noise is removed.3. Positive punishment (Punishment) (also called "Punishment by contingent stimulation")
occurs when a behavior (response) is followed by a stimulus, such as introducing a shock or
loud noise, resulting in a decrease in that behavior.
4. Negative punishment (Penalty) (also called "Punishment by contingent withdrawal")
occurs
when a behavior (response) is followed by the removal of a stimulus, such as taking away a
child's toy following an undesired behavior, resulting in a decrease in that behavior.
Al so:
y Avoidance learning is a type of learning in which a certain behavior results in the cessation
of an aversive stimulus. For example, performing the behavior of shielding one's eyes when
in the sunlight (or going indoors) will help avoid the aversive stimulation of having light inone's eyes.
y Extinction occurs when a behavior (response) that had previously been reinforced is no
longer eff ective. In the Skinner box experiment, this is the rat pushing the lever and being
rewarded with a food pellet several times, and then pushing the lever again and never
receiving a food pellet again. Eventually the rat would cease pushing the lever.
y Noncontingent reinforcement ref ers to delivery of reinforcing stimuli regardless of the
organism's (aberrant) behavior. The idea is that the target behavior decreases because it is
no longer necessary to receive the reinforcement. This typically entails time-based delivery
of stimuli identif ied as maintaining aberrant behavior, which serves to decrease the rate of
the target behavior.[2]
As no measured behavior is identif ied as being strengthened, there is
controversy surrounding the use of the term noncontingent "reinforcement".[3]
y Shaping is a form of operant conditioning in which the increasingly accurate approximations of a desired response are reinforced.
[4]
y Chaining is an instructional procedure which involves reinforcing individual responses
occurring in a sequence to form a complex behavior.[4]
[edit ] Thorndik e's law of effect
Ma
n a
: La of ff
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 10/15
O perant conditioning, sometimes called inst ument al cond itionin g or inst ument al l earnin g ,was first extensively studied by Edward L. Thorndike (187 ±19 9), who observed the
behavior of cats trying to escape from home-made puzzle boxes.[5]
When first constrained inthe boxes, the cats took a long time to escape. With experience, ineffective responses
occurred less frequently and successful responses occurred more frequently, enabling the cats
to escape in less time over successive trials. In his law of effect, Thorndike theorized that
successful responses, those producing sati s f yin g consequences, were "stamped in" by theexperience and thus occurred more frequently. Unsuccessful responses, those producing
annoyin g consequences, were st amped out and subsequently occurred less frequently. In
short, some consequences st ren gt hened behavior and some consequences weakened behavior.
Thorndike produced the first known learning curves through this procedure. B.F. Skinner
(190 ±1990) formulated a more detailed analysis of operant conditioning based on
reinforcement, punishment, and extinction. Following the ideas of Ernst Mach, Skinner
rejected Thorndike's mediating structures required by "satisfaction" and constructed a new
conceptualization of behavior without any such references. So, while experimenting with
some homemade feeding mechanisms, Skinner invented the operant conditioning chamber
which allowed him to measure rate of response as a key dependent variable using acumulative record of lever presses or key pecks.[6]
[edit ] Biological correlates of operant condit ioning
The first scientific studies identifying neurons that responded in ways that suggested theyencode for conditioned stimuli came from work by Mahlon deLong.[7][8] and by R.T. "Rusty"
Richardson and deLong.[8]
They showed that nucleus basalis neurons, which releaseacetylcholine broadly throughout the cerebral cortex, are activated shortly after a conditioned
stimulus, or after a primary reward if no conditioned stimulus exists. These neurons areequally active for positive and negative reinforcers, and have been demonstrated to cause
plasticity in many cortical regions.[9]
Evidence also exists that dopamine is activated at
similar times. There is considerable evidence that dopamine participates in both
reinforcement and aversive learning.
[10]
Dopamine pathways project much more densely ontofrontal cortex regions. Cholinergic projections, in contrast, are dense even in the posterior
cortical regions like the primary visual cortex. A study of patients with Parkinson's disease, a
condition attributed to the insufficient action of dopamine, further illustrates the role of
dopamine in positive reinforcement.[11]
It showed that while off their medication, patients
learned more readily with aversive consequences than with positive reinforcement. Patients
who were on their medication showed the opposite to be the case, positive reinforcement
proving to be the more effective form of learning when the action of dopamine is high.
[edit ] Factors that alter the effect iveness of consequences
When using consequences to modify a response, the effectiveness of a consequence can be
increased or decreased by various factors. These factors can apply to either reinforcing or punishing consequences.
1. Satiation/Deprivation: The eff ectiveness of a consequence will be reduced if the individual's
"appetite" for that source of stimulation has been satisf ied. Inversely, the eff ectiveness of a
consequence will increase as the individual becomes deprived of that stimulus. If someone is
not hungry, food will not be an eff ective reinforcer for behavior. Satiation is generally only a
potential problem with primary reinforcers, those that do not need to be learned such as
food and water.
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 11/15
2. Immediacy: After a response, how immediately a consequence is then f elt determines the
eff ectiveness of the consequence.More immediate f eedback will be more eff ective than less
immediate f eedback. If someone's license plate is caught by a traff ic camera for speeding
and they receive a speeding ticket in the mail a week later, this consequence will not be very
eff ective against speeding. But if someone is speeding and is caught in the act by an off icer
who pulls them over, then their speeding behavior is more likely to be aff ected.
3. Contingency: If a consequence does not contingently (reliably, or consistently) follow the target response, its eff ectiveness upon the response is reduced. But if a consequence follows
the response consistently after successive instances, its ability to modif y the response is
increased. The schedule of reinforcement, when consistent, leads to faster learning. When
the schedule is variable the learning is slower. Extinction is more diff icult when learning
occurs during intermittent reinforcement and more easily extinguished when learning occurs
during a highly consistent schedule.
4. Size: This is a "cost-benef it" determinant of whether a consequence will be eff ective. If the
size, or amount, of the consequence is large enough to be worth the effort, the consequence
will be more eff ective upon the behavior. An unusually large lottery jackpot, for example,
might be enough to get someone to buy a one-dollar lottery ticket (or even buying multiple
tickets). But if a lottery jackpot is small, the same person might not f eel it to be worth the
effort of driving out and f inding a place to buy a ticket. In this example, it's also useful tonote that "effort" is a punishing consequence. How these opposing expected consequences
(reinforcing and punishing) balance out will determine whether the behavior is performed or
not.
Most of these factors exist for biological reasons. The biological purpose of the Principle of
Satiation is to maintain the organism's homeostasis. When an organism has been deprived of
sugar, for example, the effectiveness of the taste of sugar as a reinforcer is high. However, as
the organism reaches or exceeds their optimum blood-sugar levels, the taste of sugar becomes
less effective, perhaps even aversive.
The Principles of Immediacy and Contingency exist for neurochemical reasons. When an
organism experiences a reinforcing stimulus, dopamine pathways in the brain are activated.This network of pathways "releases a short pulse of dopamine onto many dendrites, thus
broadcasting a rather global reinforcement signal to postsynaptic neurons."[12]
This results in
the plasticity of these synapses allowing recently activated synapses to increase their sensitivity to efferent signals, hence increasing the probability of occurrence for the recent
responses preceding the reinforcement. These responses are, statistically, the most likely to
have been the behavior responsible for successfully achieving reinforcement. But when the
application of reinforcement is either less immediate or less contingent (less consistent), the
ability of dopamine to act upon the appropriate synapses is reduced.
[edit ] Operant variability
O perant variability is what allows a response to adapt to new situations. O perant behavior is
distinguished from reflexes in that its response topo raphy (the form of the response) issubject to slight variations from one performance to another. These slight variations can
include small differences in the specific motions involved, differences in the amount of forceapplied, and small changes in the timing of the response. If a subject's history of
reinforcement is consistent, such variations will remain stable because the same successfulvariations are more likely to be reinforced than less successful variations. However,
behavioral variability can also be altered when subjected to certain controlling variables.[13]
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 12/15
[edit ] Avoidance learning
Avoidance learning belongs to negative reinforcement schedules. The subject learns that a
certain response will result in the termination or prevention of an aversive stimulus. There are
two kinds of commonly used experimental settings: discriminated and free-operant avoidance
learning.
[edit ] Discriminated avoidance learning
In discriminated avoidance learning, a novel stimulus such as a light or a tone is followed by
an aversive stimulus such as a shock (CS-US, similar to classical conditioning). During the
first trials (called escape-trials) the animal usually experiences both the CS (Conditioned
Stimulus) and the US (Unconditioned Stimulus), showing the operant response to terminate
the aversive US. During later trials, the animal will learn to perform the response already
during the presentation of the CS thus preventing the aversive US from occurring. Such trials
are called "avoidance trials."
[edit ] Free-operant avoidance learning
In this experimental session, no discrete stimulus is used to signal the occurrence of the
aversive stimulus. Rather, the aversive stimulus (mostly shocks) are presented without
explicit warning stimuli. There are two crucial time intervals determining the rate of
avoidance learning. This first one is called the S-S-interval (shock-shock-interval). This is theamount of time which passes during successive presentations of the shock (unless the operant
response is performed). The other one is called the R-S-interval (response-shock-interval)which specifies the length of the time interval following an operant response during which no
shocks will be delivered. Note that each time the organism performs the operant response, theR-S-interval without shocks begins anew.
[edit ] Two-process theory of avoidance
This theory was originally established to explain learning in discriminated avoidance
learning. It assumes two processes to take place:
a) C
a
a
ond
on
ng of f a
During the f irst trials of the training, the organism experiences both CS and aversive US
(escape-trials). The theory assumed that during those trials classical conditioning takes place
by pairing the CS with the US. Because of the aversive nature of the US the CS is supposed to
elicit a conditioned emotional reaction (CER) f ear. In classical conditioning, presenting a CS
conditioned with an aversive US disrupts the organism's ongoing behavior.
b) R
nfo
m
n
of
h
op
an
pon
by f
a
-
du
on
Because during the f irst process, the CS signaling the aversive US has itself become aversive
by eliciting f ear in the organism, reducing this unpleasant emotional reaction serves to
motivate the operant response. The organism learns to make the response during the US,
thus terminating the aversive internal reaction elicited by the CS. An important aspect of this
theory is that the term "avoidance" does not really describe what the organism is doing. It
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 13/15
does not "avoid" the aversive US in the sense of anticipating it. Rather the organism escapes
an aversive internal state, caused by the CS.
[edit ] Verbal Behavior
Ma
n a
j
k
l : V
l
bak
Bl hav
o
(book)
In 1957, Skinner published V erbal Behavior , a theoretical extension of the work he had pioneered since 1938. This work extended the theory of operant conditioning to human behavior previously assigned to the areas of language, linguistics and other areas. V erbal
Behavior is the logical extension of Skinner's ideas, in which he introduced new functionalrelationship categories such as intraverbals, autoclitics, mands, tacts and the controlling
relationship of the audience. All of these relationships were based on operant conditioningand relied on no new mechanisms despite the introduction of new functional categories.
[edit ] Four term cont ingency
Applied behavior analysis, which is the name of the discipline directly descended from
Skinner's work, holds that behavior is explained in four terms: conditional stimulus (S
C
), adiscriminative stimulus (Sd), a response (R), and a reinforcing stimulus (S
reinor S
r for
reinforcers, sometimes Save
for aversive stimuli).[1
m
]
[edit ] Operant hoarding
Operant hoardin is a referring to the choice made by a rat, on a compound schedule called
a multiple schedule, that maximizes its rate of reinforcement in an operant conditioningcontext. More specifically, rats were shown to have allowed food pellets to accumulate in a
food tray by continuing to press a lever on a continuous reinforcement schedule instead of retrieving those pellets. Retrieval of the pellets always instituted a one-minute period of
extinction during which no additional food pellets were available but those that had been
accumulated earlier could be consumed. This finding appears to contradict the usual findingthat rats behave impulsively in situations in which there is a choice between a smaller food
object right away and a larger food object after some delay. See schedules of reinforcement.
[15]
[edit ] An alternat ive to the law of effect
However, an alternative perspective has been proposed by R. Allen and Beatrix
Gardner.[16][17]
Under this idea, which they called "feedforward," animals learn during operant
conditioning by simple pairing of stimuli, rather than by the consequences of their actions.
Skinner asserted that a rat or pigeon would only manipulate a lever if rewarded for the action,
a process he called "shaping" (reward for approaching then manipulating a lever).
[18]
However, in order to prove the necessity of reward (reinforcement) in lever pressing, a
control condition where food is delivered without regard to behavior must also be conducted.
Skinner never published this control group. Only much later was it found that rats and
pigeons do indeed learn to manipulate a lever when food comes irrespective of behavior. This
phenomenon is known as autoshaping.[19]
Autoshaping demonstrates that consequence of action is not necessary in an operant conditioning chamber, and it contradicts the law of
effect. Further experimentation has shown that rats naturally handle small objects, such as alever, when food is present.
[20]Rats seem to insist on handling the lever when free food is
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 14/15
available (contra-freeloading)[21][22]
and even when pressing the lever leads to less food(omission training).[23][2
n
] Whenever food is presented, rats handle the lever, regardless if
lever pressing leads to more food. Therefore, handling a lever is a natural behavior that ratsdo as preparatory feeding activity, and in turn, lever pressing cannot logically be used as
evidence for reward or reinforcement to occur. In the absence of evidence for reinforcement
during operant conditioning, learning which occurs during operant experiments is actually
only Pavlovian (classical) conditioning. The dichotomy between Pavlovian and operantconditioning is therefore an inappropriate separation.
Conti uity Theory (E. Guthrie)
Overview:
Guthrie's contiguity theory specifies that "a combination of stimuli which has accompanied a
movement will on its recurrence tend to be followed by that movement". According toGuthrie, all learning was a consequence of association between a particular stimulus and
response. Furthermore, Guthrie argued that stimuli and responses affect specific sensory-motor patterns; what is learned are movements, not behaviors.
In contiguity theory, rewards or punishment play no significant role in learning since theyoccur after the association between stimulus and response has been made. Learning takes
place in a single trial (all or none). However, since each stimulus pattern is slightly different,many trials may be necessary to produce a general response. One interesting principle that
arises from this position is called "postremity" which specifies that we always learn the lastthing we do in response to a specific stimulus situation.
Contiguity theory suggests that forgetting is due to interference rather than the passage of
time; stimuli become associated with new responses. Previous conditioning can also bechanged by being associated with inhibiting responses such as fear or fatigue. The role of
motivation is to create a state of arousal and activity which produces responses that can be
conditioned.
Scope/Application:
Contiguity theory is intended to be a general theory of learning, although most of the researchsupporting the theory was done with animals. Guthrie did apply his framework to personality
disorders (e.g. Guthrie, 1938).
Example:
The classic experimental paradigm for Contiguity theory is cats learning to escape from a puzzle box (Guthrie & Horton, 19 6). Guthrie used a glass paneled box that allowed him to
photograph the exact movements of cats. These photographs showed that cats learned to
8/8/2019 Connection Ism
http://slidepdf.com/reader/full/connection-ism 15/15
repeat the same sequence of movements associated with the preceding escape from the box.Improvement comes about because irrelevant movements are unlearned or not included in
successive associations.
Principles:
1. In order for conditioning to occur, the organism must actively respond (i.e., do things).
2. Since learning involves the conditioning of specific movements, instruction must present
very specific tasks.
3. Exposure to many variations in stimulus patterns is desirable in order to produce ageneralized response.
. The last response in a learning situation should be correct since it is the one that will be
associated.
Summary: Piaget¶s Stage Theory of Cognitive Development is a description of cognitive
development as four distinct stages in children: sensorimotor, preoperational, concrete, and
formal.
Originator: Jean Piaget (1896-1980)
Key Terms: Sensorimotor, preoperational, concrete, formal, accommodation, assimilation.
Pia et¶s Sta e Theory of Co nitive Development
Swiss biologist and psychologist Jean Piaget (1896-1980) observed his children (and their
process of making sense of the world around them) and eventually developed a four-stage
model of how the mind processes new information encountered. He posited that children
progress through stages and that they all do so in the same order. These four stages are:
y Sensorimotor sta e (Birth to 2 years old). The infant builds an understanding of
himself or herself and reality (and how things work) through interactions with the
environment. It is able to differentiate between itself and other objects. Learning takes
place via assimilation (the organization of information and absorbing it into existingschema) and accommodation (when an object cannot be assimilated and the schemata
have to be modified to include the object.
y Preoperational sta e (ages 2 to ). The child is not yet able to conceptualize
abstractly and needs concrete physical situations. O bjects are classified in simple
ways, especially by important features.
y Concrete operations (ages 7 to 11). As physical experience accumulates,
accomodation is increased. The child begins to think abstractly and conceptualize,creating logical structures that explain his or her physical experiences.