+ All Categories
Home > Documents > Connection Ism

Connection Ism

Date post: 09-Apr-2018
Category:
Upload: aarthiashok
View: 223 times
Download: 0 times
Share this document with a friend
15
Connectionism (E. Thorndike) Overview: The learning theory of T horndike represents the original S-R framework of behavioral  psychology: Learning is the result of a ssociations forming between sti muli and responses. Such associations or "habits" become strengthened or weakened by the natur e and frequency of the S-R pairings. The paradigm for S-R theory was trial and error learning in which certain responses come to dominate others due to rewards. The hallmark of connectionism (like all  behavioral theory) was that learning could be adequately explained witho ut refering to any unobse rvable inter nal states. Thorndike's theory consists of three primary la ws: (1) law of effect - responses to a situation which are followed by a rewarding state of affa irs will be streng thened and become habitual responses to that situation, (2) law of readiness - a series of responses can be chained together to satisfy some goal which will result in annoyance if blocked, and (3) law of exercise - connection s become strengthened with pra ctice and weakened when practic e is discontinued. A corollary of the law of effect was that responses that reduce the likeliho od of achieving a rewarding state (i.e., punishments, failures) will decrease in strength. The theory suggests that transfer of learning depends upon the presence of identical elements in the original and new learning situations; i.e., transfer is always specific, never general. In later versions of the theory, the concept of "belongingness" was introduced; connections are more readily established if the person perceives that stimuli or responses go together (c.f. Gestalt principles). Another concept introduced was "polarity" which s pecifies that connections occur more easily in the direction in which they were originally for med than the opposite. Thorndike also introduced the "spread of effect" idea, i.e., rewards affect not only the connection that produced them but temporally a djacent connections as well. Scope/Application : Connectionism was meant to be a general theory of learning for animals a nd humans. Thorndike was especially interested in the application of his theory to education including mathematics (Thorndike, 1922), spelling and readi ng (Thorndike, 1921) , measurement of intelligence (Thorndike et al., 1927) and adult learning (Thorndike at al., 1928). Example: The classic example of Thorndike's S-R theo ry was a cat learning to escape from a "puzzle  box" by pressing a lever inside the box. After much trial a nd error behavior, the cat learns to associate pressing the lever (S) with opening the door (R). This S-R connection is established  because it results in a satisfying state of affairs (escape from the box). The law of exercise specifies that the connection was established because the S-R pairing occurred many times (the law of effect) and was rewarded (law of effect) as well as for ming a single sequence (law of readiness).
Transcript

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 1/15

Connectionism (E. Thorndike)

Overview:

The learning theory of Thorndike represents the original S-R framework of behavioral

 psychology: Learning is the result of associations forming between stimuli and responses.Such associations or "habits" become strengthened or weakened by the nature and frequency

of the S-R pairings. The paradigm for S-R theory was trial and error learning in which certain

responses come to dominate others due to rewards. The hallmark of connectionism (like all

 behavioral theory) was that learning could be adequately explained without refering to any

unobservable internal states.

Thorndike's theory consists of three primary laws: (1) law of effect - responses to a situationwhich are followed by a rewarding state of affairs will be strengthened and become habitual

responses to that situation, (2) law of readiness - a series of responses can be chained together to satisfy some goal which will result in annoyance if blocked, and (3) law of exercise -

connections become strengthened with practice and weakened when practice is discontinued.

A corollary of the law of effect was that responses that reduce the likelihood of achieving arewarding state (i.e., punishments, failures) will decrease in strength.

The theory suggests that transfer of learning depends upon the presence of identical elements

in the original and new learning situations; i.e., transfer is always specific, never general. In

later versions of the theory, the concept of "belongingness" was introduced; connections are

more readily established if the person perceives that stimuli or responses go together (c.f.

Gestalt principles). Another concept introduced was "polarity" which specifies that

connections occur more easily in the direction in which they were originally formed than the

opposite. Thorndike also introduced the "spread of effect" idea, i.e., rewards affect not only

the connection that produced them but temporally adjacent connections as well.

Scope/Application:

Connectionism was meant to be a general theory of learning for animals and humans.

Thorndike was especially interested in the application of his theory to education including

mathematics (Thorndike, 1922), spelling and reading (Thorndike, 1921), measurement of 

intelligence (Thorndike et al., 1927) and adult learning (Thorndike at al., 1928).

Example:

The classic example of Thorndike's S-R theory was a cat learning to escape from a "puzzle

 box" by pressing a lever inside the box. After much trial and error behavior, the cat learns to

associate pressing the lever (S) with opening the door (R). This S-R connection is established

 because it results in a satisfying state of affairs (escape from the box). The law of exercise

specifies that the connection was established because the S-R pairing occurred many times

(the law of effect) and was rewarded (law of effect) as well as forming a single sequence (law

of readiness).

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 2/15

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 3/15

Contents

[hide]

y  1 Types 

y  2 Procedure variations 

o  2.1 Classical discrimination/reversal conditioning o  2.2 Classical ISI discrimination conditioning 

o  2.3 Latent inhibition conditioning 

o  2.4 Conditioned inhibition conditioning 

o  2.5 Blocking 

y  3 Applications 

o  3.1 Little Albert 

o  3.2 Behavioral therapies 

y  4 Theories of classical conditioning 

y  5 In popular culture 

y  6 See also 

y  7 References 

y  8 Further reading y  9 External links 

[edit ] Types

Diagram representing forward conditioning. The time interval increases from left to right.

y  Forward conditioning: During forward conditioning the onset of the CS precedes the onset

of the US. Two common forms of forward conditioning are delay and trace conditioning.

y  Delay conditioning: In delay conditioning the CS is presented and is overlapped by the

presentation of the US.

y  Trace conditioning: During trace conditioning the CS and US do not overlap. Instead, the CS

is presented, a period of time is allowed to elapse during which no stimuli are presented,

and then the US is presented. The stimulus free period is called thetrace interval . It may also

be called the conditioning interval .

y  Simultaneous conditioning: During simultaneous conditioning, the CS and US are presentedand terminated at the same time.

y  Backward conditioning: Backward conditioning occurs when a conditioned stimulus

immediately follows an unconditioned stimulus. Unlike traditional conditioning models, in

which the conditioned stimulus precedes the unconditioned stimulus, the conditioned

response tends to be inhibitory. This is because the conditioned stimulus serves as a signal

that the unconditioned stimulus has ended, rather than a reliable method of predicting the

future occurrence of the unconditioned stimulus.

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 4/15

y  Tempo 

¡  l conditionin

¢  

: The£  

S¤ s

¥ ¦ ese

§  

 ̈ 

e©   

 ̈ 

¦ e

  

¦  y-

 ̈ 

¤    e

© ¤ §  

 ̈ 

e¦ v

 s 

§ ©    cq

¤ s

¤ 

 ̈ 

¤ §  

¤ s

©  e

¥  e

§ ©  e

§  

 ̈ 

¥ §    c

¦ ¦ ec

¨  ̈ 

¤ ¤ §    

  

  

  

 ̈ 

he¤ §  

 ̈ 

e¦ v

  e

 ̈ 

!   

ee§ £  

S¥ ¦ 

ese§  

 ̈ 

  

 ̈ 

¤ §  s

"  The

   c #  

  

¦ § ©    

 

¦  c §  

 ̈ 

e$  

 ̈ 

  c §   se ¦  ve    s ̈ 

he    S ¤ §  

 ̈ 

his e$  

¥  e " 

y  Unpai% ed conditionin

&  

: The CS and US are not presented together. Us'  

ally they are 

presented as independent trials that are separated by a variable(  or pse'  

do-random,

interval. This procedure is used to study non-associative behavioral responses, such as 

sensitization.y  CS-alone e

)  

tinction: The CS is presented in the absence of the US. This procedure is usually 

done after the CR has been acquired through "Forward conditioning" training. Eventually,

the CR frequency is reduced to pre-training levels.

[edi0  

] Pr1  

cedure variat i1  

2 3  

 

In addition to the simple procedures described above, some classical conditioning studies are

designed to tap into more complex learning processes. Some common variations are

discussed below.

[edit ] Classical discrimi4  

at i5 4  

/reversal c5 4  

dit i5 4  

i4  

In this procedure, two CSs and one US are typically used. The CSs may be the same modality

(such as lights of different intensity), or they may be different modalities (such as auditory

CS and visual CS). In this procedure, one of the CSs is designated CS+ and its presentation is

always followed by the US. The other CS is designated CSí and its presentation is never followed by the US. After a number of trials, the organism learns to d i6 7 8   i9    i@ A   t B   CS+ trials

and CSí trials such that CRs are only observed on CS+ trials.

During RB   vB 8 6 A    l T 8 A   i@   i@   g , the CS+ and CSí are reversed and subjects learn to suppressresponding to the previous CS+ and show CRs to the previous CSí.

[edit ] Classical ISI discrimiC  

at iD C  

 cD C  

dit iD C  

iC  

This is a discrimination procedure in which two different CSs are used to signal two different

interstimulus intervals. For example, a dim light may be presented 30 seconds before a US,while a very bright light is presented 2 minutes before the US. Using this technique,

organisms can learn to perform CRs that are appropriately timed for the two distinct CSs.

[edit ] Latent inhibit iD  

n cD  

ndit iD  

ning  

In this procedure, a CS is presented several times before paired CS-US training commences.

The pre-exposure of the subject to the CS before paired training slows the rate of CR 

acquisition relative to organisms that are not CS pre-exposed. Also see Latent inhibition for 

applications.

[edit ] Condit ioned inhibit ion condit ioning  

Three phases of conditioning are typically used:

Phase 1: 

A CS (CS+) is paired with a US until asymptotic CR levels are reached.

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 5/15

Phase 2: 

CS+E  

US trials are continued, but interspersed with trials on which the CS+ in compound with 

a second CS, but not with the US (i.e., CS+E  

CS trials). Typically, organisms show CRs on

CS+E  

US trials, but suppress responding on CS+E  

CS trials.

Phase 3: 

In this retention test, the previous CS is paired with the US. If conditioned inhibition has 

occurred, the rate of acquisition to the previous CS should be impaired relative to

organisms that did not eF  

perience Phase 2.

[edit ] Blocking 

MaG 

n aH  

P  

R  : B

oP  k 

ngR   ff 

R P I 

This form of classical conditioning involves two phases.

Phase 1: 

A CS (CS1) is paired with a US.

Phase 2: 

A compound CS (CS1+CS2) is paired with a US.

Test: 

A separate test for each CS (CS1 and CS2) is performed. The blocS   ing eff ect is observed in a

lacS  

of conditioned response to CS2, suggesting that the f irst phase of training blocS  ed the 

acquisition of the second CS.

[edit ] Applicat ions

[edit ] Little Albert 

MaT 

n aU  

W  

Y  : L

V V 

Y   A

bY  

U  

V Y   xp

Y  

U  

mY  n

John B. Watson, founder of behaviorism, demonstrated classical conditioning empirically

through experimentation using the Little Albert experiment in which a child ("Albert") was

 presented with a white rat (CS). After a control period in which the child reacted normally to

the presence of the rat, the experimentors paired the presence of the rat with a loud, jarring

noise caused by clanging two pipes together behind the child's head (US). As the trials progressed, the child began showing signs of distress at the sight of the rat, even when

unaccompanied by the frightening noise. Furthermore, the child demonstrated generalizationof stimulus associations, and showed distress when presented with any white, furry object² 

even such things as a rabbit, dog, a fur coat, and a Santa Claus mask with hair.

[edit ] Behavioral therapies

Ma ̀

n aa  

 ̀

c  

e  : B

e  hav 

 ̀

oa  

b h

e  

a  

apy 

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 6/15

In human  psychology, implications for therapies and treatments using classical conditioningdiffer from operant conditioning. Therapies associated with classical conditioning are

aversion therapy, flooding and systematic desensitization.

Classical conditioning is short-term, usually requiring less time with therapists and less effort

from patients, unlike humanistic therapies.[cit 

f ti

g h h i i  ded ]

The therapies mentioned are designed

to cause either aversive feelings toward something, or to reduce unwanted fear and aversion.

[edit ] Theories of classical condit ioning  

There are two competing theories of how classical conditioning works. The first, stimulus-

response theory, suggests that an association to the unconditioned stimulus is made with the

conditioned stimulus within the brain, but without involving conscious thought. The second,

stimulus-stimulus theory involves cognitive activity, in which the conditioned stimulus is

associated to the concept of the unconditioned stimulus, a subtle but important distinction.

Stimulus-response theory, referred to as S-R theory, is a theoretical model of behavioral

 psychology that suggests humans and other animals can learn to associate a new stimulus, the

conditioned stimulus (CS), with a pre-existing stimulus, the unconditioned stimulus (US),

and can think, feel or respond to the CS as if it were actually the US.

The opposing theory, put forward by cognitive behaviorists, is stimulus-stimulus theory (S-S

theory). S-S theory is a theoretical model of classical conditioning that suggests a cognitive

component is required to understand classical conditioning and that S-R theory is an

inadequate model. It proposes that a cognitive component is at play. S-R theory suggests that

an animal can learn to associate a conditioned stimulus (CS) such as a bell, with the

impending arrival of food termed the unconditioned stimulus, resulting in an observable

 behavior such as salivation. S-S theory suggests that instead the animal salivates to the bell

 because it is associated with the concept of food, which is a very fine but important

distinction.

To test this theory, psychologist Robert Rescorla undertook the following experiment.[2]

Rats

learned to associate a loud noise as the unconditioned stimulus, and a light as the conditioned

stimulus. The response of the rats was to freeze and cease movement. What would happen

then if the rats were habituated to the US? S-R theory would suggest that the rats wouldcontinue to respond to the CS, but if S-S theory is correct, they would be habituated to the

concept of a loud sound (danger), and so would not freeze to the CS. The experimental resultssuggest that S-S was correct, as the rats no longer froze when exposed to the signal light.

[3] 

His theory still continues and is applied in everyday life.[1]

 

[edit ] In popular cult ure

One of the earliest literary references to classical conditioning can be found in the comic

novel T p  

e Lif e and Opiniq   ns o f  T r   i st r   am Shandy, Gentl eman (1759) by Laurence Sterne. Thenarrator Tristram Shandy explains

[s  

]how his mother was conditioned by his father's habit of 

winding up a clock before having sex with his wife:

My father, [...], was, I believe, one of the most regular men in every thing he did [...] [H]e had made 

it a rule for many years of his lif e,on the f irst Sunday-night of every month throughout the whole 

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 7/15

year,as certain as ever the Sunday-night came,to wind up a large house-clock, which we had

standing on the back-stairs head, with his own hands:And being somewhere between fifty and

sixty years of age at the time I have been speaking of,he had likewise gradually brought some

other little family concernments to the same period, in order, as he would often say to my uncle

Toby, to get them all out of the way at one time, and be no more plagued and pestered with them

the rest of the month.[...]

[F]rom an unhappy association of ideas, which have no connection in nature, it so fell out at length, that my poor mother could never hear the said clock wound up, ²but the thoughts of 

some other things unavoidably popped into her head²& vice versa: ²Which strange

combination of ideas, the sagacious Locke, who cer tainly understood the nature of these

things better than most men, aff irms to have produced more wry actions than all other sources

of pre judice whatsoever.

Another example is in the dystopian novel, A Clockwork Orant   e in which the f ilm's anti hero and protagonist, Alex, is given a solution to cause severe nausea, and is forced to watch

violent acts. This renders him unable to perform any violent acts without inducing similar 

nausea.

The song "Dinner Bell" by the group "They Might Be Giants" talks about "salivating dogs".

In the song "Cenoir Studies 02" by rapper "Cani bus" (Germaine Williams), he says

"Sometimes I wonder who's listening, the auditory Pavlovian conditioning's so sickening".

Operant conditioning 

From Wikipedia, the free encyclopedia

Jump to: navigation, search 

"O perant " redirects here. For the deu  

inition ou  

the word " operant ", see Wi v  

tionar w  

x  operant .

It has been suggested thatMutual operant conditioning be merged into this article or

section. (Discussy  

Operant conditioning is the use of consequences to modify the occurrence and form of 

 behavior. Operant conditioning is distinguished fromclassical conditioning (also called

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 8/15

respondent conditioning) in that operant conditioning deals with the modification of "voluntary behavior" or operant behavior. O perant behavior "operates" on the environment

and is maintained by its consequences, while classical conditioning deals with theconditioning of respondent behaviors which are elicited by antecedent conditions. Behaviors

conditioned via a classical conditioning procedure are not maintained by consequences.[1]

 

Contents

[hide] 

y  1 Reinforcement, punishment, and e  

tinction 

o  1.1 Four conte  

ts of operant conditioning 

y  2 Thorndike s law of eff ect 

y  3 Biological correlates of operant conditioning 

y  4 Factors that alter the eff ectiveness of consequences 

y  5 Operant variability 

y  6 Avoidance learning 

o  6.1 Discriminated avoidance learning 

o  6.2 Free-operant avoidance learning y  7 Two-process theory of avoidance 

y  8 Verbal Behavior 

y  9 Four term contingency 

y  10 Operant hoarding 

y  11 An alternative to the law of eff ect 

y  12 See also 

y  13 Ref erences 

y  14 External links 

[edit ] Reinforcement, punishment, and ext inct ion

Reinforcement and punishment, the core tools of operant conditioning, are either positive

(delivered following a response), or negative (withdrawn following a response). This creates

a total of four basic consequences, with the addition of a fifth procedure known as extinction 

(i.e. no change in consequences following a response).

It's important to note that actors are not spoken of as being reinforced, punished, or 

extinguished; it is the actions that are reinforced, punished, or extinguished. Additionally,reinforcement, punishment, and extinction are not terms whose use is restricted to the

laboratory. Naturally occurring consequences can also be said to reinforce, punish, or extinguish behavior and are not always delivered by people.

y  Reinforcement is a consequence that causes a behavior to occur with greater frequency.

y  Punishment is a consequence that causes a behavior to occur with less frequency.

y  Extinction is the lack of any consequence following a behavior. When a behavior is 

inconsequential, producing neither favorable nor unfavorable consequences, it will occur

with less frequency. When a previously reinforced behavior is no longer reinforced with 

either positive or negative reinforcement, it leads to a decline in the response.

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 9/15

[edit ] Four contexts of operant condit ioning 

Here the terms positive and ne g ative are not used in their  popular sense, but rather: positive 

refers to addition, and ne g ative refers to subtraction.

What is added or subtracted may be either reinforcement or punishment. Hence positive

 puni shment is sometimes a confusing term, as it denotes the "addition" of a stimulus or increase in the intensity of a stimulus that is aversive (such as spanking or an electric shock).

The four procedures are:

1.  Positive reinforcement (Reinforcement)   occurs when a behavior (response) is followed by a

stimulus that is rewarding, increasing the frequency of that behavior. In the Skinner box 

experiment, a stimulus such as food or sugar solution can be delivered when the rat engages 

in a target behavior, such as pressing a lever.

2.  Negative reinforcement (Escape) occurs when a behavior (response) is followed by the 

removal of an aversive stimulus, thereby increasing that behavior's frequency. In the Skinner

box experiment, negative reinforcement can be a loud noise continuously sounding inside 

the rat's cage until it engages in the target behavior, such as pressing a lever, upon which the 

loud noise is removed.3.  Positive punishment (Punishment) (also called "Punishment by contingent stimulation")

  

occurs when a behavior (response) is followed by a stimulus, such as introducing a shock or

loud noise, resulting in a decrease in that behavior.

4.  Negative punishment (Penalty) (also called "Punishment by contingent withdrawal") 

occurs 

when a behavior (response) is followed by the removal of a stimulus, such as taking away a

child's toy following an undesired behavior, resulting in a decrease in that behavior.

 Al  so: 

y  Avoidance learning is a type of learning in which a certain behavior results in the cessation

of an aversive stimulus. For example, performing the behavior of shielding one's eyes when

in the sunlight (or going indoors) will help avoid the aversive stimulation of having light inone's eyes.

y  Extinction occurs when a behavior (response) that had previously been reinforced is no

longer eff ective. In the Skinner box experiment, this is the rat pushing the lever and being

rewarded with a food pellet several times, and then pushing the lever again and never

receiving a food pellet again. Eventually the rat would cease pushing the lever.

y  Noncontingent reinforcement ref ers to delivery of reinforcing stimuli regardless of the 

organism's (aberrant) behavior. The idea is that the target behavior decreases because it is 

no longer necessary to receive the reinforcement. This typically entails time-based delivery 

of stimuli identif ied as maintaining aberrant behavior, which serves to decrease the rate of 

the target behavior.[2]

As no measured behavior is identif ied as being strengthened, there is 

controversy surrounding the use of the term noncontingent "reinforcement".[3]

 

y  Shaping is a form of operant conditioning in which the increasingly accurate approximations of a desired response are reinforced.

[4] 

y  Chaining is an instructional procedure which involves reinforcing individual responses 

occurring in a sequence to form a complex behavior.[4]

 

[edit ] Thorndik e's law of effect 

Ma 

n a  

 

 

  

 

   : La    of      ff   

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 10/15

O perant conditioning, sometimes called inst    ument al cond itionin g or inst    ument al  l earnin g ,was first extensively studied by Edward L. Thorndike (187 ±19 9), who observed the

 behavior of cats trying to escape from home-made puzzle boxes.[5]

When first constrained inthe boxes, the cats took a long time to escape. With experience, ineffective responses

occurred less frequently and successful responses occurred more frequently, enabling the cats

to escape in less time over successive trials. In his law of effect, Thorndike theorized that

successful responses, those producing sati s f  yin g consequences, were "stamped in" by theexperience and thus occurred more frequently. Unsuccessful responses, those producing

annoyin g consequences, were st amped out and subsequently occurred less frequently. In

short, some consequences st ren gt hened behavior and some consequences weakened behavior.

Thorndike produced the first known learning curves through this procedure. B.F. Skinner  

(190 ±1990) formulated a more detailed analysis of operant conditioning based on

reinforcement, punishment, and extinction. Following the ideas of Ernst Mach, Skinner 

rejected Thorndike's mediating structures required by "satisfaction" and constructed a new

conceptualization of behavior without any such references. So, while experimenting with

some homemade feeding mechanisms, Skinner invented the operant conditioning chamber  

which allowed him to measure rate of response as a key dependent variable using acumulative record of lever presses or key pecks.[6] 

[edit ] Biological correlates of operant condit ioning  

The first scientific studies identifying neurons that responded in ways that suggested theyencode for conditioned stimuli came from work by Mahlon deLong.[7][8] and by R.T. "Rusty"

Richardson and deLong.[8]

They showed that nucleus basalis neurons, which releaseacetylcholine broadly throughout the cerebral cortex, are activated shortly after a conditioned

stimulus, or after a primary reward if no conditioned stimulus exists. These neurons areequally active for positive and negative reinforcers, and have been demonstrated to cause

 plasticity in many cortical regions.[9]

Evidence also exists that dopamine is activated at

similar times. There is considerable evidence that dopamine participates in both

reinforcement and aversive learning.

[10]

Dopamine pathways project much more densely ontofrontal cortex regions. Cholinergic projections, in contrast, are dense even in the posterior 

cortical regions like the primary visual cortex. A study of patients with Parkinson's disease, a

condition attributed to the insufficient action of dopamine, further illustrates the role of 

dopamine in positive reinforcement.[11]

It showed that while off their medication, patients

learned more readily with aversive consequences than with positive reinforcement. Patients

who were on their medication showed the opposite to be the case, positive reinforcement

 proving to be the more effective form of learning when the action of dopamine is high.

[edit ] Factors that alter the effect iveness of consequences

When using consequences to modify a response, the effectiveness of a consequence can be

increased or decreased by various factors. These factors can apply to either reinforcing or  punishing consequences.

1.  Satiation/Deprivation: The eff ectiveness of a consequence will be reduced if the individual's 

"appetite" for that source of stimulation has been satisf ied. Inversely, the eff ectiveness of a

consequence will increase as the individual becomes deprived of that stimulus. If someone is 

not hungry, food will not be an eff ective reinforcer for behavior. Satiation is generally only a

potential problem with primary reinforcers, those that do not need to be learned such as 

food and water.

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 11/15

2.  Immediacy: After a response, how immediately a consequence is then f elt determines the 

eff ectiveness of the consequence.More immediate f eedback will be more eff ective than less 

immediate f eedback. If someone's license plate is caught by a traff ic camera for speeding

and they receive a speeding ticket in the mail a week later, this consequence will not be very 

eff ective against speeding. But if someone is speeding and is caught in the act by an off icer

who pulls them over, then their speeding behavior is more likely to be aff ected.

3.  Contingency: If a consequence does not contingently (reliably, or consistently) follow the target response, its eff ectiveness upon the response is reduced. But if a consequence follows 

the response consistently after successive instances, its ability to modif y the response is 

increased. The schedule of reinforcement, when consistent, leads to faster learning. When

the schedule is variable the learning is slower. Extinction is more diff icult when learning

occurs during intermittent reinforcement and more easily extinguished when learning occurs 

during a highly consistent schedule.

4.  Size: This is a "cost-benef it" determinant of whether a consequence will be eff ective. If the 

size, or amount, of the consequence is large enough to be worth the effort, the consequence 

will be more eff ective upon the behavior. An unusually large lottery jackpot, for example,

might be enough to get someone to buy a one-dollar lottery ticket (or even buying multiple 

tickets). But if a lottery jackpot is small, the same person might not f eel it to be worth the 

effort of driving out and f inding a place to buy a ticket. In this example, it's also useful tonote that "effort" is a punishing consequence. How these opposing expected consequences 

(reinforcing and punishing) balance out will determine whether the behavior is performed or

not.

Most of these factors exist for biological reasons. The biological purpose of the Principle of 

Satiation is to maintain the organism's homeostasis. When an organism has been deprived of 

sugar, for example, the effectiveness of the taste of sugar as a reinforcer is high. However, as

the organism reaches or exceeds their optimum blood-sugar levels, the taste of sugar becomes

less effective, perhaps even aversive.

The Principles of Immediacy and Contingency exist for neurochemical reasons. When an

organism experiences a reinforcing stimulus, dopamine pathways in the brain are activated.This network of pathways "releases a short pulse of dopamine onto many dendrites, thus

 broadcasting a rather global reinforcement signal to postsynaptic neurons."[12]

This results in

the plasticity of these synapses allowing recently activated synapses to increase their sensitivity to efferent signals, hence increasing the probability of occurrence for the recent

responses preceding the reinforcement. These responses are, statistically, the most likely to

have been the behavior responsible for successfully achieving reinforcement. But when the

application of reinforcement is either less immediate or less contingent (less consistent), the

ability of dopamine to act upon the appropriate synapses is reduced.

[edit ] Operant variability

O perant variability is what allows a response to adapt to new situations. O perant behavior is

distinguished from reflexes in that its response topo raphy (the form of the response) issubject to slight variations from one performance to another. These slight variations can

include small differences in the specific motions involved, differences in the amount of forceapplied, and small changes in the timing of the response. If a subject's history of 

reinforcement is consistent, such variations will remain stable because the same successfulvariations are more likely to be reinforced than less successful variations. However,

 behavioral variability can also be altered when subjected to certain controlling variables.[13]

 

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 12/15

[edit ] Avoidance learning  

Avoidance learning belongs to negative reinforcement schedules. The subject learns that a

certain response will result in the termination or prevention of an aversive stimulus. There are

two kinds of commonly used experimental settings: discriminated and free-operant avoidance

learning.

[edit ] Discriminated avoidance learning 

In discriminated avoidance learning, a novel stimulus such as a light or a tone is followed by

an aversive stimulus such as a shock (CS-US, similar to classical conditioning). During the

first trials (called escape-trials) the animal usually experiences both the CS (Conditioned

Stimulus) and the US (Unconditioned Stimulus), showing the operant response to terminate

the aversive US. During later trials, the animal will learn to perform the response already

during the presentation of the CS thus preventing the aversive US from occurring. Such trials

are called "avoidance trials."

[edit ] Free-operant avoidance learning 

In this experimental session, no discrete stimulus is used to signal the occurrence of the

aversive stimulus. Rather, the aversive stimulus (mostly shocks) are presented without

explicit warning stimuli. There are two crucial time intervals determining the rate of 

avoidance learning. This first one is called the S-S-interval (shock-shock-interval). This is theamount of time which passes during successive presentations of the shock (unless the operant

response is performed). The other one is called the R-S-interval (response-shock-interval)which specifies the length of the time interval following an operant response during which no

shocks will be delivered. Note that each time the organism performs the operant response, theR-S-interval without shocks begins anew.

[edit ] Two-process theory of avoidance

This theory was originally established to explain learning in discriminated avoidance

learning. It assumes two processes to take place:

a) C  

a   

 

  a

 

  ond 

 

 

 

on 

ng of f    a

  

  

During the f irst trials of the training, the organism experiences both CS and aversive US 

(escape-trials). The theory assumed that during those trials classical conditioning takes place 

by pairing the CS with the US. Because of the aversive nature of the US the CS is supposed to

elicit a conditioned emotional reaction (CER) f ear. In classical conditioning, presenting a CS 

conditioned with an aversive US disrupts the organism's ongoing behavior.

b) R  

 

nfo  

  m

  n

 of 

 h

  op

  

  

an 

  

  

  

 pon  

  by f 

  a

  

-  

  du

 

 

on  

Because during the f irst process, the CS signaling the aversive US has itself become aversive 

by eliciting f ear in the organism, reducing this unpleasant emotional reaction serves to

motivate the operant response. The organism learns to make the response during the US,

thus terminating the aversive internal reaction elicited by the CS. An important aspect of this 

theory is that the term "avoidance" does not really describe what the organism is doing. It

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 13/15

does not "avoid" the aversive US in the sense of anticipating it. Rather the organism escapes 

an aversive internal state, caused by the CS.

[edit ] Verbal Behavior  

Ma 

n a  

 

 

j  

l  : V 

l  

  

bak 

Bl  hav 

 

o  

(book)

In 1957, Skinner published V erbal Behavior , a theoretical extension of the work he had pioneered since 1938. This work extended the theory of operant conditioning to human behavior previously assigned to the areas of language, linguistics and other areas. V erbal  

 Behavior is the logical extension of Skinner's ideas, in which he introduced new functionalrelationship categories such as intraverbals, autoclitics, mands, tacts and the controlling

relationship of the audience. All of these relationships were based on operant conditioningand relied on no new mechanisms despite the introduction of new functional categories.

[edit ] Four term cont ingency

Applied behavior analysis, which is the name of the discipline directly descended from

Skinner's work, holds that behavior is explained in four terms: conditional stimulus (S

C

), adiscriminative stimulus (Sd), a response (R), and a reinforcing stimulus (S

reinor S

r for 

reinforcers, sometimes Save

for aversive stimuli).[1

m  

[edit ] Operant hoarding  

Operant hoardin is a referring to the choice made by a rat, on a compound schedule called

a multiple schedule, that maximizes its rate of reinforcement in an operant conditioningcontext. More specifically, rats were shown to have allowed food pellets to accumulate in a

food tray by continuing to press a lever on a continuous reinforcement schedule instead of retrieving those pellets. Retrieval of the pellets always instituted a one-minute period of 

extinction during which no additional food pellets were available but those that had been

accumulated earlier could be consumed. This finding appears to contradict the usual findingthat rats behave impulsively in situations in which there is a choice between a smaller food

object right away and a larger food object after some delay. See schedules of reinforcement.

[15] 

[edit ] An alternat ive to the law of effect 

However, an alternative perspective has been proposed by R. Allen and Beatrix

Gardner.[16][17]

Under this idea, which they called "feedforward," animals learn during operant

conditioning by simple pairing of stimuli, rather than by the consequences of their actions.

Skinner asserted that a rat or pigeon would only manipulate a lever if rewarded for the action,

a process he called "shaping" (reward for approaching then manipulating a lever).

[18]

 However, in order to prove the necessity of reward (reinforcement) in lever pressing, a

control condition where food is delivered without regard to behavior must also be conducted.

Skinner never published this control group. Only much later was it found that rats and

 pigeons do indeed learn to manipulate a lever when food comes irrespective of behavior. This

 phenomenon is known as autoshaping.[19]

Autoshaping demonstrates that consequence of action is not necessary in an operant conditioning chamber, and it contradicts the law of 

effect. Further experimentation has shown that rats naturally handle small objects, such as alever, when food is present.

[20]Rats seem to insist on handling the lever when free food is

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 14/15

available (contra-freeloading)[21][22]

and even when pressing the lever leads to less food(omission training).[23][2

n  

] Whenever food is presented, rats handle the lever, regardless if 

lever pressing leads to more food. Therefore, handling a lever is a natural behavior that ratsdo as preparatory feeding activity, and in turn, lever pressing cannot logically be used as

evidence for reward or reinforcement to occur. In the absence of evidence for reinforcement

during operant conditioning, learning which occurs during operant experiments is actually

only Pavlovian (classical) conditioning. The dichotomy between Pavlovian and operantconditioning is therefore an inappropriate separation.

Conti uity Theory (E. Guthrie)

Overview:

Guthrie's contiguity theory specifies that "a combination of stimuli which has accompanied a

movement will on its recurrence tend to be followed by that movement". According toGuthrie, all learning was a consequence of association between a particular stimulus and

response. Furthermore, Guthrie argued that stimuli and responses affect specific sensory-motor patterns; what is learned are movements, not behaviors.

In contiguity theory, rewards or punishment play no significant role in learning since theyoccur after the association between stimulus and response has been made. Learning takes

 place in a single trial (all or none). However, since each stimulus pattern is slightly different,many trials may be necessary to produce a general response. One interesting principle that

arises from this position is called "postremity" which specifies that we always learn the lastthing we do in response to a specific stimulus situation.

Contiguity theory suggests that forgetting is due to interference rather than the passage of 

time; stimuli become associated with new responses. Previous conditioning can also bechanged by being associated with inhibiting responses such as fear or fatigue. The role of 

motivation is to create a state of arousal and activity which produces responses that can be

conditioned.

Scope/Application:

Contiguity theory is intended to be a general theory of learning, although most of the researchsupporting the theory was done with animals. Guthrie did apply his framework to personality

disorders (e.g. Guthrie, 1938).

Example:

The classic experimental paradigm for Contiguity theory is cats learning to escape from a puzzle box (Guthrie & Horton, 19 6). Guthrie used a glass paneled box that allowed him to

 photograph the exact movements of cats. These photographs showed that cats learned to

8/8/2019 Connection Ism

http://slidepdf.com/reader/full/connection-ism 15/15

repeat the same sequence of movements associated with the preceding escape from the box.Improvement comes about because irrelevant movements are unlearned or not included in

successive associations.

Principles:

1. In order for conditioning to occur, the organism must actively respond (i.e., do things).

2. Since learning involves the conditioning of specific movements, instruction must present

very specific tasks.

3. Exposure to many variations in stimulus patterns is desirable in order to produce ageneralized response.

. The last response in a learning situation should be correct since it is the one that will be

associated.

Summary: Piaget¶s Stage Theory of Cognitive Development is a description of cognitive

development as four distinct stages in children: sensorimotor, preoperational, concrete, and

formal.

Originator: Jean Piaget (1896-1980)

Key Terms: Sensorimotor, preoperational, concrete, formal, accommodation, assimilation.

Pia et¶s Sta e Theory of Co nitive Development 

Swiss biologist and psychologist Jean Piaget (1896-1980) observed his children (and their 

 process of making sense of the world around them) and eventually developed a four-stage

model of how the mind processes new information encountered. He posited that children

 progress through stages and that they all do so in the same order. These four stages are:

y  Sensorimotor sta e (Birth to 2 years old). The infant builds an understanding of 

himself or herself and reality (and how things work) through interactions with the

environment. It is able to differentiate between itself and other objects. Learning takes

 place via assimilation (the organization of information and absorbing it into existingschema) and accommodation (when an object cannot be assimilated and the schemata

have to be modified to include the object.

y  Preoperational sta e (ages 2 to ). The child is not yet able to conceptualize

abstractly and needs concrete physical situations. O bjects are classified in simple

ways, especially by important features.

y  Concrete operations (ages 7 to 11). As physical experience accumulates,

accomodation is increased. The child begins to think abstractly and conceptualize,creating logical structures that explain his or her physical experiences.


Recommended