Sunteți pe pagina 1din 25

See

discussions, stats, and author profiles for this publication at: https://www.researchgate.net/publication/310766633

Behavior Analysis and Neuroscience:


Complementary Disciplines

Article in Journal of the Experimental Analysis of Behavior · January 2016


DOI: 10.1002/jeab.251

CITATIONS READS

0 128

1 author:

John W. Donahoe
University of Massachusetts Amherst
85 PUBLICATIONS 1,264 CITATIONS

SEE PROFILE

Some of the authors of this publication are also working on these related projects:

A Video describing the theoretical background of the Unified Reinforcement Principle View project

Conflict behavior View project

All content following this page was uploaded by John W. Donahoe on 23 November 2016.

The user has requested enhancement of the downloaded file. All in-text references underlined in blue are added to the original document
and are linked to publications on ResearchGate, letting you access and read them immediately.
1

In press, Journal of the Experimental Analysis of Behavior. Consult Journal for pagination and bibliographic citation.

Behavior Analysis and Neuroscience: Complementary Disciplines

John W. Donahoe
University of Massachusetts/Amherst
Amherst, MA 01002

Abstract

Behavior analysis and neuroscience are disciplines in their own right but are united in that both
are subfields of a common overarching field—biology. What most fundamentally unites these
disciplines is a shared commitment to selectionism, the Darwinian mode of explanation. In selec-
tionism, the order and complexity observed in nature are seen as the cumulative products of se-
lection processes acting over time on a population of variants—favoring some and disfavoring
others—with the affected variants contributing to the population on which future selections oper-
ate. In the case of behavior analysis, the central selection process is selection by reinforcement;
in neuroscience it is natural selection. The two selection processes are inter-related in that selec-
tion by reinforcement is itself the product of natural selection. The present paper illustrates the
complementary nature of behavior analysis and neuroscience through considering their joint con-
tributions to three central problem areas: reinforcement—including conditioned reinforcement,
stimulus control—including equivalence classes, and memory—including reminding and re-
membering.

This article illustrates the complementary selection and genetics, so it may well be with se-
nature of behavior analysis and neuroscience, lection by reinforcement and its neural mecha-
each a scientific discipline in its own right, but nisms.
each informed and enriched by the other. The un- Efforts to integrate findings from behav-
derlying premise is that the advantages of the in- ior analysis and neuroscience are nothing new, of
tegration of behavior analysis and neuroscience course. Edward Thorndike’s earliest writings em-
are foreshadowed by the earlier benefits of inte- phasized the importance of not only the Law of
grating the sciences of evolution and heredity, Effect but also of what he termed the “Law of Ac-
now known as the Modern Synthesis (Dobzhan- quired Brain Connections” (Thorndike, 1903, p.
sky, 1937, cf. Donahoe, 2003). Of this earlier syn- 165; cf. Donahoe, 1999), although he lamented
thesis, the biologist and philosopher of science the absence of direct experimental knowledge of
Paul Gayon wrote, “If there is a key event in the the latter. Donald Hebb was perhaps the first
history of Darwinism, it must be its confrontation American psychologist to attempt a wide-ranging
with Mendelism. It was through its contact with integration of the two disciplines in his Organi-
the new science of heredity that the theory of se- zation of Behavior (1949), but again direct
lection became truly intelligible” (Gayon, 1992, knowledge of the neural processes remained
p. 253). As it was with evolution through natural
2

largely absent. Hebb’s proposal that the connec- different behaviors of the same organism in the
tions between two neurons increased in strength case of behavior analysis—is what most funda-
when they were co-active was simply a restate- mentally distinguishes much of psychology from
ment in neural terms of the behaviorally estab- behavior analysis. Psychology often seeks its
lished contiguity requirement. Skinner regarded principles through averaging steady-state obser-
behavior analysis as “a rigorous, extensive, and vations from different organisms, a method ap-
rapidly advancing branch of biology” (1974, p. propriate for population genetics but not neces-
255) but noted two prerequisites for such an inte- sarily for a science of behavior (e.g., Estes, 1956;
gration. Of behavior: “A rigorous description at Sidman, 1960).
the level of behavior is necessary for the demon- In subsequent sections, the relevant be-
stration of a neurological correlate.” And, of neu- havioral findings are first briefly summarized fol-
roscience: “A science of the nervous system will lowed by the neural mechanisms that mediate that
someday start from the direct observation of neu- behavior. Skinner’s admonition that a “rigorous
ral processes … It is with such a science that the description at the level of behavior” must precede
neurological point of view must be concerned if “the demonstration of a neurological correlate” is
it is to offer a convincing ‘explanation’ (sic) of amply illustrated. The first findings considered
behavior” (Skinner, 1938, p. 422). These two pre- are those related to the reinforcement process it-
requisites are becoming increasingly approxi- self. Just as natural selection informs all of evo-
mated, if not satisfied, by the present state of de- lutionary biology, so a properly formulated con-
velopment of each discipline. ception of selection by reinforcement should illu-
Before describing specific examples of minate the full range of behavior from simple
the salutary effects of integrating behavior analy- conditioning to complex human behavior. The
sis and neuroscience, two general orienting prin- discussion of selection by reinforcement, includ-
ciples are identified. First, both evolutionary bi- ing conditioned reinforcement, is followed by
ology and behavior analysis are selectionist sci- sections on stimulus control—including equiva-
ences (Catania, 1987; Donahoe, 2003; Palmer & lence classes, and memory.1
Donahoe, 1992; Staddon, 1983; McDowell,
2013a). That is, the diverse and potentially com- Selection by Reinforcement
plex outcomes of evolution and reinforcement are The approach to reinforcement described
the cumulative products of relatively simple se- here is a logical extension of Skinner’s view that
lection processes acting over time. Any given in- selection occurs at the moment when a reinforc-
stance of selection occurs under conditions that ing stimulus (that is, a reinforcer) appears in close
exist at that moment, conditions which may or proximity to a stimulus in the Pavlovian proce-
may not be shared with future moments. Strictly dure or to a response in the operant procedure.
speaking, selection processes “prepare” the or- Figure 1 depicts these two temporal relations.
ganism for conditions that existed in the past, not Note, however, that in the Pavlovian procedure
in the future except in so far as the future resem- some response necessarily occurs in proximity to
bles the past. Second, the cumulative products of the reinforcer whereas in the operant procedure
selection processes are the net effects of many in- some stimulus necessarily occurs in proximity to
dividual instances of selection, whether reproduc- the reinforcer. Thus, what truly distinguishes be-
tive fitness acting on a population of different or- tween the two procedures is not whether a stimu-
ganisms in the case of evolution or reinforcement lus or a response precedes the reinforcer, but the
acting on a population of different environment- reliability with which that event precedes the re-
behavior relations of a single organism in the case inforcer. The classical and operant procedures
of behavior (Donahoe, 2003). The nature of the clearly differ, but whether the procedural differ-
population on which the selecting events act— ence portends a fundamental difference in the
different organisms in the case of evolution or conditioning process is a separate matter.

1
The body of the article is self-contained, but supplementary technical information is occasionally
introduced by means of endnotes.
2

Figure 1. Diagram illustrating the similarity and difference between the Pavlovian and operant
procedures. In both procedures, a reinforcing stimulus (SR) is introduced that elicits a response (Relicited).
In the Pavlovian procedure SR most frequently follows a specific stimulus whereas in the operant procedure
SR most frequently follows a specific response (Rj) . Relicited typically goes unmeasured in the operant pro-
cedure. Note, however, that stimuli and responses necessarily precede the reinforcer in both procedures.

Skinner implicitly appreciated this point two processes may eventually be reduced to a sin-
in his demonstrations of superstitious condition- gle formula.”
ing. Superstitions of the first kind occurred when Behavioral Findings
the frequency of a response changed if it occurred Contiguity. The first variable identified
by chance before a reinforcing stimulus (Skinner, as critical to selection by reinforcement was the
1948; cf. Staddon & Simmelhag, 1971). Supersti- temporal relation between the reinforcer and the
tions of the second kind occurred when a rein- environment (the conditioned stimulus, or CS) in
forced response occurred by chance during a the Pavlovian procedure or the behavioral event
stimulus and the response then came under the (the operant) in the operant procedure. In the Pav-
control of that stimulus (Morse & Skinner, 1957). lovian procedure, an effective interval is typically
Viewed in this way, the distinction between Pav- described as no more than a few seconds (e.g.,
lovian and operant conditioning is procedural, Smith, Coleman, & Gormezano, 1969). The ef-
and not necessarily a difference in the condition- fect of varying the interval between the operant
ing process itself. Skinner was open to this possi- and the reinforcer is more difficult to analyze ex-
bility. However, he chose not to pursue the com- perimentally because the onset of the operant
monalities in the conditioning process in the two cannot be controlled and the events that intervene
procedures but, instead, to explore the profound between the operant and the reinforcer may in-
differences in the consequences of the procedural clude other instances of the operant, as in inter-
difference. In the Pavlovian procedures, an elic- mittent schedules of reinforcement. In addition,
ited response (the respondent) comes under the stimuli that function as conditioned reinforcers
control of arbitrary stimuli 2 whereas in the oper- may bridge the gap between the operant and the
ant procedure the entire behavioral repertoire of reinforcer (e.g., Catania, 1971; Lattal & Gleeson,
the organism potentially comes under the control 1990). In his initial operant procedure, Skinner
of arbitrary stimuli, thereby opening the path to sought to minimize these intrusions by immedi-
the emergence of complex behavior. Skinner’s ately following the operant by the “click” of a
recognition of the possible unity of the selection feeder, an auditory stimulus that had previously
process was acknowledged without dissent in his immediately preceded the delivery of food ac-
very early writings (Skinner, 1938, p. 111): “Hil- cording to a Pavlovian procedure (Skinner, 1938
gard (1937) ... points out that ... reinforcement is p. 53). A commitment to temporal contiguity was
essentially the same process in both [proce- also expressed in Ferster and Skinner’s treatment
dures]” and “Mowrer (1937) holds out that ... the of schedules of reinforcement: “The only contact
between [the schedule] and the organism occurs
2

at the moment (emphasis added) of reinforce- Wagner model has proven exceptionally fruitful
ment.... Under a given schedule of reinforcement, (Siegel & Allan, 1996, cf. Papini & Bitterman,
it can be shown that at the moment of reinforce- 1990), especially when implemented in real-time
ment a given set of stimuli will usually prevail. A computer simulations (e.g., Sutton & Barto,
schedule is simply a convenient way of arranging 1981; Donahoe et al, 1993). The model is silent,
this. (Ferster & Skinner, 1957, pp. 2-3). however, with respect to the biobehavioral mech-
The general conclusion is that a reinforc- anisms that implement it (Donahoe, 1984). The
ing stimulus produces changes in stimulus control physical events that are available to an organism
in the Pavlovian procedure and changes in re- on its first exposure to a reinforcing stimulus are
sponse frequency in the operant procedure only the stimulus itself and the responses that are elic-
over relatively brief intervals of time. 3 As Skin- ited by that stimulus. Subsequent research with
ner put it, “To say that a reinforcement is contin- the blocking procedure revealed that when the re-
gent upon a response may mean nothing more inforcing stimulus remained constant but,
than that it follows the response” (Skinner, 1948, through various means, the response elicited by
p. 168). the reinforcer was changed, conditioning oc-
Behavioral discrepancy. With the ad- curred to a stimulus that would have otherwise
vent of the blocking procedure, it became clear been blocked (Brandon, Betts, & Wagner, 1994;
that, although contiguity was necessary, it was Stickney & Donahoe, 1983). In short, blocking
not sufficient to produce selection by reinforce- was prevented when the unconditioned response
ment. As demonstrated initially with a Pavlovian (UR) was changed. Based on these findings, the
procedure (Kamin, 1968; 1969) and subsequently second requirement for selection by reinforce-
with an operant procedure (vom Saal & Jenkins, ment may be described as follows: The reinforc-
1970), something in addition to temporal contigu- ing stimulus must not only be contiguous but it
ity was required.4 Studies showed that if a stimu- must also evoke a change in the behavior that
lus or a response was followed by a putative rein- would otherwise occur at that moment (Donahoe,
forcer with a temporal relation known to be effec- Crowley, Millard, & Stickney, 1982; Donahoe,
tive for conditioning, but that the stimulus or re- Burgos, & Palmer, 1993). This is known as the
sponse was accompanied by another stimulus that behavioral-discrepancy requirement.
had previously preceded the same reinforcer, then The importance of the behavior evoked
the putative reinforcer would not be effective. by the reinforcing stimulus was further confirmed
Prior conditioning blocked the acquisition of a by other work demonstrating that the critical tem-
new environment-behavior (E-B) relation if the poral relation in the Pavlovian procedure was not
same reinforcer had previously occurred in that between the CS and the US as conventionally un-
context. Blocking has also been shown to occur derstood, but between the CS and the behavior
with conditioned reinforcers. Using a concurrent evoked by that stimulus, that is, the UR (Donahoe
schedule in which one response produced a visual & Vegas, 2004). Using the throat-movement
stimulus that had previously been paired with a preparation of the pigeon, water or air injected
reinforcer and a second response produced a dif- into the oral cavity of a restrained pigeon evokes
ferent visual stimulus that had been paired a UR of a sufficiently long latency (approxi-
equally often with the same reinforcer but only in mately 200 ms) and duration (approximately 2 s)
the presence of a previously conditioned stimu- to permit the temporal relation of the CS to the
lus, responding was maintained for only the first US and to the UR to be unconfounded. It was
response (Palmer, 1986). found that whether the CS occurred before the
What is the nature of the additional vari- US-UR (the standard forward-conditioning pro-
able? Associationist psychology proposed that a cedure), only between the CS and the UR, or only
given reinforcer would support only a fixed max- after the onsets of both the CS and UR—but dur-
imum association value and that, because prior ing the UR, the level of conditioned responding
conditioning had already attained that maximum to the CS was the same when measured on CS-
value, no further increases in associative strength alone test trials. (Appropriate controls were insti-
were possible (Rescorla & Wagner, 1972). The tuted to rule out alternative interpretations such as
formalization of this proposal in the Rescorla- sensitization.) Thus the CS-UR relation, not the
3

CS-US relation, is the critical temporal relation & Stumpf, 2015). As a result, the emphasis is
for conditioning in the classical procedure. The generally upon the net effects of the neuroana-
apparent importance of the CS-US relation was tomical and cellular components (neurons) of
based largely on procedures in which the US elic- these systems and not the intra-cellular processes
ited a short latency, brief duration UR, thereby themselves. A neural system refers to those brain
confounding the CS-US and CS-UR temporal re- structures and their interconnections that are most
lations. relevant to the behavior under consideration. The
These findings provide an interpretation neural system of concern here implements rein-
of a number of otherwise problematic findings forcement in the operant-conditioning procedure.
(Donahoe & Vegas,, 2004). Examples include: Figure 2 is a diagram of a lateral view of the hu-
(a) how the same CS-US temporal relation condi- man cerebral cortex showing (with dashed lines)
tions some responses but not others when the URs the subcortical structures and interconnections
differ in latency or duration (e.g., Betts, Brandon, that are central to implementing reinforcement.
& Wagner, 1996), (b) how a backward condition- The general flow of neural activity begins with
ing arrangement (US-CS) conditions long-dura- stimulation ofthe senses, such as vision and audi-
tion URs, such as autonomic responses, but not tion. This stimulation activates neurons in the pri-
short-duration URs, such as eye blinks (e.g., Al- mary sensory cortices subsequent to interactions
bert & Ayres, 1997; McNish, Betts, Brandon, & involving various subcortical structures (not
Wagner, 1997), and (c) how a backward US-CS shown). Activity in primary sensory neurons is
arrangement conditions components of a tempo- propagated to sensory-association cortex whose
rally extended sequence of responses that are polymodal neurons respond to various combina-
contiguous with the CS (e.g., Silva, Timberlake, tions of their inputs from neurons in primary sen-
& Cevik, 1998). The behavioral-discrepancy re- sory and sensory-association cortices. Sensory
quirement is also consistent with the a priori con- and polymodal neurons activate, in turn, neurons
jecture that the ancestral selecting environment in prefrontal and motor cortex (among others),
(that is, natural selection) would more likely pro- with the latter leading most directly to behavior
duce a conditioning process that was sensitive to via the striatum and various other subcortical and
the expressed behavior of an organism than to its spinal structures (not shown). Neurons in pre-
mere perception of a stimulus. Thomas Huxley’s frontal cortex, which includes the adjacent sup-
words resonate: “The great end of life is not plementary motor and premotor areas, coordinate
thought but action.” 5 The view that the same con- neural activity in the motor cortex. This account
ditioning process, requiring both contiguity and of the neural systems mediating E-B relations is
discrepancy, is engaged by the Pavlovian and op- simplified and does not reflect other, often recip-
erant procedures is summarized by the unified re- rocal, pathways between structures—some of
inforcement principle (Donahoe et al, 1993). which are considered in a later section. The ac-
Neuroscientific Findings count is suitable for present purposes however.
The present discussion of neural mecha-
nisms focuses on the neural systems that imple-
ment the reinforcement process (cf. Kirk, Babtie,
2

NAC

.
Figure 2. Lateral view of the human cerebral cortex showing the major cortical regions and, using
dashed lines, the subcortical structures—nucleus accumbens (NAC) and ventral tegmental area (VTA)—
and pathways central to the neural mechanisms of selection by reinforcement. (The divisions of the cortex
designated in the diagram are for heuristic purposes only. For example, the region designated primary
sensory cortex is largely the primary visual cortex and the region designated sensory-association cortex
also includes the primary auditory cortex of the temporal lobe.)

Given the many neurons in the human brain— the reinforcement process must not only affect
perhaps 1012 , the many potential connections to a the pathways of that specific reinforced E-B rela-
single neuron—perhaps 103 in the case of motor tion but also potentially a wide range of other
neurons, and the “spontaneous” activity of many pathways that might mediate other E-B relations
cortical neurons, natural selection faced a formi- under different contingencies of reinforcement.
dable challenge: What mechanism could be de- Such a mechanism is required if almost any stim-
vised whereby the strengths of connections be- ulus is to potentially control almost any operant.
tween neurons mediating a specific reinforced E- Dopamine and the selection of path-
B relation were strengthened? As with any selec- ways mediating reinforced E-B relations. The
tion process, the challenge cannot be not met on neural capability for widespread effects of rein-
a single occasion and is fallible (Donahoe, 2003; forcement is accomplished through the liberation
Palmer & Donahoe, 1992). (This challenge is and subsequent diffusion of the neuromodulator
known as the “assignment-of-credit” problem in dopamine (DA) along projections from cells in
neural-network research.) The mechanism that the ventral tegmental area (VTA) to the prefrontal
evolved has the effect of disproportionately, but and motor cortices (among others). The diffusion
not exclusively, strengthening the “right” connec- of DA allows a reinforcer to affect a wide range
tions. (Personification of selection processes— of synapses and, thereby a wide range of potential
whether natural selection or selection by rein- E-B relations mediated by the pathways involv-
forcement—is an expository device only, of ing those synapses.6 See Figure 2. For purposes
course.) of the current discussion, the neural inputs to the
What is the mechanism that permits se- prefrontal and motor cortices from sensory areas
lection by reinforcement of a wide range of spe- are presumed to be sufficient to control the rein-
cific E-B relations? When a response is followed forced behavior. This matter is considered further
by a reinforcer it must be true that pathways suf- in a later section.
ficient to enable the reinforced E-B relation are DA is generally designated a neuromod-
among the many that are active at that moment. ulator rather than a neurotransmitter because it al-
However, the neural mechanisms implementing ters the effect of other neurotransmitters on the
2

functioning of neurons. One may speculate that cf. Williams, 1994a, b). Although conditioned re-
the central role of DA in reinforcement should not inforcers may acquire additional functions, par-
be completely surprising given that DA also plays ticularly in free-operant procedures (e.g., Shahan,
an important role in the digestive process where 2010), neuroscience indicates that both uncondi-
it governs gut motility (Li, Schmauss, Cuenca, tioned and conditioned reinforcers cause DA to
Ratcliffe, & Gershon, 2006). This leads to the be liberated by VTA neurons. The route whereby
conjecture that the digestion of nutrients and re- VTA neurons are activated by conditioned rein-
inforcement were intertwined in evolutionary his- forcers differs however. Whereas unconditioned
tory. The neural activity initiated by receptors reinforcers activate VTA neurons by pathways
stimulated by taste and smell and by sexual activ- originating from receptors such as those for taste
ity activate DA neurons via pathways traversing or sexual stimulation (e.g., Smith, Liu, & Vogt,
various intermediate subcortical structures. 1996; Balfour, Yu, & Coolen, 2004), conditioned
Drugs of abuse also typically activate DA neu- reinforcers activate VTA neurons by a less direct
rons (e.g., Volkow, Fowler, & Wang, 2007; Wise, route involving the prefrontal cortex (e.g., Pears,
2002). The widely projecting outputs (axons) of Parkinson, Hopewell, Everitt, & Roberts, 2003;
VTA neurons liberate DA that diffuses from var- Wilson & Bowman, 2004)
icosities along their lengths. DA molecules re- As shown in Figure 2, the neural activity
main present for several seconds before being de- initiated by SD stimulates neurons in prefrontal
graded (Yagishita, Hayashi-Takagi, Ellis-Davies, cortex and pathways originating from prefrontal
Urakubo, Ishii, & Kasai1, 2014). The relatively neurons activate cells in the nucleus accumbens
short-lived presence of DA is consistent with the (NAC) that, in turn, cause VTA-DA neurons to
contiguity requirement. DA also plays a critical be activated. Thus both conditioned and uncondi-
role in the discrepancy requirement as described tioned reinforcers ultimately engage the same
shortly. VTA reinforcement system, but by different
Figure 3 indicates the frequency of firing routes.7 For experienced organisms in which
of VTA-DA neurons in our fellow primate, the many E-B relations have been previously ac-
crab-eating monkey. The task was a differential quired, the environment offers many opportuni-
conditioning procedure in which pressing one of ties for engaging the neural mechanisms of con-
two levers was reinforced with apple juice de- ditioned reinforcement. In this way, activities that
pending on which of two spatially separated require temporally extended sequences of behav-
lights was illuminated (Schultz, Apicella, & ior, such as problem solving, are maintained by a
Ljungberg, 1993; Schultz, Dayan, & Montague, relatively continuous stream of conditioned rein-
1997). The top panel (A) shows the frequency of forcers, which increases as the target behavior is
firing of DA neurons when the reinforcing stimu- more closely approximated (Donahoe & Palmer,
lus (SR) was presented at the outset of condition- 1993, p. 285 ff.).8
ing. Note that the baseline level of activity of DA Panel C of Figure 3 reveals an additional
neurons increased abruptly shortly after the rein- finding that relates to the behavioral-discrepancy
forcing stimulus occurred. Transition to the requirement. The lower panel shows the activity
“bursting” mode is required for DA to be liber- of DA neurons on a trial in which SD was pre-
ated along the axons of VTA neurons (Grace & sented, but SR was omitted. Note that DA activity
Bunney, 1984; Grace, Floresco, Goto, & Lodge, was actually inhibited shortly after the time when
2007; Johnson, Seutin, & North, 1992). The mid- the reinforcer was otherwise scheduled to occur.
9
dle panel (B) shows the frequency of firing of DA If another stimulus had accompanied this SD in
neurons after conditioning when the discrimina- a blocking design, then the reinforcer would not
tive stimulus (SD) was presented and a correct re- have produced an increase in DA activity and
sponse was followed by SR. The burst of firing would not have become a discriminative stimu-
now occurred at the onset of SD, not SR. This is lus. The inhibition of VTA neurons following the
consistent with the long-standing behavioral find- time at which the reinforcer previously occurred
ing that discriminative stimuli also acquire a con- is the neural basis of the discrepancy requirement
ditioned reinforcing function (Dinsmoor, 1950; (Burgos & Donahoe, 2016; Donahoe, Burgos, &
3

Palmer, 1993; Waelti, Dickinson, & Schultz,


2001).10

Figure 3. The frequency of firing of dopaminergic (DA) neurons in the ventral tegmental area (VTA) during
a differential operant-conditioning procedure. Panel A shows the DA response to the reinforcing stimulus
(SR) at the outset of conditioning. Panel B shows the DA response to the discriminative stimulus (SD) and
SR on a reinforced trial after conditioning. Panel C shows the DA response to SD on a trial after condi-
tioning in which the reinforcer was omitted. See the text for additional information, (Data from Schultz,
1997; Schultz, Apicella, & Ljungberg, 1993)

Dopamine and the cellular mecha- to an increased ability (potentiation) of receptors


nisms of selection (LTP). The liberation of DA on a neuron to be activated by a neurotransmitter
from VTA neurons by reinforcers during condi- released from the axons of other neurons. Neuro-
tioning comports well with what is known from transmitters diffuse across the synapse between
behavioral research about unconditioned rein- presynaptic and postsynaptic neurons. The neu-
forcement, conditioned reinforcement, and block- rotransmitter released by presynaptic neurons
ing. In addition, the diffusion of DA in prefrontal may either increase (excite) or decrease (inhibit)
and motor cortex (Yagishita, Hayashi-Takagi, El- the activity of postsynaptic neurons through its
lis-Davies, Urakubo, Ishii, & Kasai1, 2014) al- effect on the receptors to which the neurotrans-
lows DA to simultaneously affect the functioning mitter binds on the postsynaptic neuron. Recep-
of many neurons along the diverse pathways re- tors are typically located within the membrane of
quired to implement a wide range of reinforced the dendrites of postsynaptic neurons. Neurons
E-B relations. What is the cellular process that en- normally maintain a negative resting potential
ables this function? across the cell membrane. This potential becomes
The process that affects the strengths of more positive when neurons are activated and
connections along pathways mediating E-B rela- more negative when they are inhibited. The mem-
tions is long-term potentiation (LTP) (Bliss & brane potential is governed by the net influx of
Lømø, 1973). Before describing LTP, several negative and positive ions across the cell mem-
technical terms are briefly reviewed. LTP refers brane
.
2

Figure 4. An axon terminal of a presynaptic neuron and a dendritic spine of a postsynaptic neuron.
The excitatory neurotransmitter glutamate (Glu) released by the presynaptic neuron binds to two types of
postsynaptic Glu receptors—the fast-acting AMPA receptor and the voltage-sensitive NMDA receptor.
When the postsynaptic neuron is sufficiently depolarized by the action of Glu, the NMDA receptor allows
calcium ions (Ca2+) to enter the cell. If dopamine (DA) is simultaneously bound to DA receptors, then a
series of intracellular processes (second messengers) migrate to genetic material in the cytoplasm and/or
nucleus of the postsynaptic cell stimulating protein synthesis. The synthesized proteins diffuse down the
dendrite where they act upon those AMPA receptors to which Glu was previously bound. This produces
long-lasting changes in those Glu receptors that make them more responsive to Glu on subsequent occa-
sions. See the text for additional information.

Figure 4 is a schematic representation of enough to cause the postsynaptic neuron to “fire,”


an axon terminal of a presynaptic neuron and a those specific AMPA receptors acquire a molec-
juxtaposed spine on a dendrite of a postsynaptic ular “tag” that enhances their response to Glu for
neuron. Single neurons have many such axon ter- perhaps an hour or so (Frey, 1997; Frey & Morris,
minals and dendritic spines. Molecules of the 1997). This brief enhancement is called early
neurotransmitter glutamate (Glu) are released LTP. When the concerted action of AMPA recep-
through the cell membrane of the presynaptic tors depolarizes the postsynaptic neuron suffi-
neuron and diffuse into the synapse. (Glu is the ciently, additional effects occur: If Glu has also
primary excitatory neurotransmitter in the brain.) bound to NMDA receptors—which permit cal-
Molecules of Glu then bind to two types of Glu cium ions (Ca2+) to enter the cell—and if DA is
receptors on the postsynaptic neuron—AMPA re- present, then a series of intracellular events are
ceptors and NMDA receptors. (The abbreviations initiated that result in the synthesis of new pro-
stand for the Glu analogs, α-amino-3-hydroxy-5- teins (see Figure 4). These proteins—which re-
methyl-4-isoxazolepropionic acid and N-methyl- quire time to synthesize--diffuse down the den-
D-aspartate, respectively, that selectively bind to drite of the postsynaptic cell and produce a long-
the two types of Glu receptors.) If Glu binds to lasting enhancement in the response to Glu to
AMPA receptors, entry of positive ions into the only the previously tagged AMPA receptors. This
cell increases and the membrane potential of the produces late LTP, which is a long-lasting change
neuron becomes less negative. When the mem- in the ability of Glu to activate specific receptors
brane potential becomes less negative, even if not on the postsynaptic neuron. 11 In this way, DA
2

“forges” pathways through the frontal lobes from which subserve conditioned reinforcement. If DA
the inputs arriving from sensory and sensory-as- is not present when pre- and postsynaptic neurons
sociation cortex to motor neurons that mediate are coactive, then the enhanced response of
observable behavior (Reynolds, Hyland, & Wick- postsynaptic Glu receptors declines, producing
ens, 2001).12 The finding that DA plays an espe- long-term depression (LTD). Note that at each of
cially important role in the acquisition of effortful the various levels of analysis—the behavioral, the
responses or long sequences of responses may re- neural, and the cellular—the mode of explanation
flect that such tasks requite the potentiation of re- follows the Darwinian model of selectionism:
ceptors on large numbers of neurons (Salamone, That is, reinforcers select among E-B relations;
Correa, Nunes, Randall, & Pardo, 2013, cf. Geor- reinforcer-evoked DA selects among neural path-
gopoulos, Schwartz, & Kettner, 1986; Sanes & ways that mediate the E-B relations, and DA-in-
Donoghue, 2000). DA also potentiates Glu-ergic stigated protein synthesis selects among tagged
receptors on postsynaptic neurons in the NAC, post-synaptic receptors (Donahoe, 1997).13

Stimulus Control environment. The behavioral effects of differen-


Behavioral Findings tial conditioning are most apparent with experi-
The vast range of phenomena encompassed by mental procedures in which different operants
the field of stimulus control was foreshadowed are conditioned to the different environments.
by Skinner’s early comment that “discriminative For example, lever presses of two different dura-
stimuli are practically inevitable after condition- tions were differentially reinforced in the pres-
ing” (Skinner, 1937, p. 273; cf. Donahoe, ence of two light intensities and tests of stimulus
Palmer, & Burgos, 1997). The effects of only generalization were then conducted with various
two sets of phenomena are considered here— intermediate light intensities (Crowley, 1979).
discrimination formation and equivalence clas- The result was that responding during generali-
ses. Other areas are omitted, notably concept zation tests with the other light intensities was
formation initiated by the pioneering work of confined to the two previously reinforced re-
Richard Herrnstein (Herrnstein, Loveland, & sponse durations, their proportions varying with
Cable, 1976) and continuing as an active area of the proximity of the test intensity to the two
current research (e.g. Zentall, Wasserman, & Ur- training intensities. Findings of similar import
cuioli, 2013). have been obtained with a variety of other pro-
Discrimination formation. If a rein- cedures (e.g., Migler, 1966; Migler & Millenson,
forcer occurs in an environment, whether in a 1969; Scheuerman, Wildemann, & Holland,
Pavlovian or an operant procedure (e.g., Gyn- 1978; Wildemann & Holland, 1972). The gen-
ther, 1957 and Guttman & Kalish, 1956, respec- eral conclusion from this work is that new envi-
tively), then behavior comes under the control of ronments do not control new behavior but new
not only that environment but also other envi- mixtures of previously conditioned behavior
ronments to the extent that they share stimuli in (Bickel, & Etzel, 1985; Donahoe & Wessells,
common. This is the so-called common-ele- 1980, pp. 194-196). Any “creativity” arises from
ments account of stimulus generalization formal- variation in the environment, not from within the
ized by Skinner’s former student William Estes organism itself. As E-B relations accumulate,
(Estes, Burke, & Atkinson, 1957). If a differen- the variation increases upon which future selec-
tial conditioning procedure is instituted, then the tions by reinforcement may act, with the emer-
controlling stimuli are confined to those that are gent possibility of ever more complex E-B rela-
specific to the environment in which the rein- tions.
forcer occurred. Stimuli that are not shared with Equivalence classes. Equivalence clas-
the conditioning environment become the occa- ses (Sidman & Tailby, 1982) are typically studied
sion for whatever previously conditioned and using matching-to-sample procedures with multi-
unconditioned responses are supported by that
2

ple, arbitrary (symbolic) conditional discrimina- forcement on the environmental guidance of be-
tions. For example, conditional discriminations havior transcend the specific E-B relations that
using A-B and A-D are trained where the pairs of were explicitly reinforced, even beyond those of
letters stand for different sets of sample and com- stimulus generalization. The controlling stimuli
parison stimuli, respectively. After these condi- in the training environment (for example, A) do
tional discriminations have been acquired, unre- not have stimulus elements in common with those
inforced probe trials are occasionally introduced of the test environment (for example, D).
using previously untrained combinations of sam- Neuroscientific Findings
ple and comparison stimuli, for example D-B. An When the discriminative stimuli in con-
equivalence class is said to form if an organism ditioning and generalization-test situations are
reliably responds to the comparison stimulus as- well defined in physical terms, the findings are
sociated with a sample stimulus that was previ- relatively straightforward. For example, rabbits
ously trained with a different—but correspond- trained using a classical-conditioning procedure
ing—comparison stimulus. Equivalence classes to respond to a particular frequency of an auditory
meet the three criteria of mathematical equiva- stimulus were then tested with other frequencies.
lence—identity (A = A), symmetry (if A = B, Generalization tests included measurement of the
then B = A), and transitivity (if A = B and B = C, responses of neurons in primary auditory cortex
then A = C). In the preceding example of a B-D as well as behavioral responses. It was found that
probe trial, responding to the appropriate D com- the greater the overlap between the number of
parison stimulus requires both symmetry (only A- cells firing to the test stimuli and those firing to
B was trained, not B-A) and transitivity (B-A-D). the training stimulus, the greater the behavioral
Equivalence classes have important implications response (Thompson, 1965). That is, responding
for the interpretation of complex behavior be- occurred to the extent that the conditioning and
cause they demonstrate that the effects of rein- test environments activated neural elements in
common.

Figure 5. Lateral view of the human cerebral cortex showing the major cortical regions and sub-
cortical structures and pathways (indicated by dashed lines) that play an important role in the selection of
environment-environment relations.

But what of equivalence-class proce- conditioning environment share no readily iden-


dures in which the discriminated stimuli in the tifiable physical similarity to those in the testing
2

environment? To address this question it is nec- unreinforced pairs of stimuli, although they had
essary to consider additional neural systems, spe- occurred together equally often (Fujimichi, Naya,
cifically the relation between sensory-association & Miyashita, 2005; Sakai & Miyashita, 1991).
(S-A) cortex and the hippocampus. Figure 5 de- Neurons that responded equally to either member
picts that relation. As previously noted, neurons of a reinforced pair were designated “pair-coded
in S-A cortex receive their inputs from primary neurons.” For such neurons, the two stimuli were
sensory and other S-A neurons. Thus the activa- literally equivalent. The effect of reinforcement
tion of S-A neurons reflects the co-occurrence of on the acquisition of E-E relations is facilitated
multiple environmental inputs. The axons of S-A by DA-ergic projections from the VTA to CA1
neurons give rise to a multi-synaptic pathway neurons (see Figure 5), the output neurons of the
(among others) that innervates the hippocampus, hippocampus to SA cortex (Duncan, Ketz, Inati,
a subcortical structure located beneath the tem- & Davachi, 2012; Gasbarri, Verney, Innocenzi,
poral lobes. After interactions involving other Campana, & Pacitti, 1994; Li, Cullen, Anwyl, &
neurons within the hippocampus, neurons in the Rown, 2003; Swanson & Kohler, 1986). As al-
CA1 region of the hippocampus give rise to a ready noted, without the contribution of DA, re-
multi-synaptic pathway that projects back to S-A ceptors on CA1 neurons would be potentiated for
cortex. (CA is an abbreviation for the Latin phase, only a few hours (early LTP). With the contribu-
cornu Ammons, or horn of Ammons, that refers to tion of DA, potentiation of CA1 and S-A neurons
the shape of the hippocampus.) It has been pro- may endure indefinitely until “over-written” by
posed that the output of the hippocampus from other E-E relations.
the CA1 region projects back to the very regions With nonhuman organisms, behavioral
and/or neurons of the S-A cortex that gave rise to research provides clear evidence of both identity
the hippocampal inputs (Amaral, 1987). The syn- (generalized identify matching) and transitivity,
chronized activation of S-A neurons by their sen- but little or no compelling evidence of symmetry
sory inputs and, very shortly thereafter, by their (but see Kastak, Schusterman, & Kastak, 2001;
hippocampal inputs depolarizes those S-A neu- Urcuioli, 2008). In humans, strong evidence ex-
rons and, in this way, potentiates specific recep- ists for all three components of equivalence clas-
tors on the neuron (cf. Strycker, 1986). These re- ses (Lionello-DeNolf, 2009). What is responsible
ceptors are then tagged and potentially undergo for this apparent inter-species difference? Re-
LTP. 14 The net effect of this process produces S- search in neuroscience offers some clues. In pri-
A neurons that respond to their multi-modal in- mates, unlike other mammalian orders, the bidi-
puts. As a simplistic example, if the inputs to a rectional pathways that loop between the S-A
given S-A neuron arise from the co-occurrence of cortex and hippocampus arise exclusively from
a tone and a light, then that S-A neuron becomes and synapse upon S-A neurons (Amaral, 1987).
functionally a tone-light neuron. The net result is Organisms that have a lesser capacity for acquir-
that S-A neurons become sensitive to the con- ing pair-coded neurons—or lack homologous
junctions of spatio-temporally contiguous, envi- neural structures that achieve the same end—may
ronmental events, or environment-environment have a diminished or absent capacity to form the
(E-E) relations. pair-coded neurons that support symmetric dis-
The acquisition of E-E relations between criminations (cf. Sidman, Rauzin, Lazar, Cun-
sample and comparison stimuli has been pro- ningham, Tailby, & Carrigan, 2013). For organ-
posed to underlie the formation of equivalence re- isms whose neuroanatomy permits the formation
lations (Donahoe & Palmer, 1993, p. 148). Re- of pair-coded neurons, the development of neu-
cent neuroscientific evidence supports this pro- rons that respond to multi-sensory inputs should
posal. Rhesus monkeys that had acquired an arbi- be the norm if the contingencies of reinforcement
trary matching-to-sample task using computer- foster their selection.15As Sidman proposed,
generated, abstract visual stimuli developed sin- equivalence-class formation may be a normative
gle neurons in the S-A cortex that responded effect of reinforcement (Sidman, 2000).
equally to either the sample or the comparison
stimulus of pairs to which correct responding had
been reinforced. Such neurons were not found for
3

Memory acquaintances controls covert behavior (mne-


Behavioral Findings monic behavior) that in the past has been effec-
In a behaviorally informed approach to tive in producing covert stimuli sufficient for the
memory, as the field is conventionally known, emission of the overt response—in this case ut-
phenomena may be divided into two cases—re- tering the person’s name (Donahoe & Palmer,
minding and remembering (Donahoe & Palmer 1993, pp. 275-277). In brief, remembering re-
1993; Palmer 1991). Reminding refers to those quires engaging in behavior, whether overt or
cases in which some behavior is scheduled for re- covert, that produces stimuli that are sufficient to
inforcement and the behavior takes place in an remind oneself, so to speak, of the behavior
environment in which that behavior has been pre- needed to produce the reinforcer. However, note
viously reinforced. As an example, when asked that when the reminding behavior is covert, this
“Who was the President during the Civil War,” is an interpretation of memory not an experi-
the response “Lincoln” is scheduled for reinforce- mental analysis (cf. Skinner, 1974, p. 194;
ment and has almost certainly been previously re- Palmer, 2011). That is, the account is an extrapo-
inforced in that verbal context. This is simply a lation from behavioral observations that draws
straight-forward example of stimulus control. Us- upon only processes that have been analyzed ex-
ing language more congenial to the field of perimentally in previous contexts and that are
memory, the environment may be said to remind highly likely to have occurred in the current con-
the organism of the appropriate behavior. By con- text.
trast, remembering refers to cases in which some Neuroscientific Findings.
behavior is scheduled for reinforcement but the The behavioral interpretation of remem-
contemporaneous environment does not include a bering requires that the behavior scheduled for re-
sufficient number of the stimuli to which the tar- inforcement is already within the repertoire of the
get response was previously conditioned. Readily organism, but that stimuli adequate to control that
observable behavior can sometimes produce the behavior are not present. For example, the ques-
requisite controlling stimuli as when one consults tion “Who was the Union General who won the
a smartphone to find the telephone number of Battle of Gettysburg?” may not be sufficient to
AAA after one’s car has broken-down. Consult- control the verbal response “Meade.” However, a
ing a smartphone to find a telephone number and response of the desired form can be emitted under
then keying in that number have both been previ- other circumstances, for example, as a textual re-
ously reinforced. A greater challenge is presented sponse to the printed word “Meade.” A biobehav-
by those cases of remembering in which the be- ioral interpretation proposes that remembering is
havior required to produce the stimuli needed to accomplished at the neural level by exploiting the
control the appropriate response is covert. For ex- extensive pathways (tracts) that course back and
ample, upon seeing the approach of casual ac- forth between the S-A and prefrontal cortex (Do-
quaintances, you struggle to recall their names. nahoe & Palmer, 1993; cf. Fuster, 2015). Re-
You may then subvocally sound out the letters of peated cycles of these neural interactions can po-
the alphabet hoping to hit upon letters that, to- tentially activate motor neurons that produce the
gether with the appearance of their faces, prompt target behavior.
the correct names. The sight of the approaching
2

Figure 6. Schematic view of the upper surface of the cerebral cortex showing the left and right
hemispheres. The right hemisphere is deflected back to reveal the corpus callosum (black), the large fiber
bundle that interconnects neurons in the two hemispheres. The curved block arrow indicates the tracts that
connect the sensory-association (S-A) area of the right hemisphere to the prefrontal cortex of that same
hemisphere. (Comparable tracts exist within the left hemisphere but are not shown.) The block arrows
indicate other tracts between the two hemispheres that course through the corpus callosum. The portion of
the corpus callosum that was first severed is indicated by A; the portion that was later severed is indicated
by B (Tomita , Ohbayashi, Nakahara, Hasegawa, & Miyashita, 1999). See the text for additional infor-
mation.

A behavioral interpretation of remember- coursing through the posterior portion of the cor-
ing has received support from the experimental pus callosum (denoted by A in Figure 6). After
analysis of neuroscience. To appreciate these the subjects had acquired the matching-to-sample
findings, additional neuroanatomical information task to a high level of proficiency, the A portion
is required. Figure 6 is a schematic view looking of the corpus callosum was severed. Despite the
down on the upper surface of the cerebral cortex. absence of connections between the S-A cortices
The right hemisphere has been deflected back to of the two hemispheres, performance remained at
reveal the large fiber bundle (corpus callosum, a high level. Performance was maintained by two
shown in black) that interconnects neurons in the sets of alternate pathways: One pathway (indi-
two hemispheres. A monkey was trained with cated by a curved block arrow) connected neu-
multiple, abstract, conditional-discriminations rons in the S-A cortex to neurons in the prefrontal
using a matching-to-sample task in which visual cortex of the right hemisphere. The second set of
stimuli were presented in such a way that sample pathways (shown as a single-headed, right-an-
stimuli activated only neurons in the left visual gled, block arrow in Figure 6) connected neurons
cortex. The corresponding two comparison stim- in the prefrontal cortex of the right hemisphere to
uli activated only neurons in the right visual cor- neurons in the SA area of the left hemisphere via
tex (Tomita, Ohbayashi, Nakahara, Hasegawa, & the anterior portion of the corpus callosum (de-
Miyashita, 1999). In an intact brain, stimuli that noted B in Figure 6). In short, feedback from ac-
directly activate neurons in only one hemisphere tivity in prefrontal cortex was sufficient to control
can also activate neurons in the contralateral discriminative performance. This feedback is the
hemisphere by means of pathways (shown as an neural counterpart of mnemonic behavior pro-
inverted V-shaped, double-headed, block arrow)
2

posed in a behavioral interpretation of remember- addition, the same neural systems provide an in-
ing (Donahoe & Palmer, 1994): The test environ- tegrated account of unconditioned and condi-
ment (the sample stimulus) was not sufficient to tioned reinforcement, with the latter forming a
control responding to the reinforced comparison critical element in the emergence of complex be-
stimulus, but it instigated activity in the prefrontal havior. Our knowledge of the neural mechanisms
cortex that was sufficient to appropriately control of reinforcement remains incomplete, of course,
responding. When the anterior portion of the cor- particularly with regard to a detailed account of
pus callosum was severed (indicated by B in Fig- the interactions between neurons in prefrontal
ure 6), performance returned to chance levels and cortex, NAC, and VTA, but the discrepancy re-
could not be recovered. The experiment demon- quirement arising from behavioral research con-
strates that neural activity that does not evoke re- tinues to constrain the search for these mecha-
sponses at the behavioral level of observation can nisms.
nevertheless control behavior when the contem- For more complex behavior such as that
porary environment fails to provide stimuli in observed in the study of equivalence classes and
whose presence the response was previously re- memory, behavioral research can offer verbal in-
inforced. The behaviorally based distinction be- terpretations (e.g., Donahoe & Palmer, 1994, pp.
tween reminding and remembering is thus con- 139ff, 223ff) that draw upon observations of phe-
sistent with independent convergent evidence nomena that have been subjected to experimental
from neuroscience where the two classes of analysis in other contexts: The understanding of
memory are known as bottom-up vs. top-down equivalence classes can appeal to what is known
processing, respectively (Fujimichi et al, 2005). about joint stimulus control when conditioning
13
In humans, covert stimuli—such as those pro- serial and compound discriminative stimuli; the
duced by subvocal verbal behavior that occurred study of memory can appeal to covert behavior
at the time of acquisition—can facilitate these that is consistent with what is known about overt
neural processes if they occur later during re- behavior in otherwise comparable situations. Be
membering, Verbal behavior—whether overt or that as it may, neuroscience offers the possibility
covert—is not necessary for remembering, how- of supplementing these behavioral 15interpreta-
ever, as shown by the aforementioned findings tions with experimental analyses at the neural
with monkeys. level of observation: Pair-coded neurons provide
a mechanism whereby stimuli that bear no physi-
Conclusions cal similarity to one another become equivalent at
These findings and their implications are the neural level and feedback from activity in pre-
but a few examples illustrating the complemen- frontal cortex to sensory-association cortex pro-
tary nature of behavior analysis and neurosci- vides a mechanism whereby memories occur un-
ence. Behavioral research indicates that contigu- der circumstances in which the contemporaneous
ity and discrepancy are required for selection by environment bears no specific similarity to the
reinforcement, and neuroscience is well on its environment in which the behavior was originally
way to identifying the physiological mechanisms acquired.
that implement these requirements. The under- A strong case can be made that the inte-
standing of reinforcement arising from behav- gration of behavior analysis and neuroscience
ioral research constrains the search for its neural portends a biobehavioral science whose implica-
mechanisms to those that are capable of affecting tions for understanding complex behavior, in-
the strength of a wide range of E-B relations cluding human behavior, are as profound as the
while simultaneously confining the cumulative earlier integration of the sciences of heredity and
effects of selection to the particular class of E-B genetics for understanding complex structure.
relations that precede the reinforcer. The widely Skinner explicitly acknowledged “the spatial gap
projecting DA system from the VTA to the between behavior and the variables of which it is
frontal lobes accommodates the potentially wide- a function” and argued that this gap “can be filled
spread effects of selection by reinforcement; the only by neuroscience, and the sooner… the bet
specificity of LTP to particular postsynaptic re- ter. Now seems a propitious time to vigorously
ceptors satisfies the specificity requirement. In pursue that goal (Ortu & Vaidya, 2016).
1

References

Aggarwai, M., Hyland,B. I., & Wickens,J. R. (2012). Neural Brandon, S. E., Betts, S. L., & Wagner, A. R. (1994). Discrimi-
control of dopamine transmission: Implications for rein- nated lateralized eyeblink conditioning in the rabbit: An ex-
forcement learning. European Journal of Neuroscience, 35, perimental context for separating specific and general asso-
1115-1123. ciative influences. Journal of Experimental Psychology: An-
Albert, M. & Ayres, J. J. (1997). One-trial simultaneous and imal Behavior Processes. 20, 292-307.
backward excitatory fear conditioning in rats: Lick suppres- Burdakov, D., Liss, B., & Ashcroft, F. M. (2003). Orexin excites
sion, freezing, and rearing to CS compounds and their ele- GABAergic neurons of the arcuate nucleus by activating the
ments. Learning & Behavior, 25, 210-220. sodium–calcium exchanger. Journal of Neuroscience, 23,
Amaral, D. G. (1987). Memory: Anatomical organization of can- 4957-4951.
didate brain regions. In V. B. Mountcastle (Ed.), Handbook Burgos, J. E. (2005). Theoretical note: The C/T ratio in artificial
of physiology. Section I, The nervous system (Vol. V. Higher neural networks. Behavioural Processes, 69, 249-256.
functions of the brain. (pp. 211-294). Washington, D. C.: Burgos, J. E. & Donahoe, J. W. (2016). Unified principle of re-
Mount..
American Physiological Society. inforcement in a neural-network model: Reply to N. T. Cal-
Arntzen, E. (2006). Delayed matching to sample and stimulus vin and J. J. McDowell. Behavioural Processes, 126, 46-54.
equivalence: Probability of responding in accord with Calvin, O. L. & McDowell, J. J. (2016). Extending unified-the-
equivalence as a function of different delays. The ory-of-reinforcement neural networks to steady-state oper-
Psychological Record, 56, 135–167.. ant behavior. Behavioural Processes,127, 52-61.
Arntzen, E., Nartey, R. K., & Fields, L. (2015). Enhancing Catania, A. C. (1987). Some Darwinian lessons for behavior
responding in accordance with stimulus equivalence by the analysis: A review of Bowler’s The eclipse of Darwinism.
delayed and relational properties of meaningful stimuli. Journal of the Experimental Analysis of Behavior, 47, 249-
Journal of the Experimental Analysis of Behavior, 103, 524– 257.
541. Catania, A. C. (1971). Reinforcement schedules: The role of re-
Arntzen, E., Norbom, A., & Fields, L. (2015). Sorting: An sponses preceding the one that produces the reinforcer. Jour-
alternative measure of class formation? The Psychological nal of the Experimental Analysis of Behavior, 15, 271-287.
Record, 65, 615–625. Chambers, K. C. (1990). A neural model for conditioned taste
Balfour, M. E., Yu, L., & Coolen, L. M. (2004), Sexual behavior aversions. Annual Review of Neuroscience, 13, 373-385.
and sex-associated environmental cues activate the meso- Corbit, L. H. & Balleine, B. W. (2011). The general and out-
limbic system in male rats. Neuropsychopharmacology. 29, come-specific forms of Pavlovian-instrumental transfer are
718-30. differently mediated by the nucleus accumbens core and
Bauer, E. P., Schafe, G. E., & LeDoux, J. E. (2002). NMDA re- shell. Journal of Neuroscience, 31, 11786-11794.
ceptors and L-type voltage-gated calcium channels contrib- Creed, M. C., Ntamati, N. R., & Tan, K. R. (2014). VTA GABA
ute to long-term potentiation and different components of neurons modulate specific learning behaviors through the
fear memory formation in the lateral amygdala. Journal of control of dopamaine and cholinergic systems. Frontiers in
Neuroscience, 22, 5239-5249. Behavioral Neuroscience, 8, 1-7.
Berridge, K. C. (2007). The debate over dopamine’s role in re- Crowley, M. A. (1979). The allocation of time to temporally de-
ward: The case for incentive salience. Psychopharmacology, fined behaviors: Responding during stimulus generalization.
191, 391-431. Journal of the Experimental Analysis of Behavior, 32, 191-
Betts, S. L., Brandon, S. E., & Wagner, A. R. (1996). Disasso- 197.
ciation of the blocking of conditioned eyeblink and condi- Crowley, M. A. & Donahoe, J. W. (2004). Matching: Its acqui-
tioned fear following a shift in US locus. Animal Learning sition and generalization. Journal of the Experimental Anal-
& Behavior, 24, 459-470. ysis of Behavior, 82, 143-159.
Bickel, W. K. & Etzel, B. C. (1985). The quantal nature of con- Dichter, G. S., Felder, J. N., Green, S. R., Ritenberg, A. M., Sas-
trolling stimulus-response relations in tests of stimulus gen- son, N. J., & Bodfish, J. W. (2012). Reward circuitry func-
eralization. Journal of the Experimental Analysis of Behav- tion in autism spectrum disorders. Social Cognitive and Af-
ior, 44, 245-270. fective Neuroscience, 7, 160-172.
Bliss, T. V. & Lømø, T. (1973). Long-lasting potentiation of syn- Dinsmoor, J. A. (1950). A quantitative comparison of the dis-
aptic transmission in the dentate area of the anaesthetized criminative and reinforcing functions of a stimulus. Journal
rabbit following stimulation of the perforant path. Journal of of Experimental Psychology, 40, 458-472.
Physiology, 232, 331-356. Dobzhansky, T. G. (1937). Genetics and the origin of species.
Bradfield, L. A., Bartran-Gonzales, J., Chieng, B, & Balleine, B. New York: Columbia University Press.
W. The thalmostriatal pathway and cholinergic control of Domjan, M. & Galef, B. G. (1983). Biological constraints on in-
goal-directed action: Interlacing new with existing learning strumental and classical conditioning: Retrospect and pro-
in the striatum. Neuron, 79, 153-156. spect. Animal Learning & Behavior, 11, 151-161.
Donahoe, J. W. (1984). Skinner—The Darwin of ontogeny? Be-
havioral and Brain Sciences, 7, 487-488.
2

Donahoe, J. W. (1997). Neural plasticity. In J. W. Donahoe & V. Baker, C. I. (1997). Gaze following and joint attention in
P. Dorsel (Eds.), Neural-network models of cognition (pp. rhesus monkeys (Macaca mulatto). Journal of Comparative
80-81). New York: Elsevier Science Press Psychology, 111, 286-293.
Donahoe, J. W. (1999). Edward L. Thorndike: The selectionist Estes, W. K. (1956). The problem of inferences from curves
connectionist. Journal of the Experimental Analysis of Be- based on group data. Psychological Bulletin, 53, 134-140.
havior, 72, 451-454. Estes, W. K., Burke, C. J., Atkinson, R. C. (1957). Probabilistic
Donahoe, J. W. (2003). Selectionism. In Lattal, K.A. & Chase, discrimination learning. Journal of Experimental Psychol-
P.N. (Eds.) Behavior Theory and Philosophy (pp.103-128). ogy. 54, 233-239.
Dordrecht, Netherlands: Kluwer Academic Publishers. Estes, W. K. & Skinner, B. F. (1941). Some quantitative proper-
Donahoe, J. W. (2010). Man as machine: A review of Memory ties of anxiety. Journal of Experimental Psychology, 29,
and the computational brain: Why cognitive science will 390-400.
transform neuroscience. Behavior and Philosophy, 38, 79- Ferster, C. B. & Skinner, B. F. (1957). Schedules of reinforce-
97. ment. New York: Appleton-Century-Crofts.
Donahoe, J. W. (2012). Origins of the molar-molecular debate. Fields, L., Reeve, K. F., Varelas, A., Rosen, D., & Belanich, J.
European Journal of Behavior Analysis, 13, 195-200. (1997). Equivalence class formation using stimulus-pairing
Donahoe, J. W. (2013). Theory and behavior analysis. Behavior and yes-no. The Psychological Record, 47, 661–686.
Analyst, 36, 361-371. Frey, U. (1997). Cellular mechanisms of long-term potentiation:
Donahoe, J. W, & Burgos, J. E. (2005). Selectionism: Complex Late maintenance. In Donahoe, J. W. and Dorsel, V. P.
outcomes from simple processes. Behavioural and Brain (Eds.). Neural-network models of cognition: Biobehavioral
Sciences, 28, 429-430. foundations (pp. 105-128). New York: Elsevier Science
Donahoe, J. W., Burgos, J. E., & Palmer, D. C. (1993). A selec- Press.
tionist approach to reinforcement. Journal of the Experi- Frey, U, & Morris, R. G. M. (1997). Synaptic tagging and long-
mental Analysis of Behavior, 60, 17-40. term potentiation. Nature 385, 533–536.
Donahoe, J. W., Crowley, M. A., Millard, W. J., & Stickney, K. Fujimichi, R., Naya, Y., & Miyashita, Y. (2005). Neural basis of
A. (1982). A unified principle of reinforcement: Some im- semantic-like memory: Representation, activation, and cog-
plications for matching. In M. L. Commons, R. J. Herrnstein, nitive control. In M. S. Gazaniga (Ed.), i (pp. 905-918),
& H. Rachlin (Eds.), Quantitative analyses of behavior: Vol. Cambridge, MA: MIT Press. http://www.physiol.m.u-to-
2. Matching and maximizing accounts (pp. 493-521). Cam- kyo.ac.jp/en/review/review_en_03.html
bridge, MA: Ballinger. Fuster, J. (2015). The prefrontal cortex (9th edition). New York:
Donahoe, J. W. & Palmer, D. C. (1993/2005). Learning and Academic Press.
complex behavior. Boston: MA: Allyn & Bacon. (Reprinted Gao, X-B & Horvath, T. (2014). Function and dysfunction of
Richmond, MA: Ledgetop Publishing) http://lcb-online.org hypocretin.orexin: An enegertics point of view. Annual Re-
Donahoe, J. W. & Palmer, D. C., & Burgos, J. E. (1997). The S- view of Neuroscience, 37, 101-116.
R issue: Its status in behavior analysis and Donahoe and Gallistel, C. R. & King, A. P. (2009). Memory and the computa-
Palmer’s Learning and Complex Behavior. (1997). Journal tional brain: Why cognitive science will transform neurosci-
of the Experimental Analysis of Behavior, 67, 193-211. ence. New York: Wiley.
Donahoe, J. W., & Vegas, R. (2004). Pavlovian conditioning: Gasbarri, A., Verney, C., Innocenzi, R., Campana, E., Pacitti, C.
The CS-UR relation. Journal of Experimental Psychology: (1994). Mesolimbic dopaminergic neurons innervating the
Animal Behavior Processes, 30, 17-33. hippocampal formation in the rat: a combined retrograde
Donahoe, J. W. & Vegas, R. (in press). Respondent (Pavlovian) tracing and immunohistochemical study. Brain Research,
conditioning. In W. W. Fisher, C. C. Piazza, and H. S. Roane 668, 71–79
(Eds.), Handbook of applied behavior analysis (2nd edition). Gayon, J. (1998). Darwin’s struggle for survival: Heredity and
New York: Guilford Press. Cobb2nd
the hypothesis of natural selection. (Translated by M. Add
Donahoe, J. W. & Wessells, M. G. (1980). Learning, language, from Darwin et l’aprés-Darwin: une histoire de l’hypothèse
and memory. New York: Harper & Row.
edition
de sélection naturalle. Paris: Editions Kimé, Paris, 1992.)
Dreyer, J. K., Vander-Weele, C. M., & Lovic, V. (2016). Func- Cambridge, UK: Cambridge University Press.
tionally distinct dopamine signals in nucleus accumbens Geisler, S., Derst, C., Veh, R. W., & Zahm, D. S. (2007). Glu-
core and shell in the freely moving rat. Journal of Neurosci- tamatergic afferents of the ventral tegmental area in the rat.
ence, 36, 96-112. Journal of Neuroscience, 27, 5730–5743.
Duncan, K., Ketz, N., Inati, S. J., & Davachi, L. (2012). Evi- Georgopoulos, A. P., Schwartz, A. B., & Kettner, R. E. (1986).
dence for area CAi as a match/mismatch detector: A high- Neural population coding of movement direction. Science,
resolution fMRI study of the human hippocampus. Hippo- 233, 1416-1419.
campus, 22, 389-398. Gerfen, C. R. & Surmeier, D. J. (2011). Modulation of striatal
Dworkin, B. R. (1993). Learning and physiological regulation. projection neurons by dopamine. Annual Review of Neuro-
Chicago: University of Chicago Press. science, 34, 441-466.
Dworkin, B. R., & Miller, N. E. (1986). Failure to replicate vis- Gibbon, J. & Balsam, p. D. (1981). The spread of association in
ceral learning in the acute curarized rat preparation. Behav- time. In c. Locurto & H. S. Terrace (Eds.), Conditioning the-
ioral Neuroscience, 100, 299-314. ory. New York: Academic Press.
Eickelboom, R. & Stewart, J. (1982). Conditioning of drug-in- Gormezano, I. (1966). Classical conditioning. In J. B. Sidowski
duced physiological responses. Psychological Review, 89, (Ed.), Experimental methods and instrumentation in psy-
507-528.w chology. New York: McGraw-Hill.
Emery, N. J., Lorincz, E. N., Perrett, D. I., Oram, M. W., &
3

Grace, A. A. & Bunney, B. S. (1984). The control of firing pat- Lattal, K. A & Gleeson, S. (1990). Response acquisition with
tern in nigral dopamine neurons: Burst firing. Journal of delayed reinforcement. Journal of Experimental Psychol-
Neuroscience, 4, 2877-2890. ogy: Animal Behavior Processes, 16, 27-39.
Grace, A. A., Floresco, S. B., Goto, Y., & Lodge, D. J. (2007). Leader, G., Barnes, D., & Smeets, P. M. (1996). Establishing
Regulation of firing of dopaminergic neurons and control of equivalence relations using a respondent-type training
goal-directed behaviors. Trends in Neuroscience, 30, 220 – procedure. The Psychological Record, 46, 685–706.
227. Lee, C. R. & Tepper, J. M. (2009). Basal ganglia control of
Grisante, P. C., Galesi, F. L., Sabino, N. M., Debert, P., Arntzen, substantia nigra dopaminergic neurons. Journal of Neural
E., & McIlvane, W. J. (2013). Go/no-go procedure with Transmission Supplement, 73, 71-90.
compound stimuli: Effects of training structure on the Leyland, C. M. & Mackintosh, N. J. (1978). Blocking of first-
emergence of equivalence classes. The Psychological and second-order autoshaping in pigeons. Animal Learning
Record, 63, 63–72. & Behavior, 6, 392-394.
Guttman, N. & Kalish, H. J. (1956). Discriminabilty and stimu- Li, S., Cullen, W. K., Anwyl, R., & Rowan, M. J. (2003). Dopa-
lus generalization, Journal of Experimental Psychology, 51, mine-dependent facilitation of LTP induction in hippocam-
79-88. pal CA1 by exposure to spatial novelty. Nature Neurosci-
Gynther, M. D. (1957). Differential eyelid conditioning as a ence, 6, 326-331.
function of stimulus similarity and strength of response to Li, Z. S., Schmauss, C., Cuenca, A., Ratcliffe, E., & Gershon,
the CS. Journal of Experimental Psychology, 53, 408-416. M. D. (2006). Physiological modulation of intestinal motil-
Hebb. D. O. (1949). The Organization of Behavior. New York: ity by enteric dopaminergic neurons and the D2 receptor:
Wiley. Analysis of dopamine receptor expression, location, devel-
Herrnstein, R. J., Loveland, D. H., & Cable, C. (1976). Natural opment, and function in wild-type and knock-out mice. Jour-
concepts in pigeons. Journal of Experimental Psychology: nal of Neuroscience, 26, 2798 –2807.
Animal Behavior Processes, 4, 285-302. Lionello-DeNolf, K. M. (2009). The search for symmetry.
Jensen, G., Ward, R. D., & Balsam, P. D. (2013). Information: Learning & Behavior, 37, 18-203.
Theory, brain, and behavior. Journal of the Experimental MacDonall, J. S. (2009). The stay/switch model of concurrent
Analysis of Behavior, 100, 408-431. choice. Journal of the Experimental Analysis of Behavior,
Johnson, D. F. (1970). Determiners of selective stimulus control 91, 21-39.
in the pigeon. Journal of Comparative and Physiological Marr, M. J. (1992). Behavior dynamics: One perspective. Jour-
Psychology, 76, 298-307. nal of the Experimental Analysis of Behavior, 57, 249-266.
Johnson, S. W., Seutin, V., & North, R. N. (1992). Burst firing Martin, S. J., Grimwood, P. D., & Morris, R. G. M. (2000). Syn-
in dopamine neurons induced by N-methyl-D-Aspartate: aptic plasticity and memory: An evaluation of the hypothe-
Role of electrogenic sodium pump. Science, 258, 665-667. sis. Annual Review of Neuroscience,23, 649-711.
Ju, W., Morishita, W., Tsui, J., Gaiette, G., Deerinck, T. J., Ad- McDowell, J. J. (2013a). A quantitative evolutionary theory of
ams, S. R., Garner, C. C., Tsien, R. Y., Ellisman, M. H., & adaptive behavior dynamics. Psychological Review, 120,
Malenka, R. C. (2004). Activity-dependent regulation of 731-750.
dendritic synthesis and trafficking of AMPA receptors. Na- McDowell, J. J. (2013b). Representations of Complexity: How
ture Neuroscience,7, 244-253. Nature Appears in Our Theories. Behavior Analyst, 36, 345-
Kamin, L. J. (1968). "Attention-like" processes in classical con- 359.
ditioning. In M. R. Jones (Ed.), Miami symposium on the McNish, K. A., Betts, S. L., Brandon, S. E., & Wagner,A. R.
prediction of behavior (pp. 9-31). Coral Gables, FL: Univer- (1997). Divergence of conditioned eyeblink and conditioned
sity of Miami Press. fear in backward Pavlovian training. Animal Learning & Be-
Kamin, L. J. (1969). Predictability, surprise, attention and con- havior, 25, 43-62.
ditioning. In B. A. Campbell & R. M. Church (Eds.), Pun- Migler, B. (1964). Effects of averaging data during stimulus gen-
ishment and aversive behavior (pp. 279-296). New York: eralization. Journal of the Experimental Analysis of Behav-
Appleton-Century-Crofts. ior, 7, 303-307.
Kastak, C. R., Schusterman, R. J., & Kastak, D. (2001). Equiva- Migler, B. & Millenson, J. R. (1969). Analysis of response rates
lence classification by California sea lions using class-spe- during stimulus generalization. Journal of the Experimental
cific reinforcers. Journal of the Experimental Analysis of Be- Analysis of Behavior, 12, 81-87.
havior, 76, 131-156. Moorman, D. E. & Aston-Jones, G. (2010). Oreixin/hypocretin
Killeen, P. R. (2015). The logistics of choice. Journal of the Ex- modulates response of ventral tegmental dopamine neurons
perimental Analysis of Behavior, 104, 74-92. to prefrontal activation: Diurnal influences. Journal of Neu-
Kirk, P. D. W., Babtie, A. C., Stumpf, M. P. H. (2015). Systems roscience, 17, 15585-15599.
biology (un)certainties. Science, 350, 387–388. Morse, W. H., & Skinner, B. F. (1957). A second type of super-
Kumaran, D., Hassabis, D., & McClelland, J. L. (2016). What stition in the pigeon. American Journal of Psychology,70,
learning systems do intelligent agents need? Complementary 308-311.
Italicize
learning systems theory updated. Trends in Cognitive Sci- Oakley, D. A., & Russell, I. S. (1972). Neocortical lesions Science
and
ences, 20, 512-534 Pavlovian conditioning. Physiology & Behavior, 8, 915-926.
Lalive, A. L., Munoz, M. B., Bellone, C., Slesinger, P. A., O’Keefe, J. & Nadel, L. (1978). The hippocampus as a cognitive
Christian Luscher, P. A., & Tan, K. R. (2014). Firing Modes map. Oxford, UK: Oxford University Press.
of Dopamine Neurons Drive Bidirectional GIRK Channel Ortu, D. & Vaidya, M. (2016). The challenges of integrating be-
Plasticity. Journal of Neuroscience, 34, 5107-5114. havioral and neural data: Bridging and breaking boundaries
across levels of analysis. The Behavior Analyst, 39, 1-16.
4

Palmer, D. C. (1986). The blocking of conditioned reinforce- Seidenbecher, T., Balschun, D., & Reymann, K. G. (1995).
ment. Unpublished doctoral dissertation. University of Mas- Drinking after water deprivation prolongs unsaturated LTP
sachusetts, Amherst, MA. in the dentate gyrus of rats. Physiology & Behavior. 57,
Palmer, D. C. (1991). A behavioral interpretation of memory. In 1001–1004.
L. J. Hayes & P. N. Chase (Eds.), Dialogues in Verbal Be- Shahan, T. A. (2010). Conditioned reinforcement and response
havior.(pp. 261-279). Reno, NV: Context Press. strength. Journal of the Experimental Analysis of Behavior,
Palmer, D. C. (2011). Consideration of private events is required 93, 269–289.
in a comprehensive science of behavior. Behavior Analyst, Skinner, B. F. (1948). "Superstition" in the pigeon. Journal of
34, 201-207. Experimental Psychology, 38, 168-172.
Palmer, D. C. & Donahoe, J. W. (1992). Essentialism and selec- Sidman, M. (1960). Tactics of scientific research. New York:
tionism in cognitive science and behavior analysis. Ameri- Basic Books.
can Psychologist, 47, 1344-1358. Sidman, M. (2000). Equivalence relations and the reinforcement
Papini, M. R. & Bitterman, M. E. (1990). The role of contin- contingency. Journal of the Experimental Analysis of Behav-
gency in classical conditioning. Psychological Review, 97, ior, 74, 127-146.
396-403. Sidman, M. & Tailby, W. (1982). Conditional discrimination vs.
Pavlov, I. P. (1960). Conditioned Reflexes. New York: Dover matching tosample: An expansion of the testing paradigm.
Press. Journal of the Experimental Analysis of Behavior, 37, 5-22.
Pears, A., Parkinson, J. A., Hopewell, L., Everitt, B. J., & Rob- Sidman, M. Rauzin, R., Lazar, R., Cunningham, S., Tailby, W.,
erts, . C. (2003). Lesions of the orbitofrontal but not medial & Carrigan, P. (1982). A search for symmetry in the condi-
prefrontal cortex disrupt conditioned reinforcement in pri- tional discrimnations of rhesus monkeys, baboons, & chil-
mates. Journal of Neuroscience, 23, 11189 –1120. dren. Journal of the Experimental Analysis of Behavior, 37,
Plenz, D. & Wickens, J. (2010). The striatal skeleton: Medium 23-44.
spiny projection neurons and their lateral connections. In H. Siegel, S & Allan, L. G. (1996). The widespread influence of the
Steiner and K. Y. Tseng (Eds.). Handbook of behavioral Rescorla-Wagner model. Psychonomic Bulletin & Review,
neuroscience, Vol. 20 (pp. 99-112). New York: Elsevier Sci- 3, 314-3321.
ence Press. Silva, F. J., Timberlake, W., & Cevik, M. O. (1998). A behavior
Pliskoff, S. S. (1971). Effects of symmetrical and asymmetrical systems approach to the expression of backward associa-
changeover delays on concurrent performances. Journal of tions. Learning & Motivation, 29, 1-22.
the Experimental Analysis of Behavior, 16, 249–256. Singer, W. (1997). Development and plasticity of neocortical
Rescorla, R. A. (1968). Probability of shock in the presence and processing architecttres. In J. W. Donahoe & V. P. Dorsel
absence of CS in fear conditioning. Journal of Comparative (Eds.), Neural-network models of cognition (pp. 142-159).
and Physiological Psychology, 66, 1-5. New York: Elsevier Science Press.
Rescorla, R. A. & Wagner, A. R. (1972). A theory of Pavlovian Skinner, B. F. (1937). Two types of conditioned reflex:a REPLY
conditioning: Variations in the effectiveness of reinforce- YOT Konorski and Miller. General Psychology, 16, 22-279.
ment and nonreinforcement. In A. H. Black & W. F. Prokasy Skinner, B. F. (1948). "Superstition" in the pigeon. Journal of
(Eds.), Classical conditioning: Current research and theory Experimental Psychology, 38, 168-172.
(pp. 64-99). New York: Appleton-Century-Crofts. Skinner, B. F. (1950). The behavior of organisms. New York:
Reynolds, J. N. J., Hyland, B. I., & Wickens, J. R. (2001). A Appleton-Century-Crofts.
cellular mechanism for learning. Nature, 413, 67-70. Skinner, B. F. (1974). About behaviorism. New York: Random
Salamone, J. D., Correa, M., Nunes, E. J., Randall, P. A., & House.
Pardo, M. (2013). The behavioral pharmacology of effort- Smith, M. C., Coleman, S. P. & Gormezano, I. (1969). Classical
related choice behavior: Dopamine, adenosine, and beyond, conditioning of the rabbit nictitating membrane response at
Journal of the Experimental Analysis of Behavior, 97, 97- backward, simultaneous, and forward CS-US intervals.
125. Journal of Comparative and Physiological Psychology, 69,
Sakai, K. & Miyashita, Y. (1991). Neural organization for the 226-231.
long-term memory of paired associates. Nature, 354, 152- Smith, D. V., Liu, H., & Vogt, M. B. (1996). Responses of gus-
155. tatory cells in the nucleus of the solitary tract of the hamster
Sanes, J. N. & Donoghue, J. P. (2000). Plasticity and primary after NaCl or miloride adaptation. Journal of Neurophysiol-
motor cortex. Annual Review of Neuroscience, 23, 393-415. ogy, 76, 47–58.
Scheuerman, K. V., Wildemann, D. G., & Holland, J. G. (1978). Solomon, R. L. & Corbit, J. D. (1974). An opponent-process the-
A clarification of continuous repertoire development. Jour- ory of motivation: I. Temporal dynamics of affect. Psycho-
nal of the Experimental Analysis of Behavior, 30, 197-203. logical Review,81, 119-145.
Schultz, W. (1997). Adaptive dopaminergic neurons report the Solomon, R. L. & Turner, L. H. (1962). Discriminative classical
appetitive value of environmental stimuli. In J. W. Donahoe conditioning in dogs paralyzed by curare can later control
& V. P. Dorsel (Eds.), Neural-network models of condition- conditioned avoidance responses in the normal state. Psy-
ing (pp. 317-35). New York: Elsevier Science Press. chological Review, 69, 202-219.
Schultz, W., Apicella, P., & Ljungberg,T. (1993). Conditioned Squire, L. R. (2004). Memory systems of the brain: A brief his-
stimuli during successive steps of learning a delayed re- tory and current perspective". Neurobiology of Learning and
sponse task. Journal of Neuroscience, 13, 900-913. Memory. 82, 171–177.
Schultz, W., Dayan, P., & Montague, P. R. (1997). A neural sub- Staddon, J. E. R. (1983). Adaptive behavior and learning. New
strate of prediction and reward. Science, 275, 1593-1599. York: Cambridge University Press.
5

Staddon, J. E R. (2014). On choice and the Law of Effect. Inter- Wise, R. A. (2002). Brain reward circuitry: Insights from un-
national Journal of Comparative Psychology, 27, 569-584. sensed incentives. Neuron, 36, 229-240.
Staddon, J. E. R. & Simmelhag, V. L. (1971). The “superstition” Yagishita, S., Hayashi-Takagi, A., Ellis-Davies, G. C. R., Ura-
experiment: A reexamination of its implications for the prin- kubo, H., Ishii, S., & Kasai1, H. (2014). A critical time win-
ciples of adaptive behavior. Psychological Review, 78, 3-43. dow for dopamine actions on the structural plasticity of den-
Stein, L., Xue, B. G., & Belluzzi, J. D. (1993). A cellular ana- dritic spines. Science 345, 1616-1620.
logue of operant conditioning. Journal of the Experimental Zentall, T. R., Wasserman, E. A., & Urcuioli,, P. J. (2013). As-
Analysis of Behavior, 60, 41-55. sociative concept learning in animals. Journal of the Exper-
Stickney, K. J., Donahoe, J. W., & Carlson, N. R. (1981). Con- imental Analysis of Behavior, 101, 130-151.
ditining preparation for the nictitating membrane of the pi-
geon. Behavior Research Methods & Instrumentation, 13,
633-656.
Stickney, K., & Donahoe, J. W. (1983). Attenuation of blocking
by a change in US locus. Animal Learning & Behavior, 11,
60-66.
Stryker, M. P. (1986). Binocular impulse blockade prevents the
formation of ocular dominance columns in cat visual cortex.
Journal of Neuroscience, 6, 2117-2133.
Sutton, R. S. & Barto, A. G. (1981). Toward a modern theory of
adaptive networks: Expectation and prediction. Psychologi-
cal Review, 88, 135-170.
Swanson, L. W. & Kohler, C. (1986). Anatomical evidence for
direct projections from the entorhinal to the entire cortical
mantle in the rat. Journal of Neuroscience, 6, 3010-3023.
Thorndike, E. L. (1903). Elements of psychology. New York: A.
G. Seiler.
Tomita, H., Ohbayashi, M., Nakahara, K., Hasegawa, I. &
Miyashita, Y. (1999) Top-down signal from prefrontal cor-
tex in executive control of memory retrieval. Nature, 401,
699-703.
Tong, Z. Y., Overton, P. G., Clark, D., (1996). Stimulation of the
prefrontal cortex in the rat induces patterns of activity in
midbrain dopaminergic neurons which resemble natural
burst events. Synapse. 22, 195–208.
Trowill, J. A. (1967). Instrumental conditioning of the heart rate
in the curarized rat. Journal of Comparative and Physiolog-
ical Psychology, 63, 7-11.
Urcuioli, P. J. Associative symmetry, antisymmetry, and a the-
ory of pigeons’ equivalence-class formation. Journal of the
Experimental Analysis of Behavior, 90, 257-282.
Volkow, N. D., Fowler, J. S., & Wang, G-J. (2007). Dopamine
in drug abuse and addiction. Archives of Neurology, 64,
1575-1579.
Neurological Review,r andom Saal, W. & Jenkins, H. M. (1970).
Blocking the development of stimulus control. Learning &
Motivation, 1, 52-64.
Waelti, P., Dickinson, A., & Schultz, W. (2001). Dopamine re-
sponses comply with basic assumptions of formal learning
theory. Nature, 412, pp.43-48.
Williams, B. A. (1975). The blocking of reinforcement control.
Journal of the Experimental Analysis of Behavior, 24, 215-
226.
Williams, B. A. (1994a). Conditioned reinforcement: Neglected
or outmoded explanatory construct? Psychonomic Bulletin
& Review, 1, 457–475.
Williams, B. A. (1994b). Conditioned reinforcement: Experi-
mental and theoretical issues. Behav. Anal., 17, 261-285.
Wilson, D. I. & Bowman, M. (2004). Nucleus accumbens neu-
rons in the rat exhibit differential activity to conditioned re-
inforcers and primary reinforcers within a second-order
schedule of saccharin reinforcement. European Journal of
Neuroscience, 20, 2777-2788.
6

Endnotes

1
The body of the article is self-contained, but supplementary technical information is occasionally intro-
duced by means of endnotes.

2
“Experimental analysis,” as Skinner used the term, referred to circumstances in which all of the effects
of the variables relevant to a phenomenon are either manipulated, measured, or controlled. This idealized
set of circumstances is only approximated even in the laboratory. Classical procedures in which the condi-
tioned response and the elicited response closely resemble one another—such as the salivary response of
the dog (Pavlov, 1927/1960) and the nictitating-membrane response of the rabbit (Gormezano, 1966) or
the pigeon (Stickney, Donahoe, & Carlson, 1981)—very closely approximate these requirements. How-
ever, other classical procedures reveal the more complex outcomes that can be produced by the condition-
ing process. As one example, the conditioned-suppression procedure devised by one of Skinner’s students
(Estes & Skinner, 1941) demonstrated that an unmonitored unconditioned response (the response to
shock) may interfere with the execution of a monitored operant (bar pressing), thereby providing an indi-
rect but sensitive measure of conditioning. This procedure eventually led to the important discovery of the
discrepancy requirement, which is described subsequently (Kamin, 1968; Rescorla, 1968). As a second
example of a complex outcome produced by a classical procedure, the response measured after condition-
ing may sometimes appear to be the opposite to the response to the US: For example, a conditioned stim-
ulus paired with the injection of glucose produces a decrease in circulating glucose when the stimulus is
presented alone. However, the true unconditioned response to the injection of glucose is not an increase in
circulating glucose, but a decrease the glucose liberated from glucose-storing organs when neurons detect
an increase in exogenous glucose (see Eickelboom & Stewart, 1982). In general, pre-existing homeostatic
and other mechanisms resulting from natural selection must be taken into account when interpreting the
results of any conditioning procedure (cf. Dworkin, 1993; Solomon & Corbit, 1974).
3
The statement that conditioning in the classical procedure occurs over only brief temporal intervals be-
tween the CS and the reinforcer is based on procedures in which the reinforcer-elicited response has a
short latency and brief duration. Findings reported later indicate that it is not the temporal relation of the
CS to the reinforcing stimulus per se that is critical but the relation of the CS to the behavior produced by
the reinforcer. For example, in taste aversions a novel taste (CS) followed by the ingestion of a nonlethal
poison becomes aversive even though the behavioral effects (Relicited) of ingestion occur some hours later.
In the evolutionary history of organisms, gustatory and olfactory stimuli have inevitably preceded the gas-
tric consequences of ingestion. This has promoted the selection of neural structures relating such stimuli
to gastric responses. Neural traces of these stimuli endure and are then contiguous with the gastric conse-
quences (Chambers, 1990). In respects other than the temporal disjunction between the taste CS and the
gastric Relected to the poison US, findings from the conditioning of taste aversions are consistent with those
from other conditioning preparations (Domjan & Galef, 1983).

4
Several other studies conducted around the time of these critical experiments pointed toward the insuffi-
ciency of temporal contiguity for conditioning (e.g., Johnson, 1970), but their implications were not fully
appreciated (cf. Williams, 1975).
5
The following section of the paper takes the position that an analysis of moment-to-moment events re-
veals the dynamical processes of which relations between environmental and behavioral events observed
over more extended periods of time are the product. It is generally agreed that stable molar relations, as
useful as they may be for some purposes, are silent with respect to the more “molecular” processes of
which they are the asymptotic expression (e.g., Donahoe, 2012; Marr, 1992; Staddon, 2014). In addition,
studies indicate that at least some molar E-B relations, such as matching, may be the concerted product of
7

multiple discriminated operants (e.g., Crowley & Donahoe, 1992; Killeen, 2015; MacDonall, 2009; Plis-
koff, 1971). Indeed, computer simulations that implement moment-to-moment processes have yielded as-
ymptotic results that reproduce some molar relations such as matching (e.g., McDowell, 2013a; Calvin &
McDowell, 2016) and the effects of varying the C/T ratio on acquisition (Burgos, 2005), where C is the
length of the CS-US interval and T is the total session length (Gibbon & Balsam, 1981). A conceptually
independent but often conflated difference between theoretical approaches to behavior theory concerns
whether it is sufficient for a theory to serve as a guide for the behavior of the theorist (e.g., Jensen, Ward,
& Balsam, 2013; Gallistel & King, 2009; cf. Donahoe, 2010) or whether theory should also embody the
biobehavioral processes that underlie the functional relations captured by the theory (McDowell, 2013b;
cf., Donahoe, 2013).
6
Although dopamine plays a critical role in reinforcement, not all instances of long-lasting changes in
synaptic plasticity are dopamine-dependent (e.g., Bauer, Schafe, & LeDoux, 2002).

7
The conditioning of autonomic responses with operant procedures has been notoriously difficult to
achieve (Dworkin & Miller, 1986). Autonomic responses are mediated at the level of the midbrain and
need not involve cortical mechanisms, including the circuits involved with conditioned reinforcement de-
scribed here (e.g., Oakley & Russell, 1972). Computer simulations indicate that conditioned reinforce-
ment is needed to bridge the temporal gap between the operant and the reinforcer (Donahoe & Burgos,
2005). These circuits are not available at the level of the midbrain. It is of interest that the few studies that
have successfully demonstrated operant conditioning of autonomic response have minimized delay of re-
inforcement by using immediate electrical stimulation of the brain as a reinforcer (e.g., Trowill, 1967).

8
Deficiencies in the neural mechanisms of conditioned reinforcement have been proposed as an important
factor in the behavioral deficits found along the autism spectrum (Donahoe & Vegas, in press). In one study,
diminished activity in the NAC to a visual conditioned reinforcer was found with autistic subjects (Dichter,
Felder, Green, Ritenberg, Sasson, Bodfish, 2012). If the numbers and sources of connections from regions
in prefrontal cortex to NAC and, possibly, VTA are reduced for those with autism, then the potential for
stimuli that activate those regions to serve as conditioned reinforcers would be impaired. As an example,
suppose that neurons in the region of the prefrontal cortex that receive inputs from the sensory areas in-
volved in face perception do not have connections to NAC. Under these circumstances, seeing human faces
would not serve as a source of conditioned reinforcement. Instead, eye contact with another person would
be viewed as a threat gesture, the typical reaction in other primates. (Emery, Lorincz, Perrett, Orm, & Baker,
1997).
9
The present discussion of neural mechanisms describes the major systems underlying reinforcement.
Although much is known of these systems and their underlying processes, much remains to be known
(e.g., Gerfen & Surmeier, 2011). As an example relating to the DA-ergic mechanisms of the discrepancy
requirement, DA neurons in the VTA are maintained at their low baseline (tonic) rates of firing by inhibi-
tory neurons acting on VTA neurons from the output of NAC. (Lalive, Munoz, Bellone, & Slesinger,
2014). Unconditioned reinforcers drive the VTA neurons sufficiently to overcome this tonic inhibition.
The onset of conditioned reinforcers activate Glu pathways from prefrontal cortex to neurons in NAC and
these increase DA activity by stimulation of DA VTA neurons (Geizler, Derst, Veh, & Zahn, 2007)
and/or by inhibiting inhibitory neurons from NAC to VTA, thereby briefly liberating DA from VTA neu-
rons before tonic inhibition is reinstated (Aggarwai, Hyland, & Wickens, 2012; Creed, Ntamati, & Tan,
2014; Moorman & Aston-Jones, 2010; Tong, Overton, & Clark, 1996). As an additional complication,
inputs from the hypothalamus to the region of the VTA co-release the neuropeptide orexin/hypocretin
which provides a mechanism whereby the state of deprivation affects DA VTA activity (cf. Burdakov,
Liss, & Ashcroft, 2003; Gao & Horvath, 2014; Seidenbecher, Balschun, & Reymann, 1995). Finally,
NAC is itself not a homogeneous structure; the core and shell of the nucleus are differently innervated
8

(Dreyer, Vander-Weele, & Lovic, 2016) and have somewhat different functions (Corbit & Balleine,
2011). (For a different view of the place of DA in learned behavior see Berridge, 2007).

10
The behavioral expression of the UR is not necessary for conditioning as demonstrated by cases in
which the behavioral UR is prevented (as when transmission at the neuromuscular junction is blocked;
e.g., Solomon & Turner, 1962) or by cases in which the environment does not support expression of the
UR (as when an auditory CS is paired with food for the pecking response of the pigeon in an autoshaping
experiment; e.g., Leyland & Mackintosh, 1978). The response elicited by the reinforcer is the behavioral
event that is most closely synchronized with the occurrence of the DA-ergic neural mechanisms of rein-
forcement. (Donahoe & Vegas, 2004). However, once these neural mechanisms had been naturally se-
lected their behavioral expression is not necessary for the environment to engage them.

11
DA receptors of the type involved here are coupled to a particular G protein that catalyzes the synthesis
of cAMP (cyclic adenosine monophosphate), which is the starting point for a series of intracellular second
messengers that stimulate protein synthesis. Studies have shown that the injection of cAMP into a cell
with tagged Glu receptors produces LTP of those receptors even though the cell was not depolarized suf-
ficiently to otherwise cause LTP. LTP may result from prolonging the opening of ion channels of Glu re-
ceptors or by producing additional Glu receptors (so called “latent” receptors) in the postsynaptic mem-
brane (Ju, Morishita, Tsui, Gaiette, Deerinck, Adams, Garner, Tsien, Ellisman, & Malenka, 2004). See
Frey, 1997 and Martin, Grimwood, & Morris, 2000 for additional information. As indicated later, neuro-
modulators other than DA can enable LTP.

12
The presentation has focused on the effect of DA arising from the VTA that acts on neurons in the pre-
frontal cortex and NAC. These projections make up the mesocortical and mesolimbic DA systems, re-
spectively. Another system of DA projections arises from the substantia nigra pars compacta (SN), a
midbrain nucleus adjacent to the VTA. The SN system projects to a region called the striatum, the region
referenced in this footnote. DA-ergic neurons in the SN are also activated by reinforcing stimuli (Lee &
Tepper, 2009). In addition to these DA-ergic inputs, neurons in the striatum receive converging inputs
from motor regions of the prefrontal cortex and from the thalamus (Plenz & Wickens, 2010). Striatal neu-
rons are thus well positioned to integrate activity from their sensory thalamic inputs with their reinforced
prefrontal motor inputs (cf. Bradfield, Bertran-Gonzalez, Chieng, & Baleine, 2013). More “downstream”
motor systems then lead to observable behavior. A provocative speculation is that DA from the VTA acts
upon neural circuits in prefrontal cortex that specify those striatal neurons that are most active prior to re-
inforcement and that DA from SN then strengthens connections from sensory inputs to those co-active
striatal neurons. As the result of such a process, reinforced E-B relations could ultimately be mediated be-
fore fully engaging prefrontal activity, for example the activity mediating subvocal verbal behavior (i.e.,
consciousness). Through this “short-circuiting” process, such behavior would become “automatic” (Do-
nahoe, 1997, p. 354).

13
A wide variety of conditioning phenomena have been simulated in neural-network research that imple-
ments the foregoing mechanisms of reinforcement using networks whose architecture is consistent with
the neural systems described here (see Burgos & Donahoe, 2016).
14
DA-dependent LTP was first discovered in CA1 neurons (Bliss & Lomo, 1973). While DA enables
LTP at CA1 synapses, other agents play a critical role at synapses within the trisynaptic circuit of the hip-
pocampus (e.g., Stein, Xue, & Belluzzi, 1993). In S-A cortex, neurotransmitters such as noradrenaline,
acetylcholine, and/or serotonin play a role that is functionally similar to that of DA in fostering LTP in
prefrontal cortex and CA1 synapses (Singer, 1997). Early studies of the hippocampus documented its role
in spatial discriminations (O’Keefe & Nadel, 1978). Spatial discriminations require integrating multiple
9

sensory inputs to specify a particular location within the physical environment. However, the role of the
hippocampus is much more general in that it integrates multisensory inputs in a wide variety of situations,
with spatial discrimination being but one example.
15
A considerable array of procedures other than standard matching-to-sample procedures has provided
evidence of the formation of equivalence classes (e.g., Arntzen, 2006; Arntzen, Nartey, & Fields, 2015,
Arntzen, Norbom & Fields, 2015; Fields, Reeve, Varelas, Rosen, & Bananich, 1979; Grisante, Galesi,
Sabino, Debert, & Arntzen, 2013; Leader, Barnes, & Smeets, 1996). I thank Prof. Erik Arntzen for
providing some of these references.

View publication stats

S-ar putea să vă placă și