Beruflich Dokumente
Kultur Dokumente
GOAL-DIRECTED BEHAVIOUR The fundamental role of reward in the survival and well- Behavioural functions of rewards
Behaviour controlled by being of biological agents ranges from the control of Given the dynamic nature of the interactions between
representation of a goal or an vegetative functions to the organization of voluntary, complex organisms and the environment, it is not sur-
understanding of a causal
GOAL-DIRECTED BEHAVIOUR. The control of behaviour prising that specific neural mechanisms have evolved
relationship between behaviour
and attainment of a goal. requires the extraction of reward information from a that not only detect the presence of rewarding stimuli
large variety of stimuli and events. This information but also predict their occurrence on the basis of repre-
REINFORCERS concerns the presence and values of rewards, their pre- sentations formed by past experience. Through these
Positive reinforcers (rewards)
dictability and accessibility, and the numerous methods mechanisms, rewards have come to be implicit or explic-
increase the frequency of
behaviour leading to their and costs associated with attaining them. it goals for increasingly voluntary and intentional forms
acquisition. Negative reinforcers Various experimental approaches including brain of behaviour that are likely to lead to the acquisition of
(punishers) decrease the lesions, psychopharmacology, electrical self-stimulation goal objects.
frequency of behaviour leading and the administration of addictive drugs, have helped Rewards have several basic functions. A common
to their encounter and increase
the frequency of behaviour
to determine the crucial structures involved in reward view is that rewards induce subjective feelings of plea-
leading to their avoidance. processing14. In addition, physiological methods such as sure and contribute to positive emotions. Unfortunately,
in vivo microdialysis, voltammetry59 and neural imag- this function can only be investigated with difficulty in
ing1012 have been used to probe the structures and neu- experimental animals. Rewards can also act as positive
rotransmitters that are involved in processing reward REINFORCERS by increasing the frequency and intensity of
information in the brain. However, I believe that the behaviour that leads to the acquisition of goal objects,
temporal constraints imposed by the nature of the as described in CLASSICAL and INSTRUMENTAL CONDITIONING
reward signals themselves might be best met by studying PROCEDURES. Rewards can also maintain learned behav-
the activity of single neurons in behaving animals, and it iour by preventing EXTINCTION13. The rate of learning
is this approach that forms the basis of this article. Here, depends on the discrepancy between the occurrence
Institute of Physiology and I describe how neurons detect rewards, learn to predict of reward and the predicted occurrence of reward, the
Program in Neuroscience, future rewards from past experience and use reward so-called reward prediction error1416 (BOX 1).
University of Fribourg, information to learn, choose, prepare and execute goal- Rewards can also act as goals in their own right and
CH-1700 Fribourg,
Switzerland.
directed behaviour (FIG. 1). I also attempt to place the can therefore elicit approach and consummatory behav-
e-mail: processing of drug rewards within a general framework iour. Objects that signal rewards are labelled with posi-
Wolfram.Schultz@unifr.ch of neuronal reward mechanisms. tive MOTIVATIONAL VALUE because they will elicit effortful
PAVLOVIAN (CLASSICAL) behavioural responses. These motivational values arise Reward detection and perception
CONDITIONING either through innate mechanisms or, more often, Although there are no specialized peripheral receptors
Learning a predictive through learning. In this way, rewards help to establish for rewards, neurons in several brain structures seem
relationship between a stimulus
value systems for behaviour and serve as key references to be particularly sensitive to rewarding events as
and a reinforcer does not
require an action by the agent. for behavioural decisions. opposed to motivationally neutral events that are sig-
nalled through the same sensory modalities. A promi-
OPERANT (INSTRUMENTAL) nent example are the dopamine neurons in the pars
CONDITIONING
Reward detection Goal Relative reward value compacta of substantia nigra and the medially adjoin-
Learning a relationship between representation
a stimulus, an action and a
Reward prediction Reward expectation ing ventral tegmental area (groups A8, A9 and A10).
reinforcer conditional on an Medial Dorsolateral In various behavioural situations, including classical
action by the agent. temporal prefrontal, Orbitofrontal and instrumental conditioning, most dopamine neu-
cortex cortex
premotor, rons show short, phasic activation in a rather homo-
EXTINCTION parietal
cortex geneous fashion after the presentation of liquid and
Reduction and cessation of a
predictive relationship and solid rewards, and visual or auditory stimuli that pre-
behaviour following the dict reward1720 (FIG. 2a, b). These phasic neural responses
omission of a reinforcer are common (7080% of neurons) in medial tegmen-
(negative prediction error). tal regions that project to the nucleus accumbens and
MOTIVATIONAL VALUE
frontal cortex but are also found in intermediate and
A measure of the effort an agent Thalamus lateral sectors that project to the caudate and
is willing to expend to obtain an putamen20. These same dopamine neurons are also
object signalling reward or to activated by novel or intense stimuli that have atten-
avoid an object signalling Striatum
tional and rewarding properties2022. By contrast, only
punishment.
Reward a few dopamine neurons show phasic activations to
detection punishers (conditioned aversive visual or auditory
Goal
representation stimuli)23,24. Thus, the phasic response of A8, A9 and
A10 dopamine neurons seems to preferentially report
Dopamine SNpr
Amygdala
GP
rewarding and, to a lesser extent, attention-inducing
neurons
events.
Reward prediction The observation of a phasic dopamine response to
Error detection novel attention-inducing stimuli led to the suggestion
Figure 1 | Reward processing and the brain. Many reward that this response might reflect the salient attention-
signals are processed by the brain, including those that are inducing properties of rewards rather than their posi-
responsible for the detection of past rewards, the prediction tive reinforcing aspects25,26. However, dopamine neu-
and expectation of future rewards, and the use of information rons are rarely activated by strong attention-generating
about future rewards to control goal-directed behaviour. (SNpr, events such as aversive stimuli23,24 and they are
substantia nigra pars reticulata; GP, globus pallidus.)
depressed rather than activated by the omission of
a reward, which presumably also generates atten-
tion20,2729. Of course, the possibility that the dopamine
Box 1 | Reward prediction errors activation might encode a specific form of attention
Behavioural studies show that that is only associated with rewarding events cannot yet
reward-directed learning be completely ruled out.
depends on the predictability A closer examination of the properties of the phasic
Perform No Maintain
of the reward1416. The behavioural Error current dopamine response suggests that it might encode a
occurrence of the reward has reaction connectivity reward prediction error rather than reward per se.
to be surprising or Indeed, the phasic dopamine activity is enhanced by
Yes
unpredicted for a stimulus or surprising rewards but not by those that are fully pre-
action to be learned. The Modify dicted, and it is depressed by the omission of predicted
Generate rewards20,30 (FIG. 2c). The reward prediction error
degree to which a reward current
error signal
cannot be predicted is connectivity response also occurs during learning episodes27. In
indicated by the discrepancy view of the crucial role that prediction errors are
between the reward obtained for a given behavioural action and the reward that was thought to play during learning1416, a phasic dopamine
predicted to occur as a result of that action. This is known as the prediction error and response that reports a reward prediction error might
underlies a class of error-driven learning mechanisms33. constitute an ideal teaching signal for approach learn-
In simple terms, if a reward occurs unpredictably after a given action then the ing 31. This signal strongly resembles32 the teaching
prediction error is positive, and learning about the consequences of the action that signal used by very effective temporal difference re-
produced the reward occurs. However, once the consequences of that action have been inforcement models33. Indeed, neuronal networks
learned (so that the reward that follows subsequent repetition of the action is now
using this type of teaching signal have learned to play
predicted), the prediction error falls to zero and no new information about the
backgammon at a high level34 and have acquired serial
consequences of the action is learned. By contrast, if the expected reward is not received
movements and spatial delayed response tasks in a
after repeating a learned action then the prediction error falls to a negative value and the
manner consistent with the performance of monkeys
behaviour is extinguished.
in the laboratory35,36.
unrewarded movements
choice phase of the task. This delay seems to reflect
Learning
In situations involving a choice between different completion of behavioural responses, they could influ-
rewards, agents tend to optimize their choice behaviour ence behaviour before reward acquisition through this
according to the motivational values of the available mechanism. This retrograde action might be mediated
rewards. Indeed, greater efforts are made for rewards by the reward-predicting response of the dopamine
with higher motivational values. Neurons in the parietal neurons and by the reward-expectation neurons in the
cortex show stronger task-related activity when mon- striatum and orbitofrontal cortex.
keys choose larger or more frequent rewards over small- Neurons in the ventral striatum have more reward-
er or less frequent rewards83. The reward preference neu- related responses and reward-expectation activity than
rons in the orbitofrontal cortex might encode a related neurons in the caudate nucleus and putamen40,68.
process56. Together, these neurons appear to encode the Although these findings might provide a neurophysio-
basic parameters of motivational value that underlies logical correlate for the known motivational functions
choice behaviour. of the ventral striatum, the precise characterization of
In addition, agents tend to modify their behaviour the neurophysiology of the various subregions of nucle-
towards a reward when the motivational value of the us accumbens that have been suggested by behavioural
reward for that behaviour changes. Neurons in the studies85 remains to be explored.
cingulate motor area of the frontal cortex might pro- The striatal reward activities and dopamine reward
vide a neural correlate of this behaviour because they inputs might influence behaviour-related activity in the
are selectively active when monkeys switch to a differ- prefrontal, parietal and perirhinal cortices through the
ent movement in situations in which performing the different cortexbasal-gangliacortex loops. The
current movement would lead to less reward84. These orbitofrontal sectors of the prefrontal cortex seem to be
neurons do not fire if switching between movements the most involved in reward processing56,60 but reward-
is not associated with reduced reward or when rewards related activity is found also in the dorsolateral area79,80.
are reduced without movement switches. This might The above analysis suggests that the orbitofrontal cortex
suggest a mechanism that selects movements that is strongly engaged in the details of the detection, per-
maximize reward. ception and expectation of rewards, whereas the dorso-
lateral areas might use reward information to prepare,
Multiple reward signals plan, sequence and execute behaviour directed towards
Information about rewards seems to be processed in the acquisition of rewarding goals.
different forms by neurons in different brain struc- The different reward signals might also cooperate in
tures, ranging from the detection and perception of reward-directed learning. Dopamine neurons provide
rewards, through the expectation of imminent an effective teaching signal but do not seem to specify
rewards, to the use of information about predicted the individual reward, although this signal might modi-
rewards for the control of goal-directed behaviour. fy the activity of synapses engaged in behaviour leading
Neurons that detect the delivery of rewards provide to the reward. Neurons in the orbitofrontal cortex, stria-
valuable information about the motivational value and tum and amygdala might identify particular rewards.
identity of encountered objects. This information Some of these neurons are sensitive to reward prediction
might also help to construct neuronal representations and might be involved in local learning mechanisms29.
that permit subjects to expect future rewards according The reward-expectation activity in cortical and subcor-
to previous experience and to adapt their behaviour to tical structures adapts during learning according to
changes in reward contingencies. changing reward contingencies. In this way, different
Only a limited number of brain structures seem to neuronal systems are involved in different mechanisms
be implicated in the detection and perception of underlying the various forms of adaptive behaviour
rewards and reward-predicting stimuli (FIG. 1). The mid- directed towards rewards.
brain dopamine neurons report the occurrence of
reward relative to its prediction and emit a global rein- Drug rewards
forcement signal to all neurons in the striatum and Is it possible to use this expanding knowledge of the
many neurons in the prefrontal cortex. The dopamine neural substrates of reward processing to gain an under-
reward response might be the result of reward-detection standing of how artificial drug rewards exert their pro-
activity in several basal ganglia structures, the amygdala found influence on behaviour? Psychopharmacological
and the orbitofrontal cortex28. Some neurons in the studies have identified the dopamine system and the
dorsal and ventral striatum, orbitofrontal cortex, amyg- ventral striatum, including the nucleus accumbens, as
dala and lateral hypothalamus discriminate well some of the critical structures on which drugs of abuse
between different rewards, which might contribute to act2. These very structures are also implicated in the pro-
the perception of rewards, be involved in identifying cessing of rewards outlined above. One important ques-
particular objects and/or provide an assessment of their tion is thus whether drugs of abuse modify existing neu-
motivational value. ronal responses to natural rewards or constitute rewards
Reward-expectation activity might be an appropriate in their own right and, as such, engage existing neuronal
signal for predicting the occurrence of rewards and reward mechanisms.
might thereby provide a suitable mechanism for influ- There is evidence that the firing pattern of some neu-
encing behaviour that leads to the acquisition of rons in the nucleus accumbens can be enhanced or
rewards. Although rewards are received only after the depressed by the self-administration of cocaine86,87.
1. Fibiger, H. C. & Phillips, A. G. in Handbook of Physiology 30. Mirenowicz, J. & Schultz, W. Importance of unpredictability 54. Nakamura, K., Mikami, A. & Kubota, K. Activity of single
The Nervous System Vol. IV (ed. Bloom, F. E.) 647675 for reward responses in primate dopamine neurons. neurons in the monkey amygdala during performance of a
(Williams and Wilkins, Baltimore, Maryland, 1986). J. Neurophysiol. 72, 10241027 (1994). visual discrimination task. J. Neurophysiol. 67, 14471463
2. Wise, R. A. & Hoffman, D. C. Localization of drug reward 31. Schultz, W., Dayan, P. & Montague, R. R. A neural (1992).
mechanisms by intracranial injections. Synapse 10, substrate of prediction and reward. Science 275, 55. Burton, M. J., Rolls, E. T. & Mora, F. Effects of hunger on the
247263 (1992). 15931599 (1997). responses of neurons in the lateral hypothalamus to the
3. Robinson, T. E. & Berridge, K. C. The neural basis for drug 32. Montague, P. R., Dayan, P. & Sejnowski, T. J. A framework sight and taste of food. Exp. Neurol. 51, 668677 (1976).
craving: an incentive-sensitization theory of addiction. Brain for mesencephalic dopamine systems based on predictive 56. Tremblay, L. & Schultz, W. Relative reward preference in
Res. Rev. 18, 247291 (1993). Hebbian learning. J. Neurosci. 16, 19361947 (1996). primate orbitofrontal cortex. Nature 398, 704708 (1999).
4. Robbins, T. W. & Everitt, B. J. Neurobehavioural 33. Sutton, R. S. & Barto, A. G. Toward a modern theory of 57. Pratt, W. E. & Mizumori, S. J. Y. Characteristics of basolateral
mechanisms of reward and motivation. Curr. Opin. adaptive networks: expectation and prediction. Psychol. amygdala neuronal firing on a spatial memory task involving
Neurobiol. 6, 228236 (1996) Rev. 88, 135170 (1981). differential reward. Behav. Neurosci. 112, 554570 (1998).
5. Louilot, A., LeMoal, M. & Simon, H. Differential reactivity of This paper proposed a very effective temporal 58. Thorpe, S. J., Rolls, E. T. & Maddison, S. The orbitofrontal
dopaminergic neurons in the nucleus accumbens in difference reinforcement learning model that cortex: neuronal activity in the behaving monkey. Exp. Brain
response to different behavioural situations. An in vivo computes a prediction error over time. The teaching Res. 49, 93115 (1983).
voltammetric study in free moving rats. Brain Res. 397, signal incorporates primary reinforcers and 59. Rolls, E. T., Murzi, E., Yaxley, S., Thorpe, S. J. & Simpson,
395400 (1986). conditioned stimuli, and resembles in all aspects the S. J. Sensory-specific satiety: food-specific reduction in
6. Church, W. H., Justice, J. B. Jr & Neill, D. B. Detecting response of dopamine neurons to rewards and responsiveness of ventral forebrain neurons after feeding in
behaviourally relevant changes in extracellular dopamine conditioned, reward-predicting stimuli, although the monkey. Brain Res. 368, 7986 (1986).
with microdialysis. Brain Res. 41, 397399 (1987). dopamine neurons also report novel stimuli. 60. Rolls, E. T., Sienkiewicz, Z. J. & Yaxley, S. Hunger modulates
7. Young, A. M. J., Joseph, M. H. & Gray, J. A. Increased 34. Tesauro, G. TD-Gammon, a self-teaching backgammon the responses to gustatory stimuli of single neurons in the
dopamine release in vivo in nucleus accumbens and program, achieves master-level play. Neural Comp. 6, caudolateral orbitofrontal cortex of the macaque monkey.
caudate nucleus of the rat during drinking: a microdialysis 215219 (1994). Eur. J. Neurosci. 1, 5360 (1989).
study. Neuroscience 48, 871876 (1992). 35. Suri, R. E. & Schultz, W. Learning of sequential movements 61. Rolls, E. T., Scott, T. R., Sienkiewicz, Z. J. & Yaxley, S. The
8. Wilson, C., Nomikos, G. G., Collu, M. & Fibiger, H. C. by neural network model with dopamine-like reinforcement responsiveness of neurons in the frontal opercular gustatory
Dopaminergic correlates of motivated behaviour: signal. Exp. Brain Res. 121, 350354 (1998). cortex of the macaque monkey is independent of hunger.
importance of drive. J. Neurosci. 15, 51695178 (1995). 36. Suri, R. & Schultz, W. A neural network with dopamine-like J. Physiol. 397, 112 (1988).
9. Fiorino, D. F., Coury, A. & Phillips, A. G. Dynamic changes in reinforcement signal that learns a spatial delayed response 62. Apicella, P., Legallet, E. & Trouche, E. Responses of tonically
nucleus accumbens dopamine efflux during the Coolidge task. Neuroscience 91, 871890 (1999). discharging neurons in the monkey striatum to primary
effect in male rats. J. Neurosci. 17, 48494855 (1997). 37. Schultz, W. & Romo, R. Responses of nigrostriatal rewards delivered during different behavioural states. Exp.
10. Thut, G. et al. Activation of the human brain by monetary dopamine neurons to high intensity somatosensory Brain Res. 116, 456466 (1997).
reward. NeuroReport 8, 12251228 (1997). stimulation in the anesthetized monkey. J. Neurophysiol. 57, 63. Apicella, P., Ravel, S., Sardo, P. & Legallet, E. Influence of
11. Rogers, R. D. et al. Choosing between small, likely 201217 (1987). predictive information on responses of tonically active
rewards and large, unlikely rewards activates inferior and 38. Abercrombie, E. D., Keefe, K. A., DiFrischia, D. S. & neurons in the monkey striatum. J. Neurophysiol. 80,
orbital prefrontal cortex. J. Neurosci. 19, 90299038 Zigmond, M. J. Differential effect of stress on in vivo 33413344 (1998).
(1999) dopamine release in striatum, nucleus accumbens, and 64. Apicella, P., Legallet, E. & Trouche, E. Responses of tonically
12. Elliott, R., Friston, K. J. & Dolan, R. J. Dissociable neural medial frontal cortex. J. Neurochem. 52, 16551658 (1989). discharging neurons in monkey striatum to visual stimuli
responses in human reward systems. J. Neurosci. 20, 39. Hikosaka, O., Sakamoto, M. & Usui, S. Functional properties presented under passive conditions and during task
61596165 (2000) of monkey caudate neurons. III. Activities related to performance. Neurosci. Lett. 203, 147150 (1996).
13. Pavlov, P. I. Conditioned Reflexes (Oxford Univ. Press, expectation of target and reward. J. Neurophysiol. 61, 65. Williams, G. V., Rolls, E. T., Leonard, C. M. & Stern, C.
London, 1927). 814832 (1989). Neuronal responses in the ventral striatum of the behaving
14. Rescorla, R. A. & Wagner, A. R. in Classical Conditioning II: 40. Apicella, P., Ljungberg, T., Scarnati, E. & Schultz, W. monkey. Behav. Brain Res. 55, 243252 (1993).
Current Research and Theory (eds Black, A. H. and Prokasy, Responses to reward in monkey dorsal and ventral striatum. 66. Aosaki, T. et al. Responses of tonically active neurons in the
W. F.) 6499 (Appleton Century Crofts, New York, 1972). Exp. Brain Res. 85, 491500 (1991). primates striatum undergo systematic changes during
15. Mackintosh, N. J. A theory of attention: variations in the 41. Apicella, P., Scarnati, E. & Schultz, W. Tonically discharging behavioural sensorimotor conditioning. J. Neurosci. 14,
associability of stimulus with reinforcement. Psychol. Rev. neurons of monkey striatum respond to preparatory and 39693984 (1994).
82, 276298 (1975). rewarding stimuli. Exp. Brain Res. 84, 672675 (1991). 67. Apicella, P., Scarnati, E., Ljungberg, T. & Schultz, W. Neuronal
16. Pearce, J. M. & Hall, G. A model for Pavlovian 42. Lavoie, A. M. & Mizumori, S. J. Y. Spatial-, movement- and activity in monkey striatum related to the expectation of
conditioning: variations in the effectiveness of conditioned reward-sensitive discharge by medial ventral striatum predictable environmental events. J. Neurophysiol. 68,
but not of unconditioned stimuli. Psychol. Rev. 87, 532552 neurons of rats. Brain Res. 638, 157168 (1994). 945960 (1992).
(1980). 43. Bowman, E. M., Aigner, T. G. & Richmond, B. J. Neural 68. Schultz, W., Apicella, P., Scarnati, E. & Ljungberg, T.
17. Schultz, W. Responses of midbrain dopamine neurons to signals in the monkey ventral striatum related to motivation Neuronal activity in monkey ventral striatum related to the
behavioural trigger stimuli in the monkey. J. Neurophysiol. for juice and cocaine rewards. J. Neurophysiol. 75, expectation of reward. J. Neurosci. 12, 45954610 (1992).
56, 14391462 (1986). 10611073 (1996). 69. Schoenbaum, G., Chiba, A. A. & Gallagher, M. Orbitofrontal
18. Romo, R. & Schultz, W. Dopamine neurons of the monkey 44. Shidara, M., Aigner, T. G. & Richmond, B. J. Neuronal cortex and basolateral amygdala encode expected
midbrain: contingencies of responses to active touch during signals in the monkey ventral striatum related to progress outcomes during learning. Nature Neurosci. 1, 155159
self-initiated arm movements. J. Neurophysiol. 63, 592606 through a predictable series of trials. J. Neurosci. 18, (1998).
(1990). 26132625 (1998). 70. Hikosaka, K. & Watanabe, M. Delay activity of orbital and
19. Schultz, W. & Romo, R. Dopamine neurons of the monkey 45. Matsumura, M., Kojima, J., Gardiner, T. W. & Hikosaka, O. lateral prefrontal neurons of the monkey varying with
midbrain: contingencies of responses to stimuli eliciting Visual and oculomotor functions of monkey subthalamic different rewards. Cerebral Cortex 10, 263271 (2000).
immediate behavioural reactions. J. Neurophysiol. 63, nucleus. J. Neurophysiol. 67, 16151632 (1992). 71. Hollerman, J. R., Tremblay, L. & Schultz, W. Influence of
607624 (1990). 46. Schultz, W. Activity of pars reticulata neurons of monkey reward expectation on behavior-related neuronal activity in
20. Ljungberg, T., Apicella, P. & Schultz, W. Responses of substantia nigra in relation to motor, sensory and complex primate striatum. J. Neurophysiol. 80, 947963 (1998).
monkey dopamine neurons during learning of behavioural events. J. Neurophysiol. 55, 660677 (1986). 72. Tremblay, L., Hollerman, J. R. & Schultz, W. Modifications of
reactions. J. Neurophysiol. 67, 145163 (1992). 47. Niki, H., Sakai, M. & Kubota, K. Delayed alternation reward expectation-related neuronal activity during learning
21. Strecker, R. E. & Jacobs, B. L. Substantia nigra performance and unit activity of the caudate head and in primate striatum. J. Neurophysiol. 80, 964977 (1998).
dopaminergic unit activity in behaving cats: effect of arousal medial orbitofrontal gyrus in the monkey. Brain Res. 38, 73. Tremblay, L. & Schultz, W. Modifications of reward
on spontaneous discharge and sensory evoked activity. 343353 (1972). expectation-related neuronal activity during learning in
Brain Res. 361, 339350 (1985). 48. Niki, H. & Watanabe, M. Prefrontal and cingulate unit activity primate orbitofrontal cortex. J. Neurophysiol. 83,
22. Horvitz, J. C. Mesolimbocortical and nigrostriatal dopamine during timing behaviour in the monkey. Brain Res. 171, 18771885 (2000).
responses to salient non-reward events. Neuroscience 96, 213224 (1979). 74. Dickinson, A. & Balleine, B. Motivational control of goal-
651656 (2000). 49. Rosenkilde, C. E., Bauer, R. H. & Fuster, J. M. Single cell directed action. Animal Learn. Behav. 22, 118 (1994).
23. Mirenowicz, J. & Schultz, W. Preferential activation of activity in ventral prefrontal cortex of behaving monkeys. 75. Okano, K. & Tanji, J. Neuronal activities in the primate motor
midbrain dopamine neurons by appetitive rather than Brain Res. 209, 375394 (1981). fields of the agranular frontal cortex preceding visually
aversive stimuli. Nature 379, 449451 (1996). 50. Watanabe, M. The appropriateness of behavioural triggered and self-paced movement. Exp. Brain Res. 66,
24. Guarraci, F. A. & Kapp, B. S. An electrophysiological responses coded in post-trial activity of primate prefrontal 155166 (1987).
characterization of ventral tegmental area dopaminergic units. Neurosci. Lett. 101, 113117 (1989). 76. Romo, R. & Schultz, W. Neuronal activity preceding self-
neurons during differential Pavlovian fear conditioning in the 51. Tremblay, L. & Schultz, W. Reward-related neuronal activity initiated or externally timed arm movements in area 6 of
awake rabbit. Behav. Brain Res. 99, 169179 (1999). during gono go task performance in primate orbitofrontal monkey cortex. Exp. Brain Res. 67, 656662 (1987).
25. Schultz, W. Activity of dopamine neurons in the behaving cortex. J. Neurophysiol. 83, 18641876 (2000). 77. Kurata, K. & Wise, S. P. Premotor and supplementary motor
primate. Semin. Neurosci. 4, 129138 (1992). 52. Niki, H. & Watanabe, M. Cingulate unit activity and delayed cortex in rhesus monkeys: neuronal activity during
26. Redgrave, P., Prescott, T. J. & Gurney, K. Is the short-latency response. Brain Res. 110, 381386 (1976). externally- and internally-instructed motor tasks. Exp. Brain
dopamine response too short to signal reward? Trends 53. Nishijo, H., Ono, T. & Nishino, H. Single neuron responses in Res. 72, 237248 (1988).
Neurosci. 22, 146151 (1999). amygdala of alert monkey during complex sensory 78. Schultz, W. & Romo, R. Neuronal activity in the monkey
27. Hollerman, J. R. & Schultz, W. Dopamine neurons report an stimulation with affective significance. J. Neurosci. 8, striatum during the initiation of movements. Exp. Brain Res.
error in the temporal prediction of reward during learning. 35703583 (1988). 71, 431436 (1988).
Nature Neurosci. 1, 304309 (1998). A paper written by one of the few groups 79. Watanabe, M. Reward expectancy in primate prefrontal
28. Schultz,W. Predictive reward signal of dopamine neurons. investigating neurons in the primate amygdala in neurons. Nature 382, 629632 (1996).
J. Neurophysiol. 80, 127 (1998). relation to reward-related stimuli. They reported a The first demonstration that expected rewards
29. Schultz, W. & Dickinson, A. Neuronal coding of prediction number of different responses to the presentation of influence the behaviour-related activity of neurons in a
errors. Annu. Rev. Neurosci. 23, 473500 (2000). natural rewards. manner that is compatible with a goal-directed
account. Neurons in primate dorsolateral prefrontal movement selection and also for the influence of necessary and sufficient for phasic firing of single nucleus
cortex show different activities depending on the rewards on behavioural choices. accumbens neurons. Brain Res. 757, 280284 (1997).
expected reward during a typical prefrontal spatial 85. Kelley, A. E. Functional specificity of ventral striatal 93. West, M. O., Peoples, L. L., Michael, A. J., Chapin, J. K. &
delayed response task. compartments in appetitive behaviours. Ann. NY Acad. Sci. Woodward, D. J. Low-dose amphetamine elevates
80. Leon, M. I. & Shadlen, M. N. Effect of expected reward 877, 7190 (1999). movement-related firing of rat striatal neurons. Brain Res.
magnitude on the responses of neurons in the dorsolateral 86. Carelli, R. M., King, V. C., Hampson, R. E. & Deadwyler, S. A. 745, 331335 (1997).
prefrontal cortex of the macaque. Neuron 24, 415425 Firing patterns of nucleus accumbens neurons during 94. Peoples, L. L. & West, M. O. Phasic firing of single neurons
(1999). cocaine self-administration. Brain Res. 626, 1422 (1993). in the rat nucleus accumbens correlated with the timing of
81. Kawagoe, R., Takikawa, Y. & Hikosaka, O. Expectation of The first demonstration of drug-related changes in intravenous cocaine self-administration. J. Neurosci. 16,
reward modulates cognitive signals in the basal ganglia. neuronal activity in one of the key structures of drug 34593473 (1996).
Nature Neurosci. 1, 411416 (1998). addiction, the nucleus accumbens. Two important 95. Peoples, L. L., Gee, F., Bibi, R. & West, M. O. Phasic firing
82. Liu, Z. & Richmond, B. J. Response differences in monkey neuronal patterns were found in the rat: activity time locked to cocaine self-infusion and locomotion:
TE and perirhinal cortex: stimulus association related to preceding lever presses that led to acquisition of the dissociable firing patterns of single nucleus accumbens
reward schedules. J. Neurophysiol. 83, 16771692 (2000). drug and activity following drug delivery. These neurons in the rat. J. Neurosci. 18, 75887598 (1998).
The first demonstration of the prominent reward findings have been consistently reproduced by other 96. Calabresi, P., Maj, R., Pisani, A., Mercuri, N. B. & Bernardi, G.
relationships in neurons of temporal cortex. Neuronal goups. Long-term synaptic depression in the striatum: physiological
responses to task stimuli in the primate perirhinal 87. Chang, J. Y., Sawyer, S. F., Lee, R. S. & Woodward, D. J. and pharmacological characterization. J. Neurosci. 12,
cortex were profoundly affected by the distance to Electrophysiological and pharmacological evidence for the 42244233 (1992).
reward, whereas neuronal responses in the role of the nucleus accumbens in cocaine self- 97. Otmakhova, N. A. & Lisman, J. E. D1/D5 dopamine receptor
neighbouring TE area predominantly reflected the administration in freely moving rats. J. Neurosci. 14, activation increases the magnitude of early long-term
visual features of the stimuli. 12241244 (1994). potentiation at CA1 hippocampal synapses. J. Neurosci. 16,
83. Platt, M. L. & Glimcher, P. W. Neural correlates of 88. Carelli, R. M. & Deadwyler, S. A. A comparison of nucleus 74787486 (1996).
decision variables in parietal cortex. Nature 400, 233238 accumbens neuronal firing patterns during cocaine self- 98. Wickens, J. R., Begg, A. J. & Arbuthnott, G. W. Dopamine
(1999). administration and water reinforcement in rats. J. Neurosci. reverses the depression of rat corticostriatal synapses which
Neurons in primate parietal association cortex were 14, 77357746 (1994). normally follows high-frequency stimulation of cortex in vitro.
sensitive to two key variables of game theory and 89. Chang, J. Y., Janak, P. H. & Woodward, D. J. Comparison Neuroscience 70, 15 (1996).
decision making: the quantity and the probability of of mesocorticolimbic neuronal responses during cocaine 99. Calabresi, P. et al. Abnormal synaptic plasticity in the striatum
the outcomes. Based on studies of choice behaviour and heroin self-administration in freely moving rats. of mice lacking dopamine D2 receptors. J. Neurosci. 17,
and decision making in human economics and animal J. Neurosci. 18, 30983115 (1998). 45364544 (1997).
psychology, this is the first application of principles of 90. Carelli, R. M., Ijames, S. G. & Crumling, A. J. Evidence that
decision theory in primate neurophysiology. separate neural circuits in the nucleus accumbens encode Acknowledgements
84. Shima, K. & Tanji, J. Role for cingulate motor area cells in cocaine versus natural (water and food) reward. The crucial contributions of the collaborators in my own cited work
voluntary movement selection based on reward. Science J. Neurosci. 20, 42554266 (2000). are gratefully acknowledged, as are the collegial interactions with
282, 13351338 (1998). 91. Chang, J. Y., Paris, J. M., Sawyer, S. F., Kirillov, A. B. & Anthony Dickinson (Cambridge) and Masataka Watanabe (Tokyo).
Neurons in the primate cingulate motor area were Woodward, D. J. Neuronal spike activity in rat nucleus Our work was supported by the Swiss National Science
active when animals switched to a different accumbens during cocaine self-administration under different Foundation, European Community, McDonnellPew Program,
movement when continuing to perform the current fixed-ratio schedules. J. Neurosci. 74, 483497 (1996). Roche Research Foundation and British Council, and by postdoc-
movement would have produced less reward. This 92. Peoples, L. L., Uzwiak, A. J., Gee, F. & West, M. O. Operant toral fellowships from the US NIMH, FRS Quebec, Fyssen
study is interesting from the point of view of behaviour during sessions of intravenous cocaine infusion is Foundation and FRM Paris.