ABSTRACT

The ability to predict when and where a reward will occur enables humans and other animals to initiate behavioral responses prospectively in order to maximize the probability of obtaining that reward. Reward predictions can take a number of distinct forms depending on the nature of the associative relationship underpinning them (Balleine et al. 2008; and see Chapters 13 and 15 in this volume). The simplest form of reward prediction is one based on an associative “Pavlovian” relationship between arbitrary stimuli and rewards, acquired following experience of repeated contingent pairing of the stimulus with the reward. Subsequent presentation of the stimulus elicits a predictive representation of the reward, by virtue of the learned stimulus-reward association. This form of prediction is purely passive: it signals when a reward might be expected to occur and elicits Pavlovian conditioned re exes, but does not inform about the speci c behavioral actions that should be initiated in order to obtain it. By contrast, other forms of reward prediction are grounded in learned instrumental associations between stimuli, responses, and rewards, thereby informing about the speci c behavioral responses that, when performed by the animal, lead to a greater probability of obtaining that reward. Instrumental reward predictions can be either goal directed (based on response-outcome associations and therefore sensitive to the incentive value of the outcome), or habitual (based on stimulus-response associations and hence insensitive to changes in outcome value) (Balleine and Dickinson 1998). In this chapter, we will review evidence for the presence of multiple types of predictive reward signal in the brain. We will also outline some of the candidate computational mechanisms that might be responsible for the acquisition of these different forms of reward predictions and evaluate evidence for the presence of such mechanisms in the brain.