Open access peer-reviewed chapter

The Contribution of the Amygdala to Reward-Related Learning and Extinction

Written By

Rose Chesworth and Laura Corbit

Submitted: 26 October 2016 Reviewed: 10 February 2017 Published: 05 July 2017

DOI: 10.5772/67831

From the Edited Volume

The Amygdala - Where Emotions Shape Perception, Learning and Memories

Edited by Barbara Ferry

Chapter metrics overview

1,971 Chapter Downloads

View Full Metrics

Abstract

There has been substantial research into the role of the amygdala in fear conditioning and extinction of conditioned fear. The role of the amygdala in appetitive conditioning is relatively less explored. Here, we will review research into the role of the amygdala in reward‐related learning. Research to date suggests that the basolateral and central amygdala are responsible for learning about distinct aspects of a reinforcing event. For example, the basolateral amygdala is essential for distinguishing and choosing between specific rewards based on the specific‐sensory properties of those rewards as well as updating the relative value of specific rewarding events. In contrast, the central amygdala is involved in encoding reinforcement more generally and for regulating motivational influences on responding. We will also review what is known about the role of the amygdala in extinction of reward‐related behaviours and highlight areas for future research.

Keywords

  • reward
  • Pavlovian conditioning
  • instrumental learning
  • reinforcement
  • basolateral amygdala
  • central amygdala

1. Introduction

While less is known about the role of the amygdala in reward‐related learning compared to its role in fear conditioning where detailed circuitry has been mapped out, research to date nonetheless points to a very interesting and important function for the amygdala and distinct roles for sub‐nuclei of this structure. In this review chapter, we will focus on rodent studies (using rats and mice) examining the role of the basolateral (BLA) or central (CeA) amygdala in the formation and expression of both Pavlovian and instrumental associations, the effects of changes in reward magnitude or value on responding, and to changes in reward contingencies to define the role of the amygdala and subnuclei of this structure in learning about reward and control of reward‐seeking behaviours.

Advertisement

2. Basolateral nucleus of the amygdala

2.1. Involvement of BLA in encoding reward expectation in Pavlovian tasks

Neuronal firing in the BLA is elevated in response to reward‐predictive CS’s, occurring prior to reward delivery. Such activity is thought to drive reward‐seeking behaviours. For example, in an odour discrimination paradigm, BLA firing differs following presentation of stimuli that predict a positive outcome (sucrose) versus a negative outcome (quinine), suggesting the BLA is involved in learning about expectancy for consequences of a response [1]. Importantly, this discriminative neural activity precedes reliable behavioural discrimination, suggesting that the neural activity may support learning and behavioural change and is consistent with a role for the BLA in encoding of information about a reward predictive CS. Data regarding a causal role for this activity in Pavlovian learning are, however, more mixed.

A number of studies have found no evidence of any effect of BLA lesions on the acquisition of a Pavlovian response, with lesioned rats demonstrating food cup approach in the presence of a CS+ similar to that of sham rats ([24] see also [5]). The acquisition of Pavlovian autoshaping responses was also similar to that of controls [6]. However, other studies indicate BLA lesions or inactivation impairs acquisition of Pavlovian associations. Lesions of the BLA reduce rats’ preference for a flavoured solution paired with sucrose compared to sham controls, with no effect on consumption of CS+ and CS− solutions when these were not paired with sucrose, suggesting involvement of the BLA in Pavlovian associations between the flavour of the liquid and the sucrose reward [7]. BLA lesions impair taste‐potentiated odour aversion, and infusions of the GABA‐A agonist muscimol into the BLA indicate the BLA mediates the acquisition, but not the expression of taste‐potentiated odour aversion ([8], see also Ref. [9]). Also, lesions of the BLA impair the acquisition, but not expression of magazine approach in a task where discrete cues signal the location of a sucrose reward [10]. BLA lesions impair second‐order conditioning [3, 11, 12]; but this deficit is secondary to the role of the BLA in the assignment of motivational value to the first‐order CS+ without which motivational significance cannot then be transferred to the second‐order CS+ to produce conditioned responding [12]. Inactivation of the BLA with the NMDA antagonists AP‐5 or d,l‐2‐2‐amino‐5‐phosphonovalerate (D‐APV) impairs acquisition, but not expression of Pavlovian conditioned approach for sucrose or taste potentiated odour aversion [13, 14], while inhibition of dopamine D1 receptor activity with SCH‐23390 impairs acquisition of Pavlovian discriminative stimulus responding (i.e. approaching the food cup in the presence of a CS+, but not a CS−) [15]. D1 antagonism has no effect on responding when animals are trained further and tested drug free, suggesting a specific role for BLA D1 receptors in the performance of Pavlovian discriminative stimulus approach [15]. It appears acquisition of Pavlovian associations rely on a BLA to nucleus accumbens (NAcc) pathway, as optogenetic inactivation of the BLA to NAcc pathway using halorhodopsin impairs acquisition of licking behaviour for sucrose in response to sucrose predictive cues ([16], see also Ref. [17] for related function of this pathway). When optical stimulation was removed, licking returned to non‐stimulation levels, suggesting no long lasting effects of BLA to NAcc inhibition on subsequent task acquisition [16]. Thus, despite some inconsistencies in the literature, it appears that the BLA is required for acquisition, but not for expression of Pavlovian stimulus‐reward associations.

There is strong evidence implicating the BLA in the acquisition and expression of conditioned place preference (CPP) for a food reward. Electrolytic and neurotoxic lesions of the lateral amygdala (LA) impair acquisition of CPP for a food reward [18], while BLA lesions or inactivation performed after acquisition of CPP for a food reward impair expression of CPP [19, 20], suggesting that BLA is implicated in both the acquisition and expression of Pavlovian place learning. Furthermore, muscarinic receptors in the BLA are required for the consolidation of food CPP, as intra‐BLA scopolamine infusions following conditioning sessions impairs consolidation of food CPP [21]. Disconnection of the BLA from the NAcc also impairs expression of sucrose CPP, implicating this pathway in the expression of context‐food associations [19]. One study has demonstrated no effect of BLA lesions on place conditioning [10]; however, in this study, discrete cues were used to signal the presence of reward within a y‐maze, providing an alternative strategy by which the animals could solve the task and as discussed above, there are a number of studies demonstrating no effect of BLA lesions on Pavlovian learning using a discrete cue. As such, the majority of research suggests the BLA, and its projections to the NAcc are required for the acquisition and expression of CPP.

2.2. Involvement of BLA in instrumental learning

The involvement of the BLA in the acquisition and expression of instrumental appetitive learning has been extensively examined, with somewhat mixed results. For example, infusion of the NMDA antagonist AP‐5 [22] or the D1 antagonist SCH23390 [23] into the BLA prior to training has been reported to impair acquisition of lever pressing for sucrose pellets but once learning has occurred, BLA inactivation via AP‐5 or SCH‐23390 has no effect on the expression of action‐outcome contingencies, suggesting involvement of the BLA in task acquisition, but not expression [22, 23]. It is important to note that in these studies, performance of the lever‐press result produced not only primary reward but also a range of visual cues, e.g. offset of the houselight and onset of a stimulus light followed by sucrose at a 3 second delay. The presence of these stimuli as well as the delay in reward delivery makes it unclear whether the BLA is involved in acquisition of the instrumental response‐outcome contingency or these other aspects of the task. Indeed, other studies using a more pure instrumental design where the instrumental response produces reward without the presence of any stimuli or secondary reinforcers report no effect of BLA lesions on acquisition of instrumental responding for a single action‐outcome contingency (e.g. lever press delivers food pellets) ([2, 24], but see [25]). Furthermore, lesions of the BLA do not impair acquisition of instrumental responding when two action‐outcome contingencies earning distinct outcomes are trained (e.g. lever press delivers sucrose solution, chain pull delivers food pellets) [4, 26, 27]. In more complex discriminative stimulus tasks, where rats are required to initiate the correct action following stimulus presentation, BLA lesions or inactivation using the combined GABA A (muscimol) and B (baclofen) agonists [28, 29] or selective serotonin lesions of the BLA [30] do not impair task acquisition, suggesting that the BLA is not essential for stimulus‐response learning despite its role in stimulus‐reward learning. In contrast, BLA inactivation with muscimol or baclofen impairs expression of previously trained discriminative stimulus responding, suggesting that the BLA may contribute to task expression in a discriminative stimulus task but that when the BLA is inactivated during acquisition, the rats can solve the task, perhaps by using a different strategy [17].

2.3. Involvement of BLA in detecting changes in reward‐predictive nature of an action

In accordance with a role for the BLA in predicting reward, the BLA is also involved in detecting changes between an action and a rewarding outcome. This has been demonstrated through contingency degradation paradigms, in which the association between an action and its expected outcome is reduced through non‐contingent presentations of the reward. Lesions of the BLA performed prior to behavioural training impair contingency degradation, under conditions of extinction (when there is no opportunity to update action‐outcome contingencies) and also under conditions of partial reinforcement [26]. BLA lesions also impair contingency degradation when lesions are performed after acquisition of action‐outcome contingencies, ruling out any potential learning impairment which could impact the contingency degradation [31]. These studies suggest BLA involvement in detecting changes in the association between an action and expected reward.

2.4. Involvement of BLA in updating the value of an outcome

The BLA appears critical for generating internal representations of a reinforcer to guide choice. For example, BLA neurons are sensitive to reward magnitude. Rats in an eight arm radial maze exhibit differential BLA firing to rewards of high and low magnitude [32]. Also, BLA activity is altered in response to changes in expected reward magnitude (i.e. upshift or downshift in reward magnitude) [33]. However, once this contingency is learnt, BLA activity decreases, suggesting involvement of the BLA in the acquisition, but not expression, of reward magnitude changes [33]. Noradrenaline (NOR) is released in the BLA following an increase in the number of sucrose pellets delivered in an instrumental task, suggesting NOR in the BLA contributes to signalling changes in reward value [34].

In addition to changes in magnitude, the value of rewarding outcomes can change based on changes in the animal’s motivational state. Involvement of the BLA in updating the value of an outcome is demonstrated through incentive‐learning tasks. In these tasks, the value of a reinforcer is updated based on changes in internal states (e.g. hunger, satiety) and this information used to control goal‐directed responding. In an incentive learning paradigm, the experience of food consumption in a deprived state subsequently drives responding for that food in a future state of deprivation [35, 36]. Importantly, without the opportunity for direct contact with the reward in a novel motivational state, instrumental responding for that reward is not altered even when animals experience motivational state changes at test, e.g., responding is not increased despite an increase in hunger until the animal experiences that food while hungry. Consolidation and reconsolidation of this form of learning are blocked by intra‐BLA infusions of the protein synthesis inhibitor anisomycin [37]. Infusions of the non‐selective opioid antagonist naloxone into the BLA also impair acquisition but not retrieval of incentive learning [36]. This effect appears dependent on μ‐opioid receptors, as infusion of a μ‐receptor antagonist blocks acquisition of positive incentive learning (i.e. under conditions which enhance reinforcer value), and infusion of a μ‐receptor agonist impairs negative incentive learning (i.e. under conditions which reduce reinforcer value) [38]. K and δ antagonists have no effect on positive or negative incentive learning, suggesting a specific role for μ‐receptors in the BLA in incentive motivational processes [38].

Related results have been found in outcome devaluation studies. In this task, hungry rats are, for example, trained to perform two distinct instrumental responses each earning a unique food outcome. Rats are then pre‐fed one of these outcomes to satiety prior to a choice test where the two responses are available but no outcomes are delivered. BLA lesions or inactivation impairs sensitivity to outcome devaluation [4, 6, 26, 31, 39, 40]. This effect is observed when lesions are conducted both prior to instrumental learning, and after acquisition of instrumental learning [26, 31], indicating any potential effects of lesions on action‐outcome learning do not contribute to the loss of sensitivity to outcome devaluation. Also, BLA inactivation prior to the pre‐feeding treatment, but not after pre‐feeding but before the lever test impairs outcome devaluation, suggesting the BLA updates reinforcer value to guide choice, but that once reinforcer value has been updated, the BLA is no longer needed ([41], see also Ref. [6]). Such results have been taken as evidence that the BLA associates the specific sensory features of stimuli with motivational significance and updates this association as needed. This information can then be used to guide choice in outcome devaluation and related paradigms [26].

There are some reports of BLA lesions having no effect on outcome devaluation; however, it appears these results are due to differences in experimental parameters. When rats are trained in a Pavlovian magazine approach paradigm with a single reinforcer and devalued using lithium chloride (LiCl), BLA lesions do not impair outcome devaluation [3, 42]. There are a number of variables which could contribute to this apparent discrepancy, namely the method of training (Pavlovian or instrumental), the number of reinforcers used (one or two reinforcers) and the method of devaluation (LiCl or specific satiety). Johnson and colleagues [39] assessed the contribution of the method of training and method of devaluation to establish how these factors may help understand what aspect of learning requires BLA involvement. All rats were lesioned following training, to isolate the effect of the lesion to devaluation. All rats were trained with two reinforcers. Four groups of rats were used: rats were either given Pavlovian or instrumental training, and devalued via LiCl or specific satiety, creating the following four groups: Pavlovian‐LiCl, Pavlovian‐specific satiety, instrumental‐LiCl and instrumental‐specific satiety. BLA lesions impaired outcome devaluation in all four groups. As the only variable which was not manipulated was the number of reinforcers trained (in this case, two), this led the authors to suggest the number of reinforcers used may mediate whether the BLA is required for outcome devaluation. Indeed, if the BLA encodes sensory representations of a stimulus and associations of these with value, then successful devaluation performance may depend on the ability to generate sufficiently detailed outcome representations so that performance specifically related to the currently devalued outcome, but not other possible outcomes, being specifically affected. Thus, in the case of two reinforcers, the BLA is required to generate this specific representation so that animals can then directly respond to the two reinforcers based on specific sensory features in order to guide action selection as the motivational significance is largely overlapping. However, when only one reinforcer is present, no discrimination between reinforcers nor any association of value with the sensory features of the outcome is required, which may leave outcome devaluation intact when the BLA is offline ([39], for discussion, see Refs. [31, 43]).

2.5. The BLA, reward prediction and stimulus influences on responding

Despite some inconsistencies in the effects of BLA lesions or inactivation in the acquisition of instrumental responding, incentive learning and outcome devaluation tasks suggest that the BLA is important for assigning motivational significance to outcomes based on their specific sensory features. There is strong evidence to suggest that the BLA is important for assigning reward value to actions and external stimuli more generally. Early reports demonstrate that in an eight arm radial maze, BLA neurons exhibit enhanced firing to an anticipated reward encounter, implicating BLA activity in predicting rewarding events [32]. In operant tasks requiring rats to nose poke for sucrose, BLA firing is enhanced during reward expectation, but decreases when animals no longer anticipate reward delivery following their actions under behavioural extinction [44]. Also, LA neurons respond to reward predictive cues, and activity in LA neurons is associated with task efficiency and accuracy, as well as increased synaptic strength [45]. Finally, neurons activated in response to a discriminative stimulus fire in the BLA prior to the NAcc, suggesting the BLA drives NAcc neuronal responses to reward predictive cues to promote reward‐seeking behaviour [17].

Following detection of reward‐predictive stimuli, it appears glutamate transients in the BLA are involved in initiating reward‐seeking action. Glutamate transients in the BLA are enhanced during a seeking‐taking chain task for sucrose pellets, and glutamate transients tend to precede lever responses on both the distal lever (i.e. lever responses which gave access to the proximal lever) and proximal lever (i.e. lever responses on which are rewarded with sucrose) [46]. Furthermore, in a simple instrumental task, BLA glutamate transients are more likely to be associated with initiating the pressing bout, than with reward or non‐reward earning lever presses [47]. These data suggest glutamate signalling is critical for driving actions which lead to reward.

Experiments using outcome devaluation indicate involvement of the BLA in encoding the sensory specific properties of reinforcers. Further support for this notion is derived from Pavlovian‐instrumental transfer (PIT) experiments, where presentation of a CS previously paired with a reinforcer drives instrumental responding for the same reinforcer, despite the CS and instrumental response having never been trained together before. Importantly, when rats are trained with two reinforcers (i.e. two CSs are paired with two distinct rewards and two instrumental responses earn those same two rewards), responding during PIT can be identified as ‘specific’, with increased instrumental responding on the lever that, in training, delivered the same outcome as that predicted by the stimulus. In contrast, ‘general’ PIT is an elevation in responding that does not rely on a common outcome in instrumental and Pavlovian training phases (for a discussion, see [4, 48] #321). When rats are trained with multiple rewards, BLA lesions impair specific, but not general PIT [4, 27], suggesting the BLA is required for distinct stimuli to direct instrumental responding. Furthermore, blocking AMPA, but not NMDA receptors in the BLA inhibits PIT [47], and BLA glutamate transient frequency correlates with instrumental responding during the CS, which was trained with the same outcome, but not the different outcome [47]. Importantly, BLA glutamate transient frequency is enhanced during initiation of lever pressing, suggesting BLA engagement following detection of reward‐predictive stimuli which initiates goal‐directed responding [47]. It appears the involvement of the BLA in PIT is dependent on the number of reinforcers (stimuli and responses) trained, because when rats are trained with only one reinforcer, BLA lesions have no effect on PIT [2, 49]. When two reinforcers are trained, specific sensory properties need to be utilised to permit discrimination; this requires the BLA and is impaired following BLA lesions or inhibition. However, when only one reinforcer is trained, it is not necessary to distinguish between reinforcers via their sensory properties to direct responding; thus, this behaviour does not require the BLA and is therefore unimpaired following BLA lesions.

Consistent with PIT studies, BLA lesions abolish outcome‐guided responding in an outcome‐specific reinstatement task, in which outcome presentation selectively increases performance of a response previously associated with the same, but not a different outcome as that which was just presented [31], supporting the involvement of the BLA in the representation of sensory‐specific properties of stimuli and integration of those stimulus properties with motivational significance to direct choice behaviour.

2.6. The BLA mediates risky and effortful decision making

Experiments using delay discounting paradigms indicate the BLA guides choice towards high effort/high reward options. When rats are required to choose between high effort/high reward versus low effort/low reward options in a T‐maze, BLA lesions reduce choice for the high effort/high reward option [50, 51]. Similarly, when a high reward choice requires a longer delay, as in delay discounting paradigms, BLA lesions also bias choice to smaller, immediate reward [52, 53]. There can be some recovery from bias towards low effort/low reward options, suggesting the BLA is involved in the acquisition of the value of a reward in an effortful task [51]. Disconnection of the BLA from the medial prefrontal cortex or anterior cingulate cortex (ACC) also biases choice to smaller, immediate rewards ([50, 52] but see Ref. [54]), consistent with a role for these structures in effort‐based decision making [55].

Paradigms involving risky decision making indicate the BLA also guides choice towards high risk/high reward options. BLA inactivation with baclofen or muscimol induces a risk‐averse pattern of choice and, in a similar paradigm, reduces high effort choice, irrespective of the delay to reward [56]. It is possible that the BLA directs responding toward risk when loss is involved, as rats with BLA lesions bias their behaviour away from risk when loss was a consequence of a high risk choice, but do not alter their behaviour when potential gains are available [57]. BLA lesions or inactivation does not alter choice when two rewards are equal, or there is no risk involved [5658], suggesting a particular role for the BLA in biasing choice in the face of aversive consequences. It appears a BLA‐NAcc connection mediates BLA‐induced biasing of choice, as contralateral lesions of these structures biases choice toward a less risky option [54].

Some studies demonstrate BLA lesions enhance, rather than decrease risky decision making in a rodent gambling task [59], or when foot shock is used as punishment (instead of reward omission [58]). However, a recent study indicates individual differences may help explain these results. BLA inactivation can affect animals differently at an individual level—BLA inactivation increases effortful choice in rats which, at baseline, chose low effort/low reward options and BLA inactivation decreases effortful choice in rats which, at baseline, chose high effort/high reward options [60]. Furthermore, this biasing of choice appears dependent on BLA dopamine receptors. In risk‐averse rats, D1 agonist infusions into the BLA increase risky choice, whereas in risk‐prone rats, enhancing D1 activity reduces risky choice [61]. Also, infusions of the D2 agonist quinpirole reduce risky choice in risk‐prone rats [61]. It is possible dopamine receptors in the BLA mediate the interaction between costs and benefits in a task to generate subjective value which could differ between individuals or across experiments. Approaching the BLA as a mediator of decision making based on a cost/benefit analysis may explain why some studies report an increase in risky decision making following BLA inactivation—the effects of inactivation of this structure on behaviour may be dependent on task parameters which can bias decision making in a certain direction.

2.7. No consistent involvement of the BLA in reversal learning

Several studies have examined the role of the BLA in reversal learning; however, the results at present are inconsistent. One study demonstrates inactivation of the BLA with muscimol impairs reversal learning in an odour discrimination task [52]; however, another study demonstrates no effect of BLA lesions on reversal learning in a go/no‐go odour task [62]. Interestingly, in this study BLA lesions ameliorated impairments in reversal learning induced by orbitofrontal cortex lesions, suggesting projections between these two regions may control reversal learning [62]. In an operant nose‐poking discriminative stimulus task, BLA lesions have been shown to facilitate reversal learning, and limit the number of mistakes made following feedback on an incorrect trial [63]. However, in a similar nose‐poking discriminative stimulus paradigm, serotonin depletion in the BLA had no effect on reversal learning [30]. It appears the involvement of the BLA in reversal learning is not dependent on the task employed, as similar tasks (e.g. odour discrimination, operant nose poking) report inconsistent effects of BLA inactivation on reversal learning. Further research in this field is required to more conclusively determine the role of the BLA in reversal learning.

2.8. Involvement of the BLA in the appetitive extinction learning

A number of studies demonstrate that the BLA is critical for the acquisition of appetitive extinction learning. Excitotoxic lesions of the BLA enhance resistance to extinction learning when a magazine light and sucrose reinforcer are omitted, indicating BLA lesions impair extinction learning [28]. However, in this study, the use of excitotoxic lesions did not permit analysis of whether BLA lesions impair encoding, consolidation or retrieval of extinction learning [28]. Inactivation of the caudal BLA with bupivacaine (a sodium channel blocker) impairs acquisition, but not retrieval of extinction of instrumental responding, demonstrating BLA involvement of the acquisition of extinction learning [64]. In apparent contrast, intra‐BLA infusions of the NMDA partial agonist DCS, which should increase rather than decrease activation of BLA neurons, prior to extinction learning in an odour discrimination task has been reported to impair extinction and enhance responding at a retention session [65]. While a number of studies demonstrate DCS‐enhanced extinction learning (reviewed in Refs. [66, 67]), it appears the timing of DCS administration is critical in determining whether it enhances or impedes extinction learning (for a discussion, see Ref. [65]). Nonetheless, this study [65] demonstrates that NMDA receptors in the BLA contribute to extinction of appetitive learning. Finally, a subset of BLA neurons respond specifically during extinction of operant nose poking for sucrose; this subset does not respond during task acquisition and activity of these neurons is inversely correlated with responding during extinction [68]. These studies demonstrate a critical role for the BLA in detecting the absence of an expected reinforcer during instrumental appetitive extinction and are in agreement with the role of the BLA in detecting changes in reward value.

There is also some evidence to support a role for BLA signalling in the extinction of Pavlovian appetitive learning. For example, when rats are trained to lick for sucrose in the presence of a combined tone/light CS+, the firing of BLA neurons during extinction correlates strongly with extinction behaviour [69]. Furthermore, a subset of BLA neurons which responded during extinction also respond during reinstatement, suggesting the BLA is a site of plasticity mediating responding for motivationally significant stimuli [69]. These studies suggest that the BLA mediates aspects of Pavlovian appetitive extinction; however, further research in this field is required to determine the precise role of the BLA in appetitive extinction.

2.9. The BLA as part of a broader circuit involved in reward‐related learning

It is important to recognise that the BLA does not operate in isolation to control learning and performance. Below we highlight some example of interactions of the BLA with other structures. This section is not meant to comprehensive but to provide examples of how the BLA interacts with other brain areas. The BLA has dense projections to the posterior dorsomedial striatum (pDMS), insular cortex (IC) and NAcc [70], which, following detection of a change in reinforcer value in the BLA, mediate aspects of goal‐directed responding, such as knowledge of and engagement in action‐outcome contingencies. A BLA‐IC connection is required to encode and retrieve changes in reinforcer value [71]. BLA inactivation using the NR2B NMDA antagonist ifenprodil prior to specific satiety, but not prior to a choice test impairs outcome devaluation, suggesting BLA involvement in encoding changes in reinforcer value. However, ifenprodil infusions into IC prior to specific satiety or a choice test impair devaluation, suggesting that the IC mediates expression of devaluation. Finally, ifenprodil infused unilaterally into BLA prior to specific satiety and into the IC prior to a choice test blocks expression of devaluation, but ifenprodil infused into the IC prior to specific satiety and then into BLA prior to choice has no effect on expression of devaluation. This suggests the BLA updates and encodes information about reinforcer value during specific satiety, sending information to the IC prior to choice, and at test, the IC retrieves this information to guide choice between actions [71].

Also, connections between the BLA and posterior dorsomedial striatum (pDMS) are required to direct action‐outcome responding following a change in reinforcer value. The pDMS is critical for updating action‐outcome contingencies, as changes in response‐outcome associations are impaired following pDMS lesions [72, 73]. It appears the pDMS is required to retrieve action‐outcome associations following a change in reinforcer value, as unilateral lesions of the BLA coupled with inactivation of contralateral pDMS prior to a choice test impairs expression of outcome devaluation [73]. This suggests information from the BLA regarding the specific value of outcomes is transferred to the pDMS, which retrieves action‐outcome associations to guide instrumental performance [73].

Finally, a connection between the BLA and NAcc shell is necessary for action selection following reinforcer devaluation. Disconnection of the BLA and NAcc via contralateral excitotoxic lesions impairs outcome devaluation, without reducing overall responding [74]. It is possible the BLA conveys sensory‐specific outcome information and/or changes in reinforcer value to the NAcc, where it is used to direct outcome‐appropriate instrumental responding [74]. Previous reports demonstrate that NAcc shell lesions impair the ability for action‐outcome cues to bias action selection in Pavlovian‐instrumental transfer [75], supporting the NAcc shell being a limbic‐motor interface structure [76]. Thus, it appears sensory‐specific information from the BLA is used to drive action selection in the NAcc shell, which can direct actions through motor output structures such as the ventral pallidum and medial dorsal thalamus.

Advertisement

3. Central nucleus of the amygdala

3.1. The CeA in Pavlovian learning

The CeA is involved in conditioning with both appetitive and aversive reinforcement [77], and one proposed role for the CeA is determining the valence of reinforcing events. For example, c‐fos immunoreactivity is increased in the medial CeA following exposure to a CS+ signalling food delivery, compared to a CS which did not signal food delivery [78]. CeA c‐fos immunoreactivity is also increased following exposure to a CS+ signalling foot shock, particularly in ventral regions of the structure, suggesting sub‐regions of the CeA may detect the valence of a CS [78]. We focus here on data related to appetitive learning.

When a visual or auditory CS is paired with food, rodents can acquire distinctive behaviours to CS presentation; they may orient themselves to the CS, either by approaching or rearing to a light or startling in response to a tone (orienting responses) or approach the site of food delivery, usually a food cup or magazine (conditioned approach). There is considerable evidence to support a role for the CeA in conditioned orienting responses to a Pavlovian CS+, but not for conditioned approach. Additionally, CeA lesions do not impair second‐order Pavlovian conditioned approach [3]. Lesions of the CeA prior to training impair the acquisition of conditioned orienting responses (e.g. rearing to a light), but leave conditioned approach intact [7981]. Similarly, inactivation of the CeA with the AMPA antagonist NBQX impairs acquisition of orienting responses [80]. CeA lesions or inactivation after Pavlovian training have no effect on the expression of Pavlovian orienting responses or food cup approach, suggesting a role for the CeA in the acquisition, rather than expression of orienting responses [80]. While some studies report no effect of CeA lesions on Pavlovian learning, these studies have only assessed conditioned approach behaviour ([2, 82] see also [7]), supporting a dissociation between conditioned approach and conditioned orienting responses in the CeA.

The CeA may be involved in conditioned approach behaviour when rats are trained to approach the magazine following the presentation of a CS+, but not a CS−, and a discrimination score is created which depicts their approach following one CS presentation over another CS. In Pavlovian approach paradigms, each CS+ is associated with a reinforcer, but not with the absence of reinforcement. CeA lesions or intra‐CeA inactivation of D1 or D3 receptors reduces conditioned approach behaviour [15, 83, 84]. If the CeA is involved in discriminating positive or negative reward value (discussed below), CeA inactivation may impair this discrimination, leading to a lower discrimination score. Supporting this interpretation, Andrzejewski and colleagues reported equal nose poking rates between the CS+ and CS−, rather than an abolition of nose poking [15], which would support lack of discrimination between the two CSs but not an inhibition of nose poking following CeA inactivation. It is also possible that these effects relate specifically to dopamine function within the CeA.

3.2. Circuitry mediating conditioned orienting responses

In rats injected with fluorogold, a retrograde tracer, into the substantia nigra pars compacta (SNc) there was a greater number of c‐fos positive/fluorogold positive cells in the CeA following food‐tone pairings than unpaired food and tone presentations, implicating this pathway in conditioning [85]. Furthermore, contralateral lesions to disconnect the CeA and SNc impair orienting responses but not food cup approach, compared to an ipsilateral lesion control group [85]. Considering that the CeA has a substantial projection to the SNc that provides dopaminergic innervation to the dorsolateral straitum (DLS) [86, 87], it is possible that a CeA‐SNc‐DLS pathway mediates orienting responses to Pavlovian food CS’s. Evidence for this comes from the demonstration that unilateral lesions of the CeA coupled with dopamine depletion in the DLS in the opposing hemisphere impairs conditioned orienting responses for food pellets, while leaving food cup approach behaviour intact [88]. Similar results were obtained when the DLS was reversibly inactivated with lidocaine [88]. Recovery of conditioned orienting responses occurred on drug free days in rats previously treated with intra‐DLS lidocaine, suggesting no long lasting effects of CeA‐DLS inactivation on acquisition of conditioned orienting [88]. Together, these results suggest orienting responses to a Pavlovian cue are mediated by indirect connections between the CeA and the DLS likely via the SNc.

3.3. The CeA in instrumental learning

The CeA does not appear to be critical for the acquisition of instrumental action‐outcome contingencies. Lesions of the CeA do not impair instrumental learning when there is a single action‐outcome contingency (i.e. one lever, one reinforcer) [2, 82] or two action‐outcome contingencies (i.e. two levers, two reinforcers) [4]. There is some evidence that the CeA is involved in updating action‐outcome contingencies. The omission of an expected reward at test enhances c‐fos immunoreactivity in the CeA, suggesting CeA detects the lack of reward [89]. Furthermore, fluorogold injections in the SNc demonstrate these c‐fos positive CeA cells project to the SNc [89], implicating a CeA‐SNc pathway in the detection of changes in reward contingencies. Lesions of the CeA produce a mild impairment in performance when an expected reward of small magnitude is omitted [90]; however, lesions of the entire CeA and BLA combined substantially reduce sensitivity to omission and so the specific contribution of the CeA is somewhat unclear [91].

CeA involvement in the detection of changes in reward value appears to depend on the paradigm used to assess this change. In studies using the outcome devaluation task, where an outcome is devalued either with selective satiety or LiCl‐induced sickness, CeA lesions do not impair behavioural sensitivity to changes in reward value [3, 82], indicating no role in this evaluative process and further substantiating intact action‐outcome learning necessary for performance in this task. Of interest, CeA lesions prevent loss of sensitivity to outcome devaluation that typically occurs with over‐training, suggesting a role for the CeA in habitual behaviour [82]. Similar effects are observed following disconnection of the CeA from the DLS, produced by contralateral lesions of these structures, suggesting that the CeA sends a reinforcement signal to the DLS to strengthen the stimulus‐response (S‐R) association that is thought to underlie habit learning [82].

While the CeA is not necessary for normal sensitivity to devaluation, it is involved in learning about changes in the magnitude of reward. When rats are trained to run in a straight alley maze task for a large food reward, a downward shift in the magnitude of the food reward increases the latency of intact rats to reach the smaller reward [92]. Post‐shift lidocaine infusions into the amygdala, which were mostly aimed at the CeA, reduce the latency to reach a smaller reward, suggesting reduced sensitivity to the change in reward magnitude [92]. Similarly, pre‐training CeA lesions slow learning about a downward shift in reward magnitude in a straight alley maze, supporting a role for the CeA in detecting reward magnitude changes [93, 94]. More recently, optogenetic stimulation of CeA with channelrhodopsin was shown to enhance lever pressing for a sucrose pellet when both the delivery and consumption of this pellet were paired with laser stimulation, compared to delivery of sucrose pellets alone, suggesting CeA stimulation may enhance the perceived magnitude of a reward [95]. Finally, the μ‐opioid agonist DAMGO administered within the CeA enhances sniffing and nibbling at a food cup or reward predictive lever, suggesting enhanced reward value attributed to these stimuli following CeA μ‐opioid stimulation [96]. Collectively, these studies implicate the CeA in processing changes in reward magnitude. Performance may be spared where tasks rely on discrimination and choice between rewards based their relative value and distinguished using specific sensory properties. Such tasks rely instead on the BLA as described above. Together these findings are consistent with the idea that whereas the BLA is responsible for assigning and updating the value of specific outcomes based on their sensory properties, the CeA is responsible for a less specific reinforcement signal, accounting for its role in both habit learning and adjusting performance following changes in reward magnitude [43, 82].

3.4. CeA involvement in stimulus influences on instrumental responding

The CeA plays a role in signalling the general motivational information carried by stimuli but consistent with the studies above, not in detailed representation of the specific features of distinct rewards or their representation. Evidence for this comes from Pavlovian‐instrumental transfer (PIT) tasks, which assess control of instrumental responding by Pavlovian cues, despite the two types of training being conducted separately. PIT occurs when presentations of a CS (previously paired with a US) drives instrumental responding which was previously trained to obtain the same US. The involvement of the CeA in PIT is dependent on the type of PIT being examined. When rats are trained on one instrumental action‐outcome contingency and undergo Pavlovian training which involves one CS‐US association, CeA lesions impair PIT [2, 48, 82], while intra‐CeA infusion of the μ‐opioid agonist DAMGO enhance PIT [97]. However, when rats are trained on two instrumental action‐outcome contingencies and two Pavlovian CS‐US pairings, lesions of the CeA have no effect the outcome‐specific PIT that is generated by this type of training [4]. Importantly, in an experimental design where a third excitatory CS+ is introduced in the Pavlovian training phase and paired with a third reward not earned by either instrumental response, CeA lesions impair responding to the third CS, but leave outcome‐specific PIT intact, suggesting the CeA is involved in general appetitive arousal rather than directing outcome‐specific responding [4]. This suggestion also accounts for the experiments described above which use only one action‐outcome contingency and one CS‐US pairing; when there is no choice between CS driven responses, a reduction in general appetitive motivation reduces responding in general. These findings suggest the CeA encodes a reinforcement signal which is devoid of specific details about an outcome.

3.5. Some involvement of the CeA extinction of appetitive learning

Several electrophysiological studies implicate the CeA in the extinction of appetitive learning. For example, Toyomitsu et al. [69] recorded from neurons in the BLA, LA and CeA during extinction of Pavlovian licking for sucrose reward. CeA firing during extinction correlated with extinction of licking; however, this correlation was not as strong as BLA firing [69]. Furthermore, while there were changes in the firing rate of CeA neurons between extinction and reinstatement, this was not as pronounced as in BLA neurons [69]. Calu et al. [98] recorded from the CeA during an over‐expectation task where initially multiple stimuli are trained as independent predictors of reward (e.g. individual stimuli such as a tone, light, etc., each predict a food pellet). The critical manipulation comes when two or more of these stimuli are then presented together, as a compound. Animals typically increase responding to such a compound indicating that they expect more reward based on the multiple predictors (e.g. since tone and light alone previously predicted one pellet, the two stimuli together should predict two pellets). However, if this compound is followed only by the original reward (one pellet) behavioural responding decreases across trials, as does activity of the CeA [98]. Together with data from extinction paradigms, this suggests that the CeA may signal reward reduction in general, not just reward omission. A recent study by Iordanova et al. [99] examined this possibility by exploring the role of the CeA in updating reward expectancies following a reduction in reward achieved either through extinction, where reward was omitted entirely, or through generating over‐expectation where, due to the presence of multiple predictors, a large reward is expected but not received. In both paradigms, the majority of recorded cells showed an increase in firing during the period where food reward was delivered and also during the preceding stimulus presentations reflecting reward expectancy. Neural firing to the extinction stimulus was reduced across trials compared to a control stimulus. When a combination of previously rewarded stimuli was introduced to generate over‐expectation, neural firing to this compound was increased relative to a control compound in early trials but then equivalent in later trials, presumably as animals came to expect the reduced reward that was received. A subpopulation of the reward‐responsive cells showed a reduction in firing to both the extinction and over expectation trials, suggesting a common role in signalling reduced reward expectation. Importantly, this change in neural activity preceded and predicted the decline in behavioural responding observed under both extinction and over‐expectation conditions. Because these conditions involved the delivery of different amounts of reward (no reward in extinction whereas reward was still delivered in over‐expectation, albeit less than initially expected based on the stimuli) the similar changes in neural activity are unlikely to reflect absolute reward magnitude, but rather may signal the reduction in reward expectancy. It is possible that this reduction in reward creates an aversive motivational state, in which case these findings could be consistent with a more general role for the CeA in emotional learning [99].

Advertisement

4. Conclusion

While less is known about the role of the amygdala in reward‐related learning compared to its role in fear conditioning where detailed circuitry has been mapped out, research to date nonetheless points to a very interesting and important function for the amygdala. For example, the basolateral amygdala is involved in associating sensory‐specific aspects of different outcomes with the rewarding effects of that outcome, a function critical for choice between alternatives and behavioural control more generally. Further, the amygdala appears to be involved in updating representations of value both when the value of the outcome is changed, for example, following devaluation, or when the relationship between predictors and outcome delivery is changed, as in extinction. Thus, the amygdala plays an important role in reward‐related learning. With the advent of tools such as optogenetics, researchers can now go on to explore how these functions are achieved within the complex circuitry of the amygdala and associated structures.

References

  1. 1. Schoenbaum G, Chiba AA, Gallagher M. Orbitofrontal cortex and basolateral amygdala encode expected outcomes during learning. Nat Neurosci. 1998;1(2):155-9.
  2. 2. Hall J, Parkinson JA, Connor TM, Dickinson A, Everitt BJ. Involvement of the central nucleus of the amygdala and nucleus accumbens core in mediating Pavlovian influences on instrumental behaviour. Eur J Neurosci. 2001;13(10):1984-92.
  3. 3. Hatfield T, Han JS, Conley M, Gallagher M, Holland P. Neurotoxic lesions of basolateral, but not central, amygdala interfere with Pavlovian second‐order conditioning and reinforcer devaluation effects. J Neurosci. 1996;16(16):5256-65.
  4. 4. Corbit LH, Balleine BW. Double dissociation of basolateral and central amygdala lesions on the general and outcome‐specific forms of pavlovian‐instrumental transfer. J Neurosci. 2005;25(4):962-70.
  5. 5. Mahler SV, Berridge KC. What and when to “want”? Amygdala‐based focusing of incentive salience upon sugar and sex. Psychopharmacology (Berl). 2012;221(3):407-26.
  6. 6. Blundell P, Hall G, Killcross S. Preserved sensitivity to outcome value after lesions of the basolateral amygdala. J Neurosci. 2003;23(20):7702-9.
  7. 7. Dwyer DM. Lesions of the basolateral, but not central, amygdala impair flavour‐taste learning based on fructose or quinine reinforcers. Behav Brain Res. 2011;220(2):349-53.
  8. 8. Ferry B, Sandner G, Di Scala G. Neuroanatomical and functional specificity of the basolateral amygdaloid nucleus in taste‐potentiated odor aversion. Neurobiol Learn Mem. 1995;64(2):169-80.
  9. 9. Ferry B, Wirth S, Di Scala G. Functional interaction between entorhinal cortex and basolateral amygdala during trace conditioning of odor aversion in the rat. Behav Neurosci. 1999;113(1):118-25.
  10. 10. Ito R, Robbins TW, McNaughton BL, Everitt BJ. Selective excitotoxic lesions of the hippocampus and basolateral amygdala have dissociable effects on appetitive cue and place conditioning based on path integration in a novel Y‐maze procedure. Eur J Neurosci. 2006;23(11):3071-80.
  11. 11. Holland PC, Hatfield T, Gallagher M. Rats with basolateral amygdala lesions show normal increases in conditioned stimulus processing but reduced conditioned potentiation of eating. Behav Neurosci. 2001;115(4):945-50.
  12. 12. Setlow B, Gallagher M, Holland PC. The basolateral complex of the amygdala is necessary for acquisition but not expression of CS motivational value in appetitive Pavlovian second‐order conditioning. Eur J Neurosci. 2002;15(11):1841-53.
  13. 13. Burns LH, Everitt BJ, Robbins TW. Intra‐amygdala infusion of the N‐methyl‐D‐aspartate receptor antagonist AP5 impairs acquisition but not performance of discriminated approach to an appetitive CS. Behav Neural Biol. 1994;61(3):242-50.
  14. 14. Ferry B, Di Scala G. Basolateral amygdala NMDA receptors are selectively involved in the acquisition of taste‐potentiated odor aversion in the rat. Behav Neurosci. 2000;114(5):1005-10.
  15. 15. Andrzejewski ME, Ryals C. Dissociable hippocampal and amygdalar D1‐like receptor contribution to discriminated Pavlovian conditioned approach learning. Behav Brain Res. 2016;299:111-21.
  16. 16. Stuber GD, Sparta DR, Stamatakis AM, van Leeuwen WA, Hardjoprajitno JE, Cho S, et al. Excitatory transmission from the amygdala to nucleus accumbens facilitates reward seeking. Nature. 2011;475(7356):377-80.
  17. 17. Ambroggi F, Ishikawa A, Fields HL, Nicola SM. Basolateral amygdala neurons facilitate reward‐seeking behavior by exciting nucleus accumbens neurons. Neuron. 2008;59(4):648-61.
  18. 18. White NM, McDonald RJ. Acquisition of a spatial conditioned place preference is impaired by amygdala lesions and improved by fornix lesions. Behav Brain Res. 1993;55(2):269-81.
  19. 19. Everitt BJ, Morris KA, O’Brien A, Robbins TW. The basolateral amygdala‐ventral striatal system and conditioned place preference: further evidence of limbic‐striatal interactions underlying reward‐related processes. Neuroscience. 1991;42(1):1-18.
  20. 20. Kantak KM, Green‐Jordan K, Valencia E, Kremin T, Eichenbaum HB. Cognitive task performance after lidocaine‐induced inactivation of different sites within the basolateral amygdala and dorsal striatum. Behav Neurosci. 2001;115(3):589-601.
  21. 21. Schroeder JP, Packard MG. Posttraining intra‐basolateral amygdala scopolamine impairs food‐ and amphetamine‐induced conditioned place preferences. Behav Neurosci. 2002;116(5):922-7.
  22. 22. Baldwin AE, Holahan MR, Sadeghian K, Kelley AE. N‐methyl‐d‐aspartate receptor‐dependent plasticity within a distributed corticostriatal network mediates appetitive instrumental learning. Behav Neurosci. 2000;114(1):84-98.
  23. 23. Andrzejewski ME, Spencer RC, Kelley AE. Instrumental learning, but not performance, requires dopamine D1‐receptor activation in the amygdala. Neuroscience. 2005;135(2):335-45.
  24. 24. Coutureau E, Marchand AR, Di Scala G. Goal‐directed responding is sensitive to lesions to the prelimbic cortex or basolateral nucleus of the amygdala but not to their disconnection. Behav Neurosci. 2009;123(2):443-8.
  25. 25. Simmons DA, Neill DB. Functional interaction between the basolateral amygdala and the nucleus accumbens underlies incentive motivation for food reward on a fixed ratio schedule. Neuroscience. 2009;159(4):1264-73.
  26. 26. Balleine BW, Killcross AS, Dickinson A. The effect of lesions of the basolateral amygdala on instrumental conditioning. J Neurosci. 2003;23(2):666-75.
  27. 27. Blundell P, Hall G, Killcross S. Lesions of the basolateral amygdala disrupt selective aspects of reinforcer representation in rats. J Neurosci. 2001;21(22):9018-26.
  28. 28. Burns LH, Everitt BJ, Robbins TW. Effects of excitotoxic lesions of the basolateral amygdala on conditional discrimination learning with primary and conditioned reinforcement. Behav Brain Res. 1999;100(1-2):123-33.
  29. 29. Jones JL, Day JJ, Aragona BJ, Wheeler RA, Wightman RM, Carelli RM. Basolateral amygdala modulates terminal dopamine release in the nucleus accumbens and conditioned responding. Biol Psychiatry. 2010;67(8):737-44.
  30. 30. Ochoa JG, Stolyarova A, Kaur A, Hart EE, Bugarin A, Izquierdo A. Post‐training depletions of basolateral amygdala serotonin fail to disrupt discrimination, retention, or reversal learning. Front Neurosci. 2015;9:155.
  31. 31. Ostlund SB, Balleine BW. Differential involvement of the basolateral amygdala and mediodorsal thalamus in instrumental action selection. J Neurosci. 2008;28(17):4398-405.
  32. 32. Pratt WE, Mizumori SJ. Characteristics of basolateral amygdala neuronal firing on a spatial memory task involving differential reward. Behav Neurosci. 1998;112(3):554-70.
  33. 33. Roesch MR, Calu DJ, Esber GR, Schoenbaum G. Neural correlates of variations in event processing during learning in basolateral amygdala. J Neurosci. 2010;30(7):2464-71.
  34. 34. Young EJ, Williams CL. Valence dependent asymmetric release of norepinephrine in the basolateral amygdala. Behav Neurosci. 2010;124(5):633-44.
  35. 35. Balleine B. Instrumental performance following a shift in primary motivation depends on incentive learning. J Exp Psychol Anim Behav Process. 1992;18(3):236-50.
  36. 36. Wassum KM, Ostlund SB, Maidment NT, Balleine BW. Distinct opioid circuits determine the palatability and the desirability of rewarding events. Proc Nat Acad Sci U S A. 2009;106(30):12512-7.
  37. 37. Wang SH, Ostlund SB, Nader K, Balleine BW. Consolidation and reconsolidation of incentive learning in the amygdala. J Neurosci. 2005;25(4):830-5.
  38. 38. Wassum KM, Cely IC, Balleine BW, Maidment NT. Micro‐opioid receptor activation in the basolateral amygdala mediates the learning of increases but not decreases in the incentive value of a food reward. J Neurosci. 2011;31(5):1591-9.
  39. 39. Johnson AW, Gallagher M, Holland PC. The basolateral amygdala is critical to the expression of pavlovian and instrumental outcome‐specific reinforcer devaluation effects. J Neurosci. 2009;29(3):696-704.
  40. 40. West EA, Forcelli PA, Murnen AT, McCue DL, Gale K, Malkova L. Transient inactivation of basolateral amygdala during selective satiation disrupts reinforcer devaluation in rats. Behav Neurosci. 2012;126(4):563-74.
  41. 41. West EA, Forcelli PA, McCue DL, Malkova L. Differential effects of serotonin‐specific and excitotoxic lesions of OFC on conditioned reinforcer devaluation and extinction in rats. Behav Brain Res. 2013;246:10-4.
  42. 42. Pickens CL, Saddoris MP, Setlow B, Gallagher M, Holland PC, Schoenbaum G. Different roles for orbitofrontal cortex and basolateral amygdala in a reinforcer devaluation task. J Neurosci. 2003;23(35):11078-84.
  43. 43. Balleine BW, Killcross S. Parallel incentive processing: an integrated view of amygdala function. Trends Neurosci. 2006;29(5):272-9.
  44. 44. Tye KM, Janak PH. Amygdala neurons differentially encode motivation and reinforcement. J Neurosci. 2007;27(15):3937-45.
  45. 45. Tye KM, Stuber GD, de Ridder B, Bonci A, Janak PH. Rapid strengthening of thalamo‐amygdala synapses mediates cue‐reward learning. Nature. 2008;453(7199):1253-7.
  46. 46. Wassum KM, Tolosa VM, Tseng TC, Balleine BW, Monbouquette HG, Maidment NT. Transient extracellular glutamate events in the basolateral amygdala track reward‐seeking actions. J Neurosci. 2012;32(8):2734-46.
  47. 47. Malvaez M, Greenfield VY, Wang AS, Yorita AM, Feng L, Linker KE, et al. Basolateral amygdala rapid glutamate release encodes an outcome‐specific representation vital for reward‐predictive cues to selectively invigorate reward‐seeking actions. Sci Rep. 2016;6:20891.
  48. 48. Corbit LH, Balleine BW. Learning and motivational processes contributing to pavlovian-instrumental transfer and their neural bases: dopamine and beyond. Curr Top Behav Neurosci. 2016;27:259-89.
  49. 49. Holland PC, Gallagher M. Double dissociation of the effects of lesions of basolateral and central amygdala on conditioned stimulus‐potentiated feeding and Pavlovian‐instrumental transfer. Eur J Neurosci. 2003;17(8):1680-94.
  50. 50. Floresco SB, Ghods‐Sharifi S. Amygdala‐prefrontal cortical circuitry regulates effort‐based decision making. Cereb Cortex. 2007;17(2):251-60.
  51. 51. Ostrander S, Cazares VA, Kim C, Cheung S, Gonzalez I, Izquierdo A. Orbitofrontal cortex and basolateral amygdala lesions result in suboptimal and dissociable reward choices on cue‐guided effort in rats. Behav Neurosci. 2011;125(3):350-9.
  52. 52. Churchwell JC, Morris AM, Heurtelou NM, Kesner RP. Interactions between the prefrontal cortex and amygdala during delay discounting and reversal. Behav Neurosci. 2009;123(6):1185-96.
  53. 53. Winstanley CA, Theobald DE, Cardinal RN, Robbins TW. Contrasting roles of basolateral amygdala and orbitofrontal cortex in impulsive choice. J Neurosci. 2004;24(20):4718-22.
  54. 54. St Onge JR, Stopper CM, Zahm DS, Floresco SB. Separate prefrontal‐subcortical circuits mediate different components of risk‐based decision making. J Neurosci. 2012;32(8):2886-99.
  55. 55. Bailey MR, Simpson EH, Balsam PD. Neural substrates underlying effort, time, and risk‐based decision making in motivated behavior. Neurobiol Learn Mem. 2016;133:233-56.
  56. 56. Ghods‐Sharifi S, St Onge JR, Floresco SB. Fundamental contribution by the basolateral amygdala to different forms of decision making. J Neurosci. 2009;29(16):5251-9.
  57. 57. Tremblay M, Cocker PJ, Hosking JG, Zeeb FD, Rogers RD, Winstanley CA. Dissociable effects of basolateral amygdala lesions on decision making biases in rats when loss or gain is emphasized. Cogn Affect Behav Neurosci. 2014;14(4):1184-95.
  58. 58. Orsini CA, Trotta RT, Bizon JL, Setlow B. Dissociable roles for the basolateral amygdala and orbitofrontal cortex in decision‐making under risk of punishment. J Neurosci. 2015;35(4):1368-79.
  59. 59. Zeeb FD, Winstanley CA. Lesions of the basolateral amygdala and orbitofrontal cortex differentially affect acquisition and performance of a rodent gambling task. J Neurosci. 2011;31(6):2197-204.
  60. 60. Hosking JG, Cocker PJ, Winstanley CA. Dissociable contributions of anterior cingulate cortex and basolateral amygdala on a rodent cost/benefit decision‐making task of cognitive effort. Neuropsychopharmacology. 2014;39(7):1558-67.
  61. 61. Larkin JD, Jenni NL, Floresco SB. Modulation of risk/reward decision making by dopaminergic transmission within the basolateral amygdala. Psychopharmacology (Berl). 2016;233(1):121-36.
  62. 62. Stalnaker TA, Franz TM, Singh T, Schoenbaum G. Basolateral amygdala lesions abolish orbitofrontal‐dependent reversal impairments. Neuron. 2007;54(1):51-8.
  63. 63. Izquierdo A, Darling C, Manos N, Pozos H, Kim C, Ostrander S, et al. Basolateral amygdala lesions facilitate reward choices after negative feedback in rats. J Neurosci. 2013;33(9):4105-9.
  64. 64. McLaughlin RJ, Floresco SB. The role of different subregions of the basolateral amygdala in cue‐induced reinstatement and extinction of food‐seeking behavior. Neuroscience. 2007;146(4):1484-94.
  65. 65. Portero‐Tresserra M, Marti‐Nicolovius M, Guillazo‐Blanch G, Boadas‐Vaello P, Vale‐Martinez A. D‐cycloserine in the basolateral amygdala prevents extinction and enhances reconsolidation of odor‐reward associative learning in rats. Neurobiol Learn Mem. 2013;100:1-11.
  66. 66. Chesworth R, Corbit LH. Recent developments in the behavioural and pharmacological enhancement of extinction of drug‐seeking. Addict Biol. 2017;22(1):3-43.
  67. 67. Myers KM, Carlezon WA, Jr. D‐cycloserine effects on extinction of conditioned responses to drug‐related cues. Biol Psychiatry. 2012;71(11):947-55.
  68. 68. Tye KM, Cone JJ, Schairer WW, Janak PH. Amygdala neural encoding of the absence of reward during extinction. J Neurosci. 2010;30(1):116-25.
  69. 69. Toyomitsu Y, Nishijo H, Uwano T, Kuratsu J, Ono T. Neuronal responses of the rat amygdala during extinction and reassociation learning in elementary and configural associative tasks. Eur J Neurosci. 2002;15(4):753-68.
  70. 70. Kelley AE, Domesick VB, Nauta WJ. The amygdalostriatal projection in the rat—an anatomical study by anterograde and retrograde tracing methods. Neuroscience. 1982;7(3):615-30.
  71. 71. Parkes SL, Balleine BW. Incentive memory: evidence the basolateral amygdala encodes and the insular cortex retrieves outcome values to guide choice between goal‐directed actions. J Neurosci. 2013;33(20):8753-63.
  72. 72. Corbit LH, Janak PH. Posterior dorsomedial striatum is critical for both selective instrumental and Pavlovian reward learning. Eur J Neurosci. 2010;31(7):1312-21.
  73. 73. Corbit LH, Leung BK, Balleine BW. The role of the amygdala‐striatal pathway in the acquisition and performance of goal‐directed instrumental actions. J Neurosci. 2013;33(45):17682-90.
  74. 74. Shiflett MW, Balleine BW. At the limbic‐motor interface: disconnection of basolateral amygdala from nucleus accumbens core and shell reveals dissociable components of incentive motivation. Eur J Neurosci. 2010;32(10):1735-43.
  75. 75. Corbit LH, Muir JL, Balleine BW. The role of the nucleus accumbens in instrumental conditioning: Evidence of a functional dissociation between accumbens core and shell. J Neurosci. 2001;21(9):3251-60.
  76. 76. Mogenson GJ, Jones DL, Yim CY. From motivation to action: functional interface between the limbic system and the motor system. Prog Neurobiol. 1980;14(2-3):69-97.
  77. 77. Shabel SJ, Janak PH. Substantial similarity in amygdala neuronal activity during conditioned appetitive and aversive emotional arousal. Proc Nat Acad Sci U S A. 2009;106(35):15031-36.
  78. 78. Knapska E, Walasek G, Nikolaev E, Neuhausser‐Wespy F, Lipp HP, Kaczmarek L, et al. Differential involvement of the central amygdala in appetitive versus aversive learning. Learn Mem. 2006;13(2):192-200.
  79. 79. Gallagher M, Graham PW, Holland PC. The amygdala central nucleus and appetitive Pavlovian conditioning: lesions impair one class of conditioned behavior. J Neurosci. 1990;10(6):1906-11.
  80. 80. Groshek F, Kerfoot E, McKenna V, Polackwich AS, Gallagher M, Holland PC. Amygdala central nucleus function is necessary for learning, but not expression, of conditioned auditory orienting. Behav Neurosci. 2005;119(1):202-12.
  81. 81. Olshavsky ME, Song BJ, Powell DJ, Jones CE, Monfils MH, Lee HJ. Updating appetitive memory during reconsolidation window: critical role of cue‐directed behavior and amygdala central nucleus. Front Behav Neurosci. 2013;7:186.
  82. 82. Lingawi NW, Balleine BW. Amygdala central nucleus interacts with dorsolateral striatum to regulate the acquisition of habits. J Neurosci. 2012;32(3):1073-81.
  83. 83. Hitchcott PK, Phillips GD. Double dissociation of the behavioural effects of R(+) 7‐OH‐DPAT infusions in the central and basolateral amygdala nuclei upon Pavlovian and instrumental conditioned appetitive behaviours. Psychopharmacology (Berl). 1998;140(4):458-69.
  84. 84. Parkinson JA, Robbins TW, Everitt BJ. Dissociable roles of the central and basolateral amygdala in appetitive emotional learning. Eur J Neurosci. 2000;12(1):405-13.
  85. 85. Lee HJ, Groshek F, Petrovich GD, Cantalini JP, Gallagher M, Holland PC. Role of amygdalo‐nigral circuitry in conditioning of a visual stimulus paired with food. J Neurosci. 2005;25(15):3881-8.
  86. 86. Gonzales C, Chesselet MF. Amygdalonigral pathway: an anterograde study in the rat with Phaseolus vulgaris leucoagglutinin (PHA‐L). J Comp Neurol. 1990;297(2):182-200.
  87. 87. Lerner TN, Shilyansky C, Davidson TJ, Evans KE, Beier KT, Zalocusky KA, et al. Intact‐brain analyses reveal distinct information carried by SNc dopamine subcircuits. Cell. 2015;162(3):635-47.
  88. 88. Han JS, McMahan RW, Holland P, Gallagher M. The role of an amygdalo‐nigrostriatal pathway in associative learning. J Neurosci. 1997;17(10):3913-19.
  89. 89. Lee HJ, Gallagher M, Holland PC. The central amygdala projection to the substantia nigra reflects prediction error information in appetitive conditioning. Learn Mem. 2010;17(10):531-8.
  90. 90. Judice‐Daher DM, Tavares TF, Bueno JL. Involvement of the basolateral complex and central nucleus of amygdala in the omission effects of different magnitudes of reinforcement. Behav Brain Res. 2012;233(1):149-56.
  91. 91. Tavares TF, Judice‐Daher DM, Bueno JL. Large neurotoxic amygdala lesion impairs reinforcement omission effects. Behav Brain Res. 2014;266:1-6.
  92. 92. Salinas JA, Packard MG, McGaugh JL. Amygdala modulates memory for changes in reward magnitude: reversible post‐training inactivation with lidocaine attenuates the response to a reduction in reward. Behav Brain Res. 1993;59(1-2):153-9.
  93. 93. Salinas JA, Parent MB, McGaugh JL. Ibotenic acid lesions of the amygdala basolateral complex or central nucleus differentially effect the response to reductions in reward. Brain Res. 1996;742(1-2):283-93.
  94. 94. Salinas JA, White NM. Contributions of the hippocampus, amygdala, and dorsal striatum to the response elicited by reward reduction. Behav Neurosci. 1998;112(4):812-26.
  95. 95. Robinson MJ, Warlow SM, Berridge KC. Optogenetic excitation of central amygdala amplifies and narrows incentive motivation to pursue one reward above another. J Neurosci. 2014;34(50):16567-80.
  96. 96. Mahler SV, Berridge KC. Which cue to “want?” Central amygdala opioid activation enhances and focuses incentive salience on a prepotent reward cue. J Neurosci. 2009;29(20):6500-13.
  97. 97. Mahler SV, Aston‐Jones GS. Fos activation of selective afferents to ventral tegmental area during cue‐induced reinstatement of cocaine seeking in rats. J Neurosci. 2012;32(38):13309-26.
  98. 98. Calu DJ, Roesch MR, Haney RZ, Holland PC, Schoenbaum G. Neural correlates of variations in event processing during learning in central nucleus of amygdala. Neuron. 2010;68(5):991-1001.
  99. 99. Iordanova MD, Deroche ML, Esber GR, Schoenbaum G. Neural correlates of two different types of extinction learning in the amygdala central nucleus. Nat Commun. 2016;7:12330.

Written By

Rose Chesworth and Laura Corbit

Submitted: 26 October 2016 Reviewed: 10 February 2017 Published: 05 July 2017