- Department of Experimental Psychology, University of Oxford, Oxford, United Kingdom
The successful pursuit of future rewards requires forming an internal goal, followed by planning, decision-making, and progress-tracking over multiple steps. The initial step—forming goals and the plans for obtaining them—involves the subjective valuation of an anticipated reward, considering both the reward’s properties and associated delay and physical-effort costs. Recent findings indicate individuals similarly evaluate cognitive effort over time (Johnson and Most, 2023). Success and failure in these processes have been linked to differential life outcomes and psychiatric conditions. Here we review evidence from single-neuron recordings and neuroimaging studies that implicate the amygdala—a brain structure long associated with cue-reactivity and emotion—in decision-making and the planned pursuit of future rewards (Grabenhorst et al., 2012, 2016, 2019, 2023;Hernadi et al., 2015;Zangemeister et al., 2016). The main findings are that, in behavioral tasks in which future rewards can be pursued through planning and stepwise decision-making, amygdala neurons prospectively encode the value of anticipated rewards and related behavioral plans. Moreover, amygdala neurons predict the stepwise choices to pursue these rewards, signal progress toward goals, and distinguish internally generated (i.e., self-determined) choices from externally imposed actions. Importantly, amygdala neurons integrate the subjective value of a future reward with delay and effort costs inherent in pursuing it. This neural evidence identifies three key computations of the primate amygdala that underlie the pursuit of future rewards: (1) forming a self-determined internal goal based on subjective reward-cost valuations, (2) defining a behavioral plan for obtaining the goal, (3) executing this plan through stepwise decision-making and progress-tracking. Based on this framework, we suggest that amygdala neurons constitute vulnerabilities for dysfunction that contribute to maladaptive reward pursuit in psychiatric and behavioral conditions. Consequently, amygdala neurons may also represent potential targets for behavioral-change interventions that aim to improve individual decision-making.
Introduction
The best rewards are often distant and can only be obtained by prospective (i.e., future-oriented) valuation of goals, planning and decision-making over several steps, and the tracking of progress toward a goal. Some future rewards can be obtained over shorter timescales, such as planning a restaurant visit in the evening to consume a desired food. However, many rewards are much more distant and require planning and persistence over longer timescales, including economic saving, achieving physical or mental health outcomes, pursuing an academic degree, or finding one’s ideal partner. What neural mechanisms underlie the pursuit of future rewards? What information is encoded by neurons in the brain’s reward system when an individual forms a goal and pursues it through planning and decision-making? We argue that investigating these neurophysiological mechanisms can validate or challenge psychological conceptions of goal pursuit and advance understanding of psychiatric and behavioral conditions in which this the pursuit of future rewards is impaired.
Psychological and behavioral-economic theories identify two key principles for planned reward-guided behaviors (Miller et al., 1960; Gollwitzer, 1993; Dickinson and Balleine, 1994; Sutton and Barto, 1998; Benhabib and Bisin, 2005; Berkman and Lieberman, 2009): the initial formation of a goal based on subjective reward and cost valuations, and the pursuit of this goal through planning, decision-making and progress-tracking. The valuation step is complex, as it requires an individual to consider the rewarding properties of a potential goal (e.g., its magnitude, desirable sensory qualities, relevance to one’s long-term values) and its associated costs (e.g., the delay, physical and cognitive effort required for obtaining it), and evaluating these factors based on subjective preferences and aversions.
Success and failure in the component process underlying the pursuit of future rewards have been linked to differential health outcomes, well-being, and specific psychiatric and behavioral conditions (Bechara, 2005; Casey et al., 2011; Bickel et al., 2012; Korn et al., 2014; Gillan et al., 2020). Accordingly, behavioral-change interventions to improve physical and mental health target the formation of resilient internal goals and the commitment to pursuing them (Gollwitzer and Schaal, 1998; Diamond and Lee, 2011; Belanger-Gravel et al., 2013; Duckworth et al., 2013; Michie et al., 2013; Sheeran et al., 2013; Scholten et al., 2019). For example, ‘mental contrasting’ involves imagining both the positive outcomes of a distant goal and possible obstacles to achieving the goal (Duckworth et al., 2013). Such interventions are designed to strengthen an individual’s capacities for successful pursuit of long-term goals. These capacities are constrained by the effort and delay involved in goal achievement, as well as inconsistencies in evaluating these cost factors (Ainslie and Herrnstein, 1981; Johnson and Most, 2023).
Here, we review recent advances in understanding the neurophysiological basis for the pursuit of future rewards in one of the key brain areas implicated in reward and decision-making: the primate amygdala. We first summarize the amygdala’s anatomical organization and its functions. We then discuss psychological and behavioral-economic conceptions of goal formation and planned behavior, and introduce evidence on delay and effort cost valuation. These considerations identify some of the key processes that might be encoded by neural systems during planned reward-directed behavior. Next, we review findings from single-neuron recordings in the amygdala of monkeys performing a reward-based economic saving task, and from related human neuroimaging studies. These studies show that amygdala neurons encode internally determined goals for future rewards and the decisions and plans for obtaining them. Finally, we discuss how prospective amygdala signals may constitute vulnerabilities for dysfunction in psychiatric and behavioral conditions, and targets for interventions aiming for behavioral change. Table 1 summarizes our neuronal framework for the pursuit of future rewards, based on identified neural signals in the primate amygdala, and its implications for understanding related dysfunction and for conceptualizing behavioral interventions in terms of neural mechanisms.
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-t001.jpg)
Table 1. Neuronal framework for the pursuit of future rewards, vulnerabilities for dysfunction, and targets for behavioral interventions affecting neuronal signals.
The primate amygdala: overview of structure and function
The amygdala, a cell complex located in the anterior-medial temporal lobe (Figure 1A), has long been associated with mediating emotional reactions to sensory cues (Rolls, 2000; Baxter and Murray, 2002; Cardinal et al., 2002; Maren and Quirk, 2004; Balleine and Killcross, 2006; Murray, 2007; Ghods-Sharifi et al., 2009; Morrison and Salzman, 2010; Johansen et al., 2011; Janak and Tye, 2015; Gothard, 2020; Pujara et al., 2022). However, recent findings also implicate primate amygdala neurons in more complex cognitive functions, including the pursuit of future rewards through economic, value-based decision-making and planning (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016; Grabenhorst et al., 2019; Grabenhorst and Schultz, 2021; Grabenhorst et al., 2023).
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g001.jpg)
Figure 1. Primate amygdala anatomy and connectivity. (A) Location of the amygdala in the anterior-medial part of the temporal lobe of the primate brain (Macaca mulatta). (B) Schematic overlaid on a cresyl-violet stained macaque coronal brain slice illustrating major amygdala nuclear subdivisions considered in this review, and some of their main input/output connections with simplified functional descriptions. Note that the sensory connections are typically bidirectional (Price, 2003). Blue: lateral nucleus (LA); orange: basolateral (BL) and basomedial nuclei (BM); magenta: centromedial nucleus (Ce); green: cortical nucleus (Co); yellow: medial nuclei (Me). B: basal nucleus of Meynert; Cl: claustrum; ErC: entorhinal cortex; fSTS: fundus of the superior temporal sulcus; GP: globus pallidus; LV: lateral ventricle; OT: optic tract; PrC: perirhinal cortex. Nomenclature based on Paxinos et al. (2000).
The primate (including human) amygdala is composed of several nuclei that vary in cell types, connections, and functions [for detailed reviews, (see Price et al., 1987; Price, 2003; Schumann et al., 2016)]. In a simplified view (Figure 1B), the lateral nucleus can be conceptualized as the amygdala’s entry point for sensory signals and storage site for stimulus-reinforcer associations: it receives inputs from all sensory systems, including particularly rich visual inputs from the inferior temporal cortex (Stefanacci and Amaral, 2002), and its neurons show potentiated responses to sensory stimuli that have been associated with reward or punishment (Sah et al., 2003; Johansen et al., 2011; Duvarci and Pare, 2014). The lateral nucleus projects to the basolateral and basomedial nuclei, which receive additional inputs from the orbitofrontal cortex, medial prefrontal cortex and hippocampus (Price et al., 1987) that enrich encoded stimulus-reinforcer associations with contextual information. Recent neurophysiological evidence in primates suggests that lateral-nucleus neurons encode accurate valuations of particular sensory objects whereas basolateral nucleus are more directly implicated in value comparisons between currently attended objects and decision-making (Grabenhorst et al., 2019; Grabenhorst et al., 2023). The basolateral and basomedial nuclei send direct outputs to the prefrontal cortex and striatum—routes by which the amygdala can influence neural systems associated with decision-making and action selection. A second critical output involves the amygdala’s centromedial nuclei, which process inputs from other amygdala nuclei to regulate behavioral and autonomic responses to affective stimuli via projections to the hypothalamus, basal forebrain, periaqueductal gray, midbrain and brainstem (Price et al., 1987; Price, 2003). Complex inhibitory systems modulate information processing in these amygdala circuits (Duvarci and Pare, 2014).
Its extrinsic and elaborate intrinsic connections (Pitkanen et al., 1997; Pitkanen and Amaral, 1998) enable the amygdala to process the value of sensory stimuli based on learned stimulus-reinforcer associations, and integrate this information with current contexts, memories, and internal states to regulate emotion, attention, memory, physiological and behavioral responses (Rolls, 2000; Paton et al., 2006; Murray, 2007; Johansen et al., 2011; Duvarci and Pare, 2014; Saez et al., 2015; Gothard, 2020; Grabenhorst and Schultz, 2021). The amygdala is a complex structure composed of different subregions with many different cell types. Accordingly, its neurons have been shown to be related to a diversity of functions including not only value-coding and decision-making (Paton et al., 2006; Grabenhorst et al., 2012; Chang et al., 2015; Hernadi et al., 2015; Grabenhorst et al., 2016; Costa et al., 2019; Grabenhorst et al., 2019; Jezzini and Padoa-Schioppa, 2020; Grabenhorst et al., 2023), but also the distinct processing of stimulus intensity, salience, arousal and other processes important in emotion, motivation, and value-guided behavior (Nishijo et al., 2008; Peck et al., 2013; Leathers and Olson, 2017; Iwaoki and Nakamura, 2022; Tang et al., 2024).
As reviewed in detail below, recent evidence indicates that primate amygdala neurons are involved in the component processes underlying the pursuit of future rewards (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016; Zangemeister et al., 2016). Importantly, the amygdala is also implicated in specific psychiatric conditions and mental-health impairments, including anxiety and depression (Price and Drevets, 2012; Bernardi and Salzman, 2019; Andrews et al., 2022; Klein-Flugge et al., 2022; Fox and Shackman, 2024). Thus, the same neural mechanisms that support successful reward pursuit may also contribute to impaired cognition and emotion in psychiatric conditions in which these processes are dysfunctional.
Psychological principles underlying the pursuit of future rewards: distinct phases of goal formation and goal pursuit, and the subjective valuation of rewards and costs
Psychological concepts identify component processes underlying the pursuit of future rewards that might be implemented in neural circuits. Goals can be conceptualized as internal representations of desired future states, outcomes or events (Austin and Vancouver, 1996; Brandstatter and Hennecke, 2018). They serve to direct behavior purposefully toward these desired states, inform the selection of behavioral plans for goal-achievement, and regulate attention, effort, and perseverance during goal pursuit (Brandstatter and Hennecke, 2018). Although these concepts stem primarily from human psychology, they can be defined in simple, concrete terms to help operationalization for behavioral and neurophysiological experiments. In the present review, we use the term ‘goal’ to refer specifically to the representation of a future reward that is selected based on subjective valuation of the reward’s attributes and associated costs (i.e., delay, effort). We use the term ‘plan’ to refer to the behavioral means for pursuing and obtaining a goal, for example through a sequence of choices and the actions that execute these choices.
One key principle in psychological models of planned behavior is the distinction between the initial process of goal formation and subsequent goal pursuit (Miller et al., 1960; Berkman and Lieberman, 2009; Bargh et al., 2010; Brandstatter and Hennecke, 2018). For example, economic saving is an elaborate planned behavior that involves forming a self-defined goal for a future reward followed by dynamic, sequential decision-making to achieve the goal (Prelec and Loewenstein, 1998; Benhabib and Bisin, 2005). As we discuss below, neural data suggest that these component processes are encoded as partly distinct activity patterns in primate amygdala neurons.
A second key principle is the initiation of a valuation process that assigns subjective value to goals based on reward, delay, uncertainty, effort cost, and other factors (Sutton and Barto, 1998; Benhabib and Bisin, 2005; Brandstatter and Hennecke, 2018). According to classical expectancy-value theory (Vroom, 1964; Ajzen, 1991; Beckmann and Heckhausen, 2018) the desirability (value) and feasibility (expectancy) of a goal affect goal selection and the effort invested into pursuing the goal. More abstract determinants of goal selection include achievement motives (Brunstein and Heckhausen, 2018), needs for autonomy, competence, and belonging (Deci and Ryan, 2000), social influences (Aarts et al., 2004; Weiss et al., 2023), and external reward (Vroom, 1964). In this review, we focus on the simple and most direct cost factors delay and effort, which can be directly manipulated in neurophysiological experiments. Before we review how primate amygdala neurons integrate valuations of reward goals with delay and effort costs, we first consider how delay and effort costs are evaluated psychologically. This discussion leads to the identification of ‘temporary preferences’ as an important concept that may help explain failures in the pursuit of future rewards.
Time-sensitive valuations of delay and effort costs govern the pursuit of future rewards
The most direct cost factor involved in the pursuit of future rewards is delay. Delay discounting, i.e., the subjective devaluation of a reward as a function of time until the reward can be obtained, has been behaviorally demonstrated in humans, monkeys and other species (Ainslie, 1974; Ainslie and Herrnstein, 1981; Mazur, 1998; McClure et al., 2004; Kable and Glimcher, 2007; Kobayashi and Schultz, 2008). Consistent with the aforementioned relevance of future rewards to health and wellbeing, dysfunctional delay discounting has been proposed as a ‘trans-disease process’ that may contribute to a range of disorders (Bickel et al., 2012). Indeed, higher rates of delay discounting have been linked to behavioral problems such as obesity, smoking, alcoholism, drug abuse, risky sexual behavior, and gambling problems (Bickel et al., 2012; Scholten et al., 2019).
Delays do not only affect the valuation of future rewards but also the valuation of future costs associated with rewards. For example, in a recent study (Stillman and Woolley, 2023), participants chose between future delivery of 24 cookies or a non-food item (a fashionable bag) matched in delay and monetary value when participants were cued to consider either potential long-term costs of eating cookies (e.g., developing diabetes, heart disease, and obesity) or short-term costs (e.g., feeling nauseous or jittery after eating sugar). Despite long-term costs being more severe, participants cued with short-term costs were significantly less likely to choose the cookies compared to those cued with long-term costs. Thus, the delay of future rewards and costs can affect the selection of reward goals, and this process can be influenced by interventions cueing particular cost types.
Humans and other animals can behave in ways that reflect valuations of reward, delay or effort that are time-sensitive, i.e., valuations that depend on when in time these outcomes will occur. In describing how decisions should be made by rational agents, normative decision theory posits that this time-dependence in reward and cost valuation is irrational and violates a principle known as dynamic consistency (Ainslie and Herrnstein, 1981; Barkan and Busemeyer, 1999). The principle of dynamic consistency states that individuals deciding rationally have consistent preferences throughout time. Therefore, a dynamically consistent decision maker who prefers A to B at a certain point in time will prefer A to B at all other points in time, and their preferences can be modeled with an exponential function that assumes constant rates of discounting, and thus consistent preferences, over time. However, delay discounting has been shown to be modeled more accurately by a nonexponential, hyperbolic function that implements a change in discount rate and can thus accounts for dynamic inconsistency (Green and Myerson, 1996). In accounting for dynamic inconsistency, a hyperbolic discounting function predicts that a decision-maker may have temporary preferences for inferior choice options (Ainslie, 2005). For example, if delay between rewards is discounted more when a smaller, sooner reward is immediate, then as the rewards draw nearer in time the decision-maker will be more likely to prefer the smaller, sooner reward. Temporary preferences were first observed in animal experiments on delay discounting (Rachlin and Green, 1972; Ainslie and Herrnstein, 1981; Mazur, 1998). In a classical delay discounting paradigm in pigeons (Ainslie, 1974), when the delay between smaller, sooner and larger, later rewards was fixed and both rewards were further delayed by equal amounts, the pigeons became less likely to choose the smaller, sooner reward. These reversals of preferences demonstrated that delay-discounting rates can depend on where a delay is placed in time.
Temporary preferences have also been shown in human experiments (Benzion et al., 1989; Green et al., 1994; Kirby and Herrnstein, 1995), which demonstrated how the gap between people’s intentions (future preference) and actions (present-moment preference) may lead to inferior choices. One study (Read and van Leeuwen, 1998) found that participants were more likely to choose an unhealthy compared to a healthy snack when deciding what they would eat immediately compared to deciding what they would eat in 1 week’s time. Similarly, meta-analysis (Sheeran, 2002) highlights the prevalence of this ‘intention-action gap’ related to temporary preferences across a range of health-related behaviors including exercise, contraceptive use, and cancer screening. Across six studies, the median percentage of individuals who set an intention to execute these behaviors yet failed to do so was 47%. Evidently, people can have an initial hope to pursue valuable distant rewards instead of immediately gratifying rewards, but as the competing options of these distant rewards draw closer in time, an increase in the discounting rate of cost factors such as effort and delay causes the inferior alternative to become preferable. Accordingly, temporary preferences resulting from dynamic inconsistency can create a gap between intention and action, undermining an original goal.
Dynamic inconsistency occurs not only in regard to delay discounting but also in regard to effort discounting, i.e., the anticipation of exerted physical or mental effort. Exertion of effort in pursuit of future rewards may come in the form of cognitive effort, as in the mental work required to solve computational problems (Murawski and Bossaerts, 2016; Hong and Stauffer, 2023) or plan a healthy, nutrient-balanced meal, or in the form of physical effort (Hosokawa et al., 2013; Burrell et al., 2023), as in the completion of strenuous exercise to burn calories. Though humans are often averse to effort when anticipating it in the future, effort that has already been expended on a reward can have the opposite effect, increasing subjective value. For example, in the ‘IKEA effect’, having to exert effort assembling a product oneself increases one’s subjective valuation of that item (note that this effect is distinct from the sunk cost fallacy, which involves irrational persistence in a previously chosen option without necessarily valuing it more) (Norton et al., 2012). Thus, the impact of effort on subjective value may depend on when it occurs in time (Inzlicht et al., 2018).
Johnson and Most (2023) investigated whether discounting cognitive effort to obtain a reward depended on how soon the effort occurred in time. They gave individuals the option to receive a smaller reward for no effort or to perform the effortful task of typing words backwards to receive a larger reward (Figure 2A). Participants were informed that each choice was potentially real and they may be rewarded with a shopping voucher of an equivalent monetary value to their choice if they were willing to perform the associated backwards-typing task. Immediacy of the effort was varied by asking participants to imagine exerting varying levels of effort either immediately, in a day, or in a month. The authors observed that individuals tended to discount the effort less when they were making the decision for their future self, compared to making decisions that were immediately relevant (Figure 2B) (Johnson and Most, 2023). Thus, people exhibit dynamic inconsistency in making decisions to obtain rewards by exerting effort, such that the degree of effort discounting depends on how soon the effort occurs. Notably, this effect was only found in individuals low in ‘need for cognition’, a personality trait describing how much one generally enjoys exerting cognitive effort, which suggests that enjoyment of effort may ‘buffer’ against the tendency to be dynamically inconsistent.
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g002.jpg)
Figure 2. Time-sensitive valuation of cognitive effort: Pushing cognitive effort into the future reduces effort discounting. (A) Participants were asked to choose between receiving a smaller reward for no effort or performing the effortful task of typing words backwards at varying temporal delays to receive a larger reward. Repeated choices between options of varying reward size resulted in subjective-value estimates from empirically identified indifference points. (B) Individuals low in need for cognition (NFC) displayed dynamic inconsistency in effort-related decision-making. Subjective value of the larger, more effortful reward rose as temporal distance to the anticipated effort increased. Mean subjective values (of an objective $20) are displayed. Error bars: standard errors. *p < 0.05 Adapted with permission from Johnson and Most (2023).
Identifying instances of dynamic inconsistency could inform strategies that support goal pursuit. In recognizing the possibility of a temporary preference, an individual setting an intention may choose to implement strategies that minimize future temptations (i.e., avoiding smaller, sooner or smaller, easier rewards) and thus reduce the likelihood of the temporary preference. For example, an individual who intends to exercise in the morning may place their phone out of reach before sleeping to reduce the likelihood of being tempted to stay in bed and sleep in. Remarkably, the aforementioned experiment by Ainslie (1974) found that even pigeons may use such pre-commitment strategies to maximize reward. In a delay discounting paradigm, three out of 10 pigeons consistently pecked a different colored key that prevented the smaller, sooner reward from being offered to them, thus pre-committing to the larger, later reward before the temporary preference arose.
The above findings show that the dynamic, time-sensitive valuation and re-valuation of reward, delay and (physical and cognitive) effort is a central process underlying the pursuit of future rewards that gives rise to temporary preferences. Accordingly, the neural mechanisms involved in selecting future reward goals and the plans to obtain them should reflect the subjective valuations of reward with associated delay and effort costs and also the updating of this integrated signal to account for temporary preferences. Evidence supporting this proposal is reviewed in the next sections and summarized in see Table 1.
A behavioral paradigm to study the neurophysiology underlying the pursuit of future reward goals: economic reward-saving decisions
Some of the psychological processes reviewed above can be studied in behavioral tasks involving sequential save-spend decisions leading to future rewards. Using this paradigm, neurophysiological studies in monkeys uncovered some of the building blocks underlying the pursuit of future rewards in amygdala neurons (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016). A subsequent study translated this approach to human functional neuroimaging (Zangemeister et al., 2016, 2019). The behavioral task used in these studies, described next, modeled key aspects of planned, goal-directed behavior: (1) forming a self-determined internal goal based on subjective reward-cost valuations, (2) defining a plan for obtaining the goal, (3) executing the plan by stepwise decision-making and progress-tracking.
In the ‘save-spend task’ (Grabenhorst et al., 2012; Hernadi et al., 2015), monkeys made consecutive choices to save (i.e., accumulate) liquid rewards for future trials until they decided to spend (i.e., consume) the saved reward amount (Figure 3A). Save-spend choice options were cued with pre-trained visual conditioned stimuli on a computer monitor. Different save cues indicated different interest rates that governed differential increases in saved reward amounts over consecutive save choices, with high interest rates leading to exponential reward growths (Figure 3B, green curves). The behavioral choice-patterns of the monkeys in this task reflected accurate understanding of the task structure: the animals produced longer sequences of save choices when interest rates were high, and shorter sequences when interest rates were low (Figure 3B, black bars). Control tests confirmed that the monkeys adjusted their behavior even to uncued changes in interest rate and that they tracked saved reward amounts over consecutive trials. The monkeys’ saving behavior was therefore adaptive and internally controlled rather than reflecting fixed, conditioned responses to pretrained cues. Analysis of reaction times showed that responses early on in saving sequences predicted the final length of the saving sequence and the amount of the final saved reward. Thus, the animals anticipated final reward amounts, as early as on the first trial of a saving sequence, consistent with planned behavior directed at internally set reward goals.
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g003.jpg)
Figure 3. Economic reward-saving behavior in monkeys. (A) Schematic of the save-spend task. Monkeys made sequences of save-spend choices to save (i.e., accumulate) liquid reward for later until deciding to spend (i.e., consume) it. The task allowed the monkeys to form an internal goal to obtain a specific future reward and plan to obtain this goal by making a save-spend choice sequence of a specific length. (B) Behavior in the saving task. Monkeys produced longer saving sequences, shown by their choice probability for different sequence lengths (black bars), when reward grew exponentially (green curve; reward growth was governed by a cued interest rate). Magenta curve: Subjective value of a saving sequence of defined length estimated from choice probability incorporating reward amount, delay and effort costs. (C) Activity of a single amygdala neuron recorded in the saving task (Imp/s: neuronal response measured in impulses per second; raster plot: each line represents a recorded action potential). The neuron responded more strongly at the time of choice when the monkey was going to make a spend choice compared to a save choice on the current trial (left panel). The choice-predictive activity was specific to the free-choice task and disappeared in the instructed, forced-choice control task (right panel), confirming the activity reflected an internally generated choice rather than reward expectation. Adapted with permission from Grabenhorst et al. (2012) and Hernadi et al. (2015).
An important feature of the task was that it allowed the animals to plan their choices multiple steps in advance, to obtain specific reward amounts through saving sequences of defined lengths requiring particular delay (i.e., waiting times) and effort (i.e., number of operand responses). Accordingly, the subjective value of a reward goal depended not only on final reward amounts but also on costs related to sequence length: because larger reward amounts typically required longer saving sequences, their value was diminished by temporal delay and physical effort. These effects on subjective values were modeled by deriving value from observed behavioral choices for different sequence lengths (Figure 3B, black bars). These ‘sequence value’ functions typically increased with sequence length up to a peak and then decreased with longer sequences that the animal chose less frequently, likely owing to temporal discounting and physical-effort costs (Figure 3B, right panel, magenta curve). This nonlinearity in subjective-value functions made it possible to determine whether neuronal activities encoded subjective value or objective sequence length, as described below.
We note that this approach to reward-saving decisions shares features with ‘token economies’ (Hackenberg, 2009), in which animals earn and accumulate conditioned stimuli and exchange them against different food and liquid reward according to a defined schedule. Similar to goal-directed sequences of save-spend choices described above, token schedules can organize an animal’s behavior over extended time periods that lead to a final (‘terminal’) primary reward. For example, pigeons’ produce sequential responses to accumulate tokens before making a different response to exchange them against food or liquid reward (Yankelevitz et al., 2008; DeFulio et al., 2014). Similarly, monkeys anticipate rewards in instructed sequential reward-schedule tasks with visual cues indicating reward proximity (Shidara and Richmond, 2002; Sugase-Miyamoto and Richmond, 2005) and make reward-maximizing choices using tokens that signify gain or loss of primary reinforcers (Yang et al., 2022; Tang et al., 2024).
Primate amygdala neurons encode plans for future rewards: the value of goals, behavioral plans for obtaining them, and adaptive progress during goal pursuit
The primate amygdala has long been implicated in reward-guided behavior (Rolls, 2000; Baxter and Murray, 2002; Paton et al., 2006; Belova et al., 2007; Murray, 2007) but only recently has been shown to participate in value-based decision-making (Grabenhorst et al., 2012; Rudebeck et al., 2013; Rudebeck et al., 2017; Costa et al., 2019; Grabenhorst et al., 2019; Jezzini and Padoa-Schioppa, 2020; Grabenhorst et al., 2023) and the planned pursuit of future reward goals (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016; Zangemeister et al., 2016). During performance of the save-spend task introduced in the previous section, single-cell recordings showed that a significant number of amygdala neurons exhibited activity patterns that predicted the monkeys’ save-spend choice on individual trials (Grabenhorst et al., 2012). These choice-predictive activities (Figure 3C) were not explained by left–right actions, nor did they reflect subjective values of save-spend choices, which were encoded separately. Rather, these amygdala neurons signaled the monkey’s economic choice to either spend reward for immediate consumption or save it for the future. Choice-predictive activities were in most cases specific to freely made choices, as they disappeared in a control task involving forced, instructed save-spend choices (Figure 3C, right panel). Thus, amygdala neurons predicted the monkey’s self-determined choices to either pursue a future goal by saving reward for later or to spend the accumulated reward immediately. More recent studies implicated amygdala neurons in choices for different types of reward (Jezzini and Padoa-Schioppa, 2020), in explore-exploit decisions (Costa et al., 2019), social decisions (Chang et al., 2015; Grabenhorst et al., 2019), and in decision computations that compare currently viewed and recently viewed choice options (Grabenhorst et al., 2023).
Amygdala neurons did not simply encode save-spend choices related to single trials but also showed more complex, prospective activities that were critical for optimal task performance (Hernadi et al., 2015; Grabenhorst et al., 2016). These ‘planning activities’ preceded the end of a saving sequence by several steps and encoded information about the final reward goal, which was self-determined by the monkey and existed only internally at the time of saving. Specifically, amygdala planning activities signaled two well-defined features of the animal’s internal saving plan (Hernadi et al., 2015): some neurons encoded ‘sequence value’, defined as the subjective value of the current saving sequence (Figures 4A,B, solid magenta curve), while a separate set of neurons encoded the ‘sequence length’, defined by the number of planned, forthcoming save choices (Figure 4B, dashed magenta curve). For example, the neuron in Figure 4A had recurring phasic responses on each trial that were highest during sequences in which the monkey would eventually spend on the fifth trial and lower for shorter or longer sequences. This activity profile reflected the distribution of subjective values derived from the monkey’s choice preferences: five-trial sequences had the highest value under the current interest rate, as the animal chose them most frequently. Importantly, subjective value was a non-monotonic function of sequence length: depending on the current interest rate, subjective value was highest for intermediate sequence lengths (Figure 4B, black bars) that offered a compromise between large rewards (Figure 4B, green curve) and moderate delays and effort. These planning activities typically disappeared during instructed trials, despite comparable reward timing and anticipation, and in control analyses were shown to be unrelated to reward proximity and expectation.
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g004.jpg)
Figure 4. Amygdala neurons encode value and length of monkeys’ saving plans. (A) An amygdala neuron with prospective activity that reflected the subjective value of the monkey’s internal saving plan. The neuron’s activity depended on the subjective value of the current sequence (‘sequence value’) that would only be completed several moments into the future. Top: Activity at trial start (yellow area) was highest for the sequence in which the monkey would eventually spend on the fifth trial, as this sequence had the highest subjective value. Bottom: activity averages for all sequence lengths (for example, the light pink activation represents the mean trial-start activity for all five-trial sequences, averaged over trials 1 to 5). Activity reflected sequence value (magenta curve), rather than linear sequence length or objective reward amount (green curve). (B) Amygdala neurons with prospective activity for future rewards. Neurons signaled the length of the planned choice sequence (dashed magenta curve, population activity, N = 92 neurons) or its subjective value (solid magenta curve, N = 93 neurons). Value activity was highest during sequences lasting six trials, which had the highest subjective value (black bars), i.e., these sequences were typically preferred by the animals, because they offered large reward (green curve) for moderate delay and physical effort. Adapted with permission from Hernadi et al. (2015).
Thus, amygdala neurons in the save-spend task encoded two key components of the monkeys’ internal saving plan: the subjective value of the current reward goal (‘sequence value’) and the behavioral plan to achieve the goal through a choice sequence of defined length (‘sequence length’).
A distinct, third type of amygdala activity signaled the monkeys’ progress over the course of a saving sequence (Grabenhorst et al., 2016). These progress-tracking neurons showed gradually increasing, ‘ramping’ activity over consecutive save choices until the monkey decided to spend the saved reward (Figure 5A). Importantly, the responses occurred in the absence of external progress cues and were often specific to internally guided choices. Moreover, the slope of this ramping activity depended on the forthcoming sequence length, with steeper neuronal ramping for shorter sequences (Figure 5B). These findings suggest that amygdala progress-tracking signals adapt to the monkey’s internal plan to execute a specific sequence length. Such continual evaluation of progress by amygdala neurons seems crucial for aligning choices with internal plans and for successful goal pursuit (Johnson and Busemeyer, 2001; Benhabib and Bisin, 2005; Berkman and Lieberman, 2009).
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g005.jpg)
Figure 5. Progress-tracking in amygdala. (A) Ramping activity of an amygdala neuron in a saving sequence lasting six trials. The neuron’s responses at the time of choice increased with each consecutive step in the saving sequence. (B) The slope of the neuronal ramping activity in amygdala for different sequences (right panel) adapted to the final sequence length (‘adaptive sequence progress’, middle schematic); it did not increase linearly with elapsed time or trial number (left schematic), consistent with progress-tracking rather than time-tracking. Adapted with permission from Grabenhorst et al. (2016).
Taken together, different neurons in the primate amygdala encode important building blocks for the pursuit of future rewards: (1) the subjective value of the current reward goal (‘sequence value’), (2) the behavioral saving plan for achieving the goal (‘sequence length), (3) the save-spend choices that realized this plan at specific choice points, and (4) the internally tracked progress toward the reward goal. In relation to the psychological concepts of goal pursuit discussed above, it is possible that variation in these amygdala signals within and across individuals might underlie individual differences in the capacity to pursue long-term rewards (see Table 1). For example, fluctuations in the amygdala’s coding of current reward goals might lead to dynamic revaluation of the chosen goal and its alternatives at critical choice points, which could provide a neurophysiological mechanism for temporary preferences. Further, the initial goal valuation and goal formation could be impacted by changes in the function of sequence-value neurons: Exaggerated valuation of a goal’s rewarding aspects or diminished discounting of delay and effort costs could result in neuronal encoding and behavioral pursuit of unrealistic goals. Conversely, diminished valuation of future rewards and exaggerated delay and effort discounting could result in the absence of neuronally encoded future rewards and behavioral amotivation. Importantly, the neuronal signals for sequence progress, sequence length and sequence value described above fluctuated with behavioral performance: they were diminished on error trials, when the animal failed to complete a trial correctly, and subsequently reinstated, suggesting behavioral relevance of these amygdala signals. Thus, the amygdala signals reported above might be behaviorally relevant during an individual’s pursuit of future rewards.
Future reward goals and planning activity in the human amygdala
The monkey single-neuron investigations reviewed above inspired a human functional neuroimaging study that aimed to translate these findings to the human brain (Zangemeister et al., 2016, 2019). Healthy volunteers performed the save-spend task while their brain activity was measured using functional magnetic resonance imaging (fMRI). As in the monkey study, human participants could plan and execute economic choice sequences to drink liquid rewards—delivered in the MRI scanner—that increased in amount according to the current interest rate. The reward also varied in type, consisting either of a low-fat high-sugar milkshake or a typically preferred high-fat high-sugar milkshake (Figure 6A, left panel).
![www.frontiersin.org](https://www.frontiersin.org/files/Articles/1517231/fnins-18-1517231-HTML/image_m/fnins-18-1517231-g006.jpg)
Figure 6. Planning activity for future rewards in the human amygdala. (A) Human participants in the MRI scanner performed save-spend choice sequences to obtain liquid rewards (milkshakes) of different sizes under varying interest rates and reward types cued by visual stimuli (left). Participants produced longer saving sequences for higher interest rates and preferred rewards (right). (B) The amygdala encoded the subjective value and planned length of the current saving sequence before participants made their first choice. Time-course analyses (right, betas from multiple-regression analysis) confirmed distinct amygdala signals for sequence value and sequence length (blue shading: analysis period). Adapted with permission from Zangemeister et al. (2016).
Similar to the monkeys, the human participants produced longer saving sequences when interest rates were high, but also saved longer for the more preferred high-fat rewards, consistent with the assignment of subjective values to different saving goals (Figure 6A, right panel). At the start of a saving sequence, before participants made the first choice to save or spend, neural activity in the amygdala reflected the key planning variables identified in the monkey studies: the subjective value of the planned saving sequence and its length (Figure 6B). Importantly, the degree to which these neural planning activities depended on the current reward type (high vs. low fat) matched individual participants’ reward preferences. Thus, amygdala planning activities incorporate information about reward amount, reward type, temporal delay and physical effort. Whole-brain imaging identified other brain areas that contributed to planned reward-saving decisions: neural activities in medial and dorsolateral prefrontal cortex areas reflected planned sequence length and functional coupling with each other and with the amygdala. A distinct prospective signal recorded in the ventromedial prefrontal cortex reflected participants’ reported saving intentions at the beginning of a new saving sequence (Zangemeister et al., 2019).
Thus, the human amygdala encodes different aspects of economic saving plans as part of a functionally connected frontal-amygdala network. Importantly, planning signals for internally generated reward goals depended not only on reward amount, delay and effort cost but also on the subjective valuation of different reward types. That aspects of the single-cell activity patterns related to goal pursuit originally identified in the monkey amygdala can be measured in humans using fMRI opens up the possibility to investigate how these activities might be altered in psychiatric conditions, and how they could be affected by behavioral interventions. To stimulate research, we include in Table 1 suggestions for how changes in specific amygdala signals could be linked to failures or improvements in the pursuit of long-term goals.
Neurophysiology of planned behavior beyond amygdala
The prospective amygdala signals reviewed above likely contribute to planned behavior together with activity patterns in other brain areas, briefly reviewed here. Planning functions have traditionally been associated with frontal-lobe regions (Miller et al., 1960; Tanji and Hoshi, 2001; Stuss and Knight, 2002; Passingham and Wise, 2012). In monkeys, single neurons in specific frontal regions including lateral-prefrontal, premotor, supplementary-motor and presupplementary-motor cortices encode planned action sequences and the execution and updating of motor plans (Tanji and Shima, 1994; Shima et al., 1996; Shima and Tanji, 1998; Mushiake et al., 2006; Sohn and Lee, 2007). For example, in a remarkable study (Shima et al., 2007), monkeys performed action sequences involving different combinations of push, turn, and pull movements with a lever. Before the monkeys executed each sequence, lateral prefrontal neurons encoded planned sequences in abstract ‘categories’, i.e., according to whether they required repetition, alternation or pairings of the same action. Complementing the amygdala neurons described above that encode behavioral sequences in terms of their subjective value (Hernadi et al., 2015), these prefrontal neurons could help specify the behavioral means by which future rewards can be pursued. Further, single-cell activities in frontal-lobe areas (motor, premotor, supplementary-motor cortices), the putamen and the striatum also precede self-initiated movements (Okano and Tanji, 1987; Romo and Schultz, 1987; Kurata and Wise, 1988; Schultz and Romo, 1992; Lee et al., 2006), which represent the behavioral means for executing a plan. Further, activities in human medial frontal cortex, monkey striatum and monkey dopamine neurons encode the subjective value of delayed rewards and thereby likely contribute to prospective goal valuations (Kable and Glimcher, 2007; McClure et al., 2007; Kim et al., 2008; Kobayashi and Schultz, 2008). In multistep reinforcement learning, prefrontal-striatal systems in human imaging studies encode evaluation of reward outcomes associated with externally defined choice paths (Wunderlich et al., 2012; Doll et al., 2015). Further, when human decision-makers calibrate their behavioral persistence while waiting for rewards, neural activity in medial prefrontal cortex reflects this context-sensitive subjective valuation (McGuire and Kable, 2015).
Thus, beyond amygdala, neural activity in connected brain areas contributes to internally planned, goal-oriented behaviors. Following earlier proposals (Shima and Tanji, 1998), we suggest that amygdala neurons signaling the subjective value of reward goals and the plans to obtain them (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016; Zangemeister et al., 2016) could provide value-based inputs to other brain systems to direct planned, self-initiated behavior toward future rewards.
Amygdala neurons and self-determination
One important aspect of the findings in the save-spend task reviewed above is that amygdala activity related to choices, reward goals and progress-tracking was typically specific to self-determined (rather than externally instructed) saving behavior (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016; Zangemeister et al., 2016). This observation links the interpretation of the amygdala’s involvement in goal pursuit to psychological theories of self-determined behaviors.
Behaviors generated by oneself differ in important ways from those that are externally imposed, as conceptualized by one of the most influential theories of human motivation: Self-Determination Theory (SDT, Deci and Ryan, 2000). In the terminology of SDT, ‘internally motivated’ behaviors are more likely to be accompanied by experiences of autonomy. In contrast, behaviors driven primarily by social forces are termed ‘externally motivated’ and typically lack the feelings of agency that accompany internally motivated behaviors. Planning signals in the amygdala were shown to depend on whether the behavior was internally or externally motivated, as these signals tended to disappear when the decision maker was instructed by a visual cue which action to take (Grabenhorst et al., 2012; Hernadi et al., 2015; Grabenhorst et al., 2016). Therefore, amygdala neurons seem sensitive to whether choices are self-determined or externally imposed, which may provide a basis for experiencing decisions as autonomous. In the SDT literature, autonomy creates higher levels of behavioral persistence and satisfaction (Grolnick and Ryan, 1987; Vallerand and Bissonnette, 1992; Sheldon and Houser-Marko, 2001). As the amygdala sends projections to brain areas involved in reward-related subjective experiences, including the medial prefrontal cortex and orbitofrontal cortex (Kringelbach et al., 2003; Rolls et al., 2003; Grabenhorst et al., 2007; Rolls and Grabenhorst, 2008; Grabenhorst et al., 2010; Grabenhorst and Rolls, 2011), it is possible that amygdala signals that reflect whether behavior was performed in an autonomous, self-determined manner may contribute to the subjective experience of autonomy. Future research could therefore examine whether amygdala signals related to self-determined decisions contribute to the positive psychological outcomes associated with self-determined behavior. As past research has suggested that promoting experiences of autonomy could mitigate depressive symptoms (Moore et al., 2021), future research could also investigate the role of these amygdala signals in depression and potentially in other psychiatric conditions in which the amygdala is implicated (Price and Drevets, 2012; Bernardi and Salzman, 2019; Andrews et al., 2022; Klein-Flugge et al., 2022; Fox and Shackman, 2024).
Conclusion and perspective: amygdala neurons as possible targets for behavioral-change interventions
The successful pursuit of future rewards requires the formation of a goal based on subjective reward and cost valuations, the definition of a plan to obtain the goal, and stepwise decision-making and progress-monitoring during goal pursuit. The amygdala signals reviewed above related to the value of internal reward goals and plans for obtaining them seem well-suited to direct the self-determined pursuit of future rewards. As individuals can be inconsistent in their evaluations of cost factors such as effort and delay, they may prefer a distant goal at one point in time but then switch their preference to an inferior but immediately-gratifying reward. These temporary preferences for immediate but inferior rewards are therefore a key concept for explaining impulsive behaviors that reflect time-dependent revaluation of reward goals and cost factors. Thus, future studies could examine whether amygdala neurons encoding the value of distant rewards update their activity to reflect time-dependent revaluations of goals. Such neurons could provide a neurophysiological basis for the emergence of temporary preferences and so-called ‘self-defeating’ behaviors, i.e., a choice made during goal pursuit that deviates from an initially formed goal and prioritizes an inferior reward over a previously stated preference for a larger but distant reward. As indicated above, successes and failures in the pursuit of future reward are highly relevant to human behavior, well-being, and mental health across life domains. A valuable research program into these questions could combine single-cell recordings in monkeys making complex, primate-typical decisions about future rewards with functional neuroimaging in humans performing the same decision tasks. This translational approach could provide evidence on detailed, single-neuron activity patterns underlying reward-pursuit and temporary preferences, and on the related neural correlates in human brain areas and brain systems, which could be used to investigate neural mechanisms of behavioral-change interventions.
Although confirmation of the behavioral relevance of amygdala neurons will require further experiments, we suggest that the amygdala’s activity patterns could represent potential vulnerabilities for dysfunctional reward-guided behavior and related mental-health impairments in conditions in which the amygdala is implicated, including depression, anxiety and addiction (Price and Drevets, 2012; Bernardi and Salzman, 2019; Andrews et al., 2022; Klein-Flugge et al., 2022; Fox and Shackman, 2024). The same neural signals may also be targets for interventions (Gollwitzer and Schaal, 1998; Diamond and Lee, 2011; Belanger-Gravel et al., 2013; Michie et al., 2013; Sheeran et al., 2013; Scholten et al., 2019), in that the effects of interventions aiming for behavioral change, such as mental contrasting (Duckworth et al., 2013), may potentially be mediated by altered activity patterns of these amygdala neurons. In Table 1, we propose some specific ways in which known amygdala signals with particular functions in the pursuit of future rewards could represent vulnerabilities for dysfunction and targets for intervention. These proposals are made with the intention of stimulating further discussion and research, although we acknowledge that these suggestions remain speculative and will await confirmation by behavioral and neurophysiological data. Further to investigating the role of the amygdala in the behavior change strategies described in Table 1, future research could investigate how the complex psychological mechanisms underlying behavioral-change strategies may relate to amygdala function. For example, it has been proposed that one mechanism of mental contrasting is changing how individuals interpret specific situations (Oettingen and Schworer, 2013). After mental contrasting, an individual may begin to interpret an opportunity to eat a new flavor of ice cream as an obstacle to their weight loss progress rather than an exciting new experience. Thus, understanding the functions of amygdala neurons in the pursuit of future rewards may not only address basic biological questions concerning the neural basis of motivated behavior but may also lead to mechanistic explanations for successes and failures of behavior-change interventions.
Author contributions
SJ: Conceptualization, Writing – original draft, Writing – review & editing. FG: Conceptualization, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was funded by the Wellcome Trust and the Royal Society (Wellcome/Royal Society Sir Henry Dale Fellowship grants 206207/Z/17/Z and 206207/Z/17/A to FG). This research was funded in whole, or in part, by the Wellcome Trust. For the purpose of Open Access, the authors have applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission.
Acknowledgments
We thank S. Most, I. Hernádi, L. Zangemeister, and W. Schultz for collaboration and discussion on some of the experiments discussed in this review.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The authors declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Aarts, H., Gollwitzer, P. M., and Hassin, R. R. (2004). Goal contagion: perceiving is for pursuing. J Pers Soc Psychol 87, 23–37. doi: 10.1037/0022-3514.87.1.23
Ainslie, G. W. (1974). Impulse control in pigeons. J Exp Anal Behav 21, 485–489. doi: 10.1901/jeab.1974.21-485
Ainslie, G. (2005). Precis of breakdown of will. Behav Brain Sci 28, 635–650. doi: 10.1017/S0140525X05000117
Ainslie, G., and Herrnstein, R. J. (1981). Preference reversal and delayed reinforcement. Anim Learn Behav 9, 476–482. doi: 10.3758/BF03209777
Ajzen, I. (1991). The theory of planned behavior. Organ Behav Hum Decis Process 50, 179–211. doi: 10.1016/0749-5978(91)90020-T
Andrews, D. S., Aksman, L., Kerns, C. M., Lee, J. K., Winder-Patel, B. M., Harvey, D. J., et al. (2022). Association of Amygdala Development with Different Forms of anxiety in autism Spectrum disorder. Biol Psychiatry 91, 977–987. doi: 10.1016/j.biopsych.2022.01.016
Austin, J. T., and Vancouver, J. B. (1996). Goal constructs in psychology: structure, prcess and content. Psychol Bull 122, 338–375.
Balleine, B. W., and Killcross, S. (2006). Parallel incentive processing: an integrated view of amygdala function. Trends Neurosci 29, 272–279. doi: 10.1016/j.tins.2006.03.002
Bargh, J. A., Gollwitzer, P. M., and Oettingen, G. (2010). “Motivation” in Handbook of social psychology. eds. S. Fiske, D. Gilbert, and G. Linszey (New York: Wiley), 268–316.
Barkan, R., and Busemeyer, J. R. (1999). Changing plans: dynamic inconsistency and the effect of experience on the reference point. Psychon Bull Rev 6, 547–554. doi: 10.3758/BF03212962
Baxter, M. G., and Murray, E. A. (2002). The amygdala and reward. Nat Rev Neurosci 3, 563–573. doi: 10.1038/nrn875
Bechara, A. (2005). Decision making, impulse control and loss of willpower to resist drugs: a neurocognitive perspective. Nat Neurosci 8, 1458–1463. doi: 10.1038/nn1584
Beckmann, J., and Heckhausen, H. (2018). “Motivation as a function of expectancy and incentive” in Motivation and action. eds. J. Heckhausen and H. Heckhausen (Cham, Switzerland: Springer), 163–220.
Belanger-Gravel, A., Godin, G., Bilodeau, A., and Poirier, P. (2013). The effect of implementation intentions on physical activity among obese older adults: a randomised control study. Psychol Health 28, 217–233. doi: 10.1080/08870446.2012.723711
Belova, M. A., Paton, J. J., Morrison, S. E., and Salzman, C. D. (2007). Expectation modulates neural responses to pleasant and aversive stimuli in primate amygdala. Neuron 55, 970–984. doi: 10.1016/j.neuron.2007.08.004
Benhabib, J., and Bisin, A. (2005). Modeling internal commitment mechanisms and self-control: a neuroeconomics approach to consumption-saving decisions. Games Econ Behav 52, 460–492. doi: 10.1016/j.geb.2004.10.004
Benzion, U., Rapoport, A., and Yagil, J. (1989). Discount rates inferred from decisions: an experimental study. Manag Sci 35, 270–284. doi: 10.1287/mnsc.35.3.270
Berkman, E. T., and Lieberman, M. D. (2009). “The neuroscience of goal pursuit” in The psychology of goals. eds. G. Moskowitz and H. Grant (New York: Guilford Press), 98–126.
Bernardi, S., and Salzman, C. D. (2019). The contribution of nonhuman primate research to the understanding of emotion and cognition and its clinical relevance. Proc Natl Acad Sci USA 116, 26305–26312. doi: 10.1073/pnas.1902293116
Bickel, W. K., Jarmolowicz, D. P., Mueller, E. T., Koffarnus, M. N., and Gatchalian, K. M. (2012). Excessive discounting of delayed reinforcers as a trans-disease process contributing to addiction and other disease-related vulnerabilities: emerging evidence. Pharmacol Ther 134, 287–297. doi: 10.1016/j.pharmthera.2012.02.004
Brandstatter, V., and Hennecke, M. (2018). “Goals” in Motivation and action. eds. J. Heckhausen and H. Heckhausen. Third ed (Cham, Switzerland: Springer), 485–528.
Brunstein, J. C., and Heckhausen, H. (2018). “Achievement motivation” in Motivation and action. eds. J. Heckhausen and H. Heckhausen (Cham, Switzerland: Springer).
Burrell, M., Pastor-Bernier, A., and Schultz, W. (2023). Worth the work? Monkeys discount rewards by a subjective adapting effort cost. J Neurosci 43, 6796–6806. doi: 10.1523/JNEUROSCI.0115-23.2023
Cardinal, N., Parkinson, J. A., Hall, J., and Everitt, B. J. (2002). Emotion and motivation: the role of the amygdala, ventral striatum, and prefrontal cortex. Neurosci Biobehav Rev 26, 321–352. doi: 10.1016/S0149-7634(02)00007-6
Casey, B. J., Somerville, L. H., Gotlib, I. H., Ayduk, O., Franklin, N. T., Askren, M. K., et al. (2011). Behavioral and neural correlates of delay of gratification 40 years later. Proc Natl Acad Sci USA 108, 14998–15003. doi: 10.1073/pnas.1108561108
Chang, S. W., Fagan, N. A., Toda, K., Utevsky, A. V., Pearson, J. M., and Platt, M. L. (2015). Neural mechanisms of social decision-making in the primate amygdala. Proc Natl Acad Sci USA 112, 16012–16017. doi: 10.1073/pnas.1514761112
Costa, V. D., Mitz, A. R., and Averbeck, B. B. (2019). Subcortical substrates of explore-exploit decisions in Primates. Neuron 103:e535. doi: 10.1016/j.neuron.2019.05.017
Deci, E. L., and Ryan, R. M. (2000). The “what” and “why” of goal pursuits: human needs and the self-determination of behavior. Psychol Inq 11, 227–268. doi: 10.1207/S15327965PLI1104_01
DeFulio, A., Yankelevitz, R., Bullock, C., and Hackenberg, T. D. (2014). Generalized conditioned reinforcement with pigeons in a token economy. J Exp Anal Behav 102, 26–46. doi: 10.1002/jeab.94
Diamond, A., and Lee, K. (2011). Interventions shown to aid executive function development in children 4 to 12 years old. Science 333, 959–964. doi: 10.1126/science.1204529
Dickinson, A., and Balleine, B. (1994). Motivational control of goal-directed action. Anim Learn Behav 22, 1–18. doi: 10.3758/BF03199951
Doll, B. B., Duncan, K. D., Simon, D. A., Shohamy, D., and Daw, N. D. (2015). Model-based choices involve prospective neural activity. Nat Neurosci 18, 767–772. doi: 10.1038/nn.3981
Duckworth, A. L., Kirby, T., Gollwitzer, A., and Oettingen, G. (2013). From fantasy to action: mental contrasting with implementation intentions (MCII) improves academic performance in children. Soc Psychol Personal Sci 4, 745–753. doi: 10.1177/1948550613476307
Duvarci, S., and Pare, D. (2014). Amygdala microcircuits controlling learned fear. Neuron 82, 966–980. doi: 10.1016/j.neuron.2014.04.042
Fox, A. S., and Shackman, A. J. (2024). An honest reckoning with the amygdala and mental illness. Am J Psychiatry 181, 1059–1075. doi: 10.1176/appi.ajp.20240941
Ghods-Sharifi, S., St Onge, J. R., and Floresco, S. B. (2009). Fundamental contribution by the basolateral amygdala to different forms of decision making. J Neurosci 29, 5251–5259. doi: 10.1523/JNEUROSCI.0315-09.2009
Gillan, C. M., Kalanthroff, E., Evans, M., Weingarden, H. M., Jacoby, R. J., Gershkovich, M., et al. (2020). Comparison of the association between goal-directed planning and self-reported compulsivity vs obsessive-compulsive disorder diagnosis. JAMA Psychiatry 77, 77–85. doi: 10.1001/jamapsychiatry.2019.2998
Gollwitzer, P. M. (1993). Goal achievement: the role of intentions. Eur Rev Soc Psychol 4, 141–185. doi: 10.1080/14792779343000059
Gollwitzer, P. M., and Schaal, B. (1998). Metacognition in action: the importance of implementation intentions. Personal Soc Psychol Rev 2, 124–136. doi: 10.1207/s15327957pspr0202_5
Gothard, K. M. (2020). Multidimensional processing in the amygdala. Nat Rev Neurosci 21, 565–575. doi: 10.1038/s41583-020-0350-y
Grabenhorst, F., Baez-Mendoza, R., Genest, W., Deco, G., and Schultz, W. (2019). Primate amygdala neurons simulate decision processes of social partners. Cell 177, 986–998.e15. doi: 10.1016/j.cell.2019.02.042
Grabenhorst, F., Hernadi, I., and Schultz, W. (2012). Prediction of economic choice by primate amygdala neurons. Proc Natl Acad Sci USA 109, 18950–18955. doi: 10.1073/pnas.1212706109
Grabenhorst, F., Hernadi, I., and Schultz, W. (2016). Primate amygdala neurons evaluate the progress of self-defined economic choice sequences. eLife 5:5. doi: 10.7554/eLife.18731
Grabenhorst, F., Ponce-Alvarez, A., Battaglia-Mayer, A., Deco, G., and Schultz, W. (2023). A view-based decision mechanism for rewards in the primate amygdala. Neuron 111:e3814. doi: 10.1016/j.neuron.2023.08.024
Grabenhorst, F., and Rolls, E. T. (2011). Value, pleasure, and choice in the ventral prefrontal cortex. Trends Cogn Sci 15, 56–67. doi: 10.1016/j.tics.2010.12.004
Grabenhorst, F., Rolls, E. T., Margot, C., da Silva, M. A. A. P., and Velazco, M. I. (2007). How pleasant and unpleasant stimuli combine in different brain regions: odor mixtures. J Neurosci 27, 13532–13540. doi: 10.1523/JNEUROSCI.3337-07.2007
Grabenhorst, F., Rolls, E. T., Parris, B. A., and D’Souza, A. (2010). How the brain represents the reward value of fat in the mouth. Cereb Cortex 20, 1082–1091. doi: 10.1093/cercor/bhp169
Grabenhorst, F., and Schultz, W. (2021). Functions of primate amygdala neurons in economic decisions and social decision simulation. Behav Brain Res 409:113318. doi: 10.1016/j.bbr.2021.113318
Green, L., Fristoe, N., and Myerson, J. (1994). Temporal discounting and preference reversals in choice between delayed outcomes. Psychon Bull Rev 1, 383–389. doi: 10.3758/BF03213979
Green, L., and Myerson, J. (1996). Exponential versus hyperbolic discounting of delayed outcomes: risk and waiting time. Am Zool 36, 496–505. doi: 10.1093/icb/36.4.496
Grolnick, W. S., and Ryan, R. M. (1987). Autonomy in children's learning: an experimental and individual difference investigation. J Pers Soc Psychol 52, 890–898. doi: 10.1037/0022-3514.52.5.890
Hackenberg, T. D. (2009). Token reinforcement: a review and analysis. J Exp Anal Behav 91, 257–286. doi: 10.1901/jeab.2009.91-257
Hernadi, I., Grabenhorst, F., and Schultz, W. (2015). Planning activity for internally generated reward goals in monkey amygdala neurons. Nat Neurosci 18, 461–469. doi: 10.1038/nn.3925
Hong, T., and Stauffer, W. R. (2023). Computational complexity drives sustained deliberation. Nat Neurosci 26, 850–857. doi: 10.1038/s41593-023-01307-6
Hosokawa, T., Kennerley, S. W., Sloan, J., and Wallis, J. D. (2013). Single-neuron mechanisms underlying cost-benefit analysis in frontal cortex. J Neurosci 33, 17385–17397. doi: 10.1523/JNEUROSCI.2221-13.2013
Inzlicht, M., Shenhav, A., and Olivola, C. Y. (2018). The effort paradox: effort is both costly and valued. Trends Cogn Sci 22, 337–349. doi: 10.1016/j.tics.2018.01.007
Iwaoki, H., and Nakamura, K. (2022). Neuronal encoding of emotional valence and intensity in the monkey amygdala. J Neurosci 42, 7615–7623. doi: 10.1523/JNEUROSCI.0021-22.2022
Janak, P. H., and Tye, K. M. (2015). From circuits to behaviour in the amygdala. Nature 517, 284–292. doi: 10.1038/nature14188
Jezzini, A., and Padoa-Schioppa, C. (2020). Neuronal activity in the primate amygdala during economic choice. J Neurosci 40, 1286–1301. doi: 10.1523/JNEUROSCI.0961-19.2019
Johansen, J. P., Cain, C. K., Ostroff, L. E., and LeDoux, J. E. (2011). Molecular mechanisms of fear learning and memory. Cell 147, 509–524. doi: 10.1016/j.cell.2011.10.009
Johnson, J. G., and Busemeyer, J. R. (2001). Multiple-stage decision-making: the effect of planning horizon length on dynamic consistency. Theor Decis 51, 217–246. doi: 10.1023/A:1015507110940
Johnson, S. T., and Most, S. B. (2023). Taking the path of least resistance now, but not later: pushing cognitive effort into the future reduces effort discounting. Psychon Bull Rev 30, 1115–1124. doi: 10.3758/s13423-022-02198-7
Kable, J. W., and Glimcher, P. W. (2007). The neural correlates of subjective value during intertemporal choice. Nat Neurosci 10, 1625–1633. doi: 10.1038/nn2007
Kim, S., Hwang, J., and Lee, D. (2008). Prefrontal coding of temporally discounted values during intertemporal choice. Neuron 59, 161–172. doi: 10.1016/j.neuron.2008.05.010
Kirby, T., and Herrnstein, R. J. (1995). Preference reversals due to myopic discounting of delayed reward. Psychol Sci 6, 83–89. doi: 10.1111/j.1467-9280.1995.tb00311.x
Klein-Flugge, M. C., Jensen, D. E. A., Takagi, Y., Priestley, L., Verhagen, L., Smith, S. M., et al. (2022). Relationship between nuclei-specific amygdala connectivity and mental health dimensions in humans. Nat Hum Behav 6, 1705–1722. doi: 10.1038/s41562-022-01434-3
Kobayashi, S., and Schultz, W. (2008). Influence of reward delays on responses of dopamine neurons. J Neurosci 28, 7837–7846. doi: 10.1523/JNEUROSCI.1600-08.2008
Korn, C. W., Sharot, T., Walter, H., Heekeren, H. R., and Dolan, R. J. (2014). Depression is related to an absence of optimistically biased belief updating about future life events. Psychol Med 44, 579–592. doi: 10.1017/S0033291713001074
Kringelbach, M. L., O'Doherty, J., Rolls, E. T., and Andrews, C. (2003). Activation of the human orbitofrontal cortex to a liquid food stimulus is correlated with its subjective pleasantness. Cereb Cortex 13, 1064–1071. doi: 10.1093/cercor/13.10.1064
Kurata, K., and Wise, S. P. (1988). Premotor and supplementary motor cortex in rhesus monkeys: neuronal activity during externally- and internally-instructed motor tasks. Experimental brain research Experimentelle Hirnforschung Experimentation cerebrale 72, 237–248. doi: 10.1007/BF00250247
Leathers, M. L., and Olson, C. R. (2017). In monkeys making value-based decisions, amygdala neurons are sensitive to cue value as distinct from cue salience. J Neurophysiol 117, 1499–1511. doi: 10.1152/jn.00564.2016
Lee, I. H., Seitz, A. R., and Assad, J. A. (2006). Activity of tonically active neurons in the monkey putamen during initiation and withholding of movement. J Neurophysiol 95, 2391–2403. doi: 10.1152/jn.01053.2005
Maren, S., and Quirk, G. J. (2004). Neuronal signalling of fear memory. Nat Rev Neurosci 5, 844–852. doi: 10.1038/nrn1535
McClure, S. M., Ericson, K. M., Laibson, D. I., Loewenstein, G., and Cohen, J. D. (2007). Time discounting for primary rewards. J Neurosci 27, 5796–5804. doi: 10.1523/JNEUROSCI.4246-06.2007
McClure, S. M., Laibson, D. I., Loewenstein, G., and Cohen, J. D. (2004). Separate neural systems value immediate and delayed monetary rewards. Science 306, 503–507. doi: 10.1126/science.1100907
McGuire, J. T., and Kable, J. W. (2015). Medial prefrontal cortical activity reflects dynamic re-evaluation during voluntary persistence. Nat Neurosci 18, 760–766. doi: 10.1038/nn.3994
Michie, S., Richardson, M., Johnston, M., Abraham, C., Francis, J., Hardeman, W., et al. (2013). The behavior change technique taxonomy (v1) of 93 hierarchically clustered techniques: building an international consensus for the reporting of behavior change interventions. Ann Behav Med 46, 81–95. doi: 10.1007/s12160-013-9486-6
Miller, G. A., Galanter, E., and Pribram, K. H. (1960). Plans and the structure of behavior: Holt. New York: Rinehart and Winston, Inc.
Moore, E., Holding, A. C., Moore, A., Levine, S. L., Powers, T. A., Zuroff, D. C., et al. (2021). The role of goal-related autonomy: a self-determination theory analysis of perfectionism, poor goal progress, and depressive symptoms. J Couns Psychol 68, 88–97. doi: 10.1037/cou0000438
Morrison, S. E., and Salzman, C. D. (2010). Re-valuing the amygdala. Curr Opin Neurobiol 20, 221–230. doi: 10.1016/j.conb.2010.02.007
Murawski, C., and Bossaerts, P. (2016). How humans solve complex problems: the case of the knapsack problem. Sci Rep 6:34851. doi: 10.1038/srep34851
Murray, E. A. (2007). The amygdala, reward and emotion. Trends Cogn Sci 11, 489–497. doi: 10.1016/j.tics.2007.08.013
Mushiake, H., Saito, N., Sakamoto, K., Itoyama, Y., and Tanji, J. (2006). Activity in the lateral prefrontal cortex reflects multiple steps of future events in action plans. Neuron 50, 631–641. doi: 10.1016/j.neuron.2006.03.045
Nishijo, H., Hori, E., Tazumi, T., and Ono, T. (2008). Neural correlates to both emotion and cognitive functions in the monkey amygdala. Behav Brain Res 188, 14–23. doi: 10.1016/j.bbr.2007.10.013
Norton, M. I., Mochon, D., and Ariely, D. (2012). The IKEA effect: when labor leads to love. J Consum Psychol 22, 453–460. doi: 10.1016/j.jcps.2011.08.002
Oettingen, G., and Schworer, B. (2013). Mind wandering via mental contrasting as a tool for behavior change. Front Psychol 4:562. doi: 10.3389/fpsyg.2013.00562
Okano, K., and Tanji, J. (1987). Neuronal activities in the primate motor fields of the agranular frontal cortex preceding visually triggered and self-paced movement. Exp Brain Res 66, 155–166. doi: 10.1007/BF00236211
Passingham, R. E., and Wise, S. P. (2012). The neurobiology of the prefrontal cortex. Oxford: Oxford University Press.
Paton, J. J., Belova, M. A., Morrison, S. E., and Salzman, C. D. (2006). The primate amygdala represents the positive and negative value of visual stimuli during learning. Nature 439, 865–870. doi: 10.1038/nature04490
Paxinos, G., Huang, X.-F., and Toga, A. W. (2000). The rhesus monkey brain in stereotaxic coordinates. San Diego, CA: Academic Press.
Peck, C. J., Lau, B., and Salzman, C. D. (2013). The primate amygdala combines information about space and value. Nat Neurosci 16, 340–348. doi: 10.1038/nn.3328
Pitkanen, A., and Amaral, D. G. (1998). Organization of the intrinsic connections of the monkey amygdaloid complex: projections originating in the lateral nucleus. J Comp Neurol 398, 431–458.
Pitkanen, A., Savander, V., and LeDoux, J. E. (1997). Organization of intra-amygdaloid circuitries in the rat: an emerging framework for understanding functions of the amygdala. Trends Neurosci 20, 517–523. doi: 10.1016/S0166-2236(97)01125-9
Prelec, D., and Loewenstein, G. (1998). The red and the black: mental accounting of savings and debt. Mark Sci 17, 4–28. doi: 10.1287/mksc.17.1.4
Price, J. L. (2003). Comparative aspects of amygdala connectivity. Ann N Y Acad Sci 985, 50–58. doi: 10.1111/j.1749-6632.2003.tb07070.x
Price, J. L., and Drevets, W. C. (2012). Neural circuits underlying the pathophysiology of mood disorders. Trends Cogn Sci 16, 61–71. doi: 10.1016/j.tics.2011.12.011
Price, J. L., Russchen, F. T., and Amaral, D. G. (1987). “The amygdaloid complex” in Handbook of chemical neuroanatomy, Vol. V, integrated systems of the CNS, part I. eds. L. Swanson, C. Kohler, and A. Bjorklund (Amsterdam: Elsevier), 279–388.
Pujara, M. S., Ciesinski, N. K., Reyelts, J. F., Rhodes, S. E. V., and Murray, E. A. (2022). Selective prefrontal-amygdala circuit interactions underlie social and nonsocial valuation in Rhesus macaques. J Neurosci 42, 5593–5604. doi: 10.1523/JNEUROSCI.0794-21.2022
Rachlin, H., and Green, L. (1972). Commitment, choice and self-control. J Exp Anal Behav 17, 15–22. doi: 10.1901/jeab.1972.17-15
Read, D., and van Leeuwen, B. (1998). Predicting hunger: the effects of appetite and delay on choice. Organ Behav Hum Decis Process 76, 189–205. doi: 10.1006/obhd.1998.2803
Rolls, E. T. (2000). “Neurophysiology and functions of the primate amygdala, and the neural basis of emotion” in The amygdala: A functional analysis, second. ed. J. P. Aggleton. Edition ed (Oxford: Oxford University Press), 447–478.
Rolls, E. T., and Grabenhorst, F. (2008). The orbitofrontal cortex and beyond: from affect to decision-making. Prog Neurobiol 86, 216–244. doi: 10.1016/j.pneurobio.2008.09.001
Rolls, E. T., O'Doherty, J., Kringelbach, M. L., Francis, S., Bowtell, R., and McGlone, F. (2003). Representations of pleasant and painful touch in the human orbitofrontal and cingulate cortices. Cereb Cortex 13, 308–317. doi: 10.1093/cercor/13.3.308
Romo, R., and Schultz, W. (1987). Neuronal activity preceding self-initiated or externally timed arm movements in area 6 of monkey cortex. Exp Brain Res 67, 656–662. doi: 10.1007/BF00247297
Rudebeck, P. H., Mitz, A. R., Chacko, R. V., and Murray, E. A. (2013). Effects of amygdala lesions on reward-value coding in orbital and medial prefrontal cortex. Neuron 80, 1519–1531. doi: 10.1016/j.neuron.2013.09.036
Rudebeck, P. H., Ripple, J. A., Mitz, A. R., Averbeck, B. B., and Murray, E. A. (2017). Amygdala contributions to stimulus-reward encoding in the macaque medial and orbital frontal cortex during learning. J Neurosci 37, 2186–2202. doi: 10.1523/JNEUROSCI.0933-16.2017
Saez, A., Rigotti, M., Ostojic, S., Fusi, S., and Salzman, C. D. (2015). Abstract context representations in primate amygdala and prefrontal cortex. Neuron 87, 869–881. doi: 10.1016/j.neuron.2015.07.024
Sah, P., Faber, E. S., Lopez De Armentia, M., and Power, J. (2003). The amygdaloid complex: anatomy and physiology. Physiol Rev 83, 803–834. doi: 10.1152/physrev.00002.2003
Scholten, H., Scheres, A., de Water, E., Graf, U., Granic, I., and Luijten, M. (2019). Behavioral trainings and manipulations to reduce delay discounting: a systematic review. Psychon Bull Rev 26, 1803–1849. doi: 10.3758/s13423-019-01629-2
Schultz, W., and Romo, R. (1992). Role of primate basal ganglia and frontal cortex in the internal generation of movements. I. Preparatory activity in the anterior striatum. Exp Brain Res 91, 363–384. doi: 10.1007/BF00227834
Schumann, C. M., Vargas, M. V., and Lee, A. (2016). “A synopsis of primate amygdala neuroanatomy” in Living without an amygdala. eds. D. G. Amaral and R. Adolphs (New York: The Guilford Press), 39–71.
Sheeran, P. (2002). Intention-behavior relations: a conceptual and empirical review. Eur Rev Soc Psychol 12, 1–36. doi: 10.1080/14792772143000003
Sheeran, P., Harris, P., Vaughan, J., Oettingen, G., and Gollwitzer, P. M. (2013). Gone exercising: mental contrasting promotes physical activity among overweight, middle-aged, low-SES fishermen. Health Psychol 32, 802–809. doi: 10.1037/a0029293
Sheldon, K. M., and Houser-Marko, L. (2001). Self-concordance, goal attainment, and the pursuit of happiness: can there be an upward spiral? J Pers Soc Psychol 80, 152–165. doi: 10.1037/0022-3514.80.1.152
Shidara, M., and Richmond, B. J. (2002). Anterior cingulate: single neuronal signals related to degree of reward expectancy. Science 296, 1709–1711. doi: 10.1126/science.1069504
Shima, K., Isoda, M., Mushiake, H., and Tanji, J. (2007). Categorization of behavioural sequences in the prefrontal cortex. Nature 445, 315–318. doi: 10.1038/nature05470
Shima, K., Mushiake, H., Saito, N., and Tanji, J. (1996). Role for cells in the presupplementary motor area in updating motor plans. Proc Natl Acad Sci USA 93, 8694–8698. doi: 10.1073/pnas.93.16.8694
Shima, K., and Tanji, J. (1998). Role for cingulate motor area cells in voluntary movement selection based on reward. Science 282, 1335–1338. doi: 10.1126/science.282.5392.1335
Sohn, J. W., and Lee, D. (2007). Order-dependent modulation of directional signals in the supplementary and presupplementary motor areas. J Neurosci 27, 13655–13666. doi: 10.1523/JNEUROSCI.2982-07.2007
Stefanacci, L., and Amaral, D. G. (2002). Some observations on cortical inputs to the macaque monkey amygdala: an anterograde tracing study. J Comp Neurol 451, 301–323. doi: 10.1002/cne.10339
Stillman, P. E., and Woolley, K. (2023). Undermining desire: reducing unhealthy choices by highlighting short-term (vs. long-term) costs. J Consum Res 50, 554–575. doi: 10.1093/jcr/ucad004
Stuss, D. T., and Knight, R. T. (2002). Principles of frontal lobe function. Oxford: Oxford University Press.
Sugase-Miyamoto, Y., and Richmond, B. J. (2005). Neuronal signals in the monkey basolateral amygdala during reward schedules. J Neurosci 25, 11071–11083. doi: 10.1523/JNEUROSCI.1796-05.2005
Sutton, R. S., and Barto, A. G. (1998). Reinforcement learning, vol. 31. Cambridge, MA: MIT Press, 5.
Tang, H., Bartolo, R., and Averbeck, B. B. (2024). Ventral frontostriatal circuitry mediates the computation of reinforcement from symbolic gains and losses. Neuron 112:e3785. doi: 10.1016/j.neuron.2024.08.018
Tanji, J., and Hoshi, E. (2001). Behavioral planning in the prefrontal cortex. Curr Opin Neurobiol 11, 164–170. doi: 10.1016/S0959-4388(00)00192-6
Tanji, J., and Shima, K. (1994). Role for supplementary motor area cells in planning several movements ahead. Nature 371, 413–416. doi: 10.1038/371413a0
Vallerand, R. J., and Bissonnette, R. (1992). “Intrinsic, extrinsic, and amotivational styles as predictors of behavior: a prospective study” in 1997 Human Nature: A Critical Reader. ed. L. Betzig, vol. 60 (New York: Oxford University Press), 599–620.
Weiss, M., Gollwitzer, M., and Hewig, J. (2023). Social influence and external feedback control in humans. F1000Res 12:438. doi: 10.12688/f1000research.133295.1
Wunderlich, K., Dayan, P., and Dolan, R. J. (2012). Mapping value based planning and extensively trained choice in the human brain. Nat Neurosci 15, 786–791. doi: 10.1038/nn.3068
Yang, Y. P., Li, X., and Stuphorn, V. (2022). Primate anterior insular cortex represents economic decision variables proposed by prospect theory. Nat Commun 13:717. doi: 10.1038/s41467-022-28278-9
Yankelevitz, R. L., Bullock, C. E., and Hackenberg, T. D. (2008). Reinforcer accumulation in a token-reinforcement context with pigeons. J Exp Anal Behav 90, 283–299. doi: 10.1901/jeab.2008.90-283
Zangemeister, L., Grabenhorst, F., and Schultz, W. (2016). Neural basis for economic saving strategies in human amygdala-prefrontal reward circuits. Curr Biol 26, 3004–3013. doi: 10.1016/j.cub.2016.09.016
Keywords: amygdala, reward, temporal discounting, effort, depression, goal pursuit, planning
Citation: Johnson ST and Grabenhorst F (2025) The amygdala and the pursuit of future rewards. Front. Neurosci. 18:1517231. doi: 10.3389/fnins.2024.1517231
Edited by:
Manuela Sellitto, University of Pavia, ItalyReviewed by:
Ludise Malkova, Georgetown University, United StatesValeria Gonzalez, Reed College, United States
Copyright © 2025 Johnson and Grabenhorst. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Fabian Grabenhorst, ZmFiaWFuLmdyYWJlbmhvcnN0QHBzeS5veC5hYy51aw==