- 1Department of Psychology and Neuroscience, Duke University, Durham, NC, USA
- 2Center for Cognitive Neuroscience, Duke University, Durham, NC, USA
- 3Facultad de Economía y Empresa, Centro de Neuroeconomía, Universidad Diego Portales, Santiago, Chile
In order to control behavior in an adaptive manner the brain has to learn how some situations and actions predict positive or negative outcomes. During the last decade cognitive neuroscientists have shown that the brain is able to evaluate and learn from outcomes within a few hundred milliseconds of their occurrence. This research has been primarily focused on the feedback-related negativity (FRN) and the P3, two event-related potential (ERP) components that are elicited by outcomes. The FRN is a frontally distributed negative-polarity ERP component that typically reaches its maximal amplitude 250 ms after outcome presentation and tends to be larger for negative than for positive outcomes. The FRN has been associated with activity in the anterior cingulate cortex (ACC). The P3 (~300–600 ms) is a parietally distributed positive-polarity ERP component that tends to be larger for large magnitude than for small magnitude outcomes. The neural sources of the P3 are probably distributed over different regions of the cortex. This paper examines the theories that have been proposed to explain the functional role of these two ERP components during outcome processing. Special attention is paid to extant literature addressing how these ERP components are modulated by outcome valence (negative vs. positive), outcome magnitude (large vs. small), outcome probability (unlikely vs. likely), and behavioral adjustment. The literature offers few generalizable conclusions, but is beset with a number of inconsistencies across studies. This paper discusses the potential reasons for these inconsistencies and points out some challenges that probably will shape the field over the next decade.
Introduction
The global function of the nervous system can be characterized as the adaptive control of behavior, a process that involves learning which action is relevant in a given context and switching to a different behavioral policy or scenario when outcomes are less optimal than expected. Hence “outcome” refers to the consequences that an organism faces as direct result of its own actions (e.g., financial losses due to impulsive investments) or resulting from its situation (e.g., receiving an unexpected gift). In order to successfully adapt behavior the brain has to determine, as quickly and as accurately as possible, whether the current scenario and behavioral policy results in positive or negative outcomes. This paper reviews the extant literature on how two event-related potential (ERP) components, the feedback-related negativity (FRN), and the P3, shed light on the neural substrate of outcome processing.
The relevance of outcome processing is highlighted by the association between individual differences in its functioning and personality constructs (Kramer et al., 2008; Onoda et al., 2010; Smillie et al., 2011), economic preferences (Coricelli et al., 2005; Kuhnen and Knutson, 2005; De Martino et al., 2006; Kable and Glimcher, 2007; Venkatraman et al., 2009), and pathological conditions such as compulsive gambling (Reuter et al., 2005; Goudriaan et al., 2006), drug abuse (Shiv et al., 2005; Everitt et al., 2007; Fein and Chang, 2008; Franken et al., 2010; Fridberg et al., 2010; Park et al., 2010), depression (Foti and Hajcak, 2009), and schizophrenia (Gold et al., 2008; Morris et al., 2008).
In the past decade neuroimaging studies have contributed enormously to identifying brain regions and patterns of functional connectivity supporting outcome processing in the human brain (e.g., Delgado et al., 2000, 2003; Breiter et al., 2001; Elliott et al., 2003; O'Doherty et al., 2003b; Ullsperger and Von Cramon, 2003; Holroyd et al., 2004b; Huettel et al., 2006; Kim et al., 2006; Mullette-Gillman et al., 2011). The discovery of this underlying physiology has been accompanied by a wealth of new knowledge about the functional properties of these mechanisms. In particular, non-invasive electrophysiological methods have provided important information about the temporal properties of the neural mechanisms mediating outcome evaluation in humans. Notably, by recording ERPs while participants perform learning-guided choice tasks or simple gambling games, researchers have begun to describe how the brain processes outcomes within a few hundred milliseconds from their onset.
Among the potential neural correlates of outcome evaluation, the FRN is by far the most studied ERP component. The FRN is a frontocentral negative-going ERP component that peaks ~250 ms following outcome presentation and is typically larger for negative outcomes than for positive outcomes (Figure 1A). According to source localization, the FRN is generated in the medial prefrontal cortex (mPFC), most probably in the anterior cingulate cortex (ACC; Miltner et al., 1997; Gehring and Willoughby, 2002; Ruchsow et al., 2002; van Schie et al., 2004; Muller et al., 2005; Nieuwenhuis et al., 2005b; Hewig et al., 2007; Yu and Zhou, 2009; Yu et al., 2011). Consistent with ERP studies, results implicating the ACC in processing negative feedback have been reported using fMRI (Kiehl et al., 2000; Holroyd et al., 2004b).
Figure 1. A schematic representation of ERP waveforms typically elicited by outcomes (based on Gehring and Willoughby, 2002; Yeung and Sanfey, 2004; Goyer et al., 2008; Gu et al., 2011a). The horizontal axis represents the elapsed time relative to the onset of behavioral feedback (at 0 ms). (A) Example of ERP waveform for negative and positive outcomes recorded at a frontocentral electrode site. The scalp topography represents the contrast between the two waveforms at the time when the FRN peaks (~250 ms). (B) Example of ERP waveform for large and small magnitude outcomes recorded at a central-parietal electrode site. The scalp topography represents the contrast between the two waveforms at the time when the P3b peaks (~300–600 ms).
The P3 is another outcome-related ERP component. The P3 is a positive-polarity component most pronounced at the centroparietal recording sites at about 300–600ms after stimuli presentation (Figure 1B). According to a model proposed by Yeung and Sanfey (2004) outcome magnitude (i.e., large vs. small) and outcome valence (i.e., loss vs. gains) are coded separately in the brain, with the P3 being sensitive to outcome magnitude and the FRN to outcome valence. Despite controversial evidence (Hajcak et al., 2005, 2007; Bellebaum and Daum, 2008; Goyer et al., 2008; Wu and Zhou, 2009; Pfabigan et al., 2011) the independent coding model is presently the dominant account of the relationship between the FRN and the P3 during outcome processing.
The number of ERP studies of outcome have multiplied over the last decade. Yet the last systematic review of such studies focused exclusively on the FRN, and dates back 8 years (Nieuwenhuis et al., 2004a). This paper intends to provide an updated perspective about current knowledge from ERP research of outcome processing in the human brain. In order to achieve this goal, the paper reviews the historical antecedents and dominant theoretical accounts of the FRN and P3. These theories are evaluated in light of studies addressing how FRN and P3 are modulated by outcome valence (negative vs. positive), outcome magnitude (large vs. small), outcome probability (unlikely vs. likely), and behavioral adjustment. Finally, this paper discusses some challenges that ERP studies of outcome processing will probably have to address over the next decade in order to integrate FRN and P3 effects in a unitary account of outcome processing in the brain.
The Feedback-Related Negativity
Historical Antecedents of the FRN
The study of the neural basis of outcome evaluation and feedback-guided learning has been facilitated by the discovery of an ERP component, the FRN, which tends to distinguish between positive and negative outcomes. Miltner et al. (1997) was the first group to describe the FRN as an ERP component that is differentially sensitive to negative and positive feedback. In their study, they required participants to estimate the duration of a 1 s interval by pressing a button when they believed that 1 s had elapsed from the presentation of a cue. Their response was followed by the delivery of a feedback stimulus indicating whether their estimate was correct (positive feedback) or incorrect (negative feedback). A time window around 1 s was used to determine response accuracy and this window was adjusted so that the likelihood of positive and negative feedback stimuli for each participant were both 50%.
Miltner and colleagues found that the ERP elicited by negative feedback was characterized by a negative deflection at frontocentral recording sites with a peak latency of ~250 ms. This negativity was isolated with the method of difference waves, that is by subtracting the ERP response to positive feedback from the ERP response to negative feedback. Source localization estimates placed the generator of this difference wave near the ACC. The same results were found across different conditions in which feedback was provided in auditory, visual, and somatosensory modalities. Miltner and colleagues noted that the characteristics of this negativity corresponded in many respects (i.e., sensitivity to errors, polarity, scalp topography, and likely origin in the ACC) to those of the response-locked error-related negativity (ERN), an ERP component that reaches maximum amplitude about 100 ms following error commission in speeded response time tasks (Falkenstein et al., 1991; Gehring et al., 1993; Scheffers et al., 1996; for a review see Yeung et al., 2004). The authors suggested that both the ERN (elicited by error commission) and the FRN (elicited by negative feedback) reflect a general error detection function of the ACC. Indeed, converging lines of evidence from fMRI research (Kiehl et al., 2000; Holroyd et al., 2004b), magneto-encephalography (Miltner et al., 2003; Donamayor et al., 2011) and intracranial EEG recordings in humans (Wang et al., 2005) support the idea that the ACC is involved in performance monitoring and error detection.
The Reinforcement Learning Theory of the FRN
The error detection hypothesis (Miltner et al., 1997) was later extended by Holroyd and Coles (2002), who proposed that both the ERN and the FRN are scalp-recorded indexes of a neural system for reinforcement learning. This theory is based on research that implicates the basal ganglia and the midbrain dopamine (DA) system in reward prediction and reinforcement learning (Barto, 1995; Montague et al., 1996; Schultz et al., 1997; Schultz and Dickinson, 2000; Tobler et al., 2005; for a review see Schultz, 2002). From a computational standpoint, reinforcement learning problems involve a set of world states, a set of actions available to the agent in each state, a transition function which specifies the probability of moving from one state to another when performing a specific action, and a reward function, which indicates the reward or punishment associated with each transition (Ribas-Fernandes et al., 2011). In this context, the goal for learning is to discover, on a trial and error fashion, a policy (i.e., a stable mapping between states and actions) that maximizes the cumulative discounted long-term reward (Sutton and Barto, 1998).
According to the reinforcement learning theory of the ERN/FRN (RL-theory) the human brain solves reinforcement learning problems by implementing an “actor-critic architecture” (Barto, 1995; Joel et al., 2002). This theory assumes that several actors are implemented throughout the brain (e.g., amygdala, dorsolateral prefrontal cortex), each acting semi-independently and in parallel, and each trying to exert their influence over the motor system. According to the RL-theory, the ACC acts as a control filter, selecting a motor plan according to weighted state-action associations and communicating the corresponding response to the output layer (i.e., motor cortex) for execution.
The role of the critic in the actor-critic architecture is to evaluate ongoing events and predict whether future events will be favorable or unfavorable. When the critic revises its predictions for the better or for worse, it computes a temporal-difference reward prediction error (RPE). A positive or a negative RPE indicates that ongoing events are better than expected or worse than expected, respectively. The RPE is used to update both the value attached to the previous state and the strength of the state-action associations that determined the last response selection. The RL-theory attributes the role of the critic to the basal ganglia and assumes that the RPE signal corresponds to the phasic increase (for positive RPE, or +RPE) or decrease (for negative RPE, or −RPE) in the activity of midbrain DA neurons. Indeed, dopaminergic neurons in the monkey midbrain have been shown to code positive and negative errors in reward prediction (Schultz et al., 1997; Schultz and Dickinson, 2000; Tobler et al., 2005). According to the RL-theory, the RPE signal (i.e., phasic DA) is distributed to three parts of the network: (1) the adaptive critic itself (i.e., basal ganglia), where it is used to refine the ongoing predictions, (2) the motor controllers (e.g., amygdala), where it is used to adjust the state-action mappings, and (3) the control filter (i.e., ACC), where it is used to train the filter to select the most adaptive motor controller on a given situation. Together, the adaptive critic, the motor controllers, and the control filter learn how best to perform in reinforcement learning scenarios. The RL-theory proposes that the impact of DA signals on the ACC modulates the amplitude of the ERN/FRN (Holroyd and Coles, 2002). According to this theory, phasic decreases in DA activity enlarge ERN/FRN by indirectly disinhibiting the apical dendrites of motor neurons in the ACC, and phasic increases in DA activity reduce the ERN/FRN by indirectly inhibiting the apical dendrites of motor neurons in the ACC.
Despite a group of inconsistent findings that are reviewed in the next section (Hajcak et al., 2005; Oliveira et al., 2007; San Martin et al., 2010; Kreussel et al., 2012), the RL-theory remains as the dominant account of the FRN. Also, it has been an important factor contributing to the augment in the number of studies about this ERP component, mainly because of three reasons. First, the RL-theory speaks to one of the most explored issues in cognitive neuroscience, namely the cognitive functions implemented by the mPFC, and more specifically the ACC (Carter et al., 1998; Bush et al., 2000; Paus, 2001; Botvinick et al., 2004; Kerns et al., 2004; Ridderinkhof et al., 2004; Rushworth et al., 2004; Barber and Carter, 2005; Alexander and Brown, 2011). The RL-theory suggests that the general role of the ACC is the adaptive control of behavior, and that the ACC learns how to better perform its function from discrepancies between actual and optimal responses (for the ERN) or between actual and expected outcomes (for the FRN). This notion challenges the conflict-monitoring model (Botvinick et al., 2001; Botvinick, 2007) that characterizes the ACC as a region that can detect conflict between more than one response tendency, for example when a stimulus primes a pre-potent but incorrect response, or when the correct response is undetermined. While the conflict-monitoring model is consistent with neuroimaging evidence for ACC activation during error commission (Carter et al., 1998; Kerns et al., 2004) and with the characteristics of the ERN (Yeung et al., 2004), it does not address the FRN and is inconsistent with evidence from monkey neurophysiological studies that have not found any conflict-related activity in the ACC (Ito et al., 2003; Nakamura et al., 2005). The RL-theory, on the other hand, provides an explanation for both the ERN and the FRN, and accounts for ACC responses to outcomes in fMRI studies (Bush et al., 2002; Paulus et al., 2004), and in single-unit recording studies with monkeys (Niki and Watanabe, 1979; Amiez et al., 2005; Tsujimoto et al., 2006) and humans (Williams et al., 2004).
Second, the RL-theory proposes that the FRN reflects the evaluation of events along a general good–bad dimension, but it is non-specific about what actually constitutes a good or a bad outcome. Thus, the theory applies both to the difference between financial rewards and punishments (i.e., utilitarian feedback) and to the difference between correct trials and error trials (i.e., performance feedback). This factor makes the FRN especially relevant both for researchers interested in economic decision-making, and for researchers interested in the neural mechanisms of cognitive control. Gehring and Willoughby (2002) questioned the existence of such a highly generic mechanism by showing a medial frontal negativity (MFN) that was elicited by negative utilitarian feedback (loss of money vs. gain of money) but not by stimuli revealing that an alternative choice would have yielded a better result than the actual choice (i.e., performance feedback). However, in a subsequent study Nieuwenhuis et al. (2004b) showed that when feedback stimuli conveyed both utilitarian and performance information the FRN reflects either dimension of the information, depending on which aspect of the feedback is highlighted by the physical properties of the stimuli.
Finally, another contribution of RL-theory to the growing interest in the FRN is that this theory affirms that the events are treated as good or bad outcomes not in absolute terms but relative to their relationship with expectations. Since the FRN would reflect the difference between the actual and the expected outcomes, the FRN would potentially provide both a measure of the impact of events (i.e., experienced value) and a measure of anteceding expectations (i.e., expected value). Most of the research about outcome processing in the brain has explored these factors; several research groups have studied the modulation of the FRN by both properties of the outcome, such as valence (loss vs. win) and magnitude (large vs. small), or by properties of the context in which the outcome is presented, such as reward probability and the potential magnitude of reward. Few studies have also explored another critical component of the RL-theory, which implies that the amplitude of the FRN in a particular trial should predict the degree of behavioral adjustment in subsequent trials (Holroyd and Coles, 2002). The following section discusses the RL-theory in light of studies investigating the relationship between the FRN and outcome valence, outcome magnitude, outcome probability, and behavioral adjustment.
Modulation of the FRN by Outcomes and Contexts
The RL-theory predicts that the amplitude of the FRN is correlated with the magnitude of the RPE. If this is true, negative outcomes should elicit larger FRNs than positive outcomes; larger negative outcomes should elicit larger FRNs than smaller negative outcomes; and unexpected negative outcomes should elicit larger FRNs than expected negative outcomes. The RL-theory also predicts an interaction between valence and magnitude and between valence and probability. For example, small gains should be associated with greater FRNs than large gains. Finally, according to the RL-theory the amplitude of the FRN on a given trial should predict behavioral adjustment on subsequent trials. This section examines the RL-theory in the context of studies addressing how the FRN is modulated by outcome valence (negative vs. positive), outcome magnitude (large vs. small), outcome probability (unlikely vs. likely), and behavioral adjustment.
Outcome valence
The most consistent finding in the FRN literature is that such ERP component is larger for negative feedback than for positive feedback. This valence dependency has been confirmed using monetary rewards (Gehring and Willoughby, 2002; Holroyd and Coles, 2002; Yeung and Sanfey, 2004; Goyer et al., 2008; Yu et al., 2011) and non-monetary performance feedback (Miltner et al., 1997; Luu et al., 2003; Frank et al., 2005; Luque et al., 2012). These studies have typically measured the FRN amplitude by computing the difference between the ERP elicited after negative outcomes and the ERP elicited after positive outcomes. This approach is most common because the overlap between the FRN and the P3 may distort the FRN net amplitude. However, a disadvantage to this method is that it does not inform whether the FRN corresponds to a negative deflection after negative outcomes or to a positive deflection after positive outcomes, or both. Indeed, a recent hypothesis argues that the differences between positive and negative outcomes could be better explained by a positivity associated with better than expected outcomes, rather than a negativity associated with worse than expected ones (Holroyd et al., 2008). This theory proposes that unexpected outcomes, regardless of their valence, elicit a negative deflection known as the N200 (Towey et al., 1980) and that trials with unexpected rewards elicit a feedback correct-related positivity (fCRP) that cancels the effect of the N200 component in the scalp-recorded ERP.
Recently, several groups have begun to quantify the FRN independently for positive and negative outcomes, and most of these studies have found greater modulation of the FRN to positive outcomes than to negative outcomes (Cohen et al., 2007; Eppinger et al., 2008, 2009; San Martin et al., 2010; Foti et al., 2011; Kreussel et al., 2012). However, at the same time these effects tend to disconfirm the fCRP-hypothesis, by showing a greater positivity for expected compared with unexpected gains (Oliveira et al., 2007; Wu and Zhou, 2009; San Martin et al., 2010; Chase et al., 2011; Yu et al., 2011; Kreussel et al., 2012).
In summary, the FRN is consistently larger for negative as compared to positive feedback, but it is not clear if this effect is due to a negative deflection following negative outcomes, a positive deflection following positive outcomes, both alternatives, or some unmentioned factor. This issue is critical for future research regarding the neurocognitive role of the FRN, and as of yet remains an open question.
Outcome magnitude
According to the RL-theory the FRN should not reflect a main effect of outcome magnitude given that the magnitude of an outcome can be evaluated as positive or negative only after considering the valence of such outcome (i.e., large gains are better than small gains, but the opposite is true for losses). Confirming this observation, several studies have noted the absence of a main effect of outcome magnitude (Mars et al., 2004; Yeung and Sanfey, 2004; Toyomaki and Murohashi, 2005; Polezzi et al., 2010). In a now-classic study, Yeung and Sanfey (2004) asked participants to select between cards that were unpredictably associated with monetary gains and losses of variable magnitude. The authors conclude in this study that the FRN was larger for losses compared with gains and did not show a main effect of outcome magnitude. However, others have reported that the FRN is larger for small magnitude outcomes, regardless of valence (Wu and Zhou, 2009; Gu et al., 2011a; Kreussel et al., 2012). This second set of studies did not elaborate on this finding, but rather as Wu and Zhou (2009) have suggested, the discrepancy could be related either to different approaches to measuring the FRN, or to the use of experimental paradigms where the expectancy toward reward magnitude was emphasized.
If the FRN does indeed code the difference between expected outcomes and obtained outcomes, it should be sensitive to deviations from expected reward magnitude (e.g., if the person expected a large loss, a small loss should be treated as a positive outcome). Aligned with this notion, Goyer et al. (2008) found that a model that includes both valence and magnitude explained a larger proportion of the variance associated with the FRN compared with a model that only includes valence. They also found that the difference between the FRN elicited by monetary losses and the FRN elicited by monetary gains was greater for large magnitude outcomes (i.e., −25¢ minus +25¢) than for small magnitude outcomes (i.e. −5¢ minus +5¢). In a related study, Holroyd et al. (2004a) found that the amplitude of the FRN depends on the range of possible outcomes in a given block. For instance, winning +2.5¢ elicited a smaller FRN when it was the best possible outcome than the same result when winning +5¢ was also possible. This FRN effect provided strong support to the RL-theory, according to which the FRN is an indirect measure of a firing pattern of DA neurons in the midbrain. Indeed, it has been shown that the same reinforcement can lead to phasic DA decreases if it is smaller than expected or phasic DA increases if it is larger than expected (Tobler et al., 2005).
Other studies have found effects that are less consistent with the RL-theory. Hajcak et al. (2006) reported that the FRN amplitude did not scale with the magnitude of the loss; Bellebaum et al. (2010) reported that the size of the potential reward affected the FRN amplitude in response to non-reward, but not to positive feedback; and San Martin et al. (2010) reported that the size of the potential reward affected the FRN amplitude in response to monetary gains, but not to monetary losses. These studies suggest that under some circumstances the FRN does not mirror a graded RPE. In this sense therefore, the existing literature is not entirely consistent with the RL-theory and new research is necessary to help resolve these disparate results.
Outcome probability
The probability of the experienced outcome is yet another factor that is crucial to the RPE and the RL-theory predicts that the probability of reward would affect the FRN responses to the upcoming outcome. Unexpected losses therefore should be associated with greater negativities than expected losses, and unexpected gains should elicit greater positivities than expected gains. Several studies have supported this prediction (Holroyd and Coles, 2002; Nieuwenhuis et al., 2002; Holroyd et al., 2003, 2009, 2011; Potts et al., 2006; Hajcak et al., 2007; Goyer et al., 2008; Walsh and Anderson, 2011; Luque et al., 2012). Nevertheless, with the exception of the study by Potts et al. (2006), these studies employed a difference wave approach (i.e., loss minus gain) and therefore it is not clear if the effects were driven by large negativities associated with unexpected negative outcomes or large positivities associated with unexpected positive outcomes, or some combination of both. Moreover, the difference wave approach might find an effect in the direction predicted by the RL-theory even if unexpected outcomes, regardless of valence, are associated with larger negativities, provided that the effect is greater for losses. A study by Oliveira et al. (2007) shed further light onto this issue. They found that positive feedback and negative feedback elicited a similarly large FRN when the actual feedback and the expected feedback mismatched. They reported a tendency for people to be overly optimistic about their own performance, and they found that because of this tendency there were three times more mismatches between expectancy and the actual feedback for erroneous trials than for correct trials. Similarly, other studies have reported that the FRN is elicited by unexpected outcomes, regardless of valence (Wu and Zhou, 2009; Chase et al., 2011; Yu et al., 2011).
Oliveira et al. (2007) proposed that the FRN reflects the response of the ACC to violations of expectancy in general (i.e., unsigned RPE), and not only for unexpected negative outcomes. Similar conclusions have been reached using fMRI in humans (Walton et al., 2004; Aarts et al., 2008; Metereau and Dreher, 2012) and single-unit recordings in monkeys (Niki and Watanabe, 1979; Akkal et al., 2002; Ito et al., 2003; Matsumoto and Hikosaka, 2009; Hayden et al., 2011). Also, a recent computational model (Alexander and Brown, 2011) has been able to simulate the FRN under the assumption that the mPFC is activated by surprising events, regardless of valence. However, some studies have reported effects that are hard to reconcile with Oliveira et al.'s account of the FRN. Both Potts et al. (2006) and Cohen et al. (2007) found that unexpected gains elicited larger positivities than expected gains, not larger negativities as Oliveira and colleagues suggest. In another study, Hajcak et al. (2005) reported that the FRN was equally large for expected and unexpected negative feedback; a result that is inconsistent with both Oliveira et al.'s account and with the RL-theory.
Broadly speaking, the evidence reviewed above does not permit conclusions about the modulation of the FRN by outcome probability. Both the RL-theory and the hypothesis proposed by Oliveira et al. (2007) have received empirical support, but these are mutually exclusive accounts. New research is needed in order to understand the relationship between FRN amplitude and outcome probability.
Behavioral adjustment
Convergent evidence has implicated the ACC in the flexible adjustment of behavior on the basis of changes in reward and punishment values. For example, a recent study showed that cingulate lesions in monkeys impaired the ability to use previous reinforcements to guide choice behavior (Kennerley et al., 2006). Another study, using a reward-based reversal learning paradigm, identified cells in the ACC of the monkey brain that fired only if reward was less than anticipated and if such reduction in reward was followed by changes in action selection (Shima and Tanji, 1998). In humans, fMRI studies of reversal learning have reported the same effect (Bush et al., 2002; O'Doherty et al., 2003a).
If neural RPE signals are used to guide decision-making and the FRN reflects the impact of such signals in the ACC, as suggested by the RL-theory, FRN magnitudes in response to decision outcomes should be related to adjustments in overt behavior. Evidence for this hypothesis was demonstrated by Luu et al. (2003) using a task in which participants had to respond to a target arrow with the hand indicated by the direction in which the arrow pointed. Luu and colleagues found that the amplitude of the FRN elicited by a feedback indicating that the response was slow correlated with subsequent speed of response. In another study, Frank et al. (2005) showed that the difference between the FRN elicited by negative and positive feedback correlated, across participants, with the difference between subjects' ability to learn to avoid negative feedback versus to learn to approach positive feedback. Similarly, Bellebaum and Daum (2008) found that violations of reward predictions modulated the FRN only in participants that were able to learn, through trial and error, and use a rule determining reward probability. However, the last two studies (Frank et al., 2005; Bellebaum and Daum, 2008) do not rule out the possibility that the reported effects are by-products of learning instead of underlying causes of individual differences in overt behavior.
Other researchers have found results that conflict with the RL-theory. For example, Mars et al. (2004) found that while subjects used the information provided by error feedback to adjust their behavior in a time judgment task, no relationship between FRN amplitude and behavioral adjustments was found. Specifically, more precise adjustments were evident following more versus less informative feedback, and larger behavioral adjustments were seen when subjects received feedback that implied a large error than following feedback suggesting a small error. However, FRN amplitude was smaller in the informative conditions, and was not influenced by the degree of error. In another study, Walsh and Anderson (2011) presented a decision-making task with two conditions: a no instruction condition in which participants received feedback about whether their choices were rewarded and had to learn reward probabilities by trial and error, and an instruction condition, where they additionally received a description of the association between cues and reward probabilities before performing the task. The authors found that instruction eliminated the association between feedback and behavioral adjustment, but the FRN still changed with experience in the instruction condition. These studies suggest that, at least under some circumstances, the FRN can be elicited in the absence of behavioral adjustment, and behavior can be adjusted in the absence of a concomitant FRN effect.
Other studies do report a relationship between the FRN and behavioral adjustment, but not in the direction that the RL-theory predicts. For example, Yeung and Sanfey (2004) reported that after a large loss of money participants tended to repeat the selection of the larger magnitude (i.e., risky option), particularly if the preceding outcome elicited a large FRN. Similarly, using a computer Blackjack gambling task, Hewig et al. (2007) found that those participants exhibiting increased FRNs when losing after making risky choices showed a strong tendency to switch to even more risky choices. These results represent a challenge for the RL-theory, which predicts that participants should be less likely rather than more likely to perseverate in a response strategy after a large FRN.
One of the strengths of the RL-theory is that it provides predictions at the trial-by-trial level. Computational models can be fitted to behavioral data and the FRN can be compared with values derived from such models, such as the trial-by-trial fluctuation of the RPE during the task. Two studies found dissimilar results using that approach. In one study, participants played a competitive game called “matching pennies” against a simulated opponent (Cohen and Ranganath, 2007). On each trial, the subject and the computer opponent each selected one of two targets. If the subject and the computer opponent chose the same target, the subject lost one point, and if they chose opposite targets, the subject won one point. Supporting the RL-theory, the authors found that the FRN elicited by losses was more negative when subjects chose the opposite versus the same target on the subsequent trial. In another study, Chase et al. (2011) reported that the FRN amplitude was positively correlated with the magnitude of the negative RPE (−RPE), but negative outcomes that preceded behavioral adjustments were not accompanied by enlarged FRNs. The discrepancy between the results found by these studies could be explained by the characteristics of their experimental paradigms. The task used by Cohen and Ranganath (2007) discouraged the adoption of explicit rules or strategies. The simulated opponent was preprogrammed to choose randomly unless it was possible to find and exploit patterns in the behavior of the human participant. In contrast, in the study by Chase et al. (2011) participants were instructed to switch choice behavior only when they were sure that a rule determining the stimuli-response mapping had changed, and not after each exception to that rule. Indeed, in the study by Chase and colleagues the first violations of the rules were associated with greater −RPEs and greater FRNs, but not with behavioral adjustment.
In summary, different studies suggest that, at least in some learning situations, the processes underlying the generation of the FRN might be dissociated from the processes responsible for behavioral adjustments. It is possible, like the results from Chase et al. (2011) and Walsh and Anderson (2011) suggest, that the RL-theory does not account for behavioral adjustment under some learning context, but still predicts trial-by-trial fluctuations in FRN amplitude.
The P3
Historical Antecedents of the P3
The P3 is a positive large-amplitude ERP component with a broad, midline scalp distribution, and with peak latency between 300 and 600 ms following presentation of stimuli. First reported in 1965 (Desmedt et al., 1965; Sutton et al., 1965), the P3 is perhaps the single most studied component of the ERP, probably because it is elicited in many cognitive tasks involving any sensory modality. The antecedent conditions of the P3 have been extensively explored using the so-called “oddball” paradigm (Duncan-Johnson and Donchin, 1977; Pritchard, 1981; Murphy and Segalowitz, 2004; Campanella et al., 2012), in which low-frequency target stimuli (oddballs) are embedded in a train of non-target stimuli (standards). Typically the subject is required to actively respond to each target stimulus. Using this task, Duncan-Johnson and Donchin (1977) were the first to report a correlation between the probability of an eliciting stimulus and the P3 amplitude. Specifically, they found that P3 amplitude was inversely proportional to the frequency or probability of the target stimuli in an oddball sequence (for an extensive review, see Nieuwenhuis et al., 2005a).
The Context Updating Hypothesis and the Neural Sources of the P3
The most influential account of the P3 is the context updating hypothesis (Donchin, 1981; Donchin and Coles, 1988). In this framework, the P3 indexes brain activity underlying the stimuli-induced revision of a mental model of the task at hand. If subsequent stimuli deliver information that mismatches with part of such model, the model is updated, with the amplitude of the P3 being proportional to the amount of cognitive resources employed during the revision the model.
There is an important degree of uncertainty regarding the neural generators of the P3. Lutzenberger et al. (1987) argued that large-amplitude potentials like the P3 must have widespread sources. Consistent with this view, intracranial P3-like activity has been recorded from multiple cortical areas (for a review see Soltani and Knight, 2000). Probably the most likely neural sources of the P3b can be found in a region that includes the temporal-parietal junction (TPJ; consisting of the supramarginal gyrus and caudal parts of the superior temporal gyrus) and adjacent areas (Kiss et al., 1989; Smith et al., 1990; Halgren et al., 1995). Indeed, studies have shown that lesions of the TPJ region produce marked reductions of the P3 associated with infrequent, task-relevant stimuli (Yamaguchi and Knight, 1992; Verleger et al., 1994; Knight and Scabini, 1998).
P3-like potentials have also been observed in medial temporal lobe (MTL) structures, including hippocampus and amygdala in cats (Kaga et al., 1992), monkeys (Paller et al., 1992), and humans (Halgren et al., 1980; McCarthy et al., 1989; Smith et al., 1990). The thalamus is another deep structure that produces P3-like potentials in humans (Yingling and Hosobuchi, 1984). However, biophysical considerations indicate that the possible contributions of deep cortical structures like the MTL and thalamus to the scalp-recorded electroencephalogram (EEG) are much too small (Lutzenberger et al., 1987; Birbaumer et al., 1990) and thus, the available evidence suggests that TPJ and adjacent areas are the most likely neural sources of the scalp-recorded P3.
The Locus Coeruleus–P3 Hypothesis
A more recent hypothesis proposes that the P3 reflects the neuromodulatory effect of the locus coeruleus (LC) norepinephrine (NE) system in the neocortex (Nieuwenhuis et al., 2005a). This hypothesis was the first account of the P3 based on neuroscientific knowledge and it is supported both by similarities between the target areas of NE projections and likely P3 generators, and by similarities between the antecedent conditions for phasic increases in NE and the antecedent conditions for P3 generation. The main claim of the “LC-P3 hypothesis” is that the P3 reflects a LC-mediated enhancement of neural responsivity in the cortex to task-relevant stimuli. Indeed, it has been shown that NE increases the responsivity of target neurons, and that such enhanced gain produces an increase in the signal-to noise ratio of subsequent processing (Servan-Schreiber et al., 1990).
The main source of NE for the forebrain is provided by the LC which is a small nucleus in the pontine region of the brain stem. Within the neocortex, NE innervation is particularly high in the prefrontal cortex and parietal cortex (Levitt et al., 1984; Morrison and Foote, 1986; Foote and Morrison, 1987). Dense NE innervation for the thalamus, amygdala, and hippocampus has also been reported (Morrison and Foote, 1986). Building on this overlap between NE targets and likely P3 generations is one of the main strengths of the LC-P3 hypothesis. Also, the latency of the LC phasic response (150–200 ms post-stimulus), added to the time course of NE physiological effects (100–200 ms post-discharge) (Aston-Jones et al., 1980; Foote et al., 1983; Pineda, 1995; Berridge and Waterhouse, 2003), is consistent with the typical P3 latency.
Phasic activity of the LC-NE system is sensitive to various aspects of stimuli to which the P3 amplitude is also sensitive, including motivational significance, probability of occurrence, and attention allocation. Similarly to the P3, LC phasic activity is generally more closely related to the arousing nature of a given stimulus than to the affective valence of the stimulus (Berridge and Waterhouse, 2003). For example, phasic LC responses occur following both positive and negative outcomes, provided that such outcomes require the animals to update their model of the environment (Rasmussen et al., 1986).
Noting the similarities between the context updating hypothesis and the involvement of the LC-NE system during learning, Nieuwenhuis (2011) reinterpreted the LC-P3 hypothesis and the context updating hypothesis as being complementary rather than competing accounts of the P3; with the LC-P3 hypothesis providing a mechanistic explanation, grounded on neuroscientific evidence, to the more abstract context updating hypothesis. In this sense, an important contribution of the LC-P3 theory is to connect studies of the P3 with new accounts about the role of the NE-mediated attention during learning. For example, it has been proposed that phasic NE is a generic signal indicating the need to attend to the environment and learn from it (Bouret and Sara, 2004), or that phasic NE encodes unexpected uncertainty (i.e., surprise in supposedly non-volatile environments) about the current state within a task, and serves to interrupt the ongoing processing associated with the default task state (Yu and Dayan, 2005; Dayan and Yu, 2006). By conceptually linking the P3 with theories about the role of NE-mediated attention during learning, the LC-P3 hypothesis provides an initial framework to interpret the functional role of the P3 in tasks involving outcome evaluation.
Modulation of the P3 by Outcomes and Contexts
Applied to outcome evaluation and learning, the LC-P3 hypothesis predicts that outcomes associated with high levels of arousal or task-relevance (e.g., indicating the need for behavioral adjustment) will be associated with a large P3. The specific conditions associated with these factors, however, may change depending on the goal and the context of the task at hand. For example, losing when losses could be avoided should elicit a larger P3 than losing when losses are unavoidable. This section evaluates whether the LC-P3 hypothesis provides a plausible account of the P3 in brain studies of outcome evaluation and feedback-guided learning.
Outcome valence
Early ERP studies of outcome processing suggested that feedback indicating a bad performance (i.e., negative feedback) elicited larger P3s than positive feedback (Squires et al., 1973; Picton et al., 1976). However, subsequent studies showed that, when equated for probability of occurrence, positive and negative feedback elicited equally large P3s (Campbell et al., 1979), and that large P3s were elicited both by negative outcomes when participants thought they made a correct response, and by positive feedback when participants thought they made an incorrect response (Horst et al., 1980). These studies suggested that the effect of valence might have artificially emerged from the well-known sensitivity of the P3 to stimulus probability (see results with the oddball paradigm on section “Historical Antecedents of the P3”).
Using monetary rewards, Yeung and Sanfey (2004) found that the P3 was sensitive to reward magnitude but insensitive to reward valence. This result seems to fit with the idea that a large P3 is observed both to affectively negative and positive stimuli, provided that they are matched according to subjective ratings of arousal (Johnston et al., 1986; Keil et al., 2002). The claim that the P3 is insensitive to outcome valence has been supported by some studies (Sato et al., 2005; Yeung et al., 2005; Gu et al., 2011b), but there has been more evidence implicating that outcome valence does in fact modulate the P3. Some studies have reported that losses elicit larger P3s than gains (Frank et al., 2005; Cohen et al., 2007; Hewig et al., 2007), but surprisingly most of the studies have reported the opposite pattern, with larger P3s for gains than for losses (Toyomaki and Murohashi, 2005; Hajcak et al., 2007; Bellebaum and Daum, 2008; Wu and Zhou, 2009; Bellebaum et al., 2010; Polezzi et al., 2010; Zhou et al., 2010; Gu et al., 2011a; Kreussel et al., 2012). Importantly, in all of the studies reporting larger P3s after losses than after gains, participants could actually learn to avoid losses, so that the probability of losing decreased with practice. It is, therefore, possible that the effect of valence was artificially derived from the P3 sensitivity to stimulus probability. However, two studies reporting larger P3s for gains than for losses share the same characteristic (Bellebaum and Daum, 2008; Bellebaum et al., 2010). Moreover, Bellebaum and Daum (2008) found larger P3s for gains than for losses even when gains were more likely than losses. New studies might try to determine under what conditions the P3 is insensitive to outcome valence, and under what conditions it is increased for gains or increased for losses.
Recent studies have reported effects of the interaction between outcome valence and other factors. Wu and Zhou (2009) found that the difference between the P3 elicited by gains and losses (larger P3s for gains in this case) was eliminated when the amount of reward was inconsistent with the expectation built upon a preceding cue. Following the idea that the P3 might reflect the amount of cognitive resources allocated for stimulus processing (Donchin and Coles, 1988), the authors suggested that the inconsistency between the actual and the expected outcome magnitude might capture a large amount of attentional resources, such that the attention allocated to process outcome valence is reduced. A goal for new studies might be to replicate this effect and to specify under what circumstances the brain might need to distribute attentional resources between outcome variables.
In another study, Zhou et al. (2010) reported the same main effect of outcome valence, but in this case the difference was enlarged by action choice as compared to inaction. The authors suggest that action might increase the affective significance of gains. However, it is not clear why this would not also be true for losses.
In summary, the current evidence seems to disconfirm the hypothesis that the P3 is insensitive to outcome valence. Nevertheless, it is not clear whether the P3 is larger for gains or larger for losses. New studies are needed in order to clarify this issue. Also, and as already commented, early studies (Campbell et al., 1979; Horst et al., 1980) suggested that the effect of valence might result from the effect of another, more primitive, variable such as stimulus probability or affective involvement during the tasks. Contemporaneous studies need to take this possibility into consideration, during the experimental design, data analysis, and when forming conclusions.
Outcome magnitude
The independent coding model (Yeung and Sanfey, 2004) claims that the P3 is insensitive to valence and sensitive to outcome magnitude. Although the evidence reviewed in the previous section poses doubts on the first claim, the second claim is widely supported by the literature. Most of the studies that have tested the main effect of outcome magnitude have found more positive P3 responses to large magnitude outcomes than to small magnitude outcomes (Toyomaki and Murohashi, 2005; Goyer et al., 2008; Wu and Zhou, 2009; Bellebaum et al., 2010; Polezzi et al., 2010; Gu et al., 2011a; Kreussel et al., 2012). In order to interpret this effect, these studies have typically appealed to the concept of “motivational significance,” or the relevance of a stimulus for the current task (Duncan-Johnson and Donchin, 1977).
Motivationally significant or relevant stimuli are presumed to capture a large amount of attentional resources, and the P3 amplitude is supposed to scale with those attentional resources (Donchin and Coles, 1988; Nieuwenhuis et al., 2005a). Under this interpretation, large magnitude outcomes might receive more attention and have a greater impact on memory than small magnitude outcomes because they are more relevant for the final outcome of the session (cf. Adcock et al., 2006). Indeed, in the context of monetarily rewarded tasks, large magnitude outcomes have a greater impact in cumulative earnings, for better or for worse, depending on the outcome valence.
Recent studies have reported other effects involving outcome magnitude. Bellebaum et al. (2010) reported that not only the magnitude of the actual reward, but also the potential reward magnitude, modulated the P3. They used a task in which, on each trial, subjects had to guess the location of a coin that was hidden in one of six boxes. At the beginning of each trial, subjects were informed about the amount of money that they could win (i.e., 5¢, 20¢, or 50¢). Interestingly, the P3 elicited by non-rewarding outcomes (i.e., 0¢) scaled with the magnitude of the informed potential reward. In another study, Wu and Zhou (2009) found that the effect of outcome magnitude (i.e., larger P3s for large rewards than for small rewards) was eliminated when the reward amount was inconsistent with the expectation built upon a preceding cue. As already mentioned in the previous section, Wu and Zhou suggested that all attentional resources might be allocated by such inconsistency with expectations, leaving no resources available to process other outcome-related variables.
In summary, the P3 is consistently modulated by outcome magnitude, being more positive for large magnitude outcomes than for small magnitude outcomes. This effect may reflect that the motivational significance of outcomes scales with outcome magnitude. According to this interpretation both the actual and the expected magnitude of rewards determine the motivational significance of the outcome.
Outcome probability
Studies employing the classic oddball paradigm and manipulating the probability of the appearance of a particular stimulus showed that the P3 is more positive for infrequent stimuli than for frequent stimuli (Courchesne et al., 1977; Duncan-Johnson and Donchin, 1977; Johnson and Donchin, 1980). In the context of learning tasks, an early study reported that the largest P3s were elicited by negative feedback when participants thought they made a correct response, and by positive feedback when participants thought they made an incorrect response (Horst et al., 1980). Given this evidence, it has long been recognized that the P3 is modulated by stimulus probability, with more positive amplitudes elicited by unlikely our unexpected stimuli than to likely or expected stimuli.
Studies using monetary rewards and manipulating reward probability have widely supported the conclusion that the P3 is larger for unexpected outcomes than for expected outcomes, regardless of valence (Hajcak et al., 2005, 2007; Bellebaum and Daum, 2008; Wu and Zhou, 2009; Xu et al., 2011). These results are consistent with the context updating hypothesis stating that unexpected outcomes signal the need to update a mental model and that the P3 reflects the amount of cognitive resources allocated to this updating process. The results are also consistent with the LC-P3 hypothesis stating that the P3 reflects the impact of phasic NE in the neocortex. Indeed, it has been noted that increasing stimuli probability reduces the magnitude of phasic LC responses (Alexinsky et al., 1990; Aston-Jones et al., 1994), and phasic LC responses have been proposed to code unexpected uncertainty or surprise (Yu and Dayan, 2005; Dayan and Yu, 2006).
Two studies, however, have reported results that pose doubts into the modulatory effect of probability over the P3. While the P3s following wins were significantly affected by probability in these two studies, with unlikely wins eliciting larger P3s than likely wins, P3s for losses were either not modulated by probability (Cohen et al., 2007) or modulated in the opposite direction (Kreussel et al., 2012) (i.e., larger P3s for expected losses than for unexpected losses). After observing that the probability effect was maximal over anterior sites, Cohen et al. (2007), suggested that their results were more related with the FRN than with the P3. Indeed, and given that the FRN for unexpected losses tend to be larger (i.e., more negative) than the FRN for expected losses, the overlap between the FRN and the P3 may also explain the effect reported by Kreussel et al. (2012). Disentangling different outcome-related ERP components is one of the main challenges for ERP studies of outcome processing.
In summary, a large amount of evidence, coming both from classical studies using the oddball paradigm and from more recent studies using learning and gambling tasks, support the idea that the P3 is larger for unexpected events than for expected events, regardless of the event valence. Some studies have reported results contradicting this claim, but methodological considerations related to the overlap between the P3 and the FRN appear to provide a plausible explanation for such discrepancy.
Behavioral adjustment
Learning-guided decision-making tasks typically require storing and dynamically adjusting information about state-choice-outcome contingencies. Convergent evidence suggests that the LC-NE system contributes to learning this type of association. For example, it has been noted that the injection of a drug that increased the firing of LC neurons in rats promotes the animal's adaptation to changes in the behavioral requirements of a reinforcement-learning task (Devauges and Sara, 1990). In the monkey, LC activation has been reported to be restricted to task-relevant stimuli that require a behavioral shift (Aston-Jones et al., 1999). Building on this evidence, Bouret and Sara (2005) proposed that phasic NE could provoke or facilitate the dynamic reorganization of the neural networks determining the behavioral output. Similarly, Dayan and Yu (2006) proposed that NE signals encode unexpected surprise, serving to interrupt the ongoing processing and concentrate attentional resources in behavioral adjustment.
If, as the LC-P3 hypothesis proposes, the P3 reflects the NE-mediated enhancement of signal transmission in the cortex during the stimulus-induced revision of an internal model of the environment, a large P3 at the time of outcome processing should predict a large behavioral adjustment. In one of the few studies that have explored the relationship between the P3 and behavioral adjustment, Yeung and Sanfey (2004) found that individual differences in the P3 elicited by alternative, unchosen outcomes were related to behavioral adjustments. After dividing the participants into two groups on the basis of the size of their behavioral adjustment after trials in which they failed to select a card associated with a large win, they found that the difference in the P3 amplitude elicited by large-gain and large-loss alternative outcomes was larger in the participants showing a greater behavioral adjustment. In another study, Chase et al. (2011) found that P3 amplitude was greater for negative outcomes that preceded behavioral adjustment than for negative outcomes that did not precede behavioral adjustment in a probabilistic reversal-learning paradigm.
Evidence that dissociates the P3 from behavioral adjustment has also been reported. Specifically, Frank et al. (2005) found that, across participants, the difference between the ability to learn to avoid losses and the ability to learn to approach gains was predicted by the difference in the FRN amplitude elicited by negative and positive feedback, but not by the P3. Interestingly, the results found by Chase et al. (2011) showed the opposite pattern, with the P3 but not the FRN predicting behavioral adjustment. A possible reason for the discrepancy is that in Frank et al.'s study the need for adjustment was always signaled by losses, and as already reviewed the most consistent finding about the FRN is that it is larger for losses than for gains. In contrast, in the study by Chase and colleagues participants were explicitly instructed to switch choice behavior only when they were sure that a rule determining the stimulus-outcome contingencies had changed, and not after losing per se. Another possibility for this discrepancy is that both the FRN and the P3 code behavioral adjustment, but their relative involvement in this process depends on the goal of the task at hand or on the level of information processing that is required to adjust behavior. Evidence is scant to strongly support any of these possibilities, and future research is needed to determine the factors determining the relative involvement of the FRN and the P3 in behavioral adjustment.
Discussion
The studies reviewed here suggest that the brain mechanism underlying the FRN and the P3 are consistently involved in outcome processing, but at the same time the literature shows an important degree of scientific uncertainty regarding the factors that affect the amplitude of these ERP components. Although the studies reviewed here have enough trials per condition (Marco-Pallares et al., 2011) and sample sizes that allowed them to detect significant results, reaching conclusions that generalize across studies have proven to be difficult. The FRN tends to increase its amplitude in response to negative outcomes and the P3 tends to increase its amplitude in response to arousing or task-relevant outcomes, but there are important exceptions in the literature that weaken the generalization of these statements.
In order to advance a model that integrates FRN and P3 effects in a unitary and real-time account of outcome processing, ERP studies of outcome processing will have to address methodological, empirical, and conceptual challenges in the upcoming years. First, optimizing paradigm design will be critical for being able to generalize conclusions about the role of the FRN and the P3 during outcome processing. Second, in order to allow a reliable comparison between studies, the field will have to advance toward standard methods to measure the ERP components. Third, ERP studies have intrinsic limitations for identifying brain regions and networks involved in outcome processing. Complementary techniques should be increasingly used to overcome such limitations. Fourth, studies interested in the effect of outcome variables on the FRN and P3 have typically involved different task demands (e.g., passive observation of outcome, active decision-making, etc.). Determining the impact of task demands on these ERP components will be crucial to be able to generalize conclusions about their role. Finally, the studies reviewed here portray outcome processing as two relatively disconnected processes. Our modern view of the brain, however, suggests that outcome processing probably involves several subprocesses concurring and interacting in time. ERP studies should take advantage of their high temporal resolution to study the temporal cascade of outcome processing in the brain. These challenges are further discussed in the remainder of the article.
Methodological Challenge: Optimizing Paradigm Designs
During the last decade, ERP studies of outcome processing have produced a wealth of evidence employing a rich variety of experimental paradigms. While this heterogeneity is needed to generalize conclusions about the properties of a neural correlate beyond a particular experimental task, there is a risk associated with designing tasks that do not adequately isolate the variables of interest. For example, as already commented, early P3 studies concluded that negative feedback elicited larger P3s than positive feedback, but subsequently Campbell et al. (1979) showed that these results mostly reflected the well-known effect of stimulus probability on the P3.
More contemporaneous studies, especially some of those using feedback-guided learning, could be associated with a similar confound. These groups found that losing was associated with larger P3s than winning, but it is equally possible that this result reflects a probability effect: as learning progresses, losing becomes less likely. In order to dissociate a valence effect from a learning/probability/expectancy effect, ERP studies of outcome processing might benefit from measuring ERP components on different stages of the experimental session, or by comparing the ERP response from participants that demonstrate learning with those who do not.
Another confound that can limit the validity of the results is the potential gap between the goal that the participants really pursue during an experimental session and the goal that the experimenter is trying to elicit. For example, paradigms designed for studying the modulatory effect of participants' prediction on brain activity elicited by gains and losses might unintentionally emphasize the goal of predicting the upcoming outcome. If participants are asked about their belief on the incoming outcome, they might even perceive a predicted loss as a positive feedback (i.e., a correct prediction). To limit the effect of this confound while still being able study the effect of participants' predictions, paradigms should emphasize the goal of ensuring gains and avoiding losses. For example, experimenters could make outcomes contingent on participants' behavior and not purely probabilistic. Also, researchers could benefit from verbal reports about the task and its goals during pilot studies.
The study of outcome processing is associated with a large number of variables (i.e., outcome valence, outcome magnitude, expectancy toward magnitude, probability of winning, learning, motivation, etc.) that, depending on the task at hand, might covary in a way that undermines empirical results. Paradigms should be designed acknowledging this complexity in a way that minimizes the chances of introducing confounds.
Methodological Challenge: Advancing Toward Standard Measurement Methods
An inherent difficulty of the ERP technique is that, because of potential component overlap, the comparison of ERPs elicited by different experimental conditions is often difficult to interpret (Luck, 2005). Although this problem is inherently technical, the manner by which it is addressed in each study may determine empirical results and functional interpretations. Research focused on the FRN has traditionally tried to solve this issue by creating difference waves (e.g., loss minus win) (Miltner et al., 1997; Holroyd and Coles, 2002; Nieuwenhuis et al., 2002; Mars et al., 2004; Hajcak et al., 2005, 2007; Potts et al., 2006; Holroyd et al., 2009, 2011; Walsh and Anderson, 2011; Xu et al., 2011). The problem with this approach is that it does not resolve the question whether the FRN corresponds to a negative deflection in one condition or to a positive deflection in the other condition. More recent studies have highlighted the importance of this issue by showing that, when gains and losses are measured separately, even the definition of the FRN as negative deflection that distinguishes negative from positive outcomes can be questioned (Oliveira et al., 2007).
Two alternative methods to measure ERP components have been systematically employed in ERP studies of outcome processing. The first approach is to compute the mean amplitude in a time window defined for each ERP component (e.g., 200–300 ms for the FRN) post-onset of the outcome and to enter that mean amplitudes into statistical analyses (Gehring and Willoughby, 2002; Ruchsow et al., 2002; Nieuwenhuis et al., 2004b; Yeung et al., 2005; Cohen et al., 2007; Cohen and Ranganath, 2007; Hewig et al., 2007; Bellebaum and Daum, 2008; Goyer et al., 2008; Polezzi et al., 2010; San Martin et al., 2010; Zhou et al., 2010; Gu et al., 2011b; Kreussel et al., 2012; Luque et al., 2012). This method has the strength of increasing the signal-to-noise ratio in addition to allowing for better trial-by-trial measures. Nevertheless, it assumes equivalent baseline for each ERP component in different conditions. Given the overlap between P3 and FRN, this issue is particularly critical in outcome processing research. Indeed, the net amplitude of the FRN can be shifted to more positive values if the FRN for a particular condition is superimposed in a P3 that is particularly large. The second alternative method is to measure the base-to-peak difference for each deflection (e.g., defining the FRN as the difference between the most positive point and the most negative point in the 150–350 ms time window post-onset of feedback) (Holroyd et al., 2004a; Yeung and Sanfey, 2004; Frank et al., 2005; Toyomaki and Murohashi, 2005; Hajcak et al., 2006; Oliveira et al., 2007; Bellebaum et al., 2010; Chase et al., 2011). One problem with this measure is that it is especially susceptible to noise. Noise can be controlled by replacing the base of comparison with the mean amplitude in a time window around the base and by replacing the peak measure with the mean amplitude in a time window around the peak. A more fundamental problem is that the measure of the base (e.g., the beginning of the FRN) may be affected by the adjacent deflection (e.g., P2-like positivity).
Alternative methods that have been explored in recent years include isolating the activity associated with a particular ERP component using bandpass filtering (e.g., measuring the FRN after removing the slower frequency to which the P3 is associated) (Luu et al., 2003; Wu and Zhou, 2009; Gu et al., 2011a; Yu et al., 2011) or using temporospatial principal component analysis (PCA) (Carlson et al., 2011; Foti et al., 2011) or independent component analysis (ICA) (Gentsch et al., 2009). One limitation of these methods is that they strongly rely on decisions made by the researchers regarding the parameters used for bandpass filtering or during the identification of the PCA-derived or ICA-derived components that will be considered to represent the ERP components of interest. However, the emerging use of data-driven approaches for the selection of independent components (Wessel and Ullsperger, 2011) suggests that ICA might become a standard technique to decompose ERP components in the incoming years.
The problem of component overlap is inherent to all ERP research. In recent years, researches interested in how the brain processes outcomes have begun to consider ways to dissociate the contribution of the FRN and the P3 to the scalp-recorded ERP signal, and different methods have been proposed to achieve this goal. Given that the choice made among different measurement methods can have consequences, both in the empirical results that are found and in the conclusions that are proposed, future research should explore the strengths and limitations of different measurement methods and advance toward standard practices.
Methodological Challenge: Employing Multi-Methods Approaches (ERP/Time-Frequency, ERP/fMRI)
The temporal resolution of the EEG signal allows studying the neurocognitive mechanism of outcome processing and learning with a high level of temporal detail (milliseconds). The extraction of outcome-locked ERPs from the EEG signal is a good way to identify and study regularities in the neural processing of outcomes. However, ERP research presents two important limitations: it is relatively insensitive to the involvement of large-scale brain networks and it has a limited ability to identify the neural generators of the scalp-recorded signals (i.e., poor spatial resolution). Complementary techniques can be used to overcome such limitations. Specifically, time-frequency-based approaches could complement ERP studies by shedding light on the interactions among large-scale networks and fMRI can be used in a complementary fashion to help resolve the so-called inverse problem: a given distribution of scalp-recorded electrical activity could have been generated by any one of a large number of different sets of neural generators.
In order to provide a better account of the neural dynamics of outcome processing, EEG/ERP studies can benefit from the time-frequency information that is present in the same EEG signal from which ERPs are extracted. Event-related oscillations can be extracted using time-frequency decomposition analyses such as complex wavelet convolutions, from which one can obtain estimates of phase synchronization, spectral coherence, power-power correlations, spectral Granger causality, and cross-frequency coupling among recording sites (for a review see Cohen et al., 2011). Assessing large-scale networks is especially important to better understand the dynamics of feedback-guided learning, given that learning probably corresponds to changes in connectivity between neural populations (Hebb, 1949). Specifically, time-frequency-based approaches could be used to test hypotheses about inter-regional coupling between areas that probably interact during feedback-guided learning, such as medial prefrontal, sensory, and motor cortices.
A fundamental problem of EEG/ERP research is the inverse problem, by which a given distribution of scalp-recorded electrical activity could have been generated by any one of a large number of different sets of neural generators. Despite this problem, convergent evidence suggests that it is highly probable that the neural sources of the FRN are located in the mPFC (Miltner et al., 1997; Gehring and Willoughby, 2002; Ruchsow et al., 2002; Holroyd et al., 2004b; van Schie et al., 2004; Muller et al., 2005; Nieuwenhuis et al., 2005b; Hewig et al., 2007; Yu and Zhou, 2009; Yu et al., 2011). The picture is much less clear for the P3, for which neural sources are probably distributed over different regions of the cortex. In the same way that theories of the FRN have benefited from theories about the functional role of the mPFC (and vice versa), our understanding of the functional role of the P3 and its subcomponents during outcome processing could benefit from a more precise identification of its neural sources.
Increasingly, efforts are being made in order to use fMRI data to constrain the solution of the algorithms used for source localization analyses for ERPs. Although this approach does not completely solve the inverse problem, it increases the likelihood of identifying the actual sources of ERP activity. Especially promising is the use of joint ERP and fMRI ICA that have been used to reveal a number of cortical and subcortical areas involved in the generation of the response-locked ERN (Edwards et al., 2012).
Empirical Challenge: Determining the Impact of Task Demands
One of the primary goals of the brain is the adaptive control of behavior, but the exact definition of what constitutes an adaptive behavior may vary across situations, and different brain mechanisms can be recruited to guide behavior depending on the demands imposed by the task at hand. ERP studies of outcome processing have employed experimental paradigms whose behavioral demands range from the passive observation of monetary gains and losses in a computer screen (e.g., Yeung et al., 2005; Potts et al., 2006) to feedback-guided decision-making tasks requiring the inference of probabilistic rules governing state-outcome contingencies (e.g., Bellebaum and Daum, 2008; Chase et al., 2011; Walsh and Anderson, 2011). This breadth is a rich source of evidence, but at the same time is a likely factor underlying the difficulty for extracting generalizable conclusions about how outcome properties affect each ERP component.
The question of what ERP component better predicts behavioral adjustment is a good example of how different tasks may recruit different mechanisms to achieve the same overall goal (e.g., to accumulate monetary rewards). Behavioral adjustment might depend on the system underlying the FRN in tasks de-incentivizing the adoption of an explicit rule (cf. Cohen and Ranganath, 2007), and on the system underlying the P3 in tasks incentivizing the use of explicit probabilistic beliefs (cf. Chase et al., 2011). This possible dissociation has an interesting parallel with a distinction proposed by Daw et al. (2005) between model-free reinforcement learning (mediated by the basal ganglia in a way that is consistent with the RL-theory of the FRN) and model-based reinforcement learning (mediated by brain regions that have shown P3-like activity, such as the DLPFC, and MTL structures).
Future studies might elucidate whether different ERP components reflect the recruitment of different learning systems depending upon different tasks demands. In general, it is foreseeable that future ERP studies will, in greater proportion, be concerned with how different task demands and task contexts modulate the effect that outcomes have on ERP components.
Conceptual Challenge: Understanding the Temporal Cascade of Outcome Processing in the Brain
ERP studies of outcome processing have been focused primarily on the FRN and secondarily on the P3. There have been few attempts to present an account of outcome processing in the brain that integrates FRN and P3 effects. In this regard, the independent coding model (Yeung and Sanfey, 2004) is the dominant proposal to date. By proposing that the FRN codes valence but is insensitive to magnitude and that the P3 shows the opposite pattern, this model presents these two ERP components as measures reflecting brain processes that are completely independent from each other. Moreover, the temporal succession between the frontally distributed FRN and the parietally distributed P3 is not taken into account; for this model it does not matter if valence is evaluated before magnitude or if magnitude is evaluated before valence. Actually, the temporal cascade of ERP components is generally disregarded even in studies finding evidence that contradicts the independent coding model.
ERP effects are not always the real-time reflections of the underlying processes. For example, according to the LC-P3 hypothesis the P3 is an indirect index of LC phasic responses occurring 300–400 ms before the peak of the P3. However, the ubiquitous temporospatial succession between the frontally distributed FRN and the parietally distributed P3 probably reveals something meaningful about the manner in which the brain processes and learns from outcomes. Much of the contribution of ERP research to cognitive neuroscience has to do with describing the temporal cascade of neurocognitive processes involved in solving a particular task. ERP studies of outcome processing, in this regard, have traditionally sub-exploited the temporal resolution of the ERP technique.
There is also evidence suggesting that the FRN and the parietal P3 are not the only ERP deflections reflecting outcome processing in the brain. Studies could benefit from measuring the frontally distributed P3a that peaks 60–80 ms earlier than the P3b (Courchesne et al., 1975; Squires et al., 1975; Friedman et al., 2001), which according to a visual inspection is present in most of the ERP studies of outcome processing. These studies might also benefit from quantifying a positive deflection that typically began 150 ms after stimulus onset in frontal sites, and that according to Goyer et al. (2008) is modulated by outcome magnitude. This positivity, which in terms of latency could be referred to as “P2,” (see Figure 1A) probably represents an early stage of the slow-wave P3a on which the FRN is superimposed.
By considering the whole sequence of ERP deflections that are modulated by outcomes, ERP studies might contribute to build models of outcome processing that incorporate the inter-relationship between different cognitive processes that probably take part in outcome processing and learning, such as attention, valuation, and memory.
Outlook
In order to control the behavior in an adaptive manner the brain has to learn how certain situations predict positive or negative outcomes and what actions are appropriate in a given situation. ERP research has shown that the brain is able to evaluate and learn from outcomes within a few hundred milliseconds of their occurrence. However, the accumulated literature presents a high degree of scientific uncertainty regarding the factors that modulate different ERP components during outcome processing. The FRN, in most cases, is larger for negative than for positive outcomes, but the effect of outcome magnitude and outcome probability over the FRN is less clear and contradicting evidence has been found regarding the relationship between the FRN and behavioral adjustment. The P3 is consistently more positive for large magnitude and unexpected outcomes than for small magnitude and expected outcomes, respectively, but the modulatory effect of feedback valence and the relationship between P3 and behavioral adjustment is much less clear.
During the last decade, ERP research has accumulated rich evidence of how outcomes are processed in the human brain. The next decade of research will probably be characterized by growing efforts to reach conclusions that generalize across task scenarios demands. In doing so, this research will advance our understanding of how the brain is able produce adaptive behavior in a large variety of situations.
Conflict of Interest Statement
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
I thank my mentor, Dr Scott Huettel, as well as Dr. L. Gregory Appelbaum, Dr. Scott de Marchi, Dr. Marty Woldorff, Dr. Henry Yin, Vijeth Iyengar, and Sara Bergman, for helpful comments and discussions regarding this paper.
References
Aarts, E., Roelofs, A., and Van Turennout, M. (2008). Anticipatory activity in anterior cingulate cortex can be independent of conflict and error likelihood. J. Neurosci. 28, 4671–4678.
Adcock, R. A., Thangavel, A., Whitfield-Gabrieli, S., Knutson, B., and Gabrieli, J. D. (2006). Reward-motivated learning: mesolimbic activation precedes memory formation. Neuron 50, 507–517.
Akkal, D., Bioulac, B., Audin, J., and Burbaud, P. (2002). Comparison of neuronal activity in the rostral supplementary and cingulate motor areas during a task with cognitive and motor demands. Eur. J. Neurosci. 15, 887–904.
Alexander, W. H., and Brown, J. W. (2011). Medial prefrontal cortex as an action-outcome predictor. Nat. Neurosci. 14, 1338–1344.
Alexinsky, T., Aston-Jones, G., Rajkowski, J., and Revay, R. S. (1990). Physiological correlates of adaptive behavior in a visual discrimination task in monkeys. Soc. Neurosci. Abstr. 16, 164.
Amiez, C., Joseph, J. P., and Procyk, E. (2005). Anterior cingulate error-related activity is modulated by predicted reward. Eur. J. Neurosci. 21, 3447–3452.
Aston-Jones, G., Rajkowski, J., and Cohen, J. (1999). Role of locus coeruleus in attention and behavioral flexibility. Biol. Psychiatry 46, 1309–1320.
Aston-Jones, G., Rajkowski, J., Kubiak, P., and Alexinsky, T. (1994). Locus coeruleus neurons in monkey are selectively activated by attended cues in a vigilance task. J. Neurosci. 14, 4467–4480.
Aston-Jones, G., Segal, M., and Bloom, F. E. (1980). Brain aminergic axons exhibit marked variability in conduction velocity. Brain Res. 195, 215–222.
Barber, A. D., and Carter, C. S. (2005). Cognitive control involved in overcoming prepotent response tendencies and switching between tasks. Cereb. Cortex 15, 899–912.
Barto, A. G. (1995). “Adaptive critics and the basal ganglia,” in Models of Information Processing in the Basal Ganglia, eds J. Houk, J. Davis, and D. Beiser (Cambridge, MA: MIT Press), 215–232.
Bellebaum, C., and Daum, I. (2008). Learning-related changes in reward expectancy are reflected in the feedback-related negativity. Eur. J. Neurosci. 27, 1823–1835.
Bellebaum, C., Polezzi, D., and Daum, I. (2010). It is less than you expected: the feedback-related negativity reflects violations of reward magnitude expectations. Neuropsychologia 48, 3343–3350.
Berridge, C. W., and Waterhouse, B. D. (2003). The locus coeruleus-noradrenergic system: modulation of behavioral state and state-dependent cognitive processes. Brain Res. Brain Res. Rev. 42, 33–84.
Birbaumer, N., Elbert, T., Canavan, A. G., and Rockstroh, B. (1990). Slow potentials of the cerebral cortex and behavior. Physiol. Rev. 70, 1–41.
Botvinick, M. M. (2007). Conflict monitoring and decision making: reconciling two perspectives on anterior cingulate function. Cogn. Affect. Behav. Neurosci. 7, 356–366.
Botvinick, M. M., Braver, T. S., Barch, D. M., Carter, C. S., and Cohen, J. D. (2001). Conflict monitoring and cognitive control. Psychol. Rev. 108, 624–652.
Botvinick, M. M., Cohen, J. D., and Carter, C. S. (2004). Conflict monitoring and anterior cingulate cortex: an update. Trends Cogn. Sci. 8, 539–546.
Bouret, S., and Sara, S. J. (2004). Reward expectation, orientation of attention and locus coeruleus-medial frontal cortex interplay during learning. Eur. J. Neurosci. 20, 791–802.
Bouret, S., and Sara, S. J. (2005). Network reset: a simplified overarching theory of locus coeruleus noradrenaline function. Trends Neurosci. 28, 574–582.
Breiter, H. C., Aharon, I., Kahneman, D., Dale, A., and Shizgal, P. (2001). Functional imaging of neural responses to expectancy and experience of monetary gains and losses. Neuron 30, 619–639.
Bush, G., Luu, P., and Posner, M. I. (2000). Cognitive and emotional influences in anterior cingulate cortex. Trends Cogn. Sci. 4, 215–222.
Bush, G., Vogt, B. A., Holmes, J., Dale, A. M., Greve, D., Jenike, M. A., et al. (2002). Dorsal anterior cingulate cortex: a role in reward-based decision making. Proc. Natl. Acad. Sci. U.S.A. 99, 523–528.
Campanella, S., Delle-Vigne, D., Kornreich, C., and Verbanck, P. (2012). Greater sensitivity of the P300 component to bimodal stimulation in an event-related potentials oddball task. Clin. Neurophysiol. 123, 937–946.
Campbell, K. B., Courchesne, E., Picton, T. W., and Squires, K. C. (1979). Evoked potential correlates of human information processing. Biol. Psychol. 8, 45–68.
Carlson, J. M., Foti, D., Mujica-Parodi, L. R., Harmon-Jones, E., and Hajcak, G. (2011). Ventral striatal and medial prefrontal BOLD activation is correlated with reward-related electrocortical activity: a combined ERP and fMRI study. Neuroimage 57, 1608–1616.
Carter, C. S., Braver, T. S., Barch, D. M., Botvinick, M. M., Noll, D., and Cohen, J. D. (1998). Anterior cingulate cortex, error detection, and the online monitoring of performance. Science 280, 747–749.
Chase, H. W., Swainson, R., Durham, L., Benham, L., and Cools, R. (2011). Feedback-related negativity codes prediction error but not behavioral adjustment during probabilistic reversal learning. J. Cogn. Neurosci. 23, 936–946.
Cohen, M. X., Elger, C. E., and Ranganath, C. (2007). Reward expectation modulates feedback-related negativity and EEG spectra. Neuroimage 35, 968–978.
Cohen, M. X., and Ranganath, C. (2007). Reinforcement learning signals predict future decisions. J. Neurosci. 27, 371–378.
Cohen, M. X., Wilmes, K., and Vijver, I. (2011). Cortical electrophysiological network dynamics of feedback learning. Trends Cogn. Sci. 15, 558–566.
Coricelli, G., Critchley, H. D., Joffily, M., O'Doherty, J. P., Sirigu, A., and Dolan, R. J. (2005). Regret and its avoidance: a neuroimaging study of choice behavior. Nat. Neurosci. 8, 1255–1262.
Courchesne, E., Hillyard, S. A., and Courchesne, R. Y. (1977). P3 waves to the discrimination of targets in homogeneous and heterogeneous stimulus sequences. Psychophysiology 14, 590–597.
Courchesne, E., Hillyard, S. A., and Galambos, R. (1975). Stimulus novelty, task relevance and the visual evoked potential in man. Electroencephalogr. Clin. Neurophysiol. 39, 131–143.
Daw, N. D., Niv, Y., and Dayan, P. (2005). Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nat. Neurosci. 8, 1704–1711.
Dayan, P., and Yu, A. J. (2006). Phasic norepinephrine: a neural interrupt signal for unexpected events. Network 17, 335–350.
Delgado, M. R., Locke, H. M., Stenger, V. A., and Fiez, J. A. (2003). Dorsal striatum responses to reward and punishment: effects of valence and magnitude manipulations. Cogn. Affect. Behav. Neurosci. 3, 27–38.
Delgado, M. R., Nystrom, L. E., Fissell, C., Noll, D. C., and Fiez, J. A. (2000). Tracking the hemodynamic responses to reward and punishment in the striatum. J. Neurophysiol. 84, 3072–3077.
De Martino, B., Kumaran, D., Seymour, B., and Dolan, R. J. (2006). Frames, biases, and rational decision-making in the human brain. Science 313, 684–687.
Desmedt, J. E., Debecker, J., and Manil, J. (1965). [Demonstration of a cerebral electric sign associated with the detection by the subject of a tactile sensorial stimulus. The analysis of cerebral evoked potentials derived from the scalp with the aid of numerical ordinates]. Bull. Acad. R. Med. Belg. 5, 887–936.
Devauges, V., and Sara, S. J. (1990). Activation of the noradrenergic system facilitates an attentional shift in the rat. Behav. Brain Res. 39, 19–28.
Donamayor, N., Marco-Pallares, J., Heldmann, M., Schoenfeld, M. A., and Munte, T. F. (2011). Temporal dynamics of reward processing revealed by magnetoencephalography. Hum. Brain Mapp. 32, 2228–2240.
Donchin, E., and Coles, M. G. H. (1988). Is the P300 component a manifestation of context updating? Behav. Brain Sci. 11, 357–374.
Duncan-Johnson, C. C., and Donchin, E. (1977). On quantifying surprise: the variation of event-related potentials with subjective probability. Psychophysiology 14, 456–467.
Edwards, B. G., Calhoun, V. D., and Kiehl, K. A. (2012). Joint ICA of ERP and fMRI during error-monitoring. Neuroimage 59, 1896–1903.
Elliott, R., Newman, J. L., Longe, O. A., and Deakin, J. F. (2003). Differential response patterns in the striatum and orbitofrontal cortex to financial reward in humans: a parametric functional magnetic resonance imaging study. J. Neurosci. 23, 303–307.
Eppinger, B., Kray, J., Mock, B., and Mecklinger, A. (2008). Better or worse than expected? Aging, learning, and the ERN. Neuropsychologia 46, 521–539.
Eppinger, B., Mock, B., and Kray, J. (2009). Developmental differences in learning and error processing: evidence from ERPs. Psychophysiology 46, 1043–1053.
Everitt, B. J., Hutcheson, D. M., Ersche, K. D., Pelloux, Y., Dalley, J. W., and Robbins, T. W. (2007). The orbital prefrontal cortex and drug addiction in laboratory animals and humans. Ann. N.Y. Acad. Sci. 1121, 576–597.
Falkenstein, M., Hohnsbein, J., Hoormann, J., and Blanke, L. (1991). Effects of crossmodal divided attention on late ERP components. II. Error processing in choice reaction tasks. Electroencephalogr. Clin. Neurophysiol. 78, 447–455.
Fein, G., and Chang, M. (2008). Smaller feedback ERN amplitudes during the BART are associated with a greater family history density of alcohol problems in treatment-naive alcoholics. Drug Alcohol Depend. 92, 141–148.
Foote, S. L., Bloom, F. E., and Aston-Jones, G. (1983). Nucleus locus ceruleus: new evidence of anatomical and physiological specificity. Physiol. Rev. 63, 844–914.
Foote, S. L., and Morrison, J. H. (1987). Extrathalamic modulation of cortical function. Ann. Rev. Neurosci. 10, 67–95.
Foti, D., and Hajcak, G. (2009). Depression and reduced sensitivity to non-rewards versus rewards: evidence from event-related potentials. Biol. Psychol. 81, 1–8.
Foti, D., Weinberg, A., Dien, J., and Hajcak, G. (2011). Event-related potential activity in the basal ganglia differentiates rewards from nonrewards: temporospatial principal components analysis and source localization of the feedback negativity. Hum. Brain Mapp. 32, 2207–2216.
Frank, M. J., Woroch, B. S., and Curran, T. (2005). Error-related negativity predicts reinforcement learning and conflict biases. Neuron 47, 495–501.
Franken, I. H., Van Den Berg, I., and Van Strien, J. W. (2010). Individual differences in alcohol drinking frequency are associated with electrophysiological responses to unexpected nonrewards. Alcohol. Clin. Exp. Res. 34, 702–707.
Fridberg, D. J., Queller, S., Ahn, W. Y., Kim, W., Bishara, A. J., Busemeyer, J. R., et al. (2010). Cognitive mechanisms underlying risky decision-making in chronic cannabis users. J. Math. Psychol. 54, 28–38.
Friedman, D., Cycowicz, Y. M., and Gaeta, H. (2001). The novelty P3, an event-related brain potential (ERP) sign of the brain's evaluation of novelty. Neurosci. Biobehav. Rev. 25, 355–373.
Gehring, W. J., Goss, B., Coles, M. G. H., Meyer, D. E., and Donchin, E. (1993). A neural system of error detection and compensation. Psychol. Sci. 4, 385–390.
Gehring, W. J., and Willoughby, A. R. (2002). The medial frontal cortex and the rapid processing of monetary gains and losses. Science 295, 2279–2282.
Gentsch, A., Ullsperger, P., and Ullsperger, M. (2009). Dissociable medial frontal negativities from a common monitoring system for self- and externally caused failure of goal achievement. Neuroimage 47, 2023–2030.
Gold, J. M., Waltz, J. A., Prentice, K. J., Morris, S. E., and Heerey, E. A. (2008). Reward processing in schizophrenia: a deficit in the representation of value. Schizophr. Bull. 34, 835–847.
Goudriaan, A. E., Oosterlaan, J., De Beurs, E., and Van Den Brink, W. (2006). Psychophysiological determinants and concomitants of deficient decision making in pathological gamblers. Drug Alcohol Depend. 84, 231–239.
Goyer, J. P., Woldorff, M. G., and Huettel, S. A. (2008). Rapid electrophysiological brain responses are influenced by both valence and magnitude of monetary rewards. J. Cogn. Neurosci. 20, 2058–2069.
Gu, R., Lei, Z., Broster, L., Wu, T., Jiang, Y., and Luo, Y. J. (2011a). Beyond valence and magnitude: a flexible evaluative coding system in the brain. Neuropsychologia 49, 3891–3897.
Gu, R., Wu, T., Jiang, Y., and Luo, Y. J. (2011b). Woulda, coulda, shoulda: the evaluation and the impact of the alternative outcome. Psychophysiology 48, 1354–1360.
Hajcak, G., Holroyd, C. B., Moser, J. S., and Simons, R. F. (2005). Brain potentials associated with expected and unexpected good and bad outcomes. Psychophysiology 42, 161–170.
Hajcak, G., Moser, J. S., Holroyd, C. B., and Simons, R. F. (2006). The feedback-related negativity reflects the binary evaluation of good versus bad outcomes. Biol. Psychol. 71, 148–154.
Hajcak, G., Moser, J. S., Holroyd, C. B., and Simons, R. F. (2007). It's worse than you thought: the feedback negativity and violations of reward prediction in gambling tasks. Psychophysiology 44, 905–912.
Halgren, E., Baudena, P., Clarke, J. M., Heit, G., Marinkovic, K., Devaux, B., et al. (1995). Intracerebral potentials to rare target and distractor auditory and visual stimuli. II. Medial, lateral and posterior temporal lobe. Electroencephalogr. Clin. Neurophysiol. 94, 229–250.
Halgren, E., Squires, N. K., Wilson, C. L., Rohrbaugh, J. W., Babb, T. L., and Crandall, P. H. (1980). Endogenous potentials generated in the human hippocampal formation and amygdala by infrequent events. Science 210, 803–805.
Hayden, B. Y., Heilbronner, S. R., Pearson, J. M., and Platt, M. L. (2011). Surprise signals in anterior cingulate cortex: neuronal encoding of unsigned reward prediction errors driving adjustment in behavior. J. Neurosci. 31, 4178–4187.
Hewig, J., Trippe, R., Hecht, H., Coles, M. G., Holroyd, C. B., and Miltner, W. H. (2007). Decision-making in Blackjack: an electrophysiological analysis. Cereb. Cortex 17, 865–877.
Holroyd, C. B., and Coles, M. G. H. (2002). The neural basis of human error processing: reinforcement learning, dopamine, and the error-related negativity. Psychol. Rev. 109, 679–709.
Holroyd, C. B., Krigolson, O. E., Baker, R., Lee, S., and Gibson, J. (2009). When is an error not a prediction error? An electrophysiological investigation. Cogn. Affect. Behav. Neurosci. 9, 59–70.
Holroyd, C. B., Krigolson, O. E., and Lee, S. (2011). Reward positivity elicited by predictive cues. Neuroreport 22, 249–252.
Holroyd, C. B., Larsen, J. T., and Cohen, J. D. (2004a). Context dependence of the event-related brain potential associated with reward and punishment. Psychophysiology 41, 245–253.
Holroyd, C. B., Nieuwenhuis, S., Yeung, N., Nystrom, L., Mars, R. B., Coles, M. G., et al. (2004b). Dorsal anterior cingulate cortex shows fMRI response to internal and external error signals. Nat. Neurosci. 7, 497–498.
Holroyd, C. B., Nieuwenhuis, S., Yeung, N., and Cohen, J. D. (2003). Errors in reward prediction are reflected in the event-related brain potential. Neuroreport 14, 2481–2484.
Holroyd, C. B., Pakzad-Vaezi, K. L., and Krigolson, O. E. (2008). The feedback correct-related positivity: sensitivity of the event-related brain potential to unexpected positive feedback. Psychophysiology 45, 688–697.
Horst, R. L., Johnson, R. Jr., and Donchin, E. (1980). Event-related brain potentials and subjective probability in a learning task. Mem. Cognit. 8, 476–488.
Huettel, S. A., Stowe, C. J., Gordon, E. M., Warner, B. T., and Platt, M. L. (2006). Neural signatures of economic preferences for risk and ambiguity. Neuron 49, 765–775.
Ito, S., Stuphorn, V., Brown, J. W., and Schall, J. D. (2003). Performance monitoring by the anterior cingulate cortex during saccade countermanding. Science 302, 120–122.
Joel, D., Niv, Y., and Ruppin, E. (2002). Actor-critic models of the basal ganglia: new anatomical and computational perspectives. Neural Netw. 15, 535–547.
Johnson, R. Jr., and Donchin, E. (1980). P300 and stimulus categorization: two plus one is not so different from one plus one. Psychophysiology 17, 167–178.
Johnston, V. S., Miller, D. R., and Burleson, M. H. (1986). Multiple P3s to emotional stimuli and their theoretical significance. Psychophysiology 23, 684–694.
Kable, J. W., and Glimcher, P. W. (2007). The neural correlates of subjective value during intertemporal choice. Nat. Neurosci. 10, 1625–1633.
Kaga, K., Harrison, J. B., Butcher, L. L., Woolf, N. J., and Buchwald, J. S. (1992). Cat ‘P300’ and cholinergic septohippocampal neurons: depth recordings, lesions, and choline acetyltransferase immunohistochemistry. Neurosci. Res. 13, 53–71.
Keil, A., Bradley, M. M., Hauk, O., Rockstroh, B., Elbert, T., and Lang, P. J. (2002). Large-scale neural correlates of affective picture processing. Psychophysiology 39, 641–649.
Kennerley, S. W., Walton, M. E., Behrens, T. E., Buckley, M. J., and Rushworth, M. F. (2006). Optimal decision making and the anterior cingulate cortex. Nat. Neurosci. 9, 940–947.
Kerns, J. G., Cohen, J. D., Macdonald, A. W. 3rd. Cho, R. Y., Stenger, V. A., and Carter, C. S. (2004). Anterior cingulate conflict monitoring and adjustments in control. Science 303, 1023–1026.
Kiehl, K. A., Liddle, P. F., and Hopfinger, J. B. (2000). Error processing and the rostral anterior cingulate: an event-related fMRI study. Psychophysiology 37, 216–223.
Kim, H., Shimojo, S., and O'Doherty, J. P. (2006). Is avoiding an aversive outcome rewarding? Neural substrates of avoidance learning in the human brain. PLoS Biol. 4:e233. doi: 10.1371/journal.pbio.0040233
Kiss, I., Dashieff, R. M., and Lordeon, P. (1989). A parieto-occipital generator for P300, evidence from human intracranial recordings. Int. J. Neurosci. 49, 133–139.
Knight, R. T., and Scabini, D. (1998). Anatomic bases of event-related potentials and their relationship to novelty detection in humans. J. Clin. Neurophysiol. 15, 3–13.
Kramer, U. M., Buttner, S., Roth, G., and Munte, T. F. (2008). Trait aggressiveness modulates neurophysiological correlates of laboratory-induced reactive aggression in humans. J. Cogn. Neurosci. 20, 1464–1477.
Kreussel, L., Hewig, J., Kretschmer, N., Hecht, H., Coles, M. G., and Miltner, W. H. (2012). The influence of the magnitude, probability, and valence of potential wins and losses on the amplitude of the feedback negativity. Psychophysiology 49, 207–219.
Kuhnen, C. M., and Knutson, B. (2005). The neural basis of financial risk taking. Neuron 47, 763–770.
Levitt, P., Rakic, P., and Goldman-Rakic, P. (1984). Region-specific distribution of catecholamine afferents in primate cerebral cortex: a fluorescence histochemical analysis. J. Comp. Neurol. 227, 23–36.
Luck, S. (2005). An Introduction to the Event-Related Potential Technique. Cambridge, MA: MIT Press.
Luque, D., Lopez, F. J., Marco-Pallares, J., Camara, E., and Rodriguez-Fornells, A. (2012). Feedback-related brain potential activity complies with basic assumptions of associative learning theory. J. Cogn. Neurosci. 24, 794–808.
Lutzenberger, W., Elbert, T., and Rockstroh, B. (1987). A brief tutorial on the implications of volume conduction for the interpretation of the EEG. J. Psychophysiol. 1, 81–89.
Luu, P., Tucker, D. M., Derryberry, D., Reed, M., and Poulsen, C. (2003). Electrophysiological responses to errors and feedback in the process of action regulation. Psychol. Sci. 14, 47–53.
Marco-Pallares, J., Cucurell, D., Munte, T. F., Strien, N., and Rodriguez-Fornells, A. (2011). On the number of trials needed for a stable feedback-related negativity. Psychophysiology 48, 852–860.
Mars, R. B., De Brujin, E. R. A., Hulstijn, W., and Miltner, W. H. R. (2004). “What if I told you: “You were wrong”? Brain potentials and behavioral adjustments elicited by feedback in a time-estimation task,” in Errors, Conflict, and the Brain. Current Opinions on Performance Monitoring, eds M. Ullsperger and M. Falkenstein (Leipzig: MPI of Cognitive Neuroscience), 129–134.
Matsumoto, M., and Hikosaka, O. (2009). Two types of dopamine neuron distinctly convey positive and negative motivational signals. Nature 459, 837–841.
McCarthy, G., Wood, C. C., Williamson, P. D., and Spencer, D. D. (1989). Task-dependent field potentials in human hippocampal formation. J. Neurosci. 9, 4253–4268.
Metereau, E., and Dreher, J. C. (2012). Cerebral correlates of salient prediction error for different rewards and punishments. Cereb. Cortex. doi: 10.1093/cercor/bhs037. [Epub ahead of print].
Miltner, W. H., Lemke, U., Weiss, T., Holroyd, C., Scheffers, M. K., and Coles, M. G. (2003). Implementation of error-processing in the human anterior cingulate cortex: a source analysis of the magnetic equivalent of the error-related negativity. Biol. Psychol. 64, 157–166.
Miltner, W. H. R., Braun, C. H., and Coles, M. G. H. (1997). Event-related brain potentials following incorrect feedback in a time-estimation task: evidence for a “generic” neural system for error detection. J. Cogn. Neurosci. 9, 788–798.
Montague, P. R., Dayan, P., and Sejnowski, T. J. (1996). A framework for mesencephalic dopamine systems based on predictive Hebbian learning. J. Neurosci. 16, 1936–1947.
Morris, S. E., Heerey, E. A., Gold, J. M., and Holroyd, C. B. (2008). Learning-related changes in brain activity following errors and performance feedback in schizophrenia. Schizophr. Res. 99, 274–285.
Morrison, J. H., and Foote, S. L. (1986). Noradrenergic and serotoninergic innervation of cortical, thalamic, and tectal visual structures in Old and New World monkeys. J. Comp. Neurol. 243, 117–138.
Muller, S. V., Moller, J., Rodriguez-Fornells, A., and Munte, T. F. (2005). Brain potentials related to self-generated and external information used for performance monitoring. Clin. Neurophysiol. 116, 63–74.
Mullette-Gillman, O. A., Detwiler, J. M., Winecoff, A., Dobbins, I., and Huettel, S. A. (2011). Infrequent, task-irrelevant monetary gains and losses engage dorsolateral and ventrolateral prefrontal cortex. Brain Res. 1395, 53–61.
Murphy, T. I., and Segalowitz, S. J. (2004). Eliminating the P300 rebound in short oddball paradigms. Int. J. Psychophysiol. 53, 233–238.
Nakamura, K., Roesch, M. R., and Olson, C. R. (2005). Neuronal activity in macaque SEF and ACC during performance of tasks involving conflict. J. Neurophysiol. 93, 884–908.
Nieuwenhuis, S. (2011). “Learning, the P3, and the locus coeruleus-norepinephrine system,” in Neural Basis of Motivational and Cognitive Control, eds R. Mars, J. Sallet, M. F. Rushworth, and N. Yeung (Cambridge: MIT Press), 209–222.
Nieuwenhuis, S., Aston-Jones, G., and Cohen, J. D. (2005a). Decision making, the P3, and the locus coeruleus-norepinephrine system. Psychol. Bull. 131, 510–532.
Nieuwenhuis, S., Slagter, H. A., Von Geusau, N. J., Heslenfeld, D. J., and Holroyd, C. B. (2005b). Knowing good from bad: differential activation of human cortical areas by positive and negative outcomes. Eur. J. Neurosci. 21, 3161–3168.
Nieuwenhuis, S., Holroyd, C. B., Mol, N., and Coles, M. G. (2004a). Reinforcement-related brain potentials from medial frontal cortex: origins and functional significance. Neurosci. Biobehav. Rev. 28, 441–448.
Nieuwenhuis, S., Yeung, N., Holroyd, C. B., Schurger, A., and Cohen, J. D. (2004b). Sensitivity of electrophysiological activity from medial frontal cortex to utilitarian and performance feedback. Cereb. Cortex 14, 741–747.
Nieuwenhuis, S., Ridderinkhof, K. R., Talsma, D., Coles, M. G., Holroyd, C. B., Kok, A., et al. (2002). A computational account of altered error processing in older age: dopamine and the error-related negativity. Cogn. Affect. Behav. Neurosci. 2, 19–36.
Niki, H., and Watanabe, M. (1979). Prefrontal and cingulate unit activity during timing behavior in the monkey. Brain Res. 171, 213–224.
O'Doherty, J., Critchley, H., Deichmann, R., and Dolan, R. J. (2003a). Dissociating valence of outcome from behavioral control in human orbital and ventral prefrontal cortices. J. Neurosci. 23, 7931–7939.
O'Doherty, J. P., Dayan, P., Friston, K., Critchley, H., and Dolan, R. J. (2003b). Temporal difference models and reward-related learning in the human brain. Neuron 38, 329–337.
Oliveira, F. T., McDonald, J. J., and Goodman, D. (2007). Performance monitoring in the anterior cingulate is not all error related: expectancy deviation and the representation of action-outcome associations. J. Cogn. Neurosci. 19, 1994–2004.
Onoda, K., Abe, S., and Yamaguchi, S. (2010). Feedback-related negativity is correlated with unplanned impulsivity. Neuroreport 21, 736–739.
Paller, K. A., McCarthy, G., Roessler, E., Allison, T., and Wood, C. C. (1992). Potentials evoked in human and monkey medial temporal lobe during auditory and visual oddball paradigms. Electroencephalogr. Clin. Neurophysiol. 84, 269–279.
Park, S. Q., Kahnt, T., Beck, A., Cohen, M. X., Dolan, R. J., Wrase, J., et al. (2010). Prefrontal cortex fails to learn from reward prediction errors in alcohol dependence. J. Neurosci. 30, 7749–7753.
Paulus, M. P., Feinstein, J. S., Simmons, A., and Stein, M. B. (2004). Anterior cingulate activation in high trait anxious subjects is related to altered error processing during decision making. Biol. Psychiatry 55, 1179–1187.
Paus, T. (2001). Primate anterior cingulate cortex: where motor control, drive and cognition interface. Nat. Rev. Neurosci. 2, 417–424.
Pfabigan, D. M., Alexopoulos, J., Bauer, H., and Sailer, U. (2011). Manipulation of feedback expectancy and valence induces negative and positive reward prediction error signals manifest in event-related brain potentials. Psychophysiology 48, 656–664.
Picton, T. W., Hillyard, S. A., and Galambos, R. (1976). “Habituation and attention in the auditory system,” in Handbook of Sensory Physiology: Vol. V/3. Auditory system, Clinical and Special Topics, eds W. D. Keidel and W. D. Neff (Berlin: Springer-Verlag), 343–389.
Pineda, J. A. (1995). Are neurotransmitter systems of subcortical origin relevant to the electrogenesis of cortical ERPs? Electroencephalogr. Clin. Neurophysiol. Suppl. 44, 143–150.
Polezzi, D., Sartori, G., Rumiati, R., Vidotto, G., and Daum, I. (2010). Brain correlates of risky decision-making. Neuroimage 49, 1886–1894.
Potts, G. F., Martin, L. E., Burton, P., and Montague, P. R. (2006). When things are better or worse than expected: the medial frontal cortex and the allocation of processing resources. J. Cogn. Neurosci. 18, 1112–1119.
Rasmussen, K., Morilak, D. A., and Jacobs, B. L. (1986). Single unit activity of locus coeruleus neurons in the freely moving cat. I. During naturalistic behaviors and in response to simple and complex stimuli. Brain Res. 371, 324–334.
Reuter, J., Raedler, T., Rose, M., Hand, I., Glascher, J., and Buchel, C. (2005). Pathological gambling is linked to reduced activation of the mesolimbic reward system. Nat. Neurosci. 8, 147–148.
Ribas-Fernandes, J. J., Niv, Y., and Botvinick, M. M. (2011). “Neural correlates of hierarchical reinforcement learning,” in Neural Basis of Motivational and Cognitive Control, eds R. Mars, J. Sallet, M. F. Rushworth, and N. Yeung (Cambridge, MA: MIT Press).
Ridderinkhof, K. R., Ullsperger, M., Crone, E. A., and Nieuwenhuis, S. (2004). The role of the medial frontal cortex in cognitive control. Science 306, 443–447.
Ruchsow, M., Grothe, J., Spitzer, M., and Kiefer, M. (2002). Human anterior cingulate cortex is activated by negative feedback: evidence from event-related potentials in a guessing task. Neurosci. Lett. 325, 203–206.
Rushworth, M. F., Walton, M. E., Kennerley, S. W., and Bannerman, D. M. (2004). Action sets and decisions in the medial frontal cortex. Trends Cogn. Sci. 8, 410–417.
San Martin, R., Manes, F., Hurtado, E., Isla, P., and Ibanez, A. (2010). Size and probability of rewards modulate the feedback error-related negativity associated with wins but not losses in a monetarily rewarded gambling task. Neuroimage 51, 1194–1204.
Sato, A., Yasuda, A., Ohira, H., Miyawaki, K., Nishikawa, M., Kumano, H., et al. (2005). Effects of value and reward magnitude on feedback negativity and P300. Neuroreport 16, 407–411.
Scheffers, M. K., Coles, M. G., Bernstein, P., Gehring, W. J., and Donchin, E. (1996). Event-related brain potentials and error-related processing: an analysis of incorrect responses to go and no-go stimuli. Psychophysiology 33, 42–53.
Schultz, W., Dayan, P., and Montague, P. R. (1997). A neural substrate of prediction and reward. Science 275, 1593–1599.
Schultz, W., and Dickinson, A. (2000). Neuronal coding of prediction errors. Annu. Rev. Neurosci. 23, 473–500.
Servan-Schreiber, D., Printz, H., and Cohen, J. D. (1990). A network model of catecholamine effects: gain, signal-to-noise ratio, and behavior. Science 249, 892–895.
Shima, K., and Tanji, J. (1998). Role for cingulate motor area cells in voluntary movement selection based on reward. Science 282, 1335–1338.
Shiv, B., Loewenstein, G., and Bechara, A. (2005). The dark side of emotion in decision-making: when individuals with decreased emotional reactions make more advantageous decisions. Brain Res. Cogn. Brain Res. 23, 85–92.
Smillie, L. D., Cooper, A. J., and Pickering, A. D. (2011). Individual differences in reward-prediction-error: extraversion and feedback-related negativity. Soc. Cogn. Affect. Neurosci. 6, 646–652.
Smith, M. E., Halgren, E., Sokolik, M., Baudena, P., Musolino, A., Liegeois-Chauvel, C., et al. (1990). The intracranial topography of the P3 event-related potential elicited during auditory oddball. Electroencephalogr. Clin. Neurophysiol. 76, 235–248.
Soltani, M., and Knight, R. T. (2000). Neural origins of the P300. Crit. Rev. Neurobiol. 14, 199–224.
Squires, K. C., Hillyard, S. A., and Lindsay, P. H. (1973). Cortical potentials evoked by confirming and disconfirming feedback following an auditory discrimination. Percept. Psychophys. 13, 25–31.
Squires, N. K., Squires, K. C., and Hillyard, S. A. (1975). Two varieties of long-latency positive waves evoked by unpredictable auditory stimuli in man. Electroencephalogr. Clin. Neurophysiol. 38, 387–401.
Sutton, R. S., and Barto, A. G. (1998). Reinforcement Learning: An Introduction. Cambridge, MA: MIT Press.
Sutton, S., Braren, M., Zubin, J., and John, E. R. (1965). Evoked-potential correlates of stimulus uncertainty. Science 150, 1187–1188.
Tobler, P. N., Fiorillo, C. D., and Schultz, W. (2005). Adaptive coding of reward value by dopamine neurons. Science 307, 1642–1645.
Towey, J., Rist, F., Hakerem, G., Ruchkin, D. S., and Sutton, S. (1980). N250 latency and decision time. Bull. Psychon. Soc. 15, 365–368.
Toyomaki, A., and Murohashi, H. (2005). The ERPs to feedback indicating monetary loss and gain on the game of modified “rock–paper–scissors”. Int. Congr. Ser. 1278, 381–384.
Tsujimoto, T., Shimazu, H., and Isomura, Y. (2006). Direct recording of theta oscillations in primate prefrontal and anterior cingulate cortices. J. Neurophysiol. 95, 2987–3000.
Ullsperger, M., and Von Cramon, D. Y. (2003). Error monitoring using external feedback: specific roles of the habenular complex, the reward system, and the cingulate motor area revealed by functional magnetic resonance imaging. J. Neurosci. 23, 4308–4314.
van Schie, H. T., Mars, R. B., Coles, M. G., and Bekkering, H. (2004). Modulation of activity in medial frontal and motor cortices during error observation. Nat. Neurosci. 7, 549–554.
Venkatraman, V., Payne, J. W., Bettman, J. R., Luce, M. F., and Huettel, S. A. (2009). Separate neural mechanisms underlie choices and strategic preferences in risky decision making. Neuron 62, 593–602.
Verleger, R., Heide, W., Butt, C., and Kompf, D. (1994). Reduction of P3b in patients with temporo-parietal lesions. Brain Res. Cogn. Brain Res. 2, 103–116.
Walsh, M. M., and Anderson, J. R. (2011). Modulation of the feedback-related negativity by instruction and experience. Proc. Natl. Acad. Sci. U.S.A. 108, 19048–19053.
Walton, M. E., Devlin, J. T., and Rushworth, M. F. (2004). Interactions between decision making and performance monitoring within prefrontal cortex. Nat. Neurosci. 7, 1259–1265.
Wang, C., Ulbert, I., Schomer, D. L., Marinkovic, K., and Halgren, E. (2005). Responses of human anterior cingulate cortex microdomains to error detection, conflict monitoring, stimulus-response mapping, familiarity, and orienting. J. Neurosci. 25, 604–613.
Wessel, J. R., and Ullsperger, M. (2011). Selection of independent components representing event-related brain potentials: a data-driven approach for greater objectivity. Neuroimage 54, 2105–2115.
Williams, Z. M., Bush, G., Rauch, S. L., Cosgrove, G. R., and Eskandar, E. N. (2004). Human anterior cingulate neurons and the integration of monetary reward with motor responses. Nat. Neurosci. 7, 1370–1375.
Wu, Y., and Zhou, X. (2009). The P300 and reward valence, magnitude, and expectancy in outcome evaluation. Brain Res. 1286, 114–122.
Xu, Q., Shen, Q., Chen, P., Ma, Q., Sun, D., and Pan, Y. (2011). How an uncertain cue modulates subsequent monetary outcome evaluation: an ERP study. Neurosci. Lett. 505, 200–204.
Yamaguchi, S., and Knight, R. T. (1992). Effects of temporal-parietal lesions on the somatosensory P3 to lower limb stimulation. Electroencephalogr. Clinical Neurophysiol. 84, 139–148.
Yeung, N., Botvinick, M. M., and Cohen, J. D. (2004). The neural basis of error detection: conflict monitoring and the error-related negativity. Psychol. Rev. 111, 931–959.
Yeung, N., Holroyd, C. B., and Cohen, J. D. (2005). ERP correlates of feedback and reward processing in the presence and absence of response choice. Cereb. Cortex 15, 535–544.
Yeung, N., and Sanfey, A. G. (2004). Independent coding of reward magnitude and valence in the human brain. J. Neurosci. 24, 6258–6264.
Yingling, C. D., and Hosobuchi, Y. (1984). A subcortical correlate of P300 in man. Electroencephalogr. Clin. Neurophysiol. 59, 72–76.
Yu, R., Zhou, W., and Zhou, X. (2011). Rapid processing of both reward probability and reward uncertainty in the human anterior cingulate cortex. PloS ONE 6:e29633. doi: 10.1371/journal.pone.0029633
Yu, R., and Zhou, X. (2009). To bet or not to bet? The error negativity or error-related negativity associated with risk-taking choices. J. Cogn. Neurosci. 21, 684–696.
Keywords: economics rewards, reward system, feedback-related negativity, reinforcement learning, dopamine, anterior cingulate cortex, P3, locus coeruleus-norepinephrine system
Citation: San Martín R (2012) Event-related potential studies of outcome processing and feedback-guided learning. Front. Hum. Neurosci. 6:304. doi: 10.3389/fnhum.2012.00304
Received: 10 August 2012; Accepted: 22 October 2012;
Published online: 07 November 2012.
Edited by:
Shuhei Yamaguchi, Shimane University, JapanReviewed by:
Jun'Ichi Katayama, Kwansei Gakuin University, JapanKeiichi Onoda, Shimane University, Japan
Copyright © 2012 San Martín R. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits use, distribution and reproduction in other forums, provided the original authors and source are credited and subject to any copyright notices concerning any third-party graphics etc.
*Correspondence: René San Martín, Center for Cognitive Neuroscience, Duke University, Box 90999, Durham, NC 27708, USA. e-mail: rene.sanmartin@duke.edu