- 1Graduate Program in Neuroscience, University of Minnesota, Minneapolis, MN, USA
- 2Department of Neuroscience, University of Minnesota, Minneapolis, MN, USA
Contingency management is an effective treatment for drug addiction. The current explanation for its success is rooted in alternative reinforcement theory. We suggest that alternative reinforcement theory is inadequate to explain the success of contingency management and produce a model based on demand curves that show how little the monetary rewards offered in this treatment would affect drug use. Instead, we offer an explanation of its success based on the concept that it accesses deliberative decision-making processes. We suggest that contingency management is effective because it offers a concrete and immediate alternative to using drugs, which engages deliberative processes, improves the ability of those deliberative processes to attend to non-drug options, and offsets more automatic action-selection systems. This theory makes explicit predictions that can be tested, suggests which users will be most helped by contingency management, and suggests improvements in its implementation.
1. Contingency Management
Contingency management is a method of driving behavioral change through reinforcement with tangible rewards (1). It has been shown to significantly reduce drug-using behavior and increase continuous abstinence rates (2–9).
There are two main variations of contingency management, voucher-based and prize-based. In voucher-based treatment, patients are awarded points that accumulate for submission of drug-negative urine samples (3–5, 8). These points start out very low and can be exchanged for merchandise at any time. For example, in the Higgins et al. (5) study, points for the first clean sample were worth $2.50 and each subsequent sample was worth $1.50 more. By the end of the first month, a drug-negative sample was worth $16.50.
In prize-based treatment, patients earn a chance to win a prize with each drug-negative sample (1, 9–12). Typically, in these studies, prizes were worth around $1, $5, $20, and $100, and the probability to win higher-valued prizes was lower than lower-valued prizes (0.4% for a $100 prize and 68% for a $1 prize). Overall, the chance of the drug-negative sample having a monetary value of anything over a dollar was <7%.
2. Current Theories: Alternative Reinforcement
The success of contingency management is thought to be primarily due to the reinforcing properties of an alternative reward that is offered to patients for remaining abstinent (1, 5). The conceptualization of contingency management is that drug consumption is much like any other consumption of goods, and thus that increasing the cost of drugs should decrease use. Contingency management increases the cost of drugs because it creates an opportunity cost that is lost (the alternative reinforcer) when the user takes drugs. Reasoning for this is based on operant conditioning theories, noting that targeted behaviors increase with reinforcement and decrease in the presence of substitutes (13–16).
In economic terms, this change in use with cost can be measured as elasticity, which can be quantitatively defined as the change in the number of choices selected as cost increases (17–21). To determine this, one can measure the amount of effort an agent is willing to expend in order to gain the reward as a function of the cost. The function that results is called the demand curve (see Figure 1). A commodity that decreases quickly with cost is said to be “highly elastic,” while a commodity that decreases slowly with cost is said to be “inelastic.” (See Table 1 for definitions of the behavioral/neuroeconomic concepts used in this article.)
Figure 1. The canonical structure of a demand curve. Pmax is the point at which elasticity E = −1, and the elasticity transitions from inelastic (|E| < 1) to elastic (|E| > 1).
Quantitatively, the effectiveness of an alternative reinforcement depends on the elasticity of the drug that the alternative reinforcer is substituting for. Although early descriptions of drug use assumed that drugs were taken irrespective of cost, Becker and Murphy (17) pointed out that drugs were economic objects, and, as such, should show elasticity. While there are theoretical reasons to expect differences in the elasticity between drugs and natural rewards (17, 22), nevertheless, drugs do show elasticity both in non-human animals (23–27) and in humans (28–32). This means that increasing the cost (or increasing the size of the alternate options, which increases the opportunity cost) of taking the drug should decrease use. Alternative reinforcement theory predicts that the change in drug use from contingency management should be proportional to the elasticity of drug use.
As reviewed above, contingency management provides relatively low-value monetary rewards for abstinence (especially in the first month of treatment). For example, in voucher-based contingency management, rewards are as low as $2.50 for the very first negative urine sample and $16.25 for a negative sample after remaining abstinent the entire first month (5). The pre-clinical experiments suggest that the value of alternative reinforcement rewards used in contingency management should not reduce drug consumption as much as it does. The pre-clinical experiments suggest that either cost of the drug or magnitude of the reinforcer would need to be significantly higher than what is typically used in contingency management if alternative reinforcement alone were to account for the observed reductions of drug use in contingency management studies.
3. The Problem with the Alternative Reinforcement Theory
If we assume that drugs are economic objects, and thus are subject to change in demand or price, then one way to quantitatively measure level of consumption as a function of price is with a demand curve. The demand curve measures a fundamental concept of consumption: as price of the economic object increases, the consumption of that object will decrease (33, 34).
Figure 1 shows the structure of a typical demand curve. These curves can be well-fit with Eq. 1 measuring the relationship of the cost of some commodity (C) and the consumption of that commodity (Q) (35):
where L measures consumption at C = 1, and b and a are variables that relate to slope and acceleration of the slope, respectively. The slope of the curve predicts the elasticity of the commodity.
Pmax is the point at which the elasticity E = −1, which is the point at which elasticity transitions from <1 unit of decreased use per unit of increased cost (inelastic) to more than 1 unit of decreased use per unit of increased cost (highly elastic). Because the elasticity terms a, b, and the cost C appear in the exponents in Eq. 1, once the cost crosses Pmax [when C > (b + 1)/a)], consumption drops off very quickly. Using demand curves, we can construct a quantitative model to determine how monetary rewards should affect consumption of a drug. As mentioned previously, monetary values early in treatment are relatively low, and demand curve modeling suggests that these rewards alone would affect consumption of the drugs very little.
3.1. Modeling Contingency Management: The Monetary Value of Vouchers Early in Contingency Management Treatment Should have a Negligible Effect on the Consumption of Cocaine
Bruner and Johnson (21) constructed demand curves for individuals that regularly use cocaine by asking subjects how much cocaine they would buy as the cost increased. As noted above, providing alternative rewards increases the cost of the commodity (here the drug) through lost opportunities (an opportunity cost) – if the person takes the drug, then they do not get the alternative reward. This means that we can use these demand curves to predict how this opportunity cost should change the choices made.
Individuals in treatment get a voucher value of $2.50 the first time they provide a clean sample1. Using the assumption that individuals seeking treatment spend an average of $99/day (Petry, personal communication) during a typical day of cocaine use, and given that 1 unit of reward in the Bruner and Johnson (21) data was worth $5 on the street, a starting contingency management reward value of $2.50/day is worth approximately $0.13/unit.
A shift of $0.13/unit on the demand curve would be predicted to produce a negligible effect on cocaine consumption [the Bruner and Johnson (21) demand curve predicts a 1.6% change]. Even at the end of the first month of contingency management treatment, when patients receive a voucher worth $16.25 ($0.82/unit), there should be little change in consumption [the Bruner and Johnson (21) demand curve predicts a 17% change].
In order to quantitatively measure whether these economic changes could explain contingency management’s effects, we took the effect sizes reviewed in the meta-analysis by Lussier et al. (36) and asked how much the Bruner and Johnson (21) demand curve would predict from the economic change in cost alone. Of course, patients seeking treatment have increased costs for drug use due to many factors beyond the simple loss of the contingent alternate reward. Similarly, there is a large variability in how contingency management studies are run and what additional treatments they are paired with. Finally, the Bruner and Johnson (21) analysis is from one set of cocaine addicts, while the studies reviewed by Lussier et al. (36) range from alcoholics to stimulant addicts. Nevertheless, 21/27 studies had predicted changes less than the observed effect size, and the median ratio was that the predicted effect was less than half the observed (median ratio = 0.43). Figure 2 shows the distribution of observed effect sizes against the economically predicted changes. The predicted changes are significantly less than the observed changes (matched pairs median test, p = 0.00008).
Figure 2. Predicted and observed effect sizes of contingency management processes. From the meta-analysis by Lussier et al. (36), we calculated the expected change in demand by applying the contingent alternate reward (in $) to the average demand curve found by Bruner and Johnson (21). This gave a predicted effect size, which was dramatically less than the typical effects observed. See text for additional discussion.
This analysis suggests that the simple economic description of contingency management is inadequate – the rewards offered in contingency management are too small to have the observed effects. We suggest that this is because the microeconomic model on which the economic explanation for contingency management is based is inadequate – human decision-making depends on more than simple cost-benefit analyses. Instead, the human decision-making process is better described as an interaction between multiple competing components (37–43), each of which uses different processes to combine reward information (value) with past experiences (memory) to select actions (make decisions). We suggest that contingency management taps into certain aspects of these multiple decision-making systems to drive behavior to be more likely to reject the drug-taking choice.
4. Valuation
Early psychological and economic research postulated that reinforcers are transituational, meaning that the efficacy of the reinforcer remains consistent across different experimental conditions (44–46). However, studies have shown that reinforcers do not consistently elicit reliable behavioral outputs in different contexts (47).
In the fields of behavioral and neuroeconomics, decisions are assumed to derive from an underlying “value” or “utility” placed on outcomes. However, this value cannot be directly observed experimentally, and thus must be interpreted from experimental conditions. The two primary methods for deriving this value are willingness to pay experiments, in which an agent is given an opportunity to pay a cost for an outcome, and revealed preference experiments, in which an agent is given a choice between two or more options. In willingness-to-pay experiments, the agent has to decide whether to continue to pursue a given option or not. In revealed-preference experiments, the agent has to decide which option to pursue. Importantly, experiments in rats, monkeys, and humans all find differences between how animals value options under these two measurements, often finding incompatible outcomes (43, 47–49). Thus, converting experiments from single option (Go or don’t?) to multiple option (Which one?) can change how animals appear to value a given option.
A typical willingness-to-pay experiment would be the breakpoint procedure, in which an animal presses a lever to receive reward. The first reward is delivered with only a single-lever press, but the second requires two-lever presses, the third requires four, the fourth eight, and so on, doubling each time. At some point, the cost becomes too high and the animal stops pressing the lever (48, 50–52). In humans, willingness-to-pay can be assessed by simply asking “how much would you pay for this outcome?” (47).
By contrast, a typical revealed-preference experiment would provide an animal two levers, one of which provides one type of reward (A), while the other provides another type of reward (B) (48, 52, 53). The animal is only able to select one lever at any given time and thus must choose between the separate options. The implication is that the selected option is more valuable than the non-selected option. In humans, revealed preference can be assessed by asking “which option would you prefer?” (47).
Extensive evidence exists within the behavioral and neuroeconomics literature that these two measures can produce incompatible valuations, in which human subjects may be willing to pay more for option A than for option B, even when they would prefer to take option B when faced with the two options together (47). Recently, Ahmed (48) found in self-administering rats that measuring value by means of a breakpoint procedure (willingness-to-pay) can produce different ordering than when measuring value by means of a choice procedure (revealed preference); that is, subjects were willing to pay more for drug than saccharin but preferred saccharin to drug when given the choice (48). This strongly suggests that value is not an intrinsic (transituational) property, but is highly dependent on the contextual surrounding components.
These analyses implies that single-option experiments, in which an agent is tasked with deciding whether to pursue a given object or not may access different process than multiple-option experiments, in which an agent is tasked with deciding which option to pursue.
4.1. Valuation Inconsistencies Arise from Multiple Decision-Making Systems
Current theories suggest that this underlying lack of transsituationality arises because animals (including humans) make decisions based on several incompatible decision-making systems, each of which processes information about the decision in fundamentally different ways. Because these different systems drive behavior at different times, the same agent can show different valuations under different experimental conditions.
Classically, the idea that valuation is inconsistent and not transsituational has been addressed in terms of dual-process theories that humans (and presumably other animals as well) have two separable components of decision-making, one which is impulsive and depends on reacting to immediate, concrete rewards, and another which is more rational and capable of waiting for larger, more abstract rewards (54–57). Importantly, the impulsive (often called “reactive”) system is not necessarily always chasing positive rewards; it can also avoid negative consequences (58), nevertheless, the key difference in the two dual-process hypothesis is that the impulsive system attends to immediate consequences while the other (cognitive, often called “reflective”) system takes into account farther future consequences (59–62). In many of these discussions, the impulsive system is identified as more “emotional” and more related to an animal’s history, while the rational system is identified with more cognitive processing. In many of these theories, the rational system is assumed to be a self-control system, which inhibits the activity of the impulsive system (63–65), often referred to as a form of “self-control” (66, 67). This theory has a very long history (68–70) and there are good summaries of the modern perspectives on this dichotomy (40, 59, 63, 65, 67, 71). Anatomically, the impulsive system is associated with the nucleus accumbens and amygdala, while the rational system is associated with the prefrontal cortex (54, 56, 57, 59, 72, 73).
Recent computational work examining how agents process information to make a decision (such as taking a drug or not) suggests that multiple action-selection systems compete and interact to produce that decision. Current theories suggest that decisions arise from as many as four separable systems, each depending on different information-processing computations (37, 42, 43, 74–77). Each system uses past experience differently and processes information about the world differently, and thus each has advantages and disadvantages in different situations. An agent that correctly identifies the best action-selection system to use in a given situation will outperform a different agent that does not. Because different systems drive behavior at different times, valuation is not necessarily self-consistent.
Following these recent taxonomies (43), we identify four decision-making systems each of which selects actions through a different computation: (1) reflexes, in which evolutionarily useful stimulus–response pairs are hard-wired within a neural system (78, 79), (2) Pavlovian actions, in which an animal learns when to release a species-specific behavior (80–82), (3) procedural actions, in which arbitrary action chains are stored and released on cue (83, 84), and (4) deliberation, which entails a slow, goal-oriented search and evaluate process (42, 85–87). Each of these systems is instantiated in a different anatomical network – reflexes in spinal cord and brainstem (88), Pavlovian actions with amygdala and the periaqueductal gray (89, 90), procedural with motor cortex, cerebellum, and the basal ganglia (91–93), and deliberation with hippocampus and the prefrontal cortex (87, 94–96).
There are many similarities between the dual-process and multiple decision-making systems theories, particularly in the separation between more automatic and more cognitive systems (40, 43, 65). Both theories, for example, suggest that stress and cognitive load will disrupt the more cognitive systems, shifting behavior to more automatic systems. Both theories suggest that the more automatic systems tend to react to more immediate stimuli, while the more cognitive system is capable of incorporating information that is not immediately present.
However, there are important differences between the theories. For example, the information-processing theories do not imply that the more automatic systems are more impulsive, as hypothesized by the classical dual-process distinction. For example, a fire chief with extensive expertise is using a fast, non-deliberative process to make the right choice (83); no one would argue that a fire chief is making an impulsive choice. The more recent models have shown that intuition and developed expertise arises from a different computational process than emotion, suggesting that these are different systems (43). Additionally, the information-processing theory provides for interacting components that can make cognitive systems react differently in the face of concrete stimuli (97, 98).
In addition, the hypothesized causes of addictive behavior is different in the two theories, which has implications for how contingency management should be used and what modifications would do to its success. These subtle differences between these theories make different predictions and change some of the implications of our fundamental hypothesis (that contingency management accesses deliberative processes, see below). We will address the differences between these theories below, but first we address the main implications of our hypothesis that contingency management accesses deliberative processes, which are similar under the two theories.
5. Hypothesis: Contingency Management Accesses Deliberative Systems
Our hypothesis is that the provision of a concrete, identified, alternative reward in contingency management both engages deliberative processes and improves the ability of those deliberative processes to attend to non-drug options. In a sense, contingency management transitions the drug-valuation process from a willingness-to-pay condition to a revealed-preference condition. In addition, we propose that the concrete and more immediate rewards provided by contingency management increase the ability of deliberative systems to attend, value, and select the alternative (non-drug) reward. (This may be why the prize-based CM systems are more effective with lower value rewards than comparably more expensive monetary-based voucher systems.)
5.1. Pre-Clinical Experimental Support for This Hypothesis
Non-human animal self-administration studies have also found that drugs are economic objects and show a non-zero elasticity. As with human studies, increasing the cost (measured in terms of number of lever presses required to receive drug) decreases the number of self-administered drug-taking events (28, 99–101). Similarly, providing an alternative reinforcer reduces the amount of drug self-administration in both rats and monkeys (23–25, 27, 48, 53, 100, 102–106). These studies fall into two categories, which require dramatically different levels of alternative reward to decrease drug use.
Classically, the simplest measure of the cost-dependence of drug self-administration in non-human animals is the breakpoint analysis (52, 99). These studies find that much larger costs are required before an animal will cease drug self-administration than before an animal will cease taking non-drug rewards (51, 100). This suggests that it would require very large non-drug rewards to counteract drug self-administration. The first set of studies (24, 25, 27, 104) confirmed this hypothesis, in that they used single-response conditions and found that reductions in drug self-administration were only observed after very large alternative rewards. For example, Woolverton et al. (27) found that the opportunity cost of the drug option needed to be increased 100-fold (for low-drug concentrations) to 1000-fold (for average and high-drug concentrations) in order to significantly reduce self-administration. In these studies, animals could switch between conditions that either provided cocaine on pressing the primary lever or alternative reward on pressing the same primary lever. In other words, the animal could switch between situations that enabled non-deliberative processes. Other studies using similar techniques have found similar proportions (24, 25, 107, 108).
Interestingly, Ahmed [(48), see Ref. (100, 105, 106)] found much smaller alternatives could reduce drug self-administration. In these studies, the animals had two options directly available to them on opposite sides of the chamber – one lever provided cocaine, while the other provided saccharin. Preference was measured by whether the animals selected the saccharin lever or the cocaine lever. These studies also examined single-option breakpoints, in which only one lever was provided and cost was measured as the number of lever presses required before the animal gave up. These studies found that although the breakpoints for cocaine were much higher than the breakpoints for saccharin, animals preferred saccharin when provided with a revealed-preference two-lever choice paradigm. Similarly, LeSage (53) showed that providing a small amount of sucrose for not self-administering nicotine was sufficient to reduce the number of nicotine responses.
These studies support the proposed dichotomy between willingness-to-pay valuations (measured by single-lever breakpoint studies and situation-change studies, theoretically dependent on non-deliberative processes) and revealed-preference valuations (measured as forced choices between two explicit levers). The revealed-preference studies required much smaller rewards to decrease drug self-administration than the willingness-to-pay studies. The difference in size of alternate reward required to change behavior under the two paradigms suggests that the difference lies in fundamental processes underlying decision-making across multiple species (including at least rats, monkeys, and humans).
6. Components of Contingency Management That Affect Deliberation
The information processing that underlies deliberative decision-making processes is now beginning to be elucidated (87, 98, 109), particularly, in contrast to other decision-making systems (39, 43, 110). Deliberation requires recognition of a situation, a serial consideration of the potential actions available, and evaluation and comparison of those potential options (42, 87).
The main advantage of deliberation is that because these expected consequences are represented during the decision process, they can be evaluated during that process, in the context of the agent’s current goals (86). This means that the individual options must be found (85, 98, 111, 112) and then the valuation constructed (40, 47, 73, 113). Both the search process and the construction of value will be modulated by processes that computationally affect neural information processing (98, 114). Examples of these include working memory abilities (57, 115), whether the consequence is phrased as a win or a loss (40, 47, 116, 117), attention (113, 118), emotional state (119), surrounding options (120), and even the presence of unrelated numbers, such as in anchoring [where unrelated anchors such as one’s social security number can be used to change one’s expected cost and thus one’s willingness to pay for a reward (40, 47, 117, 118)].
The deliberative process is slow and computationally intensive, likely because of the cumbersome memory-retrieval and imagination-construction system needed to calculate the possible outcomes in order to evaluate them (83, 87, 98, 112). The evaluation achieved through deliberation depends on a number of stimulus factors, including the expected delay to the reward (121), and the concreteness of the reward (97). Deliberation also depends on a number of internal factors, such as one’s perceived needs and desires (86, 119), as well as one’s cognitive and executive-function abilities (98), such as episodic future thinking (95, 96), working memory (115, 122), and ability to hold attention (123, 124).
Valuation derived from deliberation depends on a direct imagination of expected outcomes and a comparison between choices (87, 98, 109). As the preclinical studies reviewed above show (48, 53), when an explicit choice between the drug and non-drug reward options is available, the drug option is less likely to be chosen; therefore, factors that increase the likelihood of engaging deliberative processes or that increase the deliberative valuation of a non-drug option should increase the efficacy of contingency management.
6.1. Delay to Reward
Rewards that are only available in the future are less valuable than rewards provided immediately (125–127) – something could happen between now and the time one expects to receive the reward (thus diminishing the usefulness of that reward) and immediate rewards can be invested (thus increasing the usefulness of immediate rewards). The diminishing value of future rewards relative to immediate rewards is quantifiably measurable through questionnaires in which subjects make decisions between immediate and delayed amounts of money, drug, or both (121).
Drug users reliably show faster discounting rates than non-addicts (128–132). Recovered addicts, however, show normal discounting rates (128). Although this early study was unable to determine whether this was a selection process in which the addicts with more normal discounting rates responded better to treatment, a more recent study has determined that successful treatment has the effect of normalizing over-fast discounting rates (133).
Many theoreticians have suggested that these preferences for more immediately available rewards can drive drug use because drugs provide very strong immediate rewards (euphoria, relief from dysphoria) while abstinence provides only long-term rewards (health, family, financial) (134, 135). Contingency management may have the effect of bringing the long-term rewards closer by providing more proximal rewards for abstinence (money, vouchers, draws from the prize-bowl).
Given the actual discounting rates reported in realistic subjects (128, 130, 136), $2.50 for the first drug-negative sample would be discounted quickly and seems unlikely to be able to deflect the user away from drugs, especially in the beginning of contingency management treatment. The delay-discounting rates that would be necessary to make these small rewards provided at the end of a week strong enough to affect decisions made days earlier in the week are unreasonably slow (137, 138), particularly for addicts, who have faster discounting rates than non-addicts [for review, see Ref. (28)]. Studies have shown that individuals discount smaller values more quickly than larger values [discounting curves are steeper, Ref. (139)], which would further reduce the discounted effectiveness of the small rewards provided early in treatment.
Furthermore, both human and non-human subjects tend to show hyperbolic discounting functions (121, 140, 141). Any non-exponential (including hyperbolic) discounting function will show preference reversals in which one choice is preferred when both choices are far in the future, but the other becomes preferred as the subject approaches the time of that second choice (142). Thus, even if a user decided at the beginning of the week to prefer the contingent reward ($2.50) to taking drugs, when faced with the immediate choice, the user would seem likely to choose the drug-use option.
During treatment in prize-based contingency management, upon submission of a drug-negative sample, individuals immediately earn a chance to win a tangible prize. In addition, individuals have a chance (albeit low in probability) to win a high-value prize for every draw they earn. This means that even though the average overall value of reinforcers earned by subjects tends to be lower in prize-based contingency management compared to voucher-based contingency management, the availability of a more immediate reward and the chance to win a high-value prize may cause individuals to discount less. These differences in discounting rates between the two versions of contingency management may help to explain similar treatment efficacy even with differing value of total potential reward.
6.2. Concreteness
The long-term rewards of abstinence tend to be more abstract than the short-term reinforcement provided by drug use (135). Several authors have suggested that the major difference between immediate rewards and delayed rewards is the concreteness of immediate rewards and the abstractness of delayed rewards (98, 143, 144).
Trope and Liberman (143) suggest that high-temporal distance creates difficult-to-conceptualize (high-level, more abstract) construals that are more difficult to reason about, while low-temporal distance creates easier-to-conceptualize (low-level, more concrete) construals. They hypothesize that more concrete options are considered to be more valuable than more abstract options. For an addict, abstinence is a high-level construal placed in the hard-to-imagine far future and is more abstract and less valuable than a concrete reinforcer, such as the option to use drugs in the present or near future, which is a low-level construal.
Current decision-making theories suggest that evaluating future outcomes depends on constructing episodically-imagined futures (87, 109, 113, 145). Kurth-Nelson and Redish (98) suggested that discounting rates may depend on how difficult it is for this construction process to find those potential future possibilities. Supporting this hypothesis is evidence that fronto-parietal areas are more active when people select the delayed option (56, 57), that subjects with better working memory and higher IQs tend to discount more slowly (115), and that training working memory can slow discounting rates (122, 133). Rewards placed in concrete episodic futures (35€ on vacation in Paris next month) are discounted more slowly than abstract future rewards (35€ next month) (146). Kurth-Nelson and Redish (98) suggest that the decreased discounting of concrete options is due to concrete futures being easier to find and construct in the deliberative search process.
Taken together, these theories imply that more concrete rewards have higher subjective value compared to abstract rewards. What does this mean for addiction? Typically, an addict has a choice between using a drug and not using a drug. The option of using the drug has immediate and concrete rewarding effects. Drug’s rewarding effects include subjective pleasurable effects and relief from withdrawal, and both of these effects are expected and concrete. The option of not using has immediate negative effects (147), but the primary distal rewarding effects are very abstract (135).
Contingency management changes this scenario by providing the addict with a concrete reward (money, a voucher, a specific prize) contingent upon abstinence, which is more proximal than rewards for abstinence alone. This allows the addict to achieve the goal of reducing drug consumption and increasing abstinence by focusing, not on the abstract abstinence, but rather on the concrete alternative.
This theory suggests that one effect of contingency management is to make both options immediate and concrete. The combination of the discounting/proximity and the concreteness theories suggest that contingency management creates a situation where the alternate reward (i.e., abstinence over drug use) is both more concrete and closer in temporal distance; thus, making it more equal to the drug-use option.
The importance of concreteness is highlighted by comparing voucher- and prize-based treatments. Although subjects were encouraged to imagine concrete items that the voucher could be used for (5), in prize-based studies, the prizes are physically present in a show-cabinet right there with the prize-bowl (9). Vouchers were also useable for a variety of rewards, while winning a given prize meant that that was the concrete prize you got.
Both voucher- and prize-based have been found to be similarly effective, even though the value of possible earned rewards is much lower in the prize-based studies (5, 9, 11, 12, 148). In both versions, high-value rewards have been found to be more successful than low-value rewards; however, the size of rewards offered in these conditions differs considerably. Even though the total value of possible rewards received in the high-value prize-based method was lower than the low-value voucher-based method, the high-value prize-based method was still effective for significantly reducing drug consumption, while the low-value voucher-based method was not. This not only exemplifies the importance of value but also how the concreteness of the reward affects perceived value. The presence of more concrete alternative rewards (specific prizes) appears to have more of an effect than less concrete alternative rewards (voucher exchanged for money, in turn, used for unspecified merchandise).
7. Conclusion and Further Discussion
In summary, we propose that contingency management’s success occurs because it provides an alternate reinforcer that forces the subject into a deliberative mode, which allows different valuation processes than non-deliberative modes. It also provides both a decreased time-to-reward and increased concreteness for the alternate reward, which should increase the valuation of the alternate reward relative to the valuation of the drug and move the agent from a willingness-to-pay valuation mode to a choice between/revealed-preference valuation mode.
7.1. Relationship to Classical Dual-Process Theories
Many theoreticians have suggested that addiction arises from a mismatch between the balance of two systems (typically called a “hot” or impulsive system and a “cold,” rational system) (64, 149, 150). While it is possible to place our hypotheses for contingency management within that two-system framework, we believe that the evidence suggests that addiction is more complicated than the simple out-of-balance theory proposes. Instead, we work from the theory that continued drug use can arise from computation errors in a number of places within the decision-system, of which a mismatch in balance between systems is only one potential failure mode (39, 43).
It is important to differentiate the vulnerabilities theory of addiction that arises from the multiple action-selection-system theory from the out-of-balance theory of addiction that arises from the dual-process theory. (See Table 2 for a list of these decision-concepts used in this paper.) Our proposal that contingency management drives subjects toward deliberative processes could follow from either of these two addiction/decision-making theories, but the implications are different, depending on which theory pertains.
The out-of-balance hypothesis of addiction is that addicts have a problem with the balance between the two systems in the dual-process theory (54, 55, 66, 67, 149). These systems can be driven out of balance either from hyperactivity in the impulsive system or hypoactivity in the rational system (55, 56, 151, 152). In either case, improving the strength of the rational system [for example, by providing working memory training (122) or by increasing activity in the prefrontal cortex (153)] should decrease drug use because it should shift the balance toward the more rational system. Our proposal that contingency management drives decision-making toward deliberation implies that if the dual-process and out-of-balance theories are correct, then what contingency management is doing is shifting the balance between these two systems. Evidence supporting this concept was recently published by Wesley et al. (57), who found that in an explicit cocaine-money choice, choosing money later over cocaine now produced additional activity in the dorsolateral prefrontal cortex.
The vulnerabilities hypothesis of addiction is that there are many potential “failure modes” within these systems, any of which can lead to addictive behaviors (39, 43, 77, 154). The concept that there are many vulnerabilities implies that addiction can arise from multiple causes. Our proposal that contingency management drives decision-making toward deliberation implies that if the multiple-action-selection systems and vulnerabilities theories are correct, then what contingency management is doing is twofold: (1) it is shifting the decision-making system into deliberation because it is providing two choices, and (2) it is improving the deliberation system algorithm, by making the goals more concrete and more immediate.
There are similarities and differences between these theories. Both theories include separate action-selection systems, only one of which includes an explicit planning component.
• The concrete nature of the alternative reward in contingency management is going to access that planning component, driving behavior toward it.
• Under neither hypothesis is the alternative reward fast enough to access the non-planning systems.
• In both theories, the planning-capable system depends on cognitive resources and prefrontal cortex.
However, the vulnerabilities theory further proposes that there are failure modes within the deliberative system as well, and thus suggests that only a subset of patients will be helped by contingency management, and that different aspects of contingency management will help different patients.
• For patients who have vulnerabilities in the Pavlovian or procedural systems who may express a desire to quit in the absence of drug-related cues, but find themselves unable to when faced with drug-related cues, contingency management can provide a second option to attend to, even when faced with drug-related cues, which can enable the deliberative system to retain control. This likely relates to the difference in valuation between single-option choices (go/no-go, willingness to pay) and dual-option choices (select between).
• For patients who have vulnerabilities in the evaluation step of deliberative systems, the concrete nature of the alternative reward in contingency management can make that reward easier to locate in the search-through-the-future process. This likely relates to the dependence of the search process on episodic future thinking.
• For patients for whom the drugs are simply an alternative reward option or for patients who have limited access to alternative rewards (155), then the opportunity cost provided by contingency management could be enough to make them reject the drug option.
• Because the vulnerabilities theory proposes that some patients will have vulnerabilities within the deliberative decision-making system [such as incorrect hypotheses about consequences of their actions (156, 157)], these patients will not be helped by contingency management, at least until they address those deliberative deficiencies.
7.2. Predictions and Implications
7.2.1. Identify Patients Capable of Deliberating
The idea that contingency management primarily accesses deliberative systems implies that it will be most successful in patients with viable deliberative systems. This suggests that identifying patients with intact deliberative systems would help identify patients most likely to be helped by contingency management programs. There are a number of cognitive tasks known to access deliberative systems (94, 146, 158–161). Whether these tasks are changed in addicts, however, remains unknown. The vulnerabilities theory predicts that some addicts will continue to show deliberative abilities in these tasks, and that those addicts will be best served by contingency management.
This hypothesis further suggests that patients with deficient deliberative systems would be helped by first training those systems. Working memory training, for example, decreases discounting rates as much as drug treatment (133).
7.2.2. Prediction: Contingency Management will Depend on Prefrontal Integrity
The two hypotheses that contingency management depends on deliberative processes and that deliberative processes depend on prefrontal integrity predict that contingency management will be most successful in patients with strongly active prefrontal systems. Evidence that prefrontal cortical interactions with hippocampus and other neural systems are a necessary component for deliberative decision-making processes is well-established (71, 94, 96, 145, 162, 163). For example, functional connectivity between prefrontal cortex and nucleus accumbens predicts success in drug-dependence treatment and an avoidance of relapse (152). In rats, optogenetic stimulation of prelimbic (prefrontal) cortices decreases compulsive drug seeking, while optogentic inhibition of prelimbic (prefrontal) cortices increased it (153). Similarly, in humans, repetitive transcranial magnetic stimulation (rTMS) over the dorsolateral prefrontal cortex reduced reported craving in nicotine addicts (164).
It also suggests that patients with improved cognitive abilities (115) and with prefrontal cortices more likely to play active roles in decision-making (56, 57, 152) will be more capable of using contingency management. These hypotheses imply that further improvements in cognitive resources [such as with working memory training (122, 133)] or increasing prefrontal activity (153) will make patients be more capable of using contingency management.
7.2.3. Combine Contingency Management with Working Memory Training and Cognitive Reassessment Therapy
Contingency management is often provided with synergistic treatment of pharmacological and sociological treatments (counseling, 12-step group work, methadone or nicotine-replacement treatment, etc.) (1). While these additional treatments provide potential rectification of decision-making vulnerabilities and failure modes, we suggest that they do not directly address the reasons for the success of contingency management. Under the hypothesis that contingency management depends on deliberative processes, improvements in those deliberative processes should provide additional improvements in the success of contingency management.
Deliberative decision-making entails the creation and imagination of hypothetical episodic futures and evaluation of those futures (43, 85, 109, 111, 145). As such, it requires a search process and memory to compare those evaluations (87, 98, 163). Changes in the recognition of the underlying paths through those futures affect the decisions made (138, 140, 165). For example, the famous dictum that “there is no such thing as one drink for an alcoholic” implies that decisions are not between drinking one drink and not, but between drinking many drinks and not. This process leads to bundling, in which future decisions are bundled together, which changes the underlying valuation of those future decisions (140, 165).
Changes in the ability to create, imagine, test, and remember those futures will also likely increase the ability to engage that deliberative system. It is possible to improve executive function and working memory through training (122, 166). These procedures decrease impulsivity as measured by discounting experiments. Given the data that cognitive load decreases engagement of the deliberative system (67, 124, 160), merely recognizing that patients are particularly vulnerable under stress and situations of increased cognitive load (165, 167), could suggest proactive procedures (such as increased rewards or increased reminders) during times of stress and cognitive load.
7.2.4. Increasing Value of the Alternate Option
From the very first introductions of contingency management, it has been clear that providing an increased value of the alternate rewards increases the success rate (1, 11, 148). This is a straightforward prediction of the alternate reinforcement theory. However, as expected from the discussion of the pre-clinical data (above), dramatic changes would require very large alternate rewards. For example, increasing the payout from $0.50 on the first negative drug urine sample to $7.00 produces a significant effect (168). Given the political difficulty of paying for drug treatment programs, finding ways to increase the success of contingency management without dramatically increasing costs would be particularly useful. Prize-based contingency management is one example of reducing costs without decreasing efficacy (1, 11).
7.2.5. Concrete Options are Discounted Less than Abstract Options – Provide Reminders of the Concrete Alternate Reward
If one could increase the proximity of the rewards at the moment of decision, one could further increase the value of the alternative option. Thus, one potential improvement would be to provide a concrete reminder of the alternate reward (such as what the current voucher value is) on an easily accessible place (such as a smartphone app) that could be accessed at the actual moment of decision.
Although concrete options are more valuable than abstract options, symbolic reminders of concrete options might also increase the value of alternate options. For example, simply stating a delayed reward will be delivered during an episodic event decreases discounting and increases value relative to equivalent, but less concrete rewards (146). Similarly, pictures of food rewards are more valuable than text descriptions of those rewards (169). Thus, visual symbols can improve both concreteness and deliberation. This suggests that providing the picture of the specific concrete option being worked toward is likely to further improve the reminder. Similarly, providing direct information about the values of the alternative options (such as days clean, days remaining to reward, points that would be lost due to relapsing) would make it easier for the patient to evaluate the alternative outcome, which should make it easier for the patient to attend to (and select) the alternative outcome. This could also be accomplished through a smartphone app that shows the picture of the reward being worked toward and information about the voucher points needed to achieve that goal.
7.2.6. Preventing Relapse after Contingency Management Treatment
As with any treatment, many patients relapse after treatment. The vulnerabilities theory suggests that addiction is caused by a multitude of potential failure modes (39, 43). Although contingency management is a support mechanism that can aid in a person’s recovery, other failure modes may still remain even after completion of the contingency management series. However, contingency management can be combined with other treatments (1, 5, 9). Studies have shown that the cognitive and discounting impairments that arise during drug and alcohol use improve with continued abstinence (133, 170–174). Thus, contingency management can create a span of time for an individual to repair these failure modes, while also learning important skills to increase the chance to remain abstinent in the future.
One potential solution would be to teach users to create their own contingency management process, providing their own deliberative alternatives. Changes in expectations and representations of the outcomes of potential options can change decision-making choices, even without changes in the underlying action-selection processes (135, 138, 140).
Author Contributions
The manuscript was co-written by both authors.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We would like to thank Nancy Petry, Josh Gordon, and Zeb Kurth-Nelson for discussion, and Nancy Petry for comments on an early draft of the manuscript. Funding: This work was supported by NIH grants R01-DA024080, R01-DA030672, and T32-DA007234.
Footnote
- ^In the Higgins et al. (5) study, subjects got a voucher worth $2.50 for the first clean sample. Taking this voucher as covering staying abstinent for more than 1 day would only decrease the predicted impact of the voucher. Since our analysis will show that voucher size is inadequate to drive changes in the demand curve, any increased required abstinence will not change our conclusions.
References
2. Stitzer ML, Bigelow GE, Leibson IA, Hawthorne JW. Contingent reinforcement for benzodiazepine-free urines: evaluation of a drug abuse treatment intervention. J Appl Behav Anal (1982) 15(4):493–503. doi:10.1901/jaba.1982.15-493
3. Higgins ST, Delaney DD, Budney AJ, Bickel WK, Hughes JR, Foerg F, et al. A behavioral approach to achieving initial cocaine abstinence. Am J Psychiatry (1991) 148:1218–24. doi:10.1176/ajp.148.9.1218
4. Higgins ST, Budney AJ, Bickel WK, Hughes JR, Foerg F, Badger G. Achieving cocaine abstinence with a behavioral approach. Am J Psychiatry (1993) 150:763–9. doi:10.1176/ajp.150.5.763
5. Higgins ST, Budney J, Bickel WK, Foerg FE, Donham R, Badger GJ. Incentives improve outcome in outpatient behavioral treatment of cocaine dependence. Arch Gen Psychiatry (1994) 51:568–76. doi:10.1001/archpsyc.1994.03950070060011
6. Silverman K, Wong CJ, Higgins ST, Brooner RK, Montoya ID, Contoreggi C, et al. Increasing opiate abstinence through voucher-based reinforcement therapy. Drug Alcohol Depend (1996) 41(2):157–65. doi:10.1016/0376-8716(96)01246-X
7. Bickel WK, Amass L, Higgins ST, Badger GJ, Esch RA. Effects of adding behavioral treatment to opioid detoxification with buprenorphine. J Consult Clin Psychol (1997) 65(5):803–10. doi:10.1037/0022-006X.65.5.803
8. Iguchi MY, Belding MA, Morral AR, Lamb RJ, Husband SD. Reinforcing operants other than abstinence in drug abuse treatment: an effective alternative for reducing drug use. J Consult Clin Psychol (1997) 65:421–8. doi:10.1037/0022-006X.65.3.421
9. Petry NM, Martin B, Cooney JL, Kranzler HR. Give them prizes and they will come: contingency management for treatment of alcohol dependence. J Consult Clin Psychol (2000) 68(2):250–7. doi:10.1037/0022-006X.68.2.250
10. Petry NM, Martin B. Low-cost contingency management for treating cocaine- and opioid-abusing methadone patients. J Consult Clin Psychol (2002) 70(2):398–405. doi:10.1037/0022-006X.70.2.398
11. Petry NM, Tedford J, Austin M, Nich C, Carroll K, Rounsaville B. Prize reinforcement contingency management for treating cocaine users: how low can we go, and with whom? Addiction (2004) 99(2):349–60. doi:10.1111/j.1360-0443.2003.00642.x
12. Petry NM, Alessi SM, Marx J, Austin M, Tardiff M. Vouchers versus prizes: contingency management treatment of substance abusers in community settings. J Consult Clin Psychol (2005) 73(6):1005–14. doi:10.1037/0022-006X.73.6.1005
15. Domjan M. The Principles of Learning and Behavior. 4th ed. Pacific Grove, CA: Brooks/Cole (1998).
16. Bouton ME. Learning and Behavior: A Contemporary Synthesis. Sunderland, MA: Sinauer Associates (2007).
17. Becker GS, Murphy KM. A theory of rational addiction. J Polit Econ (1988) 96(4):675–700. doi:10.1086/261558
18. Bickel WK, DeGrandpre RJ, Higgins ST, Hughes JR. Behavioral economics of drug self-administration. I. Functional equivalence of response requirement and drug dose. Life Sci (1990) 47:1501–10. doi:10.1016/0024-3205(90)90178-T
19. Hursh SR, Galuska CM, Winger G, Woods JH. The economics of drug abuse: a quantitative assessment of drug demand. Mol Interv (2005) 5:20–8. doi:10.1124/mi.5.1.6
20. Hursh SR, Silberberg A. Economic demand and essential value. Psychol Rev (2008) 115(1):186–98. doi:10.1037/0033-295X.115.1.186
21. Bruner N, Johnson M. Demand curves for hypothetical cocaine in cocaine-dependent individuals. Psychopharmacology (2014) 231(5):889–97. doi:10.1007/s00213-013-3312-5
22. Redish AD. Addiction as a computational process gone awry. Science (2004) 306(5703):1944–7. doi:10.1126/science.1102384
23. Carroll ME, Lac ST, Nygaard SL. A concurrently available nondrug reinforcer prevents the acquisition or decreases the maintenance of cocaine-reinforced behavior. Psychopharmacology (1989) 97(1):23–9. doi:10.1007/BF00443407
24. Nader MA, Woolverton WL. Effects of increasing the magnitude of an alternative reinforcer on drug choice in a discrete-trials choice procedure. Psychopharmacology (1991) 105:169–74. doi:10.1007/BF02244304
25. Nader MA, Woolverton WL. Choice between cocaine and food by rhesus monkeys: effects of conditions of food availability. Behav Pharmacol (1992) 3:635–8. doi:10.1097/00008877-199212000-00010
26. Woolverton WL. Cocaine self-administration: pharmacology and behavior. NIDA Res Monogr (1992) 124:189–202.
27. Woolverton WL, English JA, Weed MR. Choice between cocaine and food in a discrete-trials procedure in monkeys: a unit price analysis. Psychopharmacology (1997) 133:269–74. doi:10.1007/s002130050401
28. Bickel WK, Marsch LA. Toward a behavioral economic understanding of drug dependence: delay discounting processes. Addiction (2001) 96:73–86. doi:10.1046/j.1360-0443.2001.961736.x
29. Saffer H, Chaloupka F. The demand for illicit drugs. Econ Inq (1999) 37(3):401–11. doi:10.1111/j.1465-7295.1999.tb01439.x
30. Grossman M, Chaloupka FJ. The demand for cocaine by young adults: a rational addiction approach. J Health Econ (1998) 17:427–74. doi:10.1016/S0167-6296(97)00046-5
31. Liu J-L, Liu J-T, Hammit JK, Chou S-Y. The price elasticity of opium in Taiwan, 1914-1942. J Health Econ (1999) 18:795–810. doi:10.1016/S0167-6296(99)00023-5
32. Manski CF, Pepper JV, Petrie CV. Informing America’s Policy on Illegal Drugs: What We Don’t Know Keeps Hurting Us. Washington, DC: Academy Press (2001).
33. Allison J. Demand economics and experimental psychology. Behav Sci (1979) 24:403–17. doi:10.1002/bs.3830240606
35. Hursh SR. Behavioral economics of drug self-administration and drug abuse policy. J Exp Anal Behav (1991) 56(2):377–93. doi:10.1901/jeab.1991.56-377
36. Lussier JP, Heil SH, Mongeon JA, Badger GJ, Higgins ST. A meta-analysis of voucher-based reinforcement therapy for substance use disorders. Addiction (2006) 101:192–203. doi:10.1111/j.1360-0443.2006.01311.x
37. Daw ND, Niv Y, Dayan P. Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nat Neurosci (2005) 8:1704–11. doi:10.1038/nn1560
38. Rangel A, Camerer C, Montague PR. A framework for studying the neurobiology of value-based decision making. Nat Rev Neurosci (2008) 9:545–56. doi:10.1038/nrn2357
39. Redish AD, Jensen S, Johnson A. A unified framework for addiction: vulnerabilities in the decision process. Behav Brain Sci (2008) 31:415–87; discussion 437–61. doi:10.1017/S0140525X0800472X
42. van der Meer MAA, Kurth-Nelson Z, Redish AD. Information processing in decision-making systems. Neuroscientist (2012) 18(4):342–59. doi:10.1177/1073858411435128
43. Redish AD. The Mind Within the Brain: How We Make Decisions and How Those Decisions Go Wrong. New York: Oxford University Press (2013).
44. Meehl PE. On the circularity of the law of effect. Psychol Bull (1950) 47:52–75. doi:10.1037/h0058557
45. Griffiths RR, Brady JV, Bradford LD. Predicting the abuse liability of drugs with animal drug self-administration procedures: psychomotor stimulants and hallucinogens. In: Thompson T, Dews PB, editors. Advances in Behavioral Pharmacology. New York, NY: Academic Press (1979). 2 p.
46. Bickel WK, Madden GJ. A comparison of measures of relative reinforcing efficacy and behavioral economics: cigarettes and money in smokers. Behav Pharmacol (1999) 10:627–37. doi:10.1097/00008877-199911000-00009
47. Lichtenstein S, Slovic P, editors. The Construction of Preference. Cambridge, UK: Cambridge University Press (2006).
48. Ahmed SH. Validation crisis in animal models of drug addiction: beyond non-disordered drug use toward drug addiction. Neurosci Biobehav Rev (2010) 35(2):172–84. doi:10.1016/j.neubiorev.2010.04.005
49. Perry AN, Wetenbroek C, Becker JB. The development of a preference for cocaine over food identifies individual rats with addiction-like behaviors. PLoS One (2013) 8(11):e79465. doi:10.1371/journal.pone.0079465
50. Deneau G, Yanagita T, Seevers MH. Self-administration of psychoactive substances by the monkey. Psychopaharmacologia (1969) 16:30–48. doi:10.1007/BF00405254
51. Koob GF, Le Moal M. Neurobiology of Addiction. Philadelphia, PA: Elsevier Academic Press (2006).
52. Ahmed SH, Lenoir M, Guillem K. Neurobiology of addiction versus drug use driven by lack of choice. Curr Opin Neurobiol (2013) 23(4):581–7. doi:10.1016/j.conb.2013.01.028
53. LeSage MG. Toward a nonhuman model of contingency management: effects of reinforcing abstinence from nicotine self-administration in rats with an alternative nondrug reinforcer. Psychopharmacology (2009) 203(1):13–22. doi:10.1007/s00213-008-1362-x
54. Bechara A. Decision making, impulse control and loss of willpower to resist drugs: a neurocognitive perspective. Nat Neurosci (2005) 8(11):1458–63. doi:10.1038/nn1584
55. Bickel WK, Yi R. Temporal discounting as a measure of executive function: insights from the competing neuro-behavioral decision system hypothesis of addiction. Adv Health Econ Health Serv Res (2008) 20:289–309. doi:10.1016/S0731-2199(08)20012-9
56. McClure SM, Laibson DI, Loewenstein G, Cohen JD. Separate neural systems value immediate and delayed monetary rewards. Science (2004) 306(5695):503–7. doi:10.1126/science.1100907
57. Wesley MJ, Lohrenz T, Koffarnus MN, McClure SM, De La Garza R II, Salas R, et al. Choosing money over drugs: the neural underpinnings of difficult choice in chronic cocaine users. J Addict (2014) 2014:189853. doi:10.1155/2014/189853
58. Smith GT, Guller L, Zapolski TCB. A comparison of two models of urgency: urgency predicts both rash action and depression in youth. Clin Psychol Sci (2013) 1(3):266–75. doi:10.1177/2167702612470647
59. Evans JSBT. Dual-processing accounts of reasoning, judgement and social cognition. Annu Rev Psychol (2008) 59:255–78. doi:10.1146/annurev.psych.59.103006.093629
60. Carver CS, Johnson SL, Joormann J. Serotonergic function, two-mode models of self-regulation, and vulnerability to depression: what depression has in common with impulsive aggression. Psychol Bull (2008) 134(6):912–43. doi:10.1037/a0013740
61. Carver CS, Johnson SL, Joormann J. Major depressive disorder and impulsive reactivity to emotion: toward a dual-process view of depression. Br J Clin Psychol (2013) 52:285–99. doi:10.1111/bjc.12014
62. Johnson SL, Carver CS, Joormann J. Impulsive responses to emotion as a transdiagnostic vulnerability to internalizing and externalizing symptoms. J Affect Disord (2013) 150:872–3. doi:10.1016/j.jad.2013.05.004
64. Metcalfe J, Mischel W. A hot/cool-system analysis of delay of gratification: dynamics of willpower. Psychol Rev (1999) 106(1):3–19. doi:10.1037/0033-295X.106.1.3
66. Carver CS, Scheier M. On the Self-Regulation of Behavior. Cambridge, UK: Cambridge University Press (1998).
67. Baumeister RF, Heatherton TF, Tice DM. Losing Control: How and Why People Fail at Self-Regulation. Waltham, MA: Academic Press (1994).
68. Plato (4th century BCE). Phaedrus. Project Gutenberg (2008). Translated by Benjamin Jowett. Available from: http://www.gutenberg.org/ebooks/1636
69. Augustine of Hippo (Saint Augustine). The City of God. London: Penguin Classics (1972). 427 p. Translated by Henry Bettenson.
71. Damasio A. Descartes’ Error: Emotion, Reason, and the Human Brain. New York, NY: G. P. Putnam and Sons (1994).
73. Rangel A, Hare T. Neural computations associated with goal-directed choice. Curr Opin Neurobiol (2010) 20(2):262–70. doi:10.1016/j.conb.2010.03.001
75. Redish AD. Beyond the Cognitive Map: From Place Cells to Episodic Memory. Cambridge MA: MIT Press (1999).
76. van der Meer MAA, Redish AD. Expectancies in decision making, reinforcement learning, and ventral striatum. Front Neurosci (2010) 4:6. doi:10.3389/neuro.01.006.2010
77. Montague PR, Dolan RJ, Friston KJ, Dayan P. Computational psychiatry. Trends Cogn Sci (2012) 16(1):72–80. doi:10.1016/j.tics.2011.11.018
79. Wikman TS, Branicky MS, Newman WS. Reflex control for robot system preservation, reliability, and autonomy. Comput Electr Eng (1994) 20(5):391–407. doi:10.1016/0045-7906(94)90033-7
81. Seymour B, Dolan R. Emotion, decision making, and the amygdala. Neuron (2008) 58(5):662–71. doi:10.1016/j.neuron.2008.05.020
82. LeDoux J. Rethinking the emotional brain. Neuron (2012) 73:653–76. doi:10.1016/j.neuron.2012.02.004
84. Dezfouli A, Balleine B. Habits, action sequences and reinforcement learning. Eur J Neurosci (2012) 35(7):1036–51. doi:10.1111/j.1460-9568.2012.08050.x
85. Miller G, Galanter E, Pribram KH. Plans and the Structure of Behavior. Wilson, NC: Holt, Rhinehart (1960).
87. Johnson A, van der Meer MAA, Redish AD. Integrating hippocampus and striatum in decision-making. Curr Opin Neurobiol (2007) 17(6):692–7. doi:10.1016/j.conb.2008.01.003
89. Phelps EA, LeDoux JE. Contributions of the amygdala to emotion processing: from animal models to human behavior. Neuron (2005) 48(2):175–87. doi:10.1016/j.neuron.2005.09.025
90. McNally GP, Johansen JP, Blair HT. Placing prediction into the fear circuit. Trends Neurosci (2011) 34:283–92. doi:10.1016/j.tins.2011.03.005
91. Cisek P, Kalaska JF. Neural mechanisms for interacting with a world full of action choices. Annu Rev Neurosci (2010) 33:269–98. doi:10.1146/annurev.neuro.051508.135409
92. Hikosaka O, Miyashita K, Miyachi S, Sakai K, Lu X. Differential roles of the frontal cortex, basal ganglia, and cerebellum in visuomotor sequence learning. Neurobiol Learn Mem (1998) 70(1–2):137–49. doi:10.1006/nlme.1998.3844
93. Smith KS, Graybiel AM. A dual operator view of habitual behavior reflecting cortical and striatal dynamics. Neuron (2013) 79(2):361–74. doi:10.1016/j.neuron.2013.05.038
94. Voss JL, Gonsalves BD, Federmeier KD, Tranel D, Cohen NJ. Hippocampal brain-network coordination during volitional exploratory behavior enhances learning. Nat Neurosci (2011) 14:115–20. doi:10.1038/nn.2693
95. Schacter DL, Addis DR. On the nature of medial temporal lobe contributions to the constructive simulation of future events. In: Bar M, editor. Predictions in the Brain: Using Our Past to Generate a Future. Oxford, UK: Oxford University Press (2011). p. 58–69.
96. Hassabis D, Maguire EA. The construction system in the brain. In: Bar M, editor. Predictions in the Brain: Using Our Past to Generate a Future. Oxford, UK: Oxford University Press (2011). p. 70–82.
97. Peters J, Büchel C. The neural mechanisms of inter-temporal decision-making: understanding variability. Trends Cogn Sci (2011) 15(5):227–39. doi:10.1016/j.tics.2011.03.002
98. Kurth-Nelson Z, Redish AD. A theoretical account of cognitive effects in delay discounting. Eur J Neurosci (2012) 35:1052–64. doi:10.1111/j.1460-9568.2012.08058.x
99. Hodos W. Progressive ratio as a measure of reward strength. Science (1961) 134(3483):943–4. doi:10.1126/science.134.3483.943
100. Cantin L, Lenoir M, Augier E, Vanhille N, Dubreucq S, Serre F, et al. Cocaine is low on the value ladder of rats: possible evidence for resilience to addiction. PLoS One (2010) 5(7):e11592. doi:10.1371/journal.pone.0011592
101. Breton Y-A, Mullett A, Conover K, Shizgal P. Validation and extension of the reward-mountain model. Front Behav Neurosci (2013) 7:125. doi:10.3389/fnbeh.2013.00125
102. Carroll ME, Carmona GG, May SA. Modifying drug-reinforced behavior by altering the economic conditions of the drug and a nondrug reinforcer. J Exp Anal Behav (1991) 56(2):361–76. doi:10.1901/jeab.1991.56-361
103. Carroll ME. The economic context of drug and non-drug reinforcers affects acquisition and maintenance of drug-reinforced behavior and withdrawal effects. Drug Alcohol Depend (1993) 33(2):201–10. doi:10.1016/0376-8716(93)90061-T
104. Nader MA, Woolverton WL. Cocaine vs. food choice in rhesus monkeys: effects of increasing the response cost for cocaine. NIDA Res Monogr (1990) 105:621.
105. Lenoir M, Serre F, Cantin L, Ahmed SH. Intense sweetness surpasses cocaine reward. PLoS One (2007) 2(8):e698. doi:10.1371/journal.pone.0000698
106. Lenoir M, Ahmed SH. Supply of a nondrug substitute reduces escalated heroin consumption. Neuropsychopharmacology (2008) 33:2272–82. doi:10.1038/sj.npp.1301602
107. Anderson KG, Velkey AJ, Woolverton WL. The generalized matching law as a predictor of choice between cocaine and food in rhesus monkeys. Psychopharmacology (2002) 163(3–4):319–26. doi:10.1007/s00213-002-1012-7
108. Negus SS. Rapid assessment of choice between cocaine and food in rhesus monkeys: effects of environmental manipulations and treatment with d-amphetamine and flupenthixol. Neuropsychopharmacology (2003) 28(5):919–31.
109. Buckner RL, Carroll DC. Self-projection and the brain. Trends Cogn Sci (2007) 11(2):49–57. doi:10.1016/j.tics.2006.11.004
110. van der Meer MAA, Johnson A, Schmitzer-Torbert NC, Redish AD. Triple dissociation of information processing in dorsal striatum, ventral striatum, and hippocampus on a learned spatial decision task. Neuron (2010) 67(1):25–32. doi:10.1016/j.neuron.2010.06.023
111. Newell A, Shaw JC, Simon HA. Report on a general problem-solving program. Proceedings of the International Conference on Information Processing. Santa Monica, CA: Rand Corporation (1959). p. 256–64.
112. Redish AD, Johnson A. A computational model of craving and obsession. Ann N Y Acad Sci (2007) 1104(1):324–39. doi:10.1196/annals.1390.014
113. Hill C. The rationality of preference construction (and the irrationality of rational choice). Minn J Law Sci Technol (2008) 9(2):689–742.
114. Phelps E, Lempert KM, Sokol-Hessner P. Emotion and decision making: multiple modulatory circuits. Annu Rev Neurosci (2014) 37:263–87. doi:10.1146/annurev-neuro-071013-014119
115. Burks SV, Carpenter JP, Goette L, Rustichini A. Cognitive skills affect economic preferences, strategic behavior, and job attachment. Proc Natl Acad Sci U S A (2009) 106(19):7745–50. doi:10.1073/pnas.0812360106
116. Tversky A, Kahneman D. The framing of decisions and the psychology of choice. Science (1981) 211(4481):453–8. doi:10.1126/science.7455683
117. Kahneman D, Tversky A, editors. Choices, Values, and Frames. Cambridge, UK: Cambridge University Press (2000).
119. Andrade EB, Ariely D. The enduring impact of transient emotions on decision making. Organ Behav Hum Decis Process (2009) 109(1):1–8. doi:10.1186/s12875-014-0174-9
120. Simonson I, Tversky A. Choice in context: tradeoff contrast and extremeness aversion. J Market Res (1992) 29(3):281–95. doi:10.2307/3172740
121. Madden GJ, Johnson PS. A delay-discounting primer. In: Madden G, Bickel W, editors. Impulsivity: The Behavioral and Neurological Science of Discounting. Washington, DC: APA books (2010). p. 11–37.
122. Bickel WK, Yi R, Landes RD, Hill PF, Baxter C. Remember the future: working memory training decreases delay discounting among stimulant addicts. Biol Psychiatry (2011) 69(3):260–5. doi:10.1016/j.biopsych.2010.08.017
123. Franco-Watkins AM, Pashler H, Rickard TC. Does working memory load lead to greater impulsivity? commentary on Hinson, Jameson and Whitney (2003). J Exp Psychol Learn Mem Cogn (2006) 32(2):443–7. doi:10.1037/0278-7393.32.2.443
124. Vohs KD, Faber RJ. Spent resources: self-regulatory resource availability affects impulse buying. J Consum Res (2007) 33:537–47. doi:10.1086/510228
125. Madden G, Bickel W, editors. Impulsivity: The Behavioral and Neurological Science of Discounting. Washington, DC: APA books (2010).
126. Samuelson PA. A note on measurement of utility. Rev Econ Stud (1937) 4(2):155–61. doi:10.2307/2967612
128. Bickel WK, Odum AL, Madden GJ. Impulsivity and cigarette smoking: delay discounting in current, never, and ex-smokers. Behav Pharmacol (1999) 4:447–54.
129. Coffey SF, Gudleski GD, Saladin ME, Brady KT. Impulsivity and rapid discounting of delayed hypothetical rewards in cocaine-dependent individuals. Exp Clin Psychopharmacol (2003) 11:18–25. doi:10.1037/1064-1297.11.1.18
130. Kirby KN, Petry NM, Bickel WK. Heroin addicts have higher discount rates for delayed rewards than non-drug-using controls. J Exp Psychol Gen (1999) 128(1):78–87. doi:10.1037/0096-3445.128.1.78
131. Petry NM. Delay discounting of money and alcohol in actively using alcoholics, currently abstinent alcoholics, and controls. Psychopharmacology (2001) 154(3):243–50. doi:10.1007/s002130000638
132. Richards JB, Sabol KE, de Wit H. Effects of methamphetamine on the adjusting amount procedure, a model of impulsive behavior in rats. Psychopharmacology (1999) 146(4):432–9. doi:10.1007/PL00005488
133. Bickel W, Landes RD, Kurth-Nelson Z, Redish AD. A quantitative signature of self-control repair: rate-dependent effects of successful addiction treatment. Clin Psychol Sci (2014) 2(6):685–95. doi:10.1177/2167702614528162
134. Ainslie G, Herrnstein RJ. Preference reversal and delayed reinforcement. Anim Learn Behav (1981) 9(4):476–82. doi:10.3758/BF03209777
136. Bickel WK, Yi R, Kowal BP, Gatchalian KM. Cigarette smokers discount past and future rewards symmetrically and more than controls: is discounting a measure of impulsivity? Drug Alcohol Depend (2008) 96(3):256–62. doi:10.1016/j.drugalcdep.2008.03.009
137. Kurth-Nelson Z, Redish AD. Don’t let me do that! models of precommitment. Front Neurosci (2012) 6:138. doi:10.3389/fnins.2012.00138
138. Kurth-Nelson Z, Redish AD. Chapter 6: modeling decision-making systems in addiction. In: Gutkin B, Ahmed SH, editors. Computational Neuroscience of Drug Addiction. New York: Springer (2012). p. 163–88.
139. Green L, Myerson J, McFadden E. Rate of temporal discounting decreases with amount of reward. Mem Cognit (1997) 25(5):715–23. doi:10.3758/BF03211314
141. Mazur J. Choice, delay, probability and conditioned reinforcement. Anim Learn Behav (1997) 25(2):131–47. doi:10.3758/BF03199051
142. Frederick S, Loewenstein G, O’Donoghue T. Time discounting and time preference: a critical review. J Econ Lit (2002) 40(2):351–401. doi:10.1257/jel.40.2.351
143. Trope Y, Liberman N. Temporal construal. Psychol Rev (2003) 110(3):403–21. doi:10.1037/0033-295X.110.3.403
144. Loewenstein G, Scott S, Cohen J. Neuroeconomics. Annu Rev Psychol (2008) 59(1):647–72. doi:10.1146/annurev.psych.59.103006.093710
145. Schacter DL, Addis DR, Buckner RL. Remembering the past to imagine the future: the prospective brain. Nat Rev Neurosci (2007) 8:657–61. doi:10.1038/nrn2213
146. Peters J, Buchel C. Episodic future thinking reduces reward delay discounting through an enhancement of prefrontal-mediotemporal interactions. Neuron (2010) 66(1):138–48. doi:10.1016/j.neuron.2010.03.026
147. Koob GF, Le Moal M. Plasticity of reward neurocircuitry and the ‘dark side’ of drug addiction. Nat Neurosci (2005) 8(11):1442–4. doi:10.1038/nn1105-1442
148. Higgins ST, Heil SH, Dantona R, Donham T, Matthews M, Badger GJ. Effects of varying the monetary value of voucher-based incentives on abstinence achieved during and following treatment among cocaine-dependent outpatients. Addiction (2007) 102(2):271–81. doi:10.1111/j.1360-0443.2006.01664.x
149. Bernheim BD, Rangel A. Addiction and cue-triggered decision processes. Am Econ Rev (2004) 94(5):1558–90. doi:10.1257/0002828043052222
150. McClure SM, Bickel WK. A dual-systems perspective on addiction: contributions from neuroimaging and cognitive training. Ann N Y Acad Sci (2014) 1327:62–78. doi:10.1111/nyas.12561
151. Galvan A, Hare TA, Parra CE, Penn J, Voss H, Glover G, et al. Earlier development of the accumbens relative to orbitofrontal cortex might underlie risk-taking behavior in adolescents. J Neurosci (2006) 26(25):6885–92. doi:10.1523/JNEUROSCI.1062-06.2006
152. Camchong J, MacDonald AW III, Mueller BA, Nelson B, Specker S, Slaymaker V, et al. Changes in resting functional connectivity during abstinence in stimulant use disorder: a preliminary comparison of relapsers and abstainers. Drug Alcohol Depend (2014) 139:145–51. doi:10.1016/j.drugalcdep.2014.03.024
153. Chen BT, Yau H-J, Hatch C, Kusumoto-Yoshida I, Cho SL, Hopf FW, et al. Rescuing cocaine-induced prefrontal cortex hypoactivity prevents compulsive cocaine seeking. Nature (2013) 496(7445):359–62. doi:10.1038/nature12024
154. Tiffany ST. A cognitive model of drug urges and drug-use behavior: role of automatic and nonautomatic processes. Psychol Rev (1990) 97(2):147–68. doi:10.1037/0033-295X.97.2.147
155. Hart C. High Price: A Neuroscientist’s Journey of Self-Discovery That Challenges Everything You Know About Drugs and Society. New York: HarperCollins (2013).
156. Goldman MS, Boca FKD, Darkes J. Alcohol expectancy theory: the application of cognitive neuroscience. In: Leonard KE, Blane HT, editors. Psychological Theories of Drinking and Alcoholism. New York: Guilford (1999). p. 203–46.
157. Goldstein A. Addiction: From Biology to Drug Policy. Oxford, UK: Oxford University Press (2000).
158. Hassabis D, Kumaran D, Vann SD, Maguire EA. Patients with hippocampal amnesia cannot imagine new experiences. PNAS (2007) 104:1726–31. doi:10.1073/pnas.0610561104
159. Spiers HJ, Maguire EA. The neuroscience of remote spatial memory: a tale of two cities. Neuroscience (2007) 149(1):7–27. doi:10.1016/j.neuroscience.2007.06.056
160. Otto AR, Gershman SJ, Markman AB, Daw ND. The curse of planning: dissecting multiple reinforcement-learning systems by taxing the central executive. Psychol Sci (2013) 24(5):751–61. doi:10.1177/0956797612463080
161. Simon DA, Daw ND. Neural correlates of forward planning in a spatial decision task in humans. J Neurosci (2011) 31(14):5526–39. doi:10.1523/JNEUROSCI.4647-10.2011
162. Fellows LK, Farah MJ. Different underlying impairments in decision-making following ventromedial and dorsolateral frontal lobe damage in humans. Cereb Cortex (2005) 15(1):58–63. doi:10.1093/cercor/bhh108
163. Wang JX, Cohen NJ, Voss JL. Covert rapid action-memory simulation (crams): a hypothesis of hippocampal-prefrontal interactions for adaptive behavior. Neurobiol Learn Mem (2015) 117:22–33. doi:10.1016/j.nlm.2014.04.003
164. Li X, Hartwell KJ, Owens M, Lematty T, Borckardt JJ, Hanlon CA, et al. Repetitive transcranial magnetic stimulation of the dorsolateral prefrontal cortex reduces nicotine cue craving. Biol Psychiatry (2013) 73(8):714–20. doi:10.1016/j.biopsych.2013.01.003
166. Mischel W, Shoda Y, Rodriguez ML. Delay of gratification in children. Science (1989) 244(4907):933–8. doi:10.1126/science.2658056
167. Baumeister RF, Vohs KD, Tice DM. The strength model of self-control. Psychol Sci (2007) 16(6):351–5.
168. Packer RR, Howell DN, McPherson S, Roll JM. Investigating reinforcer magnitude and reinforcer delay: a contingency management analog study. Exp Clin Psychopharmacol (2012) 20(4):287–92. doi:10.1037/a0027802
169. Bushong B, King LM, Camerer CF, Rangel A. Pavlovian processes in consumer choice: the physical presence of a good increases willingness-to-pay. Am Econ Rev (2010) 100(4):1556–71. doi:10.1257/aer.100.4.1556
170. Bartels C, Kunert H-J, Stawicki S, Kröner-Herwig B, Ehrenreich H, Krampe H. Recovery of hippocampus-related functions in chronic alcoholics during monitored long-term abstinence. Alcohol Alcohol (2007) 42(2):92–102. doi:10.1093/alcalc/agl104
171. Bates ME, Voelbel GT, Buckman JF, Labouvie EW, Barry D. Short-term neuropsychological recovery in clients with substance use disorders. Alcohol Clin Exp Res (2005) 29(3):367–77. doi:10.1097/01.ALC.0000156131.88125.2A
172. Winward JL, Hanson KL, Bekman NM, Tapert SF, Brown SA. Adolescent heavy episodic drinking: neurocognitive functioning during early abstinence. J Int Neuropsychol Soc (2014) 20(02):218–29. doi:10.1017/S1355617713001410
173. Goldman MS. Experience-dependent neuropsychological recovery and the treatment of chronic alcoholism. Neuropsychol Rev (1990) 1(1):75–101. doi:10.1007/BF01108859
Keywords: decision-making, deliberation, addiction, contingency management, neuroeconomics, impulsivity, addiction treatment
Citation: Regier PS and Redish AD (2015) Contingency management and deliberative decision-making processes. Front. Psychiatry 6:76. doi: 10.3389/fpsyt.2015.00076
Received: 27 February 2015; Accepted: 07 May 2015;
Published: 01 June 2015
Edited by:
Mark Walton, University of Oxford, UKCopyright: © 2015 Regier and Redish. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: A. David Redish, Department of Neuroscience, University of Minnesota, 6-145 Jackson Hall, 321 Church Street SE, Minneapolis, MN 55455, USA, redish@umn.edu
†Present address: Paul S. Regier, Center for Studies of Addiction, Department of Psychiatry, University of Pennsylvania, Philadelphia, PA, USA