Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Featured ArticleResearch Articles, Behavioral/Cognitive

Worth the Work? Monkeys Discount Rewards by a Subjective Adapting Effort Cost

Mark Burrell, Alexandre Pastor-Bernier and Wolfram Schultz
Journal of Neuroscience 4 October 2023, 43 (40) 6796-6806; https://doi.org/10.1523/JNEUROSCI.0115-23.2023
Mark Burrell
Department of Physiology, Development and Neuroscience, University of Cambridge, Cambridge CB2 3DY, United Kingdom
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Alexandre Pastor-Bernier
Department of Physiology, Development and Neuroscience, University of Cambridge, Cambridge CB2 3DY, United Kingdom
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Wolfram Schultz
Department of Physiology, Development and Neuroscience, University of Cambridge, Cambridge CB2 3DY, United Kingdom
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Wolfram Schultz
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

All life must solve how to allocate limited energy resources to maximize benefits from scarce opportunities. Economic theory posits decision makers optimize choice by maximizing the subjective benefit (utility) of reward minus the subjective cost (disutility) of the required effort. While successful in many settings, this model does not fully account for how experience can alter reward-effort trade-offs. Here, we test how well the subtractive model of effort disutility explains the behavior of two male nonhuman primates (Macaca mulatta) in a binary choice task in which reward quantity and physical effort to obtain were varied. Applying random utility modeling to independently estimate reward utility and effort disutility, we show the subtractive effort model better explains out-of-sample choice behavior when compared with parabolic and exponential effort discounting. Furthermore, we demonstrate that effort disutility depends on previous experience of effort: in analogy to work from behavioral labor economics, we develop a model of reference-dependent effort disutility to explain the increased willingness to expend effort following previous experience of effortful options in a session. The result of this analysis suggests that monkeys discount reward by an effort cost that is measured relative to an expected effort learned from previous trials. When this subjective cost of effort, a function of context and experience, is accounted for, trial-by-trial choices can be explained by the subtractive cost model of effort. Therefore, in searching for net utility signals that may underpin effort-based decision-making in the brain, careful measurement of subjective effort costs is an essential first step.

SIGNIFICANCE STATEMENT All decision-makers need to consider how much effort they need to expend when evaluating potential options. Economic theories suggest that the optimal way to choose is by cost-benefit analysis of reward against effort. To be able to do this efficiently over many decision contexts, this needs to be done flexibly, with appropriate adaptation to context and experience. Therefore, in aiming to understand how this might be achieved in the brain, it is important to first carefully measure the subjective cost of effort. Here, we show monkeys make reward-effort cost-benefit decisions, subtracting the subjective cost of effort from the subjective value of rewards. Moreover, the subjective cost of effort is dependent on the monkeys’ experience of effort in previous trials.

  • adaptation
  • choice
  • preference
  • utility
  • value

Introduction

Reward follows work, but what reward is worth the work? This subjective decision is fundamental to understanding animal and human behaviors during effort-based decision-making. Notably effort requirements can vary wildly, whether it be walking to the kitchen, running for a train, or hiking up a mountain. The brain must be able to compare all the effort costs in choosing these actions against their potential rewards and therefore should possess flexible mechanisms for effort-based decision-making.

When consumer choice theory and labor supply theories are used to understand consumer and worker behavior, it is assumed decision-makers balance the possible utility (subjective benefit) against the disutility (subjective cost) of effort, with consumer choice theory predicting choices are made simply on net utility (reward minus cost; Fehr and Goette, 2007). Experimental results from optimal foraging theory, psychology and economics all support this conclusion (Bautista et al., 2001; Fehr and Goette, 2007; Zipf, 2016). However, in several contexts this model fails to explain why humans and animals select more effortful options (Friedman et al., 1968; Clement et al., 2000; Lydall et al., 2010). In animals, both context and order of experience serve as references that significantly affect reward-effort trade-offs (Knauss et al., 2020).

An attractive explanation of these phenomena is the reference dependency of effort-based decision-making developed in human behavioral labor economics (Fehr and Goette, 2007; Abeler et al., 2011; Crawford and Meng, 2011): choices are not made on the absolute effort costs but relative to a conditional expectation of effort. In this paradigm, there exists an internal expectation of effort (“effort reference-point”) informed by context and experience. Increases in this reference point increase willingness to pick effortful options as the relative magnitude of the reference-dependent effort cost decreases. Furthermore, responses may differ dramatically on either side of the reference point: the work-decisions of New York taxicab drivers were found to be 6-fold more sensitive to above-expected workloads as to below-expected workloads (Crawford and Meng, 2011). Similar findings of asymmetric work responses have been made in the economic literature (Akerlof and Yellen, 1990) and in animals (van Wolkenten et al., 2007).

However, little economic work has studied how effort reference-points change over the course of time in a single task, rather focusing on identifying what forms an appropriate reference point or modeling specific situations using reference-dependent preferences (De Giorgi and Post, 2011; Allen et al., 2017). Within prospect theory, a related theory of reference-dependent reward preferences, a few studies have documented changes in reference points in single-task contexts in response to wins/loss history, supporting a role for experience in determining reference points (Köszegi and Rabin, 2006; Arkes et al., 2008).

Here, we were interested in measuring effort disutility in rhesus macaques as a foundation to establishing the neurobiological basis of the reward-effort trade-offs. Previous tasks used to measure effort-costs have often additional confounds that make it difficult to directly assess the subjective value of effort. For example, in both commonly used designs in rodent studies, lever-pressing and the T-maze, high effort options are correlated with longer delays from trial onset (Salamone et al., 1994; Cousins et al., 1996). Given the extensive work demonstrating the subjective cost of time (temporal discounting) in both behavioral experiments and encoding by various brain regions (Critchfield and Kollins, 2001; Kobayashi and Schultz, 2008; Prévost et al., 2010) such designs are not suitable for isolating the subjective cost of effort from time. We apply a random utility model to independently model reward utility and effort disutility from safe options (McFadden, 1973; McFadden and Train, 2000). From the estimates of reward utility and effort disutility we demonstrated through out-of-sample testing that the monkeys’ choices closely follow those of a utility maximiser employing a subtractive model of effort costs. However, the willingness to exert effort was greater in both animals after effort expenditure, which can be explained by a model in which the reference point is learnt from recent experience.

Materials and Methods

The experiments were conducted on two purpose-bred male rhesus macaques (Macaca mulatta): Monkey U (weighing 14.5 kg) and Monkey W (weighing 11.5 kg). This research (PPL 70/8295) was ethically reviewed, approved, regulated and supervised by the following United Kingdom (UK) and University of Cambridge institutions and individuals: UK Home Office, implementing the Animals (Scientific Procedures) Act 1986, Amendment Regulations 2012, and represented by the local UK Home Office Inspector; the UK Animals in Science Committee; UK National Center for Replacement, Refinement and Reduction of Animal Experiments (NC3Rs); the University of Cambridge Animal Welfare and Ethical Review Body (AWERB), the University’s Biomedical Service (UBS) Certificate Holder, Welfare Officer, Governance and Strategy Committee, Named Veterinary Surgeon (NVS) and Named Animal Care and Welfare Officer (NACWO).

Equipment

To study the reward-effort trade-off it was necessary to develop hardware to implement different levels of effort. Approaches in the literature include lever pressing, grip force and torqued exoskeletons (Cousins et al., 1996; Pasquereau and Turner, 2013; Varazzani et al., 2015) These approaches were rejected for, respectively, the confound of delay to reward, incompatibility with binary choice designs and difficulty in training. Rather, we developed a custom effort joystick in which the resistance to movement could be computer controlled (Fig. 1A).

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Task design. A, Joystick with variable effort. B, Frictional resistance as function of voltage controlling magnet strength. C, Task design. Use of joystick for binary choice between options that differ in effort required. D, Fractal stimuli indicating effort required (in Newton), value bars indicating juice reward amount (in milliliters).

The joystick was constructed to allow motion in a single plane of movement, from left to right, by rotation on an axis. The different effort levels were made by altering the resistance to motion by varying the strength of an electromagnet that brought the joystick into contact with a brake pad. These forces were calibrated using fixed weights under gravity: there was a linear relationship between the applied voltage and static frictional resistance (Fig. 1A). Six levels (0–10 V in 2-V increments, corresponding to <1 to 9.8 N; for convenience, these will be considered as 0–10 N in 2 N steps in the analyses) of effort were used throughout the experiments. We incorporated sensors into the joystick to measure kinematic variables: capacitive touch detection, a potentiometer adapted to detect position, and paired strain gauges for detecting the directional strain on the joystick that reflects the force the monkey applied to the joystick.

For each daily experiment, each monkey was sat in its specifically adjusted primate chair (Crist Instruments) in a darkened room. A computer monitor placed 50 cm in front of the animal served to display visual stimuli. An opening on the front panel of the chair allowed the animal to express its choice by moving the joystick to pseudorandomly left-right alternating fixed positions.

The behavioral task was controlled by custom MATLAB software running on a Windows 10 computer with the Psychophysic Toolbox (Pelli, 1997) used to code the visual stimuli. Joystick position, strain and touch-state were digitized through a National Instruments Data Acquisition Card (PCI-6229) at a rate of 1 kHz. The electromagnet was similarly controlled through the data acquisition card at a rate of 250 Hz. Various amount of blackcurrant juice rewards (0–1 ml) were delivered by opening a solenoid (SCB262C068; ASCO) which gated a gravity-fed system.

Binary choice task

The animals were trained to work for liquid rewards. The animals were mildly fluid deprived for 6 d a week during which the totality of the fluid available was provided by the performance in the behavioral task, with additional subsequent daily access to ensure the minimum daily fluid requirements were met.

The monkeys were trained to associate two-dimensional visual stimuli (Fig. 1C,D) with blackcurrant juice rewards and experienced effort over the course of >5000 trials, including single-outcome trials. Following this training, all experimental data were gathered during a binary choice where the animal choose between two options presented simultaneously. Following option presentation, a yellow dot appeared on the screen indicating the position of the joystick and the begin of the choice phase. The animal then moved the joystick to the left or right to indicate its choice and received the reward after a 1 s delay. The chosen effort level was applied during the monkeys’ movement. Failure to make a choice resulted in the trial being repeated. Failure to complete a trial could arise for several reasons: (1) the animal did not center the joystick in the pretrial period, (2) the animal let go of the joystick during the trial, and (3) the animal did not move the joystick to either option within a 2-s time period following the go cue. After the animal had made its choice, the stimulus of the chosen option reappeared to reinforce its association with the reward. A variable intertrial period of 1–3 s with a blank screen ensued after each trial.

Experimental design and statistical analyses

To determine the relevant variables the monkeys used in their decision, we used a logistic model to estimate the probability of choosing a given side dependent on option variables (reward, effort) and control variables (side, previous choice location). The linear combination of these explanatory variables was related to the probability of a chosen side with the logistic function (Eq. 1) and parameters estimated within a generalized linear model. p(Left) ∼ β0 + β1(Reward) + β2(Effort) + β3(LastChoiceLeft) + ε.

P(left):probability left, Beta0 captures side bias, Beta1 the sensitivity to reward (left minus right value), Beta2 the sensitivity to effort (left minus right), and beta3 the probability of choosing the previously chosen side, where LastChoiceLeft is a binary variable that is 1 if the left option was previously chosen, and 0 otherwise.

Further logistic regressions were used to construct a common currency between reward and effort values. Parameters were independently estimated for each day’s session.

Logistic regressions were performed using the glm (generalized linear model) function from the stats package in R, assuming an underlying binomial distribution. Relevant variables were modeled as a linear combination related to the probability of choice by the standard logistic function (y = 1/(1 + e−x)), with e being Euler’s number. The same function returned the standard error of the mean. Coefficients were standardized using the β function in the reghelper package. Comparison of fits was done by comparing the Akaike Information Criterion (AIC; calculated by the glm.summary function), which penalizes the likelihood function by the number of terms in the model, and McFadden’s R2 (McFadden, 1973), a nonrelative metric of fit for logistic regressions, calculated using the pR2 function of the pscl package.

Reward utility and effort disutility functions were constructed by applying random utility models on choices that differed in reward only or effort only respectively. Random utility models, or discrete choice modeling (McFadden, 1973), assume choice is dependent on the underlying utility (subjective value) of the choices and random utility (error), with decision-makers choosing the highest total (underlying plus random) utility each time. Consequently, the probability of choosing a given option is related to the difference in the underlying utility of the two options, which in the models used here they are related by the logistic function. We modeled several parametrics forms of the utility functions, including linear, power-models and quadratic, and modeled the functions nonparametrically by representing the reward/effort levels as dummy variables corresponding to the discrete levels used (0.05 ml increments for reward, 2 N for effort). In all models, daily side bias and previous side choice were included as additional explanatory variables. Parameters were estimated using the Python package BIOGEME (Bierlaire, 2020), which uses maximum likelihood estimation methods to estimate the underlying utility functions from the observed choices. To model satiety effects, we split the data of a given session into consecutive quartiles of total liquid consumption during the session; then we estimated the utility function for reward separately for each quartile. Similarly, the effort disutility function was estimated for each session quartile split by total effort exertion. In all cases, utility was normalized such that the reward utility function spanned the range 0–1 (by factoring out a common slope parameter).

To test whether these utility functions accurately described the animals’ choice behavior, we used the functions to predict out-of-sample choices, which compromised all options sets where both the reward and effort were different between the options. We compared the subtractive model [reward utility minus effort disutility; u (reward) – u (cost)], hyperbolic discounting [reward multiplied by (1/(1+k*effort), where k is a scaling factor] and exponential discounting (reward multiplied by e−k* effort). Model comparison was by AIC and McFadden’s R2 (McFadden, 1973).

To determine how trial history influenced the sensitivity to effort in future trials, the logistic regression used in the first model (Eq. 1) was extended in the following terms: ∑ikBetai⋅Effortt⋅PreviousEffortt−i.

Where Effort is the effort difference on the current trial, and previous effort is the chosen effort on the ith previous trial.

To determine whether each additional previous choice term improved the fit, we compared models by computing the Akaike Information Criterion (AIC) at each step.

Given the notion that past effort experience influences an internal reference point against which current effort options are compared asymmetrically, we sought to implement a model which could identify this asymmetry. We adapted Crawford and Meng’s (2011) implementation of gain-loss utility into a random utility model as follows: p(Left) ∼ Beta0 + Beta1(JuiceUtility) + Beta2(EffortaboveRef) +Beta3(EffortbelowRef) + Beta4(LastChoseLeft) EffortbelowRef={Effort−Effortref,Effort < Effortref0,Effort≥Effortref Effort_aboveRef{0,Effort < EffortrefEffort−Effortref,Effort≥Effortref.

This model allows to statistically test the hypothesis that subjective effort cost is asymmetric around the reference point by testing for a difference between the standardized effort slope coefficients β2 and β3. In the first implementation of this model, effort reference was the moving average of the past 30 trials of effort.

With a view to accounting for trial-by-trial variation in the effort reference point, we developed a model of choice in which the effort reference point was learnt under a Rescorla–Wagner type reinforcement learning rule: Effortrefn+1=Effortrefn + alpha(Effortn−Effortrefn) alpha={alpha1,Effort > Effortrefalpha2,Effort≤Effortref.

The learning rate (α) was modeled as two constants, one for efforts above reference and one for efforts below reference to allow for similar asymmetry and hypothesis testing as discussed in the estimate of the effort sensitivity parameters. Optimal learning rates were determined by grid search, with optimality defined by minimized AIC.

Results

Task design

We tested choices between options that differed in the magnitude of reward the animal would obtain and the amount of effort to obtain that reward.

Following extensive training on stimuli (>5000 trials), we presented the animals with sequences of binary choices which differed in both reward and effort levels. The choice ratio indicated how often one option was chosen over its alternative and was used to measure the relative value of the options. We were particularly interested to determine options with equal value, as indicated by their choice at equal rate (i.e., same probability of p = 0.5 of choosing each option). Through this process, we estimated and modeled the monkeys’ value of reward and effort on a single, “common currency” scale.

The monkeys used the visual information to guide their decisions across all testing sessions, as demonstrated both through analysis of their choices and kinematic data. To determine the information the relevant variables used in the monkeys’ choices, we performed a logistic regression analysis on the probability of choosing the left option on the option variables (juice reward and effort) as well as other control variables (previous choice location, intercept showing left-right bias; Eq. 1).

The monkeys chose options with higher juice rewards when effort to obtain was equal (Fig. 2A,B) and chose low effort options over high effort options when juice rewards were equal (Fig. 2C,D). A positive coefficient of reward magnitude indicated that both monkeys preferred options with greater rewards (Monkey W: β = 3.47, z = 160.01, p < 0.001; Monkey U: β = 1.69, z = 42.63, p < 0.001; Fig. 2E,F). The negative coefficient of effort level indicated that the monkeys preferred less effort to more effort (Monkey W: β = −0.26, z = −31.16, p < 0.001; Monkey U: β = −1.02, z = −56.66, p < 0.001). These results are consistent with rational behavior where an agent makes choices “as if” maximizing subjective value by optimizing rewards (juice) at minimal cost (effort to obtain).

Figure
  • Download figure
  • Open in new tab
  • Download powerpoint

Behavioral and kinematic analysis validation and characterization of task design. A, B, Both animals chose more reward over less reward, given effort being equal. C, D, Monkeys chose less effort over more effort, given equal reward. E, F, Standardized coefficients from logistic regression on choosing left side, including side bias, reward and effort quantity, and bias toward previously chosen side. Column represents overall average, dots represent per-session estimates G, Average profile of position, velocity, and strain for Monkey U at three separate effort levels. Time zero indicates the appearance of the Go cue. H, Average strain in the 100 ms before the Go cue (shaded area) separated by chosen side and effort level. In general, both animals were applying greater strain to the joystick for higher effort levels, reflecting preparation for the greater resistance to movement and therefore demonstrating an understanding of the effort cues. Two-way ANOVA on effort level and chosen direction; chosen effort effect: Monkey U: F(5,6021) = 17.5, p < 0.001; Monkey W: F(5,4707) = 24.57, p < 0.001. Error bars are SEM.

On some days, there was a significant coefficient representing side bias. This left-right bias may reflect the distinct movements required for the left and right targets. Both monkeys exclusively used their right arm; therefore, a movement to the left involved wrist flexion and pronation. and elbow flexion; conversely, a movement to the right involved wrist extension and supination and elbow extension. Some forceful left movements also involved moderate shoulder adduction (right movements involved shoulder abduction) that was restricted by the walls of the primate chair. The left-right bias did not change within individual sessions, as determined by estimating the bias on a block-by-block basis and therefore in all subsequent analysis side-bias was accounted for on a day-by-day basis. Changes in side-bias may reflect the relative position of the chair to the joystick, which varied slightly day-to-day but not within sessions, with the relative position altering the balance of the movements required for left and right choices. There was a small but statistically significant bias to choose the side chosen in the previous trial (side choice hysteresis), potentially reflecting an attention bias to the previously chosen and thus rewarded side.

The monkeys’ understanding of the task was confirmed by the study of kinematic data (Fig. 2G). The joystick contained strain gauges that measured the preparatory efforts the monkeys made before the “Go” cue (Fig. 2H). Averaging the directional strain in the 100 ms before the Go cue, the animals prepared to move the joystick in their chosen direction and with force that changed with the effort level of their chosen option (two-way ANOVA on effort level and chosen direction; chosen effort effect: Monkey U: F(5,6021) = 17.5, p < 0.001; Monkey W: F(5,4707) = 24.57, p < 0.001). Thus, the animals seemed to be preparing to move with more force when the resistance to movement (Fig. 2H), as cued by the fractal, was higher.

Psychometric construction of common value scale

In order to test effort cost discounting at both behavioral and neuronal levels, it is necessary to establish a common scale of value between the reward and effort. Here, we first established the viability of this method through a standard psychometric approach before extending it through random utility modeling.

To establish the equivalent value of an effort level in milliliters of juice (in the “common currency” of reward), we established the indifference point (p = 0.5 choice probability) in the following way: Against a fixed magnitude of reward, at the lowest effort level, we presented a variable option (Fig. 3A). This variable option varied in reward magnitude between trials, with the same effort-level within blocks. The resulting data were fitted with a logistic function, with reward quantity as the explanatory variable. From this function, we estimated the magnitude of reward at which the variable option was chosen as frequently as the fixed option (p = 0.5 each option; Fig. 3B). The difference in reward quantity between the indifference point and the fixed option was taken to be of the same subjective value to the monkey as the difference in the effort levels between the two options. Consistent with effort as a cost (or negative reward), the effort needed to be compensated with more juice rewards; thus, effort was considered to be equivalent to negative reward quantity. To check the consistency of these measures, we repeated and measured variation across days and against different fixed options. In general, increased effort corresponded to more negative reward equivalents (Fig. 3C). The smaller monkey (Monkey W) had more negative reward equivalents for the same effort levels compared with the larger monkey (Fig. 3C, left vs right), which likely reflects the difference in their strength and thus the relative value of the effort.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Psychometric construction of common value scale. A, Task design. To establish the equivalent value of effort in units of reward the indifference point between two levels of effort was titrated. Against a fixed option (here, 0.4-ml reward, lowest effort level), a variable option was presented in which the effort level and reward was changed between trials. B, Example psychometric function fit for a single trial block in Monkey W (160 trials). Lines and shaded area represents fit and fit SEM. Value of effort determined from distance on reward-axis at the indifference point (choice p = 0.5 of each option) for each effort level. C, Average equivalent reward value of each effort level for Monkey U (left) and Monkey W (right). In general, higher effort values corresponded to more negative reward values. Effort values were more negative for the smaller monkey (Monkey W), potentially reflecting a different in strength. Red dots are average over all sessions, error bars are SEM, transparent dots are session averages.

Random utility modelling

Psychometric functions provide a good basis for establishing subjective value but do not completely capture subjective value. For example, if the marginal utility of reward decreases with increasing reward, the apparent cost of effort could appear lower when tested using a higher reward value as the fixed value. Furthermore, the apparent value of reward may shift over the course of a testing session if satiety reduces the reward value as the animal became less thirsty. Risky-choice methods (e.g., fractile procedure; Genest et al., 2016) often reveal nonlinear utility functions that may contain a mixture of concavity and convexity within the tested range. In order to obtain a more complete understanding of the subjective value of effort, it would be necessary to elicit the utility functions for gain and effort independently.

To limit the number of experimental parameters, we refrained from extensive testing of gamble options, as the additional confounds of risk preference and probability distortion might complicate the model and the understanding of the data. Therefore, we used a riskless method to establish functions for the utility of juice reward and the disutility of effort (Fig. 4A,B). In short, random utility modeling (or discrete choice modeling; McFadden, 1973) extends psychometric testing by using the choice ratio at each sampled option pair to infer the utility difference between options. Random utility modeling has been successful in exploring economic decision-making in animals and humans (Webb et al., 2019; Bujold et al., 2022). In practice, we established the choice ratio between options that varied in reward but not effort, thus estimating a reward utility function. Similarly, we varied effort but not reward to estimate an effort disutility function. We used random utility modeling to fit several common utility functional forms for the reward utility and effort disutility: linear and quadratic polynomials, power function as well as nonparametrically fitting to the discrete levels of reward/effort.

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Random utility modeling and results. A, Theory of random utility modeling for reward and effort. Random utility assumes that choice probabilities are determined by the different in utility of (ΔU), not the objective quantity difference (Δx) plus some fixed error. This shows the relationship between three choice sets that have the same objective quantity difference in reward (top) and effort (bottom) but different utility differences and the expected choice probabilities for these six pairs of choices. Random utility model thus uses these observed binary choice probabilities to infer the underlying utility functions. Utility differences are exaggerated for clarity. B, Monotonically increasing utility with reward amount; consistency across four different fitting functions. C, Estimation of reward utility and effort disutility functions from choice options that only differed in reward or in effort. Out-of-sample validation was conducted by predicting choices using the best fitting reward and effort disutility functions to predict choices that were not used to estimate either. D, Near-monotonically decreasing effort (dis)utility with effort (N, Newton). E, Out-of-sample psychometric assessment of utility difference between leftward and rightward choices. F, Quantitative comparisons of fits for the four utility functions tested. G, Model fits for three commonly used combined utility models.

Through this approach, we established independent estimates of reward utility functions (Fig. 4C) and effort (dis)utility functions (Fig. 4D). For both animals and all functional forms, reward utility increased monotonically as predicted by basic economic theory. Comparing the various functional utility forms using the AIC, which penalizes model complexity, we determined that the quadratic model was the best model for both subjects. Effort disutility decreased monotonically, reflecting the choice of low effort options over high effort options. Effort disutility was best modeled nonparametrically; Monkey U was relatively insensitive to the lower efforts, whereas Monkey W was more sensitive to the lower efforts (likely owing to his smaller size), but treated the two highest effort levels similarly, as seen in the nonparametric fit (Fig. 4D). These data were replicated with all utility functions fitted (Fig. 4F).

To validate these utility estimates, we tested how well they predicted out-of-sample choices, specifically all choices that differed in both effort and reward simultaneously which therefore were not used for estimating the utility functions. We combined the reward utility and effort disutility using common methods from the literature (subtractive, hyperbolic and exponential (Białaszek et al., 2017). The subtractive model [reward utility less effort (dis)utility] provided a better fit to the empirical data than the other models we tested (Fig. 4G). Furthermore, the difference in utility between options predicted the choice ratio between options in a sigmoid function, as expected, confirming the model’s validity (Fig. 4E).

Satiety effects

The law of diminishing marginal utility captures a common-sense notion: as we have more of one thing the subjective value of an additional unit decreases. In this experiment, the advancing reward consumption of the monkeys would induce satiety and predict a decrease in the value of the reward. We inferred such a value reduction from the choice changes with the same option set between the start of the daily session and its end. To achieve this, we split each session into four consecutive quartiles by the total reward volume consumed and examined choice probabilities in each. In Figure 5A, an example choice (0.2- vs 0.25-ml reward, fixed effort) is plotted. Early, when the animal is thirsty, there is a strong preference for the higher reward, but as the session continues, the animal chooses the lower option more frequently.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Juice consumption induces satiety effects. A, With increasing reward consumption (quartiles 1–4), the choice difference between the same options narrowed, possibly reflecting satiety. For each quartile, the left column displays the probability of choosing the option yielding 0.20 ml, whereas the right column shows the choice probability for the option yielding 0.25 ml. Data from Monkey W. Error bars are SEM. B, Utilities for different amounts of consumed reward, as modeled by including total reward consumption as a latent variable in the reward utility model, thus quantifying the reduced reward utility as reward consumption increased. The four lines refer to the four quartiles (top to bottom).

The random utility model allowed us to quantify the change in choice ratio by estimating the reward utility function for each quartile. In both animals, the utility of reward decreased during the session (Fig. 5B), as observed in other studies (Pastor-Bernier et al., 2021).

Adaptation to effort

In a similar manner to satiety having an effect on the reward utility, we expected to observe a fatigue to develop over the session and increase the sensitivity to effort. Rather, we found that both subjects became less sensitive to effort over the course of each session (Fig. 6A). For example, when choosing between options requiring 0 versus 8 N of effort, Monkey W showed significantly stronger preference for the low effort option in earlier as compared with later daily sessions (Fig. 6A), suggesting the value difference of effort had subjectively decreased.

Figure 6.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 6.

Effort choice history influences sensitivity to effort. A, For each quartile of daily trials, the left column displays the probability of choosing the low effort option (0 N), whereas the right column shows the choice probability for the high effort option (8 N) in Monkey W. B, Quadratic fits of effort disutility for the four quartiles suggest effort disutility decreases with total effort expenditure over each session. The four curves refer to the four quartiles (top to bottom). Error bars show standard errors of the parameter estimate.

Splitting sessions by the cumulative effort into quartiles and using choices that only varied in effort, we estimated the effort disutility curves with a quadratic function (Fig. 6B). In both subjects, the effort disutility diminished through the session, rather than increase as we would predict if the animals were fatiguing. While other studies have demonstrated that fatigue affects effort-based decision-making (Hogan et al., 2020), we did not clearly observe this effect, rather the reverse.

Effect of previously experienced effort on effort disutility

Following this observation that sensitivity to effort was decreasing during each session, we wanted to test whether the changes in effort were dependent on the experience of previous effort, as opposed to just time in the session. Previous studies in rodents, nonhuman primates and humans have identified that choice adaptation effects can explain seemingly irrational choices (Ferrari-Toniolo et al., 2019; Beron et al., 2022; Glimcher, 2022). To quantify the effect of previous trials, we performed a logistic regression adding the previously chosen effort level as interaction term for the variable on effort difference in Eq. 3. In a stepwise process, the explanatory power of further lagged terms was tested. Additional explanatory power was taken to be a reduction in the AIC over the previous model. This process revealed a small but significant dependence on previously chosen effort over at least the previous 27 trials in both animals (Monkey U: 27 trials, Monkey W: 30 trials).

Asymmetric effort valuation relative to effort history

A possible model to explain the diminishing effort disutility observed is the reference-dependent model. Reference-dependent models pose that value is determined relative to an internal, subjective reference point. A common finding of reference models is that there is asymmetry on either side of the reference point, for example, the gain-loss asymmetry in prospect theory. The evidence from the behavioral economics literature suggests that individuals are more sensitive to efforts above the reference point than to efforts of the same magnitude below the reference point (Card et al., 2012).

A simple implementation of a reference point is to use the moving mean effort of previous trials. To test this implementation, we adapted the gain-loss utility model used by Crawford and Meng (2011), with effort disutility being modeled as a piece-wise linear function, allowing for a change in slope at the reference point (Eq. 3; Fig. 7A) and using the random utility modeling techniques described above to fit the function to the data. In fitting this model to both animals, there was significant evidence of an asymmetric response to effort (Fig. 7C). For Monkey U, the parameter capturing effort sensitivity above the reference point [β, −0.44 95% CI: (−0.41, −0.47)] was 3.6 times greater than the parameter capturing effort sensitivity below the reference point [β, −0.13; 95% CI: (−0.10, −0.16)]. In Monkey W the change in slope was even greater, with the sensitivity above the reference point [β, −1.23; 95% CI: (−1.17, −1.28)] being 4.5 times the sensitivity below the reference point [β, −0.27; 95% CI: (−0.23, 0.31)].

Figure 7.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 7.

Reference-dependent effort preferences. A, Piecewise-linear effort model. Effort disutility is the difference between effort and the effort reference point, scaled by a slope constant that is dependent on whether the effort is above or below the reference point. B, Reference point models. The effort reference point was either modeled as the average of the previous 30 efforts (moving average model) or learnt from previous efforts in with a learning rate dependent on whether the effort was above or below the reference point (asymmetric Rescorla–Wagner model). C, Effort disutility for Monkeys U and W, as scaled to the same utility as reward utility in Figure 4. Reference was modeled as a moving average of previous chosen effort. D, Reference point model from Rescorla–Wagner reinforcement learning model. In all cases, the slope above the reference point was significantly greater than below the reference point, suggesting reference-dependent preferences. Shaded ribbon is 1 SD of model fit.

Rescorla–Wagner learning model of effort reference point

While the above finding is evidence for a reference-dependent effort valuation, it does not sufficiently address how the animals change their reference on a trial-by-trial basis, as the previous experiments established high effort trials have an effect on effort disutility through the session. Yet is necessary, as a basis for the neurophysiology, to understand how value changes on a trial-by-trial basis. To address this, we developed a model in which the effort-reference reflects an expectation of future evidence learned from previously experienced effort (Fig. 7B). To model learning of expected effort, we used the Rescorla–Wagner reinforcement learning model because of its inherent compatibility with temporal difference learning models, as it is based on prediction-error driven learning but has a reduced set of parameters that require estimating. In this model, the expectation of effort was updated on a trial-by-trial basis by the effort error (expected effort minus actual) multiplied by a learning rate. To allow for a difference in saliency of efforts above and below reference, analogous to the difference in saliency of losses and gains in prospect theory, the optimal learning rate was estimated separately for efforts above and below the reference point. We then used the same random utility modeling techniques with the piecewise linear function to estimate the effort disutility function. Optimal learning rates were established by a grid search method to minimize AIC.

In the optimal fits for both animals there was strong evidence that efforts above the reference point were more negatively valued than corresponding positive valued efforts the same magnitude below the reference point (Fig. 7D). The optimal learning rates also reflected that efforts above the reference point were significantly more salient than efforts below the reference point (Monkey U optimal rate above reference point: 0.089, below reference point: 0.039; Monkey W, optimal learning rate above reference point 0.29, below reference point: 0.09). Notably the ratio of learning rate above and below reference point and the slope of the utility function above and below was approximately equivalent in both animals (Monkey U: slope above reference = 2.87 × slope below reference, learning rate above reference = 2.28 × learning rate below reference; Monkey W: slope above reference = 3.26 × slope below reference, learning rate above reference = 3.22 × learning rate below reference). This suggests that the utility prediction error drives the learning of the effort reference point. Indeed, using a single learning rate for the effort reference point update performs very similar to the model where the reference point is learnt from objective effort difference with different learning rates for positive and negative errors (Monkey U, McFadden’s R2: 0.401 vs 0.400 for one learning rate, subjective update model vs two learning rates, objective update model; Monkey W: McFadden’s R2: 0.397 vs 0.397).

Furthermore, in this model, in which effort costs are referenced against a learned expectation of effort, better explained both monkeys’ choices versus previous models explored. It outperformed the model in which effort history was used as the reference point (Monkey U, effort history: AIC: 36,791, McFadden’s R2: 0.405, learnt effort AIC: 36,156, McFadden’s R2: 0.415; Monkey W effort history: AIC: 35,594, McFadden’s R2: 0.397, learnt effort: 35 451, McFadden’s R2: 0.401). This is potentially explained by the asymmetry of the learning rates as high efforts raise the reference effort in the learned model faster and the reference effort remained elevated even after long periods of low effort choices. This model also explained choice better in both animals versus the previous models discussed, in which reward and effort disutility were independently estimated, despite simplifying the effort disutility function to a piecewise linear function (learnt effort model: McFadden’s R2 Monkey U: 0.415, Monkey W: 0.401; subtractive model: Monkey U: 0.342, Monkey W: 0.372).

Discussion

In this study, we examined the cost of effort in a binary choice task and used random utility modeling to infer reward utility and effort disutility from riskless choice data. In line with numerous studies in humans, macaques and other animal models, we found evidence that effort is treated as a cost weighed against potential rewards during decision-making. Moreover, this effort cost function is not fixed, but rather dependent on an internal reference point, which is learned from previous experience.

Specifically, the subtractive model of effort discounting was the best fitting model, as predicted by economic theory (Fehr and Goette, 2007). Moreover, in demonstrating that the independently estimated utility functions can be used to predict out-of-sample choices with the subtractive model, we provide direct evidence for the underlying assumption that a cost-benefit calculation underlies effort-based decision-making.

Random utility models have been used across several species and humans (Webb et al., 2019; Pastor-Bernier and Schultz, 2021; Bujold et al., 2022). Here, we demonstrate that the random utility methods better fit the choices than psychometric approaches alone, as has been previous demonstrated in other tasks in nonhuman primates and humans (Webb, 2019; Bujold et al., 2022). It remains to be seen whether this model will be useful in explaining neuronal data, as previous attempts to use random utility modeling in this way suggests it does not necessarily explain more neuronal variance, potentially because of measurement issues (Webb et al., 2019). Nevertheless, these methods provided useful as the satiety and effort adaptation effects could be examined using nonoverlapping choice data.

The demonstrated satiety and effort adaptation effects explain why the random utility model produced a monotonically decreasing effort function whereas the estimates from the psychometric analysis did not; as the monkeys consumed more juice, they were less sensitive to differences in reward quantity, reflecting diminishing utility. Similarly, increasing amounts of effort experienced diminished effort disutility. The psychometric method does not account for these changes and therefore add significantly to the variance.

The effort adaptation effect is captured by the learning model, in which the reference point is an expectation of effort learnt from previous trials driven by a prediction-error mechanism, improved over models in which effort was a fixed-cost throughout the session. This model, which is an adapted Rescorla–Wagner type learning rule, is a model of reinforcement learning. Here, the similar ratio of the learning rates above and below the reference point, with the relative utilities of effort above and below the reference point suggests that effort utility errors drive the learning of expected effort, and ultimately can explain the trial-history effects.

Indeed, much recent work has demonstrated context effects, particularly temporal context effects, that affect value-based decision-making. For example, Constantinople et al. (2019) used a similar asymmetric behavioral response model (prospect theory) to characterize choice behavior in rats, demonstrating similar trial history effects that were explained by a reinforcement learning model. The behavioral inconsistencies such models predict have suggested to be potentially optimal given the fundamental limitations of the brain to represent, store and calculate value (Bhui et al., 2021; Glimcher, 2022).

Reference dependent efforts may explain why previous studies investigating the link between phasic dopamine and effort costs have not found consistent coding of effort costs following presentation of effort-predicting cues (Gan et al., 2010; Pasquereau and Turner, 2013; Varazzani et al., 2015). If the subjective value of effort is reference-dependent in the manner suggested here, then pseudorandom ordering of effort levels will bias the reference effort level toward the high end of the range of efforts, because of the asymmetry in value driving the learning of expected effort. When efforts are compared with a high effort reference point, most of the effort range will lie below the reference point, in which the relative value between effort levels is diminished and therefore little-to-no encoding of effort found in phasic dopamine. This is supported by the finding of Gan et al. (2010), in which there were sudden changes in the effort level were accompanied by marked modulation of dopamine release in the nucleus accumbens, suggesting dopamine may encode value relative to the expected value of effort.

Returning to the example of the introduction, walking to the kitchen and hiking up a mountain involve wildly different amounts of effort expenditure, yet the brain is able to decide to undertake these actions thus implying flexible decision mechanisms. Expectation is likely a major component of this flexibility: in economics, the expectation of reward and cost has been demonstrated to influence choice behavior (Kahneman and Tversky, 1979). In biology, adaptive patterns in neuronal codes are common, as range-adaptation is an inherent feature of the brain. Sensory neurons tune their firing to match the distribution of sensory signals, and recent evidence suggests similar patterns exist in neurons encoding value, including midbrain dopamine neurons (Tobler et al., 2005; Kobayashi et al., 2010; Louie et al., 2015; Burke et al., 2016). Behaviorally, many psychological experiments have demonstrated that effort generalizes and adapts across different behaviors like the adaptation to effort observed here (Eisenberger et al., 1979, 1989).

A learned effort reference point is evolutionary useful: theory and modeling suggest that an appropriate reference point that reflects realistic effort requirements supports optimal effort-based decision-making (Rigoli and Pezzulo, 2022). Other recent work has examined suggested similar biases are explainable as optimally efficient (Bhui et al., 2021; Glimcher, 2022). Unlike in this task, expectations may be learned from more than previous experience alone; the human literature provides good evidence that effort expectations learned from social information influence effort-based decisions (Card et al., 2012).

Similar reinforcement learning models have been a very successful framework for understanding neuronal activity within brain areas involved in decision-making, including the midbrain dopamine neurons. Midbrain dopamine neurons encode reward prediction errors and optogenetic experiments, in which dopamine neuron activity was manipulated to enhance or reduce prediction errors, observe changes in reward learning behavior that demonstrate a causal link between dopamine prediction errors and reward valuation (Stauffer et al., 2016). We did not exhaustively model all possibilities of adaptive functions, only showing that the learning rule, which accounts for potential asymmetry outperforms a rolling-average: we did not consider any model-based learning paradigms, nor were we able to examine intersession effects. One possibility not accounted for here is that low effort options, particularly the lowest effort option, may have positive utility. If the average effort required to obtain a reward is high, a low effort option may be perceived as a rest from effort and therefore may have increasing or even positive utility that is not revealed in these analyses because of anchoring the zero-effort option to zero utility. It is difficult to assess this effect with this task as the animals may also rest by skipping trials and it is not possible to correlate this effect with fatigue as the effort-adaptation effects appear to obscure any fatigue effects. As a technical point, random utility methods only estimate utility difference between choices and not absolute utility and therefore we cannot definitively establish zero utility in this task design.

In neuroscience, computational modeling, notably of activity in the anterior cingulate cortex, an area known to be involved in effort valuation, have directly addressed adaptive effort effects (Verguts et al., 2015). These models suggest that effort is both learnable and context-dependent and therefore expectations of future effort influence the value of effort, potentially suggesting the anterior cingulate cortex is a neural substrate for effort expectation calculations. The anterior cingulate cortex is involved in decision-making circuitry, for example, it exerts influence over dopamine neurons in the ventral tegmental area via monosynaptic glutamatergic connections, suggesting such adaptive effort values may be represented in dopamine neurons as well (Beier et al., 2015).

Overall, these experiments suggest that monkeys’ subjective value of effort is reference-dependent, and that the reference point is, at least in part, an expectation of the average effort level formed from experience in previous trials. Such effects are necessary to control and understand when probing the relationship between midbrain dopamine neurons and effort cost encoding. 

Footnotes

  • This work was supported by Wellcome Trust Grants WT 095495, WT 204811, and WT 206207; the European Research Council (ERC) Grant 293549); and the National Institutes of Mental Health (NIMH) Caltech Conte Center Grant P50MH094258. For the purpose of Open Access, the authors have applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission. We thank Aled David and Christina Thompson for animal and technical support and Philipe Bujold and Fabian Grabenhorst for discussions on the design and statistics of this experiment.

  • The authors declare no competing financial interests.

  • Correspondence should be addressed to Wolfram Schultz at Wolfram.Schultz{at}protonmail.com

SfN exclusive license.

References

  1. ↵
    1. Abeler J,
    2. Falk A,
    3. Goette L,
    4. Huffman D
    (2011) Reference points and effort provision. Am Econ Rev 101:470–492. https://doi.org/10.1257/aer.101.2.470
    OpenUrlCrossRef
  2. ↵
    1. Akerlof GA,
    2. Yellen JL
    (1990) The fair wage-effort hypothesis and unemployment. Q J Econ 105:255–283. https://doi.org/10.2307/2937787
    OpenUrlCrossRef
  3. ↵
    1. Allen EJ,
    2. Dechow PM,
    3. Pope DG,
    4. Wu G
    (2017) Reference-dependent preferences: evidence from Marathon runners. Management Sci 63:1657–1672. https://doi.org/10.1287/mnsc.2015.2417
    OpenUrl
  4. ↵
    1. Arkes HR,
    2. Hirshleifer D,
    3. Jiang D,
    4. Lim S
    (2008) Reference point adaptation: tests in the domain of security trading. Organ Behav Hum Decis Process 105:67–81. https://doi.org/10.1016/j.obhdp.2007.04.005
    OpenUrlCrossRef
  5. ↵
    1. Bautista LM,
    2. Tinbergen J,
    3. Kacelnik A
    (2001) To walk or to fly? How birds choose among foraging modes. Proc Natl Acad Sci U S A 98:1089–1094. https://doi.org/10.1073/pnas.98.3.1089 pmid:11158599
    OpenUrlAbstract/FREE Full Text
  6. ↵
    1. Beier KT,
    2. Steinberg EE,
    3. DeLoach KE,
    4. Xie S,
    5. Miyamichi K,
    6. Schwarz LA,
    7. Gao XJ,
    8. Kremer EJ,
    9. Malenka RC,
    10. Luo L
    (2015) Circuit architecture of VTA dopamine neurons revealed by systematic input-output mapping. Cell 162:622–634. https://doi.org/10.1016/j.cell.2015.07.015 pmid:26232228
    OpenUrlCrossRefPubMed
  7. ↵
    1. Beron CC,
    2. Neufeld SQ,
    3. Linderman SW,
    4. Sabatini BL
    (2022) Mice exhibit stochastic and efficient action switching during probabilistic decision making. Proc Natl Acad Sci U S A 119:e2113961119. https://doi.org/10.1073/pnas.2113961119 pmid:35385355
    OpenUrlPubMed
  8. ↵
    1. Bhui R,
    2. Lai L,
    3. Gershman SJ
    (2021) Resource-rational decision making. Curr Opin Behav Sci 41:15–21. https://doi.org/10.1016/j.cobeha.2021.02.015
    OpenUrl
  9. ↵
    1. Białaszek W,
    2. Marcowski P,
    3. Ostaszewski P
    (2017) Physical and cognitive effort discounting across different reward magnitudes: tests of discounting models. PLoS One 12:e0182353. https://doi.org/10.1371/journal.pone.0182353 pmid:28759631
    OpenUrlCrossRefPubMed
  10. ↵
    1. Bierlaire M
    (2020) A short introduction to PandasBiogeme, Technical report TRANSP-OR 200605. Transport and Mobility Laboratory, ENAC, EPFL. Lausanne, Switzerland.
  11. ↵
    1. Bujold PM,
    2. Seak LCU,
    3. Schultz W,
    4. Ferrari-Toniolo S
    (2022) Comparing utility functions between risky and riskless choice in rhesus monkeys. Anim Cogn 25:385–399. https://doi.org/10.1007/s10071-021-01560-x pmid:34568979
    OpenUrlPubMed
  12. ↵
    1. Burke CJ,
    2. Baddeley M,
    3. Tobler PN,
    4. Schultz W
    (2016) Partial adaptation of obtained and observed value signals preserves information about gains and losses. J Neurosci 36:10016–10025. https://doi.org/10.1523/JNEUROSCI.0487-16.2016 pmid:27683899
    OpenUrlAbstract/FREE Full Text
  13. ↵
    1. Card D,
    2. Mas A,
    3. Moretti E,
    4. Saez E
    (2012) Inequality at work: the effect of peer salaries on job satisfaction. Am Econ Rev 102:2981–3003. https://doi.org/10.1257/aer.102.6.2981
    OpenUrlCrossRef
  14. ↵
    1. Clement TS,
    2. Feltus JR,
    3. Kaiser DH,
    4. Zentall TR
    (2000) “Work ethic” in pigeons: reward value is directly related to the effort or time required to obtain the reward. Psychon Bull Rev 7:100–106. https://doi.org/10.3758/bf03210727 pmid:10780022
    OpenUrlCrossRefPubMed
  15. ↵
    1. Constantinople CM,
    2. Piet AT,
    3. Brody CD
    (2019) An analysis of decision under risk in rats. Curr Biol 29:2066–2074.e5. https://doi.org/10.1016/j.cub.2019.05.013 pmid:31155352
    OpenUrlCrossRefPubMed
  16. ↵
    1. Cousins MS,
    2. Atherton A,
    3. Turner L,
    4. Salamone JD
    (1996) Nucleus accumbens dopamine depletions alter relative response allocation in a T-maze cost/benefit task. Behav Brain Res 74:189–197. https://doi.org/10.1016/0166-4328(95)00151-4 pmid:8851929
    OpenUrlCrossRefPubMed
  17. ↵
    1. Crawford VP,
    2. Meng J
    (2011) New York City cab drivers’ labor supply revisited: reference-dependent preferences with rational-expectations targets for hours and income. Am Econ Rev 101:1912–1932. https://doi.org/10.1257/aer.101.5.1912
    OpenUrlCrossRef
  18. ↵
    1. Critchfield TS,
    2. Kollins SH
    (2001) Temporal discounting: basic research and the analysis of socially important behavior. J Appl Behav Anal 34:101–122. https://doi.org/10.1901/jaba.2001.34-101 pmid:11317983
    OpenUrlCrossRefPubMed
  19. ↵
    1. De Giorgi EG,
    2. Post T
    (2011) Loss aversion with a state-dependent reference point. Management Sci 57:1094–1110. https://doi.org/10.1287/mnsc.1110.1338
    OpenUrl
  20. ↵
    1. Eisenberger R,
    2. Carlson J,
    3. Guile M,
    4. Shapiro N
    (1979) Transfer of effort across behaviors. Learn Motiv 10:178–197. https://doi.org/10.1016/0023-9690(79)90043-2
    OpenUrlCrossRef
  21. ↵
    1. Eisenberger R,
    2. Weier F,
    3. Masterson FA,
    4. Theis LY
    (1989) Fixed-ratio schedules increase generalized self-control: preference for large rewards despite high effort or punishment. J Exp Psychol Anim Behav Process 15:383–392. https://doi.org/10.1037/0097-7403.15.4.383
    OpenUrlCrossRef
  22. ↵
    1. Fehr E,
    2. Goette L
    (2007) Do workers work more if wages are high? Evidence from a randomized field experiment. Am Econ Rev 97:298–317. https://doi.org/10.1257/000282807780323479
    OpenUrlCrossRef
  23. ↵
    1. Ferrari-Toniolo S,
    2. Bujold PM,
    3. Schultz W
    (2019) Probability distortion depends on choice sequence in rhesus monkeys. J Neurosci 39:2915–2929. https://doi.org/10.1523/JNEUROSCI.1454-18.2018 pmid:30705103
    OpenUrlAbstract/FREE Full Text
  24. ↵
    1. Friedman H,
    2. Tarpy RM,
    3. Komelski P
    (1968) The preference of rats for a more difficult task. Psychon Sci 13:157–158. https://doi.org/10.3758/BF03342458
    OpenUrl
  25. ↵
    1. Gan JO,
    2. Walton ME,
    3. Phillips PEM
    (2010) Dissociable cost and benefit encoding of future rewards by mesolimbic dopamine. Nat Neurosci 13:25–27. https://doi.org/10.1038/nn.2460 pmid:19904261
    OpenUrlCrossRefPubMed
  26. ↵
    1. Genest W,
    2. Stauffer WR,
    3. Schultz W
    (2016) Utility functions predict variance and skewness risk preferences in monkeys. Proc Natl Acad Sci U S A 113:8402–8407. https://doi.org/10.1073/pnas.1602217113 pmid:27402743
    OpenUrlAbstract/FREE Full Text
  27. ↵
    1. Glimcher PW
    (2022) Efficiently irrational: deciphering the riddle of human choice. Trends Cogn Sci 26:669–687. https://doi.org/10.1016/j.tics.2022.04.007 pmid:35643845
    OpenUrlCrossRefPubMed
  28. ↵
    1. Hogan PS,
    2. Chen SX,
    3. Teh WW,
    4. Chib VS
    (2020) Neural mechanisms underlying the effects of physical fatigue on effort-based choice. Nat Commun 11:4026. https://doi.org/10.1038/s41467-020-17855-5 pmid:32788604
    OpenUrlCrossRefPubMed
  29. ↵
    1. Kahneman D,
    2. Tversky A
    (1979) Prospect theory: an analysis of decision under risk. Econometrica 47:263–291. https://doi.org/10.2307/1914185
    OpenUrlCrossRefPubMed
  30. ↵
    1. Knauss ZT,
    2. Filipovic M,
    3. Smith KA,
    4. Queener MM,
    5. Lubera JA,
    6. Bolden-Hall NM,
    7. Smith JP,
    8. Goldsmith RS,
    9. Bischoff JE,
    10. Miller MK,
    11. Cromwell HC
    (2020) Effort-reward balance and work motivation in rats: effects of context and order of experience. Behav Processes 181:104239. https://doi.org/10.1016/j.beproc.2020.104239 pmid:32966870
    OpenUrlPubMed
  31. ↵
    1. Kobayashi S,
    2. Schultz W
    (2008) Influence of reward delays on responses of dopamine neurons. J Neurosci 28:7837–7846. https://doi.org/10.1523/JNEUROSCI.1600-08.2008 pmid:18667616
    OpenUrlAbstract/FREE Full Text
  32. ↵
    1. Kobayashi S,
    2. Carvalho OD,
    3. Schultz W
    (2010) Adaptation of reward sensitivity in orbitofrontal neurons. J Neurosci 30:534–544. https://doi.org/10.1523/JNEUROSCI.4009-09.2010 pmid:20071516
    OpenUrlAbstract/FREE Full Text
  33. ↵
    1. Köszegi B,
    2. Rabin M
    (2006) A model of reference-dependent preferences. Q J Econ 121:1133–1165. https://doi.org/10.1162/qjec.121.4.1133
    OpenUrlCrossRef
  34. ↵
    1. Louie K,
    2. Glimcher PW,
    3. Webb R
    (2015) Adaptive neural coding: from biological to behavioral decision-making. Curr Opin Behav Sci 5:91–99. https://doi.org/10.1016/j.cobeha.2015.08.008 pmid:26722666
    OpenUrlCrossRefPubMed
  35. ↵
    1. Lydall ES,
    2. Gilmour G,
    3. Dwyer DM
    (2010) Rats place greater value on rewards produced by high effort: an animal analogue of the “effort justification” effect. J Exp Soc Psychol 46:1134–1137. https://doi.org/10.1016/j.jesp.2010.05.011
    OpenUrlCrossRef
  36. ↵
    1. McFadden D
    (1973) Conditional logit analysis of qualitative choice behavior. In: Frontiers in econometrics (Zarembka P, ed), pp 105–142. San Diego: Academic Press.
  37. ↵
    1. McFadden D,
    2. Train K
    (2000) Mixed MNL models for discrete response. J Appl Econ 15:447–470. https://doi.org/10.1002/1099-1255(200009/10)15:5<447::AID-JAE570>3.0.CO;2-1
    OpenUrlCrossRef
  38. ↵
    1. Pasquereau B,
    2. Turner RS
    (2013) Limited encoding of effort by dopamine neurons in a cost–benefit trade-off task. J Neurosci 33:8288–8300. https://doi.org/10.1523/JNEUROSCI.4619-12.2013 pmid:23658169
    OpenUrlAbstract/FREE Full Text
  39. ↵
    1. Pastor-Bernier A,
    2. Schultz W
    (2021) Mechanisms underlying economic revealed preference of multi-component bundles in primate orbitofrontal cortex. bioRxiv 433629. https://doi.org/10.1101/2021.03.02.433629.
  40. ↵
    1. Pastor-Bernier A,
    2. Stasiak A,
    3. Schultz W
    (2021) Reward-specific satiety affects subjective value signals in orbitofrontal cortex during multicomponent economic choice. Proc Natl Acad Sci U S A 118:e2022650118. https://doi.org/10.1073/pnas.2022650118 pmid:34285071
    OpenUrlAbstract/FREE Full Text
  41. ↵
    1. Pelli DG
    (1997) The VideoToolbox software for visual psychophysics: transforming numbers into movies. Spat Vis 10:437–442. pmid:9176953
    OpenUrlCrossRefPubMed
  42. ↵
    1. Prévost C,
    2. Pessiglione M,
    3. Météreau E,
    4. Cléry-Melin M-L,
    5. Dreher J-C
    (2010) Separate valuation subsystems for delay and effort decision costs. J Neurosci 30:14080–14090. https://doi.org/10.1523/JNEUROSCI.2752-10.2010 pmid:20962229
    OpenUrlAbstract/FREE Full Text
  43. ↵
    1. Rigoli F,
    2. Pezzulo G
    (2022) A reference-based theory of motivation and effort allocation. Psychon Bull Rev 29:2070–2082. https://doi.org/10.3758/s13423-022-02135-8 pmid:35768658
    OpenUrlPubMed
  44. ↵
    1. Salamone JD,
    2. Cousins MS,
    3. Bucher S
    (1994) Anhedonia or anergia? Effects of haloperidol and nucleus accumbens dopamine depletion on instrumental response selection in a T-maze cost/benefit procedure. Behav Brain Res 65:221–229. https://doi.org/10.1016/0166-4328(94)90108-2 pmid:7718155
    OpenUrlCrossRefPubMed
  45. ↵
    1. Stauffer WR,
    2. Lak A,
    3. Yang A,
    4. Borel M,
    5. Paulsen O,
    6. Boyden ES,
    7. Schultz W
    (2016) Dopamine neuron-specific optogenetic stimulation in rhesus macaques. Cell 166:1564–1571.
    OpenUrlCrossRefPubMed
  46. ↵
    1. Tobler PN,
    2. Fiorillo CD,
    3. Schultz W
    (2005) Adaptive coding of reward value by dopamine neurons. Science 307:1642–1645. https://doi.org/10.1126/science.1105370 pmid:15761155
    OpenUrlAbstract/FREE Full Text
  47. ↵
    1. van Wolkenten M,
    2. Brosnan SF,
    3. de Waal FBM
    (2007) Inequity responses of monkeys modified by effort. Proc Natl Acad Sci U S A 104:18854–18859. https://doi.org/10.1073/pnas.0707182104 pmid:18000045
    OpenUrlAbstract/FREE Full Text
  48. ↵
    1. Varazzani C,
    2. San-Galli A,
    3. Gilardeau S,
    4. Bouret S
    (2015) Noradrenaline and dopamine neurons in the reward/effort trade-off: a direct electrophysiological comparison in behaving monkeys. J Neurosci 35:7866–7877. https://doi.org/10.1523/JNEUROSCI.0454-15.2015 pmid:25995472
    OpenUrlAbstract/FREE Full Text
  49. ↵
    1. Verguts T,
    2. Vassena E,
    3. Silvetti M
    (2015) Adaptive effort investment in cognitive and physical tasks: a neurocomputational model. Front Behav Neurosci 9:57. https://doi.org/10.3389/fnbeh.2015.00057 pmid:25805978
    OpenUrlCrossRefPubMed
  50. ↵
    1. Webb R
    (2019) The (neural) dynamics of stochastic choice. Management Sci 65:230–255. https://doi.org/10.1287/mnsc.2017.2931
    OpenUrl
  51. ↵
    1. Webb R,
    2. Levy I,
    3. Lazzaro SC,
    4. Rutledge RB,
    5. Glimcher PW
    (2019) Neural random utility: relating cardinal neural observables to stochastic choice behavior. J Neurosci Psychol Econ 12:45–72. https://doi.org/10.1037/npe0000101
    OpenUrl
  52. ↵
    1. Zipf GK
    (2016) Human behavior and the principle of least effort: an introduction to human ecology. Ravenio Books: Cambridge, MA.
Back to top

In this issue

The Journal of Neuroscience: 43 (40)
Journal of Neuroscience
Vol. 43, Issue 40
4 Oct 2023
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Masthead (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Worth the Work? Monkeys Discount Rewards by a Subjective Adapting Effort Cost
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Worth the Work? Monkeys Discount Rewards by a Subjective Adapting Effort Cost
Mark Burrell, Alexandre Pastor-Bernier, Wolfram Schultz
Journal of Neuroscience 4 October 2023, 43 (40) 6796-6806; DOI: 10.1523/JNEUROSCI.0115-23.2023

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Worth the Work? Monkeys Discount Rewards by a Subjective Adapting Effort Cost
Mark Burrell, Alexandre Pastor-Bernier, Wolfram Schultz
Journal of Neuroscience 4 October 2023, 43 (40) 6796-6806; DOI: 10.1523/JNEUROSCI.0115-23.2023
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • adaptation
  • choice
  • preference
  • utility
  • value

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Research Articles

  • Medial Frontal Theta Reduction Impairs Rule Switching via Prediction Error
  • Chemogenetic Inhibition and Optogenetic Depotentiation of the Prelimbic Cortex to Paraventricular Thalamus Pathway Attenuate Abstinence-Induced Plasticity and Heroin Seeking in Rats
  • Neural Tracking of the Maternal Voice in the Infant Brain
Show more Research Articles

Behavioral/Cognitive

  • Medial Frontal Theta Reduction Impairs Rule Switching via Prediction Error
  • Neural Tracking of the Maternal Voice in the Infant Brain
  • Switching between newly learned motor skills
Show more Behavioral/Cognitive
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.