Abstract
The orbitofrontal cortex (OFC) has been implicated in both the representation of “state,” in studies of reinforcement learning and decision making, and also in the representation of “schemas,” in studies of episodic memory. Both of these cognitive constructs require a similar inference about the underlying situation or “latent cause” that generates our observations at any given time. The statistically optimal solution to this inference problem is to use Bayes' rule to compute a posterior probability distribution over latent causes. To test whether such a posterior probability distribution is represented in the OFC, we tasked human participants with inferring a probability distribution over four possible latent causes, based on their observations. Using fMRI pattern similarity analyses, we found that BOLD activity in the OFC is best explained as representing the (log-transformed) posterior distribution over latent causes. Furthermore, this pattern explained OFC activity better than other task-relevant alternatives, such as the most probable latent cause, the most recent observation, or the uncertainty over latent causes.
SIGNIFICANCE STATEMENT Our world is governed by hidden (latent) causes that we cannot observe, but which generate the observations we see. A range of high-level cognitive processes require inference of a probability distribution (or “belief distribution”) over the possible latent causes that might be generating our current observations. This is true for reinforcement learning and decision making (where the latent cause comprises the true “state” of the task), and for episodic memory (where memories are believed to be organized by the inferred situation or “schema”). Using fMRI, we show that this belief distribution over latent causes is encoded in patterns of brain activity in the orbitofrontal cortex, an area that has been separately implicated in the representations of both states and schemas.
- Bayes' rule
- context
- posterior distribution
- schemas
- state representation
- ventromedial prefrontal cortex
Introduction
In recent years, cognitive neuroscientists studying reinforcement learning and decision making have recognized the importance of specifying representations of the environmental “state” that capture the structure of the world in a predictive way (Courville et al., 2004; Gershman and Niv, 2010). At the same time, there has been renewed interest among cognitive neuroscientists in how memory encoding and retrieval are shaped by situation-specific prior knowledge (“schemas”; Tse et al., 2007). As work in this area progresses, it is important to clarify exactly what constitutes a schema and how schemas are formed.
Whether inferring the current state or the currently relevant schema, agents are making inferences about the hidden variables that underlie and generate our observations in the world. This inference can be concretely formulated in terms of Bayesian latent cause models (e.g., Gershman et al., 2010). According to this framework, states and schemas can be viewed as hidden (latent) causes that give rise to observable events. For example, if you arrive late to a lecture, the situation (whether this is indeed the department colloquium or you have accidentally walked in on an undergraduate class) determines your observations about the average age of the audience, the proportion of audience members that are taking notes, the type of information being presented, and so on. To decide whether you are in the right place, you can use Bayesian inference to infer a belief distribution over the possible situations that might have generated the current observations, i.e., a posterior probability distribution over latent causes, p(latent cause | observations) (Fig. 1A).
We hypothesized, based on the similarity of the underlying computations, that the inference related to these two cognitive constructs (states and schemas) might be implemented using the same neural hardware. Indeed, there is one area of the brain that has separately been implicated in representing states (Wilson et al., 2014) and also schemas (Tse et al., 2011; Ranganath and Ritchey, 2012; van Kesteren et al., 2012; Ghosh and Gilboa, 2014; Richards et al., 2014; Schlichting and Preston, 2015)—the orbitofrontal cortex (OFC). Furthermore, previous univariate analyses in fMRI have implicated this region in encoding various summary statistical measures that are related to or are components of the posterior distribution, e.g., the posterior mean, likelihood of the current stimulus, and prior uncertainty (Vilares et al., 2012; d'Acremont et al., 2013; Ting et al., 2015). However, these studies have not investigated representations of a full probability distribution.
Here, we used fMRI to investigate representation in the OFC of posterior probability distributions over latent causes. In our experiment, we created a probabilistic environment in which participants were required to make inferences about the hidden causes that generated their observations. Participants viewed sequences of animal “photographs” from one of four “sectors” of a virtual “animal reserve.” Participants were asked to judge the probability that the photographs were taken in each of the sectors, based on their previous experience observing animals in each sector. Using multivariate pattern similarity analyses of fMRI activity, we found that BOLD activity in the OFC was better explained by the posterior distribution over sectors (latent causes) than by a wide range of related signals, including the current stimulus, the most probable sector (the maximum a posteriori latent cause), or the uncertainty over latent causes (operationalized as the entropy of the posterior distribution). The present result advances our understanding of the function of the OFC. It also unifies results from two different fields of cognitive neuroscience, inviting further investigation into the relationship between probabilistic inference, states, and schemas.
Materials and Methods
Participants
Thirty-two participants (age, 18–34 years; 22 female) from the Princeton University community participated in exchange for monetary compensation ($20 per hour plus up to $15 performance-related bonus). All participants were right-handed. Participants provided informed written consent. The study was approved by the Princeton University Institutional Review Board.
Experimental design
The safari.
Participants were told that they were going on a virtual safari—a visit to an animal reserve divided into four different sectors. Each sector was associated with a different color, background image, background music, and location on a map (randomized across participants).
There were five different kinds of animals in the animal reserve. Every animal appeared in every sector, but with different likelihoods P(animal | sector). The likelihoods (not shown directly to the participants) were chosen so that none of the sectors was strongly identified with a single animal, and so that none of the animals was strongly identified with a single sector (Fig. 1B; colors and animals were randomly assigned across participants).
Procedure overview.
The experiment consisted of two parts. In the first part, participants “toured” through the animal reserve to learn (through experience) the likelihoods P(animal | sector) for each animal and each sector. In the second part of the experiment, participants were shown sequences of drawings that represented “photographs” of animals, which were taken in an unknown sector. Given the animals shown in each sequence, participants were asked to infer the posterior probabilities of different sectors, P(sector | animals shown). For each participant, the experiment took place across 2 consecutive days (Table 1).
Tours task.
In the tours task (Fig. 1C), participants were instructed that they would “tour” through the animal reserve, one sector at a time, to learn the animal frequencies in each sector (the animal likelihoods). Each tour consisted of 10–40 trials through a single sector. One animal appeared on each trial, pseudorandomly chosen according to the likelihoods for that sector. Before each animal appeared, participants were shown a prompt, asking them to make a prediction about which of two animals (one correct and one randomly chosen) would appear next. The alternative (incorrect) option was chosen with uniform probability from the four other animals. To distinguish between the animals in the question prompt (which were not representative of the sector's likelihood distribution) and the animals that were actually drawn from the sector's likelihood distribution, the question prompts were shown as text while the animals drawn from the safari sector were shown as pictures.
In order for the sectors to form rich contexts, each sector was associated with a different color, background image, background music, and location on a two-by-two map (randomized across participants). Before the first trial of a tour through a sector, participants were shown the sector's location on the map. Also, for the duration of a tour through a sector, animals were displayed on the sector's color-matched backdrop image, and the music associated with that sector was played in the background.
Photographs task.
On each trial of the photographs task (Fig. 1D), participants were shown a sequence of animal “photographs,” without being told which sector the photographs were taken from. At the end of the sequence, participants were prompted to indicate which of two sectors (randomly chosen) was more (or less) probable. The two sector options for each question were chosen uniformly from the four sectors of the safari (and did not necessarily include the most or least likely sector). So, to perform well on the task, participants had to maintain a full posterior distribution over all four sectors (as opposed to estimating only the most probable sector, for instance).
Participants received 10 cents for every correctly answered question, and they received feedback on every trial. So that more probable sectors were not consistently associated with higher monetary value, we asked which of the two sectors was more probable on half of the trials, and which was less probable on the other half of the trials. To eliminate confounds with motor plan, the positions of the two response options were pseudorandomly assigned between left and right.
To encourage participants to update their inference of the sector probabilities after every animal presentation (as opposed to waiting until the time of the question to integrate over the animals observed), we varied the length of the sequences between one and six animals (so that the appearance of the question prompt was unpredictable), and participants were only allowed 2.5 s to give a response after the appearance of the question.
The posterior probability of each sector P(sector | animals seen) can be straightforwardly computed from the animal likelihoods using Bayes' rule (all sectors were equally likely a priori), as follows (Eq. 1): Feedback for the responses was generated based on these posterior probabilities. Due to a bug in the code that was undetected during data collection, the feedback was incorrectly generated for some of the trials containing only one animal presentation (this affected ∼10% of the trials). In our fMRI analyses, to account for learning from the incorrect feedback, we used each participant's estimates of the likelihoods (collected at the end of the experiment) instead of the real likelihoods, and we also performed trial-by-trial behavioral model fitting, in order to model learning from feedback (see next section).
To familiarize themselves with the photographs task, participants first performed two sessions (20 trials each) of the task outside the MR scanner. They then performed four sessions (30 trials, ∼11 min per session) inside the scanner.
Behavioral model fitting
To model participants' posterior inference on the photographs task, as well as any learning from feedback, we performed trial-by-trial model fitting of participants' responses. We tested several classes of behavioral models (note that we will explicitly refer to these as “behavioral models,” to distinguish them from the neural models that we later test against the neural data). These classes of models (Bayesian_nolearning, Additive, Most/least voter, and Bayesian_feedbackRL) are described in the following.
Bayesian_nolearning.
This behavioral model assumed that participants were correctly computing the posterior distribution over sectors P(sector | animals seen) using Bayesian inference (as in Eq. 1). To obtain the model-derived likelihood of each behavioral response (and to capture stochasticity in participants' behavior), we used a softmax on the posterior probabilities of the two options in each question prompt (Eq. 2): where β is an inverse temperature parameter (β = 0 implies equal likelihood for both options).
Additive.
In this behavioral model, rather than correctly multiplying the animal likelihoods together to obtain the posterior distribution over sectors (as in Eq. 1), we assumed that participants instead added the likelihoods together to obtain an “additive posterior” (normalized to sum to 1; Eq. 3): While statistically suboptimal, we might expect this from a simple associative mechanism that brings the sectors to mind in proportion to their association strength with the animals seen. Again, to determine response probabilities, we applied a softmax operator to the additive “posterior” probabilities for the two options in each question prompt.
Most/least voter.
These behavioral models assumed that participants were only paying attention to the most common (and/or least common) animals in each sector, a similar strategy having been previously observed in a similar task (Gluck et al., 2002). During the trials, each animal appearance “voted” for (or against) the sectors in which it was the most common (or least common). To obtain the model-derived likelihood of each behavioral response, we used a softmax on the final tally at the end of each sequence.
We tested several variants of this behavioral model, e.g., tallying only the positive votes, and/or allowing an animal to “vote” for (or against) a sector if it was one of the two most (or least) common animals in that sector. The magnitude of the positive and negative votes were either allowed to be two separate free parameters or constrained to be equal to each other. Because the magnitude of the vote already served as a scaling parameter for the input to the softmax operator, the inverse temperature of the softmax was kept constant at 1.
Bayesian_feedbackRL.
These behavioral models were designed to account for learning from feedback during the photographs task (including the incorrectly generated feedback). Here we assumed a reinforcement learning process, in which participants adjusted their internal estimates of the animal likelihoods after feedback about the two sectors in the question. These likelihoods were then used to compute the posterior distribution over sectors via Bayes' rule.
For the sector that feedback indicated to be more probable, likelihoods were adjusted upward for all animals seen on that trial. For the sector indicated to be less probable, likelihoods were adjusted downward for all animals seen on the trial (Fig. 2; Eq. 4): Estimates of the likelihoods were renormalized after each adjustment. The learning rates αpos and αneg were either allowed to be two separate free parameters or they were constrained to be equal.
For the initialization of the likelihoods, we tested two versions: initialization at the true animal likelihoods or initialization according to the participants' subjective estimates of the likelihoods (collected at the end of the experiment; see below).
Finally, the likelihoods were used to compute the posterior distribution over sectors via Bayes' rule. Thus, posterior inference in the FeedbackRL behavioral model also used Bayes' rule. The only difference from the “Bayesian_nolearning” behavioral model above is that the likelihoods (which enter into the posterior inference computation from Eq. 1) were adjusted on each trial according to feedback.
We tested several additional variants of this behavioral model. In one variant, participants only adjusted their likelihoods in response to “You are incorrect” feedback (instead of in response to all feedback). In another variant of the behavioral model, we scaled the learning rates separately for each animal according to how much that animal contributed to the final posterior distribution (Eq. 5): In this variant, animals appearing multiple times in a trial would have higher effective learning rates, having contributed more to the final decision.
For all behavioral models.
In a postexperiment questionnaire, we asked participants to provide their estimates for the animal likelihoods in each sector. For each of the behavioral models above, we tested versions using (1) the actual animal likelihoods and (2) subjective estimates of the animal likelihoods. For the few participants who provided likelihood estimates that did not sum to 1, we normalized the estimates. To avoid taking logarithms of 0, we converted estimated likelihoods of 0 into 0.01 (and renormalized).
For each of the behavioral models, we also tested versions in which the earlier and/or later animals in each sequence were given extra weight. To model these primacy/recency effects, we fit a power law function for each participant to give more weight to the earlier and/or later animals in each sequence (e.g., 1w, 2w, … for animal 1, animal 2, …). The likelihoods were exponentiated by this weighting and renormalized. If modeling both recency and primacy, the weightings for each were summed. We tested versions in which the recency and primacy free parameters w were either allowed to be two free parameters or they were constrained to be equal.
Free parameters for each behavioral model were fit to each participant's behavioral data separately, using Matlab's “fmincon” function, with ≥10 random initializations for each behavioral model and each participant. The best-fitting parameters (the maximum likelihood estimates) were used to evaluate, for each participant and each behavioral model, the (geometric) mean likelihood per trial (i.e., the exponentiated log likelihood per trial, without any penalization for number of parameters), the Akaike information criterion (AIC), and the Bayesian information criterion (BIC), to compare the behavioral models and determine which best accounted for participants' behavior.
fMRI acquisition and preprocessing
Functional brain images were acquired using a 3 T MRI scanner (Skyra, Siemens) and preprocessed using FSL [Functional MRI of the Brain (FMRIB) Software Library; http://fsl.fmrib.ox.ac.uk/fsl/]. An echoplanar imaging sequence was used to acquire 36 slices (3 mm thickness with 1 mm gap; TR = 2 s; TE = 27 ms; flip angle, 71°). To increase signal in the OFC, slices were angled ∼30° from the axial plane toward a coronal orientation (Deichmann et al., 2003). For each participant, there were four scanning runs in total (∼11 min each). The functional images were spatially filtered using a Gaussian kernel (full width at half maximum of 5 mm), and temporally filtered using a high-pass cutoff of 0.0077 Hz. We performed motion correction using a six-parameter rigid body transformation to coregister functional scans, and then registered the functional scans to an anatomical scan using a six-parameter affine transformation.
The motion regressors (and their derivatives) were residualized out from the functional images, as were the mean time courses for CSF and white matter [segmentation was performed using FSL's FAST (FSL Automated Segmentation Tool) function], and also the mean time course for blood vessels (estimated by taking voxels with the top 1% in SD across time). Then, the functional images were z-scored over time. All analyses were performed for each participant in participant space, and then spatially normalized by warping each participant's anatomical image to MNI space using a 12-parameter affine transformation.
Region of interest—suborbital sulcus
Our region of interest (ROI) was determined as the intersection of two sets of brain areas. The first set of areas, the OFC, has been postulated to be involved in the representation of state, due to evidence from studies of human and animal reinforcement learning and decision making (Wilson et al., 2014). The second set of areas, sometimes referred to as the “posterior medial network,” has been postulated to be involved in the computation and representation of schemas or context (Ranganath and Ritchey, 2012), as the set of areas with high connectivity with parahippocampal cortex (PHC). The intersection of these sets of areas is the suborbital sulcus, a medial subregion of the OFC (see Fig. 6A). Using Freesurfer (Destrieux et al., 2010), the ROI was drawn as the anatomically parcellated cortical region centered on the voxel with maximal resting-state functional connectivity to the PHC (Libby et al., 2012; the ROI comprised 105 voxels in MNI 3 mm space and 97.3 ± 2.6 voxels in subject space).
Representational similarity analysis
If the suborbital ROI contains a multivariate representation of the posterior distribution over latent causes, then patterns of neural activity in this area should be more similar for pairs of time points at which the posterior distribution was similar, and they should be dissimilar for pairs of time points at which the posterior distribution was dissimilar. Therefore, to test whether multivariate patterns of activity in the ROI might be representing the posterior distribution over sectors, we performed a representational similarity analysis (Kriegeskorte et al., 2008).
We first computed the similarity of the posterior distribution over sectors for every pair of time points during which we expected the posterior distribution to be updated (i.e., at the times of the animal appearances). This provided us with the similarity matrix for the posterior. We also computed the similarity of the neural pattern in the ROI for every pair of time points—the similarity matrix for the ROI. Then we computed the Spearman rank correlation of these two matrices (taking only the upper triangle and excluding the diagonal). We denote this Spearman correlation as the similarity match between the posterior and the ROI (Fig. 3). We expected the similarity match to be positive, i.e., that the neural patterns in the ROI should be more similar for pairs of time points at which the posterior distribution over sectors was more similar.
We also computed the similarity match for the ROI with other signals (henceforth called “neural models”), to compare with the similarity match between the ROI and the posterior distribution over latent causes. This is important because the similarity structure for the ROI could potentially be correlated with the similarity structure of the posterior distributions for reasons other than that the posterior distribution is represented in this area. For example, the posterior distribution is, on average, more similar for pairs of time points at which the same animal is presented; if the suborbital ROI represents the animal currently presented, we would also find a positive similarity match between the ROI and the posterior distribution. We therefore compared the similarity match between the ROI and each alternative neural model, to determine the neural model that best explained the similarity structure of the neural data.
The set of alternative neural models used for this comparison included the log-transformed posterior distribution (because many signals in the brain are known to be represented in log space; e.g., Gibbon, 1977; Yang and Shadlen, 2010; Longo and Lourenco, 2007), the current stimulus, the maximum a posteriori (MAP) sector (most probable sector), the entropy of the posterior distribution (a proxy for overall uncertainty), the probability of the MAP sector (a proxy for confidence, acting approximately as the converse of the entropy), the rank ordering of sectors in the posterior (since the task tests participants' ability to rank order the sectors), and temporal distance between measurements (because fMRI pattern similarity is known to vary as a function of the temporal distance between measurements). We also included neural models of the posterior and MAP that were instead derived using the Bayesian_feedbackRL behavioral model (given that this was the best behavioral model after Bayesian_nolearning, as determined from behavioral model fitting, described above). See Table 3 for a full list of neural models tested.
To investigate the specificity of the result to our ROI, we also performed a whole-brain “searchlight” analysis, using 25-voxel spherical searchlights. As with the ROI, we computed the similarity of the neural patterns in each searchlight to obtain the neural similarity matrix for the searchlight. We then computed the Spearman correlation of the similarity matrix for each searchlight with each of our neural models. The analysis was repeated for a searchlight centered on every voxel in the brain.
For both the ROI and searchlight analyses, the neural pattern for each animal appearance was averaged over the two TRs during which the animal appeared on the screen (after correcting for the hemodynamic lag with a 4 s shift). Similarity for neural patterns was computed using normalized correlation, to accord with the similarity measure used for the posterior-based neural models (similar results are obtained when using Pearson correlation instead). Searchlight results are displayed on an inflated brain, using the AFNI (Analysis of Functional NeuroImages) SUMA (Surface Mapping with AFNI) surface mapper (http://afni.nimh.nih.gov/afni/suma).
Statistics and confidence intervals
Unless stated otherwise, all statistics were computed using random-effects bootstrap distributions on the mean by resampling participants with replacement (Efron and Tibshirani, 1986). All confidence intervals in the text are given as SEM.
To test the reliability of searchlight results across participants, we used the “randomize” function in FSL (http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/randomize) to perform permutation tests and generate a null distribution of cluster masses for multiple-comparisons correction (using FSL's “threshold-free cluster enhancement,” p < 0.05 two tailed).
Results
Participants learned the animal likelihoods in the tours task
We evaluated participants' final learning of the likelihood of each animal in each sector using performance from the last set of tours on the last day. In those tours, the participants chose the more likely animal 73 ± 3% of the time. Note that even if participants had perfect knowledge of the animal likelihoods, we might not expect participants to choose the more likely animal 100% of the time, due to noise in the decision process or probability matching (the tendency to match choice probabilities to the probability of each option being correct; Vulkan, 2000; Erev and Barron, 2005).
In a postexperiment questionnaire, we asked participants to estimate the animal likelihoods P(animal | sector) for every animal and every sector. These estimates were close to the true likelihoods, on average (Fig. 4A). The mean Kullback–Leibler divergence of the estimated likelihoods from the real likelihoods was 0.13 ± 0.015. As discussed below, we used these participant-estimated likelihoods in our neural analyses, in lieu of the correct likelihoods.
Performance on photographs task suggested maintenance of posterior distributions over sectors
During the fMRI scan sessions, participants correctly chose the more (or less) probable sector 67 ± 1% of the time, which is significantly above chance (t(31) = 13.1, p < 1e-12). Moreover, logistic regression on participants' responses showed that, the larger the difference in posterior probability between the correct and incorrect options, the more likely participants were to choose the correct answer (Fig. 4B). Again, as in the tours task, we expected stochasticity in participants' behavior due to noise and probability matching.
Note that the two sector options in each question were chosen at random, and therefore required participants to discriminate between posterior probabilities for any possible pair of sectors. Interestingly, participants performed similarly well whether or not questions included the MAP (most probable) sector (accuracy 69 ± 2% for questions including the MAP; 66 ± 1% for questions not including the MAP; not significantly different). This result further indicates that participants were tracking the full posterior distribution, and not just the MAP sector.
Trial-by-trial behavioral model fitting suggested that participants were approximately Bayesian
The relative performance of the behavioral models is shown in Figure 5, and the mean parameter fits are shown in Table 2. For model comparison, we used the best-performing version from each class of behavioral models (Table 2).
The two Bayesian behavioral models (with and without feedbackRL) performed best, explaining the data about equally well. Overall, the model with feedbackRL was the best behavioral model according to AIC, but the Bayesian model without learning was the best behavioral model according to BIC, which penalizes more strongly for extra parameters.
The additive behavioral model performed worse than the Bayesian behavioral models, indicating that participants were accumulating evidence multiplicatively, in accordance with the optimal strategy (Eq. 1). None of the heuristic inference models that we tested (the “most–least voter” class of models) could successfully outperform the Bayesian behavioral models. Nor did we identify any significant effect of recency or primacy (any small improvements in the model likelihoods were not justified by the increased number of parameters). We therefore concluded that participants were Bayesian or near-Bayesian in their inference.
As shown in Figure 5, using the participants' subjective estimates of the animal likelihoods (from the postexperiment questionnaire) provided a better fit for all behavioral models, compared with using the real animal likelihoods. This may be surprising for the feedbackRL behavioral model, given that the participant estimates were elicited at the end of the experiment, but were used in the model to initialize estimates of the likelihoods. However, the low learning rates (Table 2, average fit learning rates; also, 19% of participants had fitted learning rates of 0) suggest that changes in the likelihoods throughout the experiment were small relative to the differences between the real and estimated likelihoods. The low learning rates also explain why the feedbackRL behavioral model fit the data similarly well to a Bayesian behavioral model that did not allow for changes of the likelihood during the task: the models are nested (identical for learning rates of zero) and similar for low learning rates.
Representational similarity analysis suggests that the suborbital sulcus contains a representation of the (log) posterior distribution over latent causes
Figure 6B shows the representational similarity match of the suborbital sulcus with each of the neural models, relative to the representational similarity match with the best neural model—the logposterior (absolute values of the representational similarity match are shown in Table 4). For all of the alternative neural models tested, 95% or more of our bootstrap samples showed better representational similarity match for the logposterior than for the alternative neural model.
Because the posterior distribution tends to be more similar for neighboring time points compared with more distant time points, and that might also be the case for neural patterns, we took special care to verify that the logposterior neural model was superior to the alternative (control) time neural model. This was indeed the case. Moreover, we found that the temporal model displayed a negative representational similarity match with the neural patterns, because BOLD patterns for neighboring time points tended to be anticorrelated. This result was not dependent on our linear model for temporal distances; because we used Spearman's rank correlation to compute representational similarity match, the negative similarity match result would be observed for any other model of temporal distance that falls off monotonically (e.g., an exponential model). Therefore, since the posterior distribution showed a positive similarity match while the temporal neural model showed a negative similarity match, we can conclude that any positive correlations between the similarity matrices for the posterior distribution and time cannot be responsible for the representational similarity result for the posterior distribution.
Searchlight results for the representational similarity analysis are shown in Figure 7. The OFC and ventromedial prefrontal cortex showed a significantly greater representational similarity match for the logposterior model compared with every other neural model (p < 0.05 corrected, for every comparison), except for the entropy, posterior ranking, and posterior models. It also showed a greater representational similarity match for the logposterior than entropy using a more liberal threshold of p < 0.05 uncorrected.
Discussion
Because the underlying structure of the world is often not directly observable, we must make inferences about the underlying situations or “latent causes” that generate our observations. The statistically optimal way to do this is to use Bayes' rule to infer the posterior distribution over latent causes. Based on previous studies implicating the OFC in the representation of the current context or situation (related to the idea of state in studies of reinforcement learning and decision making, and to the idea of schemas in studies of episodic memory), we hypothesized that the OFC might represent a posterior probability distribution over latent causes, computed using approximately Bayesian inference. To test this, we asked participants to make inferences about the probability of possible situations in an environment where the situation probabilistically generated their observations.
Using representational similarity analysis of fMRI activity during the inference task, we found that patterns of activity in the suborbital sulcus within the OFC were indeed best explained as representing a posterior distribution over latent causes. Searchlight analyses implicated the OFC more generally in this representation. Furthermore, participants' behavioral performance showed that they had access to a full posterior distribution over the latent causes for their choices; using trial-by-trial model fitting, we showed that participants' behavior was best explained as using Bayesian inference.
Our study provides evidence that the OFC represents a full posterior distribution over situations, as opposed to the best guess of the situation (the MAP) or other summary measures of the distribution, such as the overall uncertainty. We operationalized uncertainty as the entropy of the distribution; the highest entropy occurs when the distribution is completely flat (i.e., the participant is maximally uncertain about which latent cause generated the observations), and the lowest entropy occurs when the distribution is fully loaded on one latent cause (i.e., the participant is absolutely certain about which latent cause generated the observations). Our similarity analyses showed the entropy to have a widespread positive similarity match in many areas of the cortex, which we might expect because entropy should be correlated with the difficulty of the task, and so entropy might therefore be correlated with greater overall activity in many regions of the brain. Nonetheless, in >95% of our bootstrap samples, activity in the OFC was better explained by the posterior distribution than by the entropy. Furthermore, searchlight analyses showed the specificity of this result.
Our results, using multivariate analysis, build on previous fMRI studies that have used univariate analyses in the OFC to investigate a range of summary statistical quantities related to the posterior distribution, but which do not capture the full distribution. These studies have shown that univariate activation of the ventromedial prefrontal cortex (vmPFC; which includes or is similar to our ROI) is correlated with a variety of summary statistics, e.g., expected reward (Ting et al., 2015), reward uncertainty (Critchley et al., 2001; Tobler et al., 2007), variance of the prior distribution in a sensory task (Vilares et al., 2012), and marginal likelihood of the current stimulus (d'Acremont et al., 2013). Our experiment benefited from several key features: (1) multivariate neural analysis, (2) four different latent causes, and (3) dissociation of latent cause from both reward and motor plan. These features enabled us to identify orbitofrontal representation of a full posterior distribution over latent causes that was separate from value, and which also explained neural activity in the area better than any single summary statistic we tried. Our result may therefore explain why evidence for different summary statistics was found in different studies: these are all components of the full posterior distribution or correlates of it.
Our study also builds on previous work in the fields of decision making and episodic memory that has implicated the OFC in representations of the current situation or context. In decision making, a belief distribution over states enables one to optimally learn or compute a behavioral policy even when the state of the world is not directly observable (as in partially observable Markov decision processes; Kaelbling et al., 1998). The OFC has long been implicated in reinforcement learning and decision making in a wide range of settings; a recent review provides a unifying explanation for these results by postulating that the OFC represents inferred states in partially observable situations (Wilson et al., 2014). In theories of episodic memory, it is believed that we organize our memories according to an inferred schema that specifies the situation and stores previously learned relationships that a new memory can be incorporated into (Tse et al., 2007; Hupbach et al., 2008). These schemas seem to be represented or processed in the vmPFC (Ranganath and Ritchey, 2012; van Kesteren et al., 2012; for review, see Schlichting and Preston, 2015), an area of the brain that is similar to our ROI. For example, Tse et al. (2011) showed evidence that activation of rat mPFC is highest immediately after memory encoding that should involve incorporating new information into existing schemas. Ezzyat and Davachi (2011) showed that greater activation of vmPFC in humans during memory encoding is correlated with how strongly those memories are associated with other memories in the same “event,” which is consistent with the idea that the vmPFC is involved in schemas that are bound to memories. Our results confirm the involvement of the OFC in representations of the current situation, and additionally show that this representation in the OFC takes the form of a distribution over possible situations.
Finally, our work also builds on previous studies investigating neural circuits involved in the “weather prediction” task, very similar to ours, in which one of two “weather” outcomes is probabilistically predicted by sequences of cards. Knowlton et al. (1996) implicated the striatum in the learning of these probabilistic associations. In our task, participants learned the animal likelihoods outside the MR scanner, and thus we could not assess the brain areas involved in the learning phase. However, our results are compatible with those of Knowlton et al. insofar as the OFC may use associations learned by the striatum (in our experiment, the animal likelihoods) to make inferences when presented with new observations (in our experiment, the photographs task). More recently, Yang and Shadlen (2007) used the weather-prediction task to show representation of a decision variable in the parietal cortex that took the form of the log likelihood ratio between two options. In our experiment, we decorrelated the posterior probability from both decision variables and stimulus–reward associations, and we also investigated representations of the posterior probability over latent causes before the decision period. We conjecture that the OFC contains representations of the current state or situation in terms of a posterior distribution over the possible states, a representation that is likely used by downstream areas, e.g., parietal cortex, for decision making.
Previous work on the weather-prediction task also showed that most individuals used heuristic strategies in inferring the weather (Gluck et al., 2002). In our experiment, we explored several heuristic behavioral models of participants' inference, but were not able to find any that predicted participants' behavior better than the optimal Bayesian models. There are several reasons why our task may have discouraged the use of heuristics. First, the animal likelihoods in our experiment were designed to avoid one-to-one mappings between observations and latent causes. Second, the task environment had four possible latent causes (instead of two), and the task itself required rank ordering all four latent causes rather than just estimating the MAP, thus increasing complexity and leading to the inadequacy of simple heuristics. Finally, we provided participants with extensive training on the probabilistic relationships in the safari, so that heuristics may have been less necessary.
Neurally, the posterior distribution we found in the OFC was best modeled as being represented in log space. Representation in log space may be advantageous because addition can then replace the multiplicative operation required to accumulate evidence in nonlog space (e.g., across animal presentations, in our experiment); the ability of neurons to add is well characterized, while it is less clear to what extent neurons can multiply (Yuste and Tank, 1996; Peña and Konishi, 2001; Gabbiani et al., 2002). Indeed, neural representation in log space is common in many domains, e.g., decision variables (Yang and Shadlen, 2007), time (Gibbon, 1977), and numbers (Longo and Lourenco, 2007).
To summarize, we designed a task in which participants' observations were probabilistically generated by unobserved “situations” or “latent causes,” and found evidence that the OFC represents a probability distribution over possible latent causes. A representation of the log posterior distribution explained OFC activity better than alternatives such as the best guess of the current situation or overall uncertainty in the current situation. This finding was further supported by behavioral evidence that participants had access to the full probability distribution for decision making, and used Bayesian inference to compute the probability distribution. Our results may explain why previous studies of the OFC have found evidence for representation of various summary statistical quantities in the OFC (these are in fact components of the full posterior probability distribution). Our results may also unify findings from disparate literature on reinforcement learning and episodic memory, which separately implicate the OFC in representations of the current situation.
Footnotes
- Received February 26, 2016.
- Revision received April 26, 2016.
- Accepted May 9, 2016.
This work was supported by National Science Foundation/National Institutes of Health Collaborative Research in Computational Neuroscience Grant NSF IIS-1009542, National Institutes of Health Grant 2T32MH065214, and U.S. Army Research Office Grant W911NF1410101. We thank sciencewithme.com for the animal illustrations.
The authors declare no competing financial interests.
- Correspondence should be addressed to Stephanie Chan, Princeton Neuroscience Institute, Princeton University, Princeton, NJ 08544. s.chan{at}alum.mit.edu
- Copyright © 2016 the authors 0270-6474/16/367817-12$15.00/0