Abstract
Threat cues have been widely shown to elicit increased sensory and attentional neural processing. However, whether this enhanced recruitment leads to measurable behavioral improvements in perception is still in question. Here, we adjudicate between two opposing theories: that threat cues do or do not enhance perceptual sensitivity. We created threat stimuli by pairing one direction of motion in a random dot kinematogram with an aversive sound. While in the MRI scanner, 46 subjects (both men and women) completed a cued (threat/safe/neutral) perceptual decision-making task where they indicated the perceived motion direction of each moving dot stimulus. We found strong evidence that threat cues did not increase perceptual sensitivity compared with safe and neutral cues. This lack of improvement in perceptual decision-making ability occurred despite the threat cue resulting in widespread increases in frontoparietal BOLD activity, as well as increased connectivity between the right insula and the frontoparietal network. These results call into question the intuitive claim that expectation automatically enhances our perception of threat and highlight the role of the frontoparietal network in prioritizing the processing of threat-related environmental cues.
- classical/pavlovian conditioning
- expectation
- fMRI
- frontoparietal network
- perceptual decision-making
- threat perception
Significance Statement
Threatening information receives enhanced priority processing in the brain. Evidence of increased neural activity to threat has fostered the current view that such selective processing leads to a boost in perception, suggesting that motivationally relevant top-down effects can directly change what we see. In the real world, danger is often preceded by an environmental cue that predicts its imminent approach. Here we used an aversive conditioning paradigm to test whether threat cues can change subjects' ability to visually distinguish between threat and safe stimuli. Our results provide strong evidence for the lack of an effect of threat expectation on perceptual sensitivity, supporting the theory that perception is impenetrable by top-down cognitive influences despite robust neural attentional priority.
Introduction
Our survival depends on the ability to quickly and accurately identify threats in our environment. Consequently, our brain has evolved to be exceptionally sensitive to danger. Humans detect threat-related stimuli faster (Hansen and Hansen, 1988; Öhman et al., 2001; Horstmann, 2007), make quicker saccades to them (Bradley et al., 2000; Mogg et al., 2007; Gerdes et al., 2009), and look at them more often (Wieser et al., 2009) compared with neutral stimuli. Importantly, this prioritization of threat has been shown to occur not only in the frontoparietal association cortex (Mohanty et al., 2009; Petro et al., 2017) but even as early as the primary visual cortex (Pourtois et al., 2004; Stolarova et al., 2006; Keil et al., 2007), suggesting that it is built into the basic properties of our visual system.
However, while the perception of threat-related stimuli is clearly enhanced, much less is known about the effects of expecting a threat. In the real world, dangerous stimuli are often at least partially predictable, such as when we hear a rustling in the bushes before a wild animal attacks or the screeching of tires before a car collision. Therefore, one important factor in understanding how the brain processes threat involves examining the mechanisms of brain function when expecting danger.
Two competing theories exist about the effects of threat expectation. On the one hand, given that threat stimuli receive enhanced sensory processing, it is natural to assume that expecting a threat also enhances perceptual sensitivity (Sussman et al., 2017). On the other hand, perception is often conceptualized as a separate module that cannot be influenced directly by cognition (Firestone and Scholl, 2015), which would indicate that a mere expectation of threat should not enhance perceptual sensitivity.
Surprisingly, while these two hypotheses appear to coexist in the literature, very little work has been done to adjudicate between them. In the only paper on the subject, Sussman et al. (2017) compared perceptual sensitivity between conditions where subjects were instructed to detect fearful versus neutral faces. They found enhanced performance for the condition where subjects detected fearful faces and interpreted their results as evidence that threat expectation enhances perception. However, their design did not involve true expectation since neutral and fearful faces were equally likely in both conditions. Instead, it is possible that subjects made more mistakes in the potentially less natural condition of having to detect neutral stimuli (as opposed to the arguably more intuitive condition of detecting a fearful face). Therefore, it remains unknown whether perceptual sensitivity is enhanced in the presence of true expectation, defined as modifying the probability of subsequent stimuli (Summerfield and De Lange, 2014).
Here, we adjudicated between the two theories that threat expectation does or does not enhance perception. We employed a fear conditioning paradigm to associate neutral perceptual stimuli (dot motion of different directions) with the likely presence or absence of an aversive outcome (loud noise). We found that expecting a particular type of sensory stimulus strongly biased subjects' responses toward that stimulus. However, perceptual sensitivity remained constant regardless of whether subjects were led to expect the threat stimulus or the safe stimulus or were given no expectation at all. Similarly, expecting a threat stimulus had no effect on processing in the visual cortex. Instead, threat cues strongly modulated both the activity and the connectivity profile in frontoparietal and salience networks, suggesting that threat expectation has strong postperceptual effects but no direct influence on perception itself.
Materials and Methods
Subjects
Forty-nine adult subjects were recruited for the study. We excluded three subjects from the analysis due to chance performance (accuracy, <55%), extreme startle reflex, or forgetting to wear contact lenses, respectively. Therefore, a total of 46 subjects (22 females; mean age, 21.7 years; SD, 4.5 years) were included in the analysis. The sample size was determined based on similar fMRI studies on visual perception (Sussman et al., 2017; Di Luzio et al., 2022). Subjects were compensated $20/h or 1 course credit/hour for a total of 2.5 h. All subjects were right-handed with normal hearing, normal or corrected-to-normal vision, and no history of neurological disorders, brain trauma, psychiatric illness, or illicit drug use. The study was approved by the Georgia Tech Institutional Review Board. All subjects were screened for MRI safety and provided written informed consent.
Stimuli and task
Subjects judged the motion direction (left or right) of white dots (density, 2 dots/degree2; speed, 5°/s) presented in a black circle (3° radius) in front of a gray background (Fig. 1A). A proportion of dots moved coherently in the right or left direction, while the rest of the dots moved randomly. Each dot had a lifetime between three and five frames (refresh rate, 60 Hz), and the coherent motion was carried out by a random subset of dots on each frame. Each dot motion stimulus was preceded by a letter cue (“L,”left; “R,” right; “N,” neutral). The letters L and R predicted the forthcoming stimulus with 75% validity, whereas the letter N was not predictive (both left and right motion were equally likely). We explicitly informed subjects of these contingencies in the instructions. Each trial began with cue presentation for 2, 4, or 6 s (chosen randomly), followed by a 3 s dot motion stimulus and an untimed response. A screen with a fixation dot was then presented between trials for a period of 1 or 2 s. All stimuli were created in MATLAB using Psychtoolbox-3 (Brainard, 1997; Pelli, 1997; Kleiner et al., 2007).
Experimental details. A, The main perceptual task performed in the scanner. Each trial began with a predictive cue (L, left; R, right; N, neutral) that indicated the likely direction of motion. A random dot kinematogram stimulus was then presented and subjects indicated the direction of motion. B, The aversive conditioning paradigm. Subjects learned that one direction (left or right, counterbalanced across subjects) was paired with a loud aversive sound delivered through headphones. C, Depiction of the sequence of events and respective locations of each phase of the experiment.
We used an aversive conditioning paradigm to assign threat (CS+) and safe (CS−) conditions to the otherwise neutral right/left moving dot stimuli (Fig. 1B). The aversive sound consisted of a female scream combined with high-pitched microphone feedback and was delivered through headphones for 1 s (at 90 dB in the lab and 100 dB in the scanner). The dot motion direction (right/left) that represented the aversive stimulus was counterbalanced across participants. The aversive sound was presented at the end of the dot motion stimulus such that they ended at the same time. In order to determine whether aversive conditioning was successful, subjects were given a postexperiment questionnaire where they rated the aversive sound [unconditioned stimulus (US)] on valence and arousal (1−5 scale from least to most unpleasant/arousing) using the Self-Assessment Manikin (SAM) scale (Bradley and Lang, 1994). Additionally, we verified that conditioning was successful by examining whole-brain activity in known fear conditioning regions [insula, dorsolateral prefrontal cortex (DLPFC), and middle frontal gyrus] for CS+ valid versus CS− valid trials (Taschereau-Dumouchel et al., 2020).
Experimental design
We split the experiment into three phases that were carried out in the lab, mock scanner, and MRI scanner (Fig. 1C).
Practice and thresholding outside the scanner
Subjects were first introduced to the task through a 62-trial training outside the MRI scanner in which they learned to perform the task at progressively lower coherence levels. These practice coherence levels were the same for all subjects (50, 30, 20, 12, 8, 7, and 6%), starting with a higher coherence of dot motion that gradually decreased, making the task progressively more difficult. Subjects were then familiarized with the three letter cues that indicated whether the motion direction of the upcoming stimulus was most likely to be left (“L”), right (“R”), or neutral (“N,” which gave no information about direction). All subjects received feedback on their accuracy as they went through the practice session. They then completed a two-up-one-down staircase with a step size of 2% that determined their individual threshold. The staircase was run until there were 10 reversals, and the final coherence level was the average of the last six reversals. This procedure results in an overall accuracy of ∼71% accuracy.
Conditioning inside the mock scanner
After being familiarized with the task, subjects underwent aversive conditioning in the mock scanner that was as similar as possible to the MRI scanner in which they would complete the main task. The conditioning phase consisted of 16 CS+ and 16 CS− interleaved trials (3 s each; ISI = 1, 2, or 3 s), where 75% of the CS+ stimulus trials coterminated with a loud aversive sound (i.e., 75% reinforcement rate). We used a high coherence level (30%) during conditioning to establish a clear association with threat and safe directions. Following conditioning, subjects completed a 24-trial practice session (8 trials of each cue) that included the aversive sound. During this practice, 50% of the valid CS+ trials coterminated with the aversive sound. The CS− and neutral cues were never paired with the aversive sound throughout the experiment.
Thresholding, reconditioning, and main task inside the scanner
Following the practice in the mock scanner, we prepared subjects for the MRI scanner. In the scanner, they completed the same thresholding task to ensure there were no substantial changes due to the new environment. Despite our best efforts to match them, the lab and scanner environments were slightly different, which can lead to slightly different thresholds in some subjects. If there was a difference in lab and scanner thresholds, we used the higher threshold. While it is possible that the conditioning itself can affect the thresholds, we were unable to directly test this because we only collected the threshold twice: in the lab (before conditioning) and in the scanner (after conditioning). Therefore, we cannot determine whether there was an effect of conditioning independent of the change in the environment. The motion coherence level (mean = 8.36%; SD = 4.09%) remained constant for all trials of the main task. Before beginning the main task, subjects completed a shorter version of aversive conditioning (24 trials total) to further strengthen the threat/safety association in the scanner environment, as well as a short practice run.
The main experiment included five runs each consisting of four blocks of 15 trials (60 trials/run; 300 trials total). Subjects were given a 12 s break between blocks and could take longer breaks between runs. In addition to the 60 experimental trials in each run (where no aversive sound was presented), we interleaved eight extra CS+ valid trials in which the aversive sound was played. We added these extra trials to ensure the conditioned response was not extinguished. These eight trials were removed from the main analyses.
At the end of each run, subjects reported their level of anxiety via button press (scale of 1–4). They also completed the trait version of the State–Trait Anxiety Inventory at the very beginning of the experiment immediately after the consenting process (Spielberger et al., 1983). These two anxiety measures are outside the scope of the current paper and will be communicated in a separate report.
fMRI acquisition and preprocessing
We collected the BOLD signal data on a 3 T MRI system (Prisma-Fit MRI system; Siemens) using a 32-channel head coil. We acquired anatomical images using T1-weighted sequences (MEMPRAGE sequence, FoV = 256 mm; TR = 2,530 ms; TE = 1.69 ms; 176 slices; flip angle = 7°; voxel size = 1.0 × 1.0 × 1.0). We acquired functional images using T2*-weighted gradient echoplanar imaging sequences (FoV = 220 mm; slice thickness = 2.5, TR = 1,200 ms; TE = 30 ms; 51 slices; flip angle = 65°; voxel size = 2.5 × 2.5 × 2.5; multiband factor = 3; interleaved slices).
We used SPM12 (Wellcome Department of Imaging Neuroscience) to analyze the MRI data. Functional images were first converted from DICOM to NIFTI and then preprocessed with the following steps: despiking, slice-timing correction, realignment, segmentation, coregistration, and normalization. We used the default settings in SPM12 for all of these steps. Finally, the functional images were smoothed with a 6 mm full-width at half-maximum (FWHM) Gaussian kernel.
Behavioral statistical analyses
We computed signal detection theoretic parameters for perceptual sensitivity and response bias (Green and Swets, 1966; Macmillan and Creelman, 2005). We used stimulus sensitivity (d′) to determine subjects’ ability to distinguish between left and right motion directions (i.e., CS+ and CS−). We also calculated response criterion (c) to determine the degree of response bias subjects had for one stimulus type over the other (CS+/CS−). To calculate these measures, we used the standard SDT formulas based on the observed hit rate (HR) and false alarm rate (FAR) as follows:
fMRI statistical analyses
We analyzed BOLD signal activity using a general linear model (GLM) in SPM12 (Wellcome Department of Imaging Neuroscience). We modeled regressors of interest separately for cue (L, R, N) and stimulus (CS+, CS−), obtained by convolving the unit impulse time series for each condition with the canonical hemodynamic response function. The regressor for each trial included the whole period between the onset of the cue and the offset of the moving dot stimulus. Twelve nuisance regressors related to head motion were included: three regressors related to translation and three regressors related to rotation of the head, as well as their derivatives (Lund et al., 2005). Group-level analyses compared BOLD signal activity when combining the cue and stimulus time periods, as well as when considering the cue and stimulus periods separately. While the main analysis included both valid and invalid trials, an additional analysis was used to isolate the cuing effect by examining the following contrast: (CS+ cue and CS+ stim trials − neutral cue and CS+ stim trials) − (CS− cue and CS− stim trials − neutral cue and CS− stim trials). Lastly, we defined the left and right motion areas (MT+) as regions of interest (ROIs) using the MarsBaR toolbox in SPM based on the contrast stimulus > fixation (Brett et al., 2002). The ROIs were defined as 5 mm spheres centered around peak activity MNI coordinates in right and left MT+ separately (right MT, 46, −68, 0; left MT, −44, −72, 2). β-Values were then extracted and compared across cues using an ANOVA.
We further examined whether the functional connectivity between different brain regions was dependent on the threat expectation in a trial (CS+ cue vs CS− and neutral cue trials). We used a generalized psychophysiological interaction (gPPI) analysis to determine whether changes in the time course of whole-brain BOLD activity matches that of a specified seed region in the brain. We conducted the gPPI analysis in the CONN 7 toolbox (Whitfield-Gabrieli and Nieto-Castanon, 2012) using the right insular cortex (coordinates, [47, 14, 0])—a critical site in the salience network for threat processing (Fullana et al., 2016)—as a seed region.
Data and code accessibility
All data and codes for the behavioral analyses are freely available at https://osf.io/79x42/. Unthresholded fMRI maps have been uploaded to NeuroVault60 and can be accessed at http://neurovault.org/images/798801/.
Results
We adjudicated between the hypotheses that threat expectations do or do not enhance perceptual sensitivity. Subjects completed a discrimination task where one of the two directions of motion was coupled with an aversive stimulus (a loud sound). On each trial, cues indicated whether the threat (CS+) or safe (CS−) stimulus was more likely or whether both stimuli were equally likely. We compared both the behavior and brain activity in the presence of threat expectation (i.e., for trials with a CS+ cue) compared with its absence (i.e., trials with CS− or neutral cues).
Manipulation check for the conditioned response
To determine that the aversive conditioning was successful, we first examined whole-brain activity in CS+ valid compared with CS− valid trials. Evidence of a strong conditioning effect was found based on increased activation in the insula (MNI, −38, 16, 4), DLPFC (MNI, 30, 52, 24), and posterior middle frontal gyrus (pMFG; MNI, 46, 8, 48). The insula region is known to be associated with physiological reactivity as measured by skin conductance response to threatening stimuli (Taschereau-Dumouchel et al., 2020). Similarly, the DLPFC and posterior middle frontal gyrus have been shown to be associated with subjective fear states in the presence of threatening stimuli (Taschereau-Dumouchel et al., 2020). The increased activation in these known fear conditioning regions for the CS+ valid compared with CS− valid trials provides evidence that our conditioning paradigm was successful. Additionally, in a postexperiment questionnaire, subjects were also asked to rate the aversive sound using the SAM scale on valence and arousal (1–5 scale from least to most unpleasant/arousing; Bradley and Lang, 1994). The average valence was 3.98, and the average arousal was 3.53, suggesting that the sound was perceived as aversive and therefore successfully functioned as an US.
Effect of CS± and CS− cues on response bias
As an initial manipulation check, we examined whether subjects took the predictive cues into account by exhibiting a response bias toward choosing the expected stimulus. We quantified response bias using the signal detection theoretic measure for response criterion, c, where negative c values indicate a bias toward the CS− stimulus (see Materials and Methods). We found that the CS+ cue led to a significant bias toward choosing the CS+ stimulus (average c = 0.30; t(45) = 4.0; p =
Behavioral effects of threat expectation. A, Both the CS+ and CS− cues produced a significant response bias toward the corresponding stimulus. Negative c values here indicate a bias toward CS+ responses. B, Similar criterion shift was observed for CS+ and CS− cues (compared with neutral cues). C, Perceptual sensitivity (d′) was similar for CS+ and CS− cues. Sensitivity was slightly higher for neutral cues presumably due to the noise associated with shifting a criterion from its default value. D, RT was similar across all three types of cues. Error bars show SEM; dots indicate individual subjects.
Importantly, we compared the strength of the bias induced by the CS+ and CS− cues. If expecting a threat (but not a safe) cue leads to altered sensory processing, then one may predict that, compared with neutral cues, there would be a larger response bias shift for the CS+ than the CS− cues. To check for such an effect, we compared the deviation scores
Threat expectation does not enhance perceptual sensitivity
Having established that subjects made equivalent adjustments to their response criterion for CS+ and CS− cues, we turned to our main goal of adjudicating between the two hypotheses regarding the influence of threat expectation on perceptual sensitivity (i.e., that threat expectation does or does not improve perceptual sensitivity). To adjudicate between the two hypotheses, we compared the perceptual sensitivity (d′) for each of the CS+, CS−, and neutral cue trials. We found that subjects showed no change in perceptual sensitivity based on cue type (CS+, CS−, or neutral; F(2,135) = 1.03; p = 0.36;
Threat expectation does not enhance sensory processing
The behavioral results above strongly suggest that threat expectation does not result in enhanced perceptual discrimination but do not clarify whether threat expectation leads to heightened neural responses in sensory areas. To address this question, we examined the neural activations in the motion-sensitive area MT+ as a function of cue type. We defined the left and right MT+ ROIs (see Materials and Methods) and compared the activity in each of the two ROIs for trials with CS+, CS−, and neutral cues. We found no significant difference in the strength of activation as a function of cue type in either the left MT+ (F(2,135) = 0.06; p = 0.94;
MT+ activations. BOLD signal activations for each cue type (CS+, neutral, CS−) in both the left and right MT+. The three cue types produced equivalent activation in both motion-sensitive areas. n.s., not significant.
Threat cues enhance activity in the frontoparietal and salience networks
Having established that expecting a threat stimulus has no effect on perception or early visual processing, we examined the effects of threat expectation on neural activity in the rest of the brain. Specifically, we performed a whole-brain analysis contrasting CS+ cue trials to CS− and neutral cue trials, using a contrast encompassing both the cue and stimulus periods to observe the entire timespan where expectation processing might occur. This allowed us to examine the influence of threat expectancy-related brain activation at a comprehensive level. We found that threat expectation led to increased activity in regions of both the frontoparietal network [including the right frontal eye field (FEF), DLPFC, bilateral inferior parietal lobule (IPL), bilateral midcingulate cortex, dorsal precuneus, and cerebellum] and the salience network [including the bilateral anterior insula and the bilateral dorsal anterior cingulate cortex (dACC)] (Fig. 4A, Table 1). These results are consistent with the notion that threat cues produce a strong affective response and may indicate high-level prioritizing of threat-related environmental cues (Fullana et al., 2016). On the other hand, the opposite contrast (CS− and neutral cues >CS+ cues) produced strong activations in the somatomotor cortex (Fig. 4B, Table 2), possibly indicating a freezing response for threat cues but not for safe or neutral cues. For completeness, we also conducted the same analyses for the cue and stimulus periods separately, finding similar results in each time period (Extended Data Figs. 4-1 and 4-2).
Whole-brain comparison between activation produced by CS+ versus CS− and neutral cues. A, Brain activations obtained from the contrast CS+ cue trials >CS− and neutral cue trials. (B) Brain activations obtained from the contrast CS− and neutral cue trials > CS+ cue trials. The maps were thresholded at p < 0.001 for display purposes. Colors indicate t-values. Extended Data Figures 4-1 and 4-2 represent analyses on the cue-only and stimulus-only period, respectively. aPFC, anterior prefrontal cortex; dACC, dorsal anterior cingulate cortex; DLPFC, dorsolateral prefrontal cortex; dmPFC, dorsomedial prefrontal cortex; FEF, frontal eye fields; FFG, fusiform gyrus; IFG, iinferior frontal gyrus; ITG, inferior temporal gyrus; MCC, midcingulate cortex; MTG, middle temporal gyrus; PPC, posterior parietal cortex; pMFG, posterior middle frontal gyrus; SPL, superior parietal lobule.
Figure 4-1
Whole brain comparison between activation produced by CS + cue vs. CS- cue for the cue period only. (A) Brain activations obtained from the contrast CS + cue > CS- cue. (B) Brain activations obtained from the contrast CS- cue > CS + cue. Activation is very similar to CS + cue trials > CS- & Neutral cue trials and vice versa. Heat map indicates t-values at p > .001. ACC, Anterior cingulate cortex; aPFC, Anterior prefrontal cortex; dACC, Dorsal anterior cingulate cortex; DLPFC, Dorsolateral prefrontal cortex; FFG, Fusiform gyrus; IFG, Inferior frontal gyrus; ITG, Inferior temporal gyrus; MCC, mid-cingulate cortex; MTG, Middle temporal gyrus; pMFG Posterior middle frontal gyrus; PPC, Posterior Parietal Cortex. Download Figure 4-1, EPS file.
Figure 4-2
Whole brain comparison between activation produced by CS + cue vs. CS- cue for the cue period only. (A) Brain activations obtained from the contrast CS + cue > CS- cue. (B) Brain activations obtained from the contrast CS- cue > CS + cue. Activation is very similar to CS + cue trials > CS- & Neutral cue trials and vice versa. Heat map indicates t-values at p > .001. aPFC, Anterior prefrontal cortex; dACC, Dorsal anterior cingulate cortex; DLPFC, Dorsolateral prefrontal cortex; dmPFC, Dorsomedial prefrontal cortex; FEF, Frontal eye-fields; IFG, Inferior frontal gyrus; ITG, Inferior temporal gyrus; MCC, mid-cingulate cortex; MTG, Middle temporal gyrus; Posterior inferior temporal gyrus; pMFG, Posterior middle frontal gyrus; PPC, Posterior Parietal Cortex; SPL, Superior parietal lobule. Download Figure 4-2, EPS file.
Whole-brain BOLD signal activation for CS+ cue trials versus CS− and neutral cue trials
Whole-brain BOLD signal activation for CS− and neutral cue trials versus CS+ cue trials
While the previous contrasts provided a comprehensive examination of threat expectancy-related brain activity, further isolation of the cuing effect can help control for any possible stimulus confounds of motion direction. We therefore conducted an analysis that isolated the cuing effect by subtracting trials in which the neutral cue was followed by the CS+ stimulus from trials in which the CS+ cue was followed by the CS+ stimulus. This was compared with an equivalent subtraction based on the CS− trials. Altogether, the contrast we used was (CS+ cue and CS+ stim trials − neutral cue and CS+ stim trials) − (CS− cue and CS− stim trials − neutral cue and CS− stim trials). The resulting brain activation differences between the CS+ cuing effect and the CS− cuing effect were similar to previous analyses, albeit with smaller cluster sizes. Namely, we found increased activity in the frontoparietal network (pMFG, bilateral DLPFC, bilateral IPL, bilateral midcingulate cortex, dorsal precuneus, and cerebellum) and the salience network (bilateral anterior insula and bilateral dACC; Fig. 5 and Extended Data Tables 5-1 and 5-2). As with the results from Figure 5, the results here are consistent with the prioritization of threat-related environmental cues.
Effects of cuing controlling for stimulus confounds. Whole-brain comparison between activation produced by the CS+ cuing effect and the CS− cuing effect, specifically (CS+ cue and CS+ stim trials − neutral cue ad CS+ stim trials) − (CS− cue and CS− stim trials − neutral cue and CS− stim trials). A, Brain activations obtained from the contrast CS+ cuing effect > CS− cuing effect. B, Brain activations obtained from the contrast CS− cuing effect > CS+ cuing effect. Activation is very similar to CS+ cue trials > CS− and neutral cue trials and vice versa. Heat map indicates t-values at p < 0.001. Tables 5-1 and 5-2 show activated clusters for the cuing effect. aPFC, anterior prefrontal cortex; DLPFC, dorsolateral prefrontal cortex; FFG, fusiform gyrus; IFG, inferior frontal gyrus; ITG, inferior temporal gyrus; MTG, middle temporal gyrus; PPC, posterior parietal cortex; pMFG, posterior middle frontal gyrus; SPL, superior parietal lobule.
Table 5-1
Whole brain BOLD-signal activation for CS cuing effect. Activated clusters for the contrast CS + cuing effect > CS- cuing effect. Download Table 5-1, DOCX file.
Table 5-2
Whole brain BOLD-signal activation for CS cuing effect. Activated clusters for the contrast CS- cuing effect > CS + cuing effect. Download Table 5-2, DOCX file.
The interpretation that the increased activity in the frontoparietal and salience networks elicited by the CS+ cues is due to the prioritization of threat-related environmental stimuli leads to the prediction that CS+ cues should also increase the communication between the two networks. Specifically, one may expect increased effective connectivity between areas of the frontoparietal and salience networks in the presence of CS+ cues. To check for such an effect, we compared the strength of the connectivity between the insular cortex (an area known to be highly reactive to threat stimuli, Fullana et al., 2016) and the rest of the brain for CS+ versus CS− and neutral cues. We found that CS+ cues increased the connectivity between the right insula and both the anterior cingulate gyrus and the right angular gyrus (Fig. 6, Table 3). Further, consistent with the possibility of a freezing response for CS+ cues, we also found decreased connectivity between the right insula and motor areas including the somatomotor cortex and the left putamen. Overall, these results demonstrate that threat expectation had a substantial effect on neural activity in higher levels of processing despite the fact that it did not affect either perceptual sensitivity or early sensory activity.
Connectivity differences between CS+ and CS−/neutral cues. We examined the connectivity differences between the right insular cortex and the rest of the brain. We found that the right insular cortex exhibited increased connectivity with the posterior parietal cortex (PPC) and the pregenual anterior cingulate cortex (pgACC) during trials with CS+ cues. On the other hand, the right insular cortex showed decreased connectivity with the somatomotor cortex and the putamen during trials with CS+ cues.
Functional connectivity changes between the right insula and the rest of the brain
Discussion
Higher-level cognitive states, such as those generated by threatening stimuli, are often thought to penetrate perception and directly change what we see (Pourtois et al., 2013). This view that top-down effects modulate perception and, as a result, change the physical percept is common in studies of threat, emotion, motivation, and beliefs (Firestone and Scholl, 2015). Here, we provide evidence that threat information does not enhance nor hinder visual perception. This lack of modulation occurs even though the threat status of an oncoming stimulus is preceded by predictive cues that should, in theory, provide sufficient opportunity for top-down effects to exert their influence. Neurally, we found a lack of modulation of early sensory processing in the MT+ regardless of whether a threat stimulus was expected or not, despite the presence of widespread activations related to the attentional priority that was greater for threat-related compared with safe and neutral cues. Taken together, our findings highlight the prioritized neural processing of threat that is consistent with the literature while showing that such enhanced recruitment does not modulate the early visual areas or change the nature of our visual percepts.
In the real world, dangerous stimuli are often preceded by environmental signals that predict the oncoming threat. Given that many studies have demonstrated enhanced processing and superior performance for threat stimuli (Hansen and Hansen, 1988; Bradley et al., 2000; Öhman et al., 2001; Horstmann, 2007; Gerdes et al., 2009; Stegmann et al., 2020), it appears plausible that the mere expectation of threat stimuli should also lead to better performance. Here, we tested this hypothesis using a well-established paradigm where a cue provided probabilistic information regarding the likely identity of an oncoming stimulus (Rahnev et al., 2011; de Lange et al., 2013; Kok et al., 2017; Hu and Rahnev, 2019). We found strong evidence that expecting a threat versus a safe stimulus had no effect on behavioral performance, as measured both by perceptual sensitivity (d′) and RT. Our behavioral results thus indicate that the expectation of threat may not enhance perception.
At first glance, our results seem contrary to a previous study that reported improved perceptual sensitivity to threat compared with neutral cues (Sussman et al., 2017). However, a closer examination of the two studies shows that the two findings are not contradictory to each other. In Sussman et al.’s (2017) study, subjects discriminated between degraded fearful and neutral faces. At the beginning of each trial, subjects saw a cue that specified the mapping between stimuli and responses. Specifically, the letter F meant that subjects should indicate if the face was fearful or not, such that a “yes” response indicated a fearful face, whereas the letter N meant that subjects should indicate if the face was neutral or not, such that a “yes” response indicated a neutral face. Critically, the cues were not predictive of the actual probabilities of fearful/neutral face presentations (which was always set to 50/50). As such, the cues established a change in stimulus–response mapping by using two different response patterns (fearful faces mapping to “yes” for an F cue but “no” for an N cue) rather than a change in the expectation of the oncoming stimuli. Sussman et al. (2017) found better performance when subjects were detecting fearful compared with neutral faces. However, given that the cues were not predictive (e.g., the letter F was not followed by a fearful face any more than by a neutral face), these findings show the effects of changing the stimulus–response mapping rather than expectation. Specifically, Sussman et al.’s (2017) results show that the task set of detecting fearful faces among neutral ones produces better performance, perhaps because it is more intuitive and thus less error-prone, compared with a task set of detecting neutral faces among fearful ones. Our study provided a true expectation cue, such that the CS+ and CS− cues were 75% predictive of the upcoming stimulus and did not use different task sets (i.e., unlike in Sussman et al. (2017), we had a constant stimulus–response mapping between stimuli and responses for the whole experiment). Therefore, Sussman et al.’s (2017) results do not contradict our findings here that true stimulus expectation (defined as probabilistic information about the forthcoming stimulus; Summerfield and De Lange, 2014) does not improve perceptual sensitivity.
Our behavioral results were mirrored by a lack of modulation in early sensory areas responsible for stimulus processing. This was true not only in a whole-brain analysis but also when we defined the MT+ (the sensory area specialized for processing motion stimuli) as an ROI and tested for differences between the CS+ and CS− cues. Not only did we fail to find a significant difference between the cues, but also Bayes factor analyses showed very strong evidence for the null hypothesis of no difference for both the left and right MT+. These results are especially striking in the context of previous studies that have shown clear modulations in sensory areas to the threat stimuli themselves (Keil et al., 2007; McTeague et al., 2015). These previous studies demonstrate that conditioning paradigms like the current one can lead to significant changes in early sensory processing. Since early visual areas are capable of threat/safe differentiation (Keil et al., 2007), their lack of activation in our paradigm provides evidence that perceptual sensitivity remains unchanged by the expectation of threat. Our results are in line with established psychophysics findings related to the well-known oblique effect, where vertical and horizontal line orientations are detected faster than oblique orientations, likely due to their prevalence in nature (Appelle, 1972). Despite the consistent increase in perceptual sensitivity for cardinal versus oblique orientations, when high-probability predictive cues are introduced, there is no enhancement in performance based on stimulus orientation (Stein and Peelen, 2015). Similarly, our data show that even though the visual system is more sensitive to threat stimuli, expecting such stimuli leads to no additional enhancement of perceptual sensitivity.
Critically, the lack of behavioral improvement or changes in sensory areas was coupled with strongly increased activity in the frontoparietal and salience networks. These results complement previous findings that threat stimuli themselves tend to elicit strong activations in these networks (Mohanty et al., 2009; Petro et al., 2017). In our case, these activations are likely to reflect heightened priority processing that is cognitive rather than perceptual in nature.
Overall, we adjudicated between two theories that threat expectation does or does not change perception. We found clear support for the theory that threat expectation does not alter perception, even though it leads to widespread activation in the frontoparietal association cortex. Our results suggest that expectation does not automatically enhance the perception of threat and highlight the role of the frontoparietal network in prioritizing the processing of threat-related environmental cues.
Footnotes
We thank Ashna Bhardwaj for her help with data collection. This work was supported by the National Institutes of Health (Grant R01MH119189) and the Office of Naval Research (Grant N00014-20-1-2622).
The authors declare no competing financial interests.
- Correspondence should be addressed to Nadia Haddara at nadiahaddara{at}gmail.com.