Abstract
Humans with amygdalar lesions show proportional reductions of the emotional response to facial expressions in the fusiform face area as well as deficits in emotion recognition from facial expressions. While processing of bodily expressions shares many similarities with facial expressions, there is no substantial evidence that lesions of the amygdala result in similar behavioral and neural sequelae. We combined behavioral assessment with functional neuroimaging in a group of male and female humans with unilateral anterior temporal lobe (ATL) resections, including the amygdala (right: n = 10; left: n = 10) and 12 matched controls. The objective was to assess whether the amygdala is crucial for the recognition of body expressions and for modulatory effects on distant areas during perception of body expressions. The behavioral results revealed normal performance in both patient groups on emotion categorization of body expressions. The neuroimaging results showed that ATL patients displayed no enhanced activations in right fusiform body area and left extrastriate body area and that left ATL patients additionally displayed no enhanced activations in right posterior superior temporal sulcus and right extrastriate body area, respectively. Multivoxel pattern analysis revealed altered categorization capacity between emotional and neutral stimuli in right posterior superior temporal sulcus in right ATL patients. In addition, we also found emotional enhancement in frontal, parietal, occipital, and cingulate regions in controls. Together, our data show that the amygdala and ATLs are not necessary for recognition of dynamic body expressions, but suggest that amygdala lesions affect body emotion processing in distant brain areas.
SIGNIFICANCE STATEMENT For humans, information from emotional expressions of others is crucial to support social interactions. The majority of emotion studies has focused on facial expressions; however, in daily life, we also use information from body postures and body movement. Visual processing of body expressions relies on a brain network, including body-specific visual areas and visuomotor areas. Even though the importance of the amygdala and its modulatory effects on distant brain regions have been documented, it remains unclear whether the amygdala plays a crucial role in emotional body processing. By combining behavioral and neuroimaging data in patients with amygdalar lesions, we provide further evidence for its modulatory effect on distant areas during the perception of body expressions.
Introduction
Converging evidence from functional imaging studies and neuropsychological studies in clinical cohorts indicate that the anterior temporal lobe (ATL) and the amygdala in particular are key components of the “emotional brain” (Adolphs et al., 1994; Morris et al., 1996; Olson et al., 2007; Fusar-Poli et al., 2009; Adolphs, 2010; Haxby and Gobbini, 2011; Bickart et al., 2014). The role of the amygdala in emotional face processing has been well established by means of functional imaging (Vuilleumier and Pourtois, 2007) and lesion studies. Patients with bilateral amygdalar lesions show impaired recognition of facial expression, particularly fear, and this deficit appears more pronounced compared with unilateral lesions (Adolphs et al., 1994, 1999). Also, neurodegenerative disorders characterized by amygdala atrophy, such as behavioral variant frontotemporal dementia (bvFTD), are associated with deficits in facial expression recognition, and this deficit is proportional to amygdalar volume (Rosen et al., 2002; De Winter et al., 2016; Kumfor et al., 2017). Unilateral anterior temporal lobectomy typically includes the amygdala and results in facial emotion recognition impairments (Adolphs et al., 1994; Monti and Meletti, 2015). Amygdalar modulatory effects at distance, especially in face-selective areas, further corroborate the central role of the amygdala in the brain's response to emotional faces. Activity in amygdala correlates with activity in FFA during perception of facial expressions (Morris et al., 1998; Herrington et al., 2011; Diano et al., 2017), and fMRI studies in patients with both unilateral and bilateral amygdalar lesions show reductions of emotion effects in fusiform face area proportional to the lesion extent (Vuilleumier et al., 2004; De Winter et al., 2016), supporting the notion of a functional coupling between amygdala and fusiform face area. Recent work on bodily expressions seems to confirm the importance of this region in the processing of emotional body actions as well (Hadjikhani and de Gelder, 2003; de Gelder et al., 2004; de Gelder, 2006; Peelen et al., 2007; Pichon et al., 2008; Goldberg et al., 2015; Van den Stock, Hortensius, et al., 2015). Amygdala responses to body expressions correlate with responses in fusiform body area (FBA) and extrastriate body area (EBA) (Peelen et al., 2007). Furthermore, the amygdala, as a node in the emotion network, has strong connections with the action observation network and functions primarily as a source region (i.e., it modulates activation in other regions) (Jastorff et al., 2015; Huang et al., 2018). However, there is currently no evidence that the amygdala is directly involved in decoding or representing emotion-specific information independent of the mode of expression (e.g., Peelen et al., 2010; Skerry and Saxe, 2014) or correlations with emotion-recognition performance (Jastorff et al., 2015), although this may be related to multivariate pattern analysis (MVPA) limitations (Skerry and Saxe, 2014).
Activation patterns in other regions of the emotion or action observation network, such as the posterior STS (pSTS), reliably discriminate emotional expressions in body postures and other modalities (Peelen et al., 2010; Skerry and Saxe, 2014; Jastorff et al., 2015).
At the behavioral level, there is currently no strong evidence that lesions of the amygdala result in body expression recognition deficits (Atkinson et al., 2007; de Gelder et al., 2014). To study the role of the amygdala in bodily emotion recognition and its modulatory effect on distant areas during the perception of body expressions, we performed a series of behavioral and neuroimaging experiments in patients with unilateral ATL resections and matched controls. The aims of the study were as follows: (1) to reveal whether intact amygdala are necessary for body expression recognition; and (2) to reveal effects at distance in neural activity and decoding accuracy of ATL resections during body expression perception in body-responsive areas as well as at the whole-brain level.
Materials and Methods
Participants
A total of 20 patients who had previously undergone unilateral surgical resection due to mesial temporal lobe epilepsy were recruited through the epileptic surgery database of University Hospitals Leuven, Belgium. Demographics are presented in Table 1. All patients had an extensive history of uncontrollable epilepsy before surgery, with an average onset of seizure at 12.4 years of age (SD = 2.3 years). Ten (4 male and 6 female) and 7 (6 male and 1 female) participants underwent, respectively, a left and right unilateral anterior temporal lobectomy. Two male participants underwent a right amygdalo-hippocampectomy. One female patient had a right-sided tumor resection, including the amygdala and hippocampus. The lesion overlap map is displayed in Figure 1. The time since surgery was variable with a mean length of 10.5 years (SD = 0.9 years). All patients were seizure free at the time of scanning. Six patients were medication free at the time of scanning. The remaining patients were prescribed levetiracetam (n = 7), valproate (n = 2), carbamazepine (n = 7), lamotrigine (n = 2), topiramate (n = 1), pregabaline (n = 1), and clonazepam (n = 2). Exclusion criteria for patients were generalized cognitive impairment (IQ < 80), major DSM axis I psychopathology, other neurological conditions than epilepsy, or major somatic disorders. The healthy control group (5 male and 7 female) was recruited using the hospital coworkers' internal advertisement service. Controls were matched for mean age and sex ratio with the patient group (Table 1), had no self-reported major DSM axis I psychopathology, neurological, or somatic disorder, and did not take any psychotropic medication.
To document any cognitive changes following ATL resection, we performed a cognitive assessment, including global cognitive ability (Mini Mental State Examination [MMSE]), verbal memory (Rey's Auditory Verbal Learning Test), categorical verbal fluency (Animal Verbal Fluency), and visual divided attention and task shifting (Trail Making Test A and B). Participants also completed a mood (Beck Depression Inventory) and affect (Positive and Negative Affect Scale [PANAS]) questionnaire. All but 1 right resection patient and 1 control were right handed as assessed through the Edinburgh Handedness Inventory. All participants had normal or corrected-to-normal visual acuity.
The study was conducted in accordance with the Declaration of Helsinki, and all participants gave written informed consent. The ethical committee of University Hospitals Leuven approved the study.
Experimental design
Stimuli.
The stimuli have been described in detail previously (Jastorff et al., 2015, 2016). In short, they consisted of video clips displaying puppet models of walking humans. They were constructed using motion capture data of emotionally expressive (angry, fearful, sad, and happy) and neutral gaits. An animated custom-built volumetric puppet model was built and rendered in MATLAB (The MathWorks) (for details, see Roether et al., 2009). These stimuli served as prototypes for subsequent emotion morphing (Giese and Poggio, 2000; Jastorff et al., 2006). The complete set of prototypical stimuli contained 30 stimuli, six examples of 5 conditions (angry, happy, sad, fearful and neutral). We used a continuum of expressions ranging from 100% neutral prototype to 125% emotional prototype. For frame examples of stimulus prototypes and different morph levels, see Figure 2.
Scrambled versions of the stimuli were created by decomposing the figure and rearranging the individual geometric shapes. A constant random offset in horizontal or vertical direction was added to the original movement of the geometric shapes forming the puppet model. In half of the stimuli, the shapes translated horizontally, whereas in the other half the shapes translated vertically. The speed of each shape was the average speed of all shapes in the original stimulus. Thus, in both conditions, the global motion energy was matched (for stimulus examples, see Movies 1–3).
Behavioral experiments.
Morphed stimuli were displayed on an LCD screen (60 Hz frame rate; 1600 × 1200 pixels resolution) that was viewed binocularly from a distance of 40 cm, producing a stimulus size of ∼7 degrees visual angle. Stimulus presentation and recording of the participants' responses were implemented with the MATLAB Psychophysics Toolbox (Brainard, 1997). The stimuli were shown against a uniform gray background (Fig. 2).
A trial consisted of presentation of a stimulus at the center of the screen for 10 s. No fixation requirements were imposed. Participants were instructed to categorize the stimulus as happy, angry, fearful, or sad by pressing 1 of 4 labeled buttons, indicating their choice for the corresponding emotion category. Participants were instructed to respond as soon as they had made their decision, but we did not emphasize responding quickly. If the participant answered within 10 s, stimulus presentation was terminated immediately; otherwise, it halted after 10 s and a uniform gray screen was shown until the subject entered a response. After a 1.5 s intertrial interval, the next trial started. The experiment started with a demonstration session where subjects were familiarized with the stimuli for a maximum of 12 trials (3 trials per emotion). No feedback regarding performance was provided during the entire experiment.
We used a 1-up-1-down staircase procedure to calculate individual emotion detection thresholds for every emotion separately. The initial test level was set at 50% emotion for each staircase. Emotions were shown in a random order. When participants answered correctly, test level increased with 5% (a 5% less emotional stimulus was shown), when they answered incorrectly, test level decreased with 5% (a 5% more emotional stimulus was shown). The minimal test level for each emotion was 0%, and the maximal test level was 125%. These motion morphs were derived from 6 different prototypical stimuli for each emotion category, as was described above. When the boundaries were hit 4 times for a particular emotion, the presentation of this emotion was stopped. Otherwise, it stopped after 10 reversals of the staircase or after 35 trials per emotion. Hence, the number of trials and selection of stimuli per emotion category were dependent on the responses of the participants. Individual emotion detection thresholds were calculated for each emotion based on the reversal values of the 4 staircases. Each threshold represents the mean of 10 reversals and is an estimate of the 50% correct performance level. To obtain a general emotion detection threshold, thresholds were averaged over the four emotions for each subject.
We added a control experiment similar to the emotion task in terms of stimuli and display parameters. Stimuli consisted of motion morphs derived from neutral walking and neutral running prototypes and the instruction stated to categorize the stimuli as walking or running. This task allows investigating the specificity of any emotion processing deficits (Jastorff et al., 2016; Van den Stock, 2018).
fMRI task.
Prototypical avatars (i.e., 100% emotional) expressing the four emotions and a neutral condition were presented against a black background in an event-related oddball paradigm (Fig. 2). As an oddball, we added a neutral backwards-walking avatar. Stimuli appeared centrally on the screen for 1.5–5.0 s followed by an interstimulus interval (ISI) of variable length (2.3–5.0 s, determined by an exponential function) (Dale, 1999), displaying only a red fixation dot. In each run, each stimulus was presented at two different sizes. Two different sizes were used to minimize the effect of low-level features, such as retinal position on the fMRI activations. Twelve null trials (baseline) were randomly interspersed presenting only the fixation dot for 3 s. The oddball was shown 6 times. Participants were asked to press a button every time the oddball appeared. Four functional runs were recorded per participant. To familiarize with the stimuli, all participants watched some stimuli before the experiment. The order of experiments (behavioral/fMRI) was counterbalanced between participants.
Localizer experiment.
To study emotion effects in independently defined body-responsive ROIs, we performed one localizer run for each participant in addition to the experimental runs. The localizer consisted of two stimulus conditions: intact prototypical (neutral and emotional) stimuli and scrambled versions of these prototypical stimuli, both organized in separate blocks with duration of 15 s. The stimulus duration varied across stimuli but was matched between scrambled and intact blocks with a 500 ms ISI. Blocks were separated by an ISI of 6–7 s (baseline). The run consisted of 16 stimulation blocks, eight of every condition. Participants were instructed to fixate a red fixation dot in the center of the screen during the entire run.
MRI data acquisition.
Structural and functional images were collected on a 3T Philips achieva dstream scanner with a 32-channel receive-only headcoil. A standard EPI sequence was used to acquire the functional runs (TR = 2 s, TE = 30 ms, flip angle = 90°, 36 slices, 0.3 mm interslice gap, 2.75 mm × 2.75 mm × 3.5 mm voxel size). All functional scans were preceded by 4 dummy scans to allow for magnetization to reach equilibrium state. During the scan session, a high-resolution T1-weighted anatomical image (TR = 9.7, TE = 4.6, flip angle = 8°, 182 slices, matrix size 256 × 256, 1 mm × 1 mm × 1 mm voxel size) was acquired.
Statistical analysis
Behavioral data analysis.
Neuropsychological data as well as data from the control task were analyzed by means of one-way ANOVAs using IBM SPSS Statistics, version 24. To assess group differences in emotion recognition, linear mixed-models analyses with random intercept and fixed effects for group, emotion, and group × emotion were performed on the emotion detection thresholds. All post hoc tests were performed with a Bonferroni correction.
fMRI data analysis.
All data were analyzed using statistical parametric mapping (SPM12, Wellcome Trust Center for Neuroimaging, London) within MATLAB. Data were slice time-corrected, motion-corrected (aligned to the mean image across runs), and coregistered to the subject's high-resolution anatomical image. Images were then normalized to MNI space with a voxel size of 2 mm × 2 mm × 2 mm using New Segment with an extra prior (the mean of white matter and CSF Tissue probability maps) (Ripollés et al., 2012) and DARTEL. Data were smoothed using an 8 mm isotropic Gausian Kernel.
For each participant, all conditions with durations and onsets were entered in a GLM, yielding a design matrix containing 6 regressors modeling the 6 conditions (four emotions + neutral + fixation (baseline)). Head movement parameters, obtained from the motion correction during preprocessing, were entered as a regressor-of-no-interest for each run. To exclude variance due to the oddball and the subjects' response, two additional regressors were added to the model: the appearance of the oddball and the subsequent button press during the ISIs. On the subject level, all regressors were convolved with the canonical hemodynamic response function. We calculated contrasts for each participant for each of the four emotional conditions versus neutral and the average of all four emotional conditions versus neutral. Data were further analyzed at second level through ANOVA using a full factorial design. Factors consisted of group (3 levels: controls, right resection, and left resection) and emotion (4 levels: angry vs neutral, fear vs neutral, sad vs neutral, and happy vs neutral). Statistical threshold was set at Pheight < 0.05, FWE-corrected, cluster extent of 20 voxels. The following effects were investigated: the effect of emotion (i.e., a one-sample t test on each of the four emotions vs neutral contrasts) in each group separately; effect of emotion in controls versus right resected patients (i.e., a two-sample t test on each of the four emotions vs neutral contrasts between controls and right resected patients); effect of emotion in controls versus left resected patients; and effect of emotion in left versus right resected patients. In addition, all inverse group effects were also investigated.
Data from the localizer were preprocessed in a similar way to the data from the experimental runs, yielding a design matrix containing three regressors modeling the three conditions (bodies, scrambles, and baseline). The contrast “bodies vs scrambles” was calculated for each subject. Subsequently, a random-effects analysis was performed on the parameter estimates of activity for this contrast across participants over the whole group (controls and patients) using a one-sample t test. Threshold was set at Pheight < 0.001, uncorrected with a minimal cluster extent of 100 voxels. Subsequently, ROI analyses by means of small volume corrected (SVC) contrasts were performed to investigate group differences in emotion response in the body-responsive ROIs defined by the localizer. The mean percent signal change (PSC) was calculated for the average of the four emotional conditions relative to baseline as well as for the neutral condition relative to baseline by averaging the PSC within each of these predefined ROIs using Marsbar (Brett et al., 2002). Group differences were investigated using a linear mixed model with random intercept and fixed effects for group, ROI, and group × ROI. All post hoc tests were performed with a Bonferroni correction.
To investigate functional coupling between amygdala and body-responsive regions, we correlated fitted responses in right and left amygdala with fitted responses in body-responsive ROIs in the healthy control group only. Fitted responses were computed for all contrasts at the peak voxels in the ROIs as derived from the localizer experiment. Consequently, we calculated Spearman correlations between the fitted responses in every ROI to investigate connectivity between the ROIs in healthy controls.
MVPA.
To investigate whether body-responsive brain regions contain information for emotion categorization, MVPA analyses were performed in MNI space on unsmoothed data. We used the decoding toolbox (Hebart et al., 2014) to assess performance in classification between emotional and neutral stimuli as well as between the four emotional conditions based on predefined t contrasts and β values, respectively, in the ROIs defined by the localizer.
To this end, we first calculated a contrast for the average of all four emotional conditions versus baseline and the neutral condition versus baseline per run per subject. The resulting t maps were loaded into the toolbox, which then subsequently used all runs to test and to train on a leave-1-run-out basis. To assess classification between the four emotions, the parameter estimates of each emotional condition per run were loaded into the toolbox. The toolbox used these parameter estimates of all four runs to test and train in a leave-1-run-out basis.
Subsequently, we conducted a linear mixed-models analysis with random intercept and fixed effects for group, ROI, and group × ROI on the accuracy-minus-chance values. All post hoc tests were performed with a Bonferroni correction.
Parametric testing of all nonvoxelwise analyses always depended on the results of a normality check by means of Shapiro–Wilk test and visual inspection of the QQ plot. These were either performed on the variables or on the unstandardized residuals of linear mixed-models analyses. If homogeneity of variances could not be assumed based on Levene's test, a Welch test was performed and 95% CIs are reported as an estimate of effect size.
Results
Neuropsychological testing
Results are presented in Table 1. There was a significant effect of group for MMSE (F(2,14.155) = 7.374, p = 0.006), Boston Naming Test (F(2,13.338) = 9.003, p = 0.003), the Auditory Verbal Learning Test (% Recall: F(2,14.191) = 5.860, p = 0.013), and the Animal Verbal Fluency (F(2,28) = 3.642, p = 0.039, η2 = 0.206). Post hoc tests revealed that, compared with controls, left ATL patients were moderately impaired on MMSE (t(2) = 3.583, p = 0.01, 95% CI [0.4051, 2.7616]), Boston Naming Test (t(2) = 3.910, p = 0.008, 95% CI [2.236, 13.152]), Auditory Verbal Learning Test (% Recall: t(2) = 2.990, p = 0.032, 95% CI [0.0324, 0.6746]), and the Animal Verbal Fluency (t(2) = 2.663, p = 0.038, 95% CI [0.2548, 11.3563]). These findings are in accordance with the function of the resected area (i.e., semantic memory) and were used as regressors of no interest in further analysis.
There was no significant group effect of Beck Depression Inventory score (F(2,28) = 0.927, p = 0.408), but a significant group effect on the PANAS for negative affect (F(2,28) = 5.503, p = 0.10, η2 = 0.282). Post hoc tests revealed that patients with left ATL scored significantly higher on the PANAS for negative affect than patients with right ATL(t(2) = 2.647, p = 0.04, 95% CI [0.3412, 17.6810]) and controls (t(2) = 3.068, p = 0.014, 95% CI [1.6539, 17.8128]). This score was also used as a regressor of no interest in further analysis.
Body emotion recognition experiment
Linear mixed-model analysis on the emotion detection thresholds was performed with MMSE, AVLT, BNT, AVF, and PANAS-negative affect score as variables of no interest. This did not reveal a significant group effect in the emotion recognition task (F(2,96) = 0.195, p = 0.823), nor any group × emotion interaction (F(6,49) = 0.456, p = 0.837). There was, however, a significant effect of emotion (F(3,49) = 11.522, p < 0.001). Post hoc tests revealed that better performance on “happy” than on the three other emotions: happy versus angry (t(40.956) = −4.83, p < 0.001, 95% CI [−29.753, −7.240]), happy versus sad (t(44.018) = −3.60, p = 0.004, 95% CI[−23.761, −3.454]), and happy versus fear (t(44.451) = −4.29, p < 0.001, 95% CI [−25.325, −6.060]). Results are presented in Figure 3.
Imaging results
Because of technical failure, 1 patient with a left-sided resection was excluded from further analysis. For the oddball detection task, all subjects obtained a maximum of 6 hits (6 oddballs) per run.
Body-responsive regions
Based on the localizer data of the entire sample, we defined bilateral EBA and right pSTS (Fig. 4, yellow outline). A logical conjunction analysis (requiring significant activations in all groups) revealed that both bilateral EBA and right pSTS were activated in the three groups (Pheight < 0.01, uncorrected).
Within-group emotion effects
Control subjects.
An emotion effect (i.e., a one-sample t test on the average of the four emotions vs neutral contrasts) in healthy controls was observed in a large bilateral cluster in lateral occipitotemporal cortex, overlapping EBA and pSTS, in addition to smaller distributed clusters: bilateral inferior frontal gyrus, precuneus, superior parietal lobule, supplementary motor areas, right fusiform, postcentral and superior temporal gyrus, left insula, and middle frontal gyrus. At the subcortical level, there was an effect in the amygdala bilaterally. As the cluster in the right fusiform gyrus overlapped with coordinates reported for FBA (Peelen and Downing, 2005), we selected this as an additional ROI. Within the predefined body-responsive regions, we found large clusters in right EBA (x = 46, y = −64, z = 8, 316 voxels), in left EBA (x = −52, y = −66, z = 4, 175 voxels), and rpSTS (x = 62, y = −38, z = 16, 348 voxels) (Figure 4; Table 2).
Amygdalar connectivity with body-responsive regions.
We calculated the fitted responses in peak voxels of left and right amygdala and left and right EBA, right pSTS, and right FBA. Spearman correlation analyses were performed and revealed a significant result for the mean of the four emotions compared with neutral between the right amygdala and right EBA (ρ = 0.783, N = 12, p = 0.0075, one-tailed, Bonferroni-corrected, 95% CI [0.333, 1]). Spearman correlation analysis between right amygdala and left amygdala (ρ = 0.615, N = 12, p = 0.0825, one-tailed, Bonferroni-corrected) and between right amygdala and right FBA (ρ = 0.587, N = 12, p = 0.1225, one-tailed, Bonferroni uncorrected) were not significant. These results confirm previous findings (Peelen et al., 2007) and provide support for the paradigm as a valid tool to investigate modulatory influences between amygdala and body-responsive regions.
Right TLE.
The effect of emotion (a one-sample t test on the average of the four emotions vs neutral contrasts) in the right resection group revealed significant results in the left superior temporal gyrus, inferior frontal gyrus, and right middle temporal gyrus. ROI analysis revealed only one small cluster in right EBA (x = 56, y = −66, z = 4, 28 voxels), but not in the other two body-selective ROIs (Fig. 4, red outline).
Left TLE.
In the left ATL group, we found an effect of emotion in the left parahippocampal gyrus. We did not find any significant activation in the body-responsive ROIs (Fig. 4, blue outline).
Between-group comparisons
Controls versus right resection group.
Compared with the right resection group, healthy controls showed a stronger emotion effect (i.e., a two-sample t test on the average of each of the four emotions vs neutral contrasts) in the bilateral precuneus, the right fusiform, cingulate, and parahippocampal and middle occipital gyrus, as well as in the right caudate nucleus, amygdala, and superior parietal lobule (Fig. 5; Table 3). ROI analyses showed only a small cluster in left EBA (x = −50, y = −72, z = 10, 5 voxels). The reverse contrast (right resection groups vs controls) revealed no significant differences in activation.
Controls versus left resection group.
Compared with the left resection group, healthy controls showed a stronger emotion effect (i.e., a two-sample t test on the average of each of the four emotions vs neutral contrasts) in the left superior occipital, postcentral, middle frontal, and superior temporal gyrus, as well as in the left precuneus and cerebellum (Fig. 5; Table 3). In the right hemisphere, there were significant differences in middle, medial, and inferior frontal, precentral, middle temporal, and fusiform gyrus, and the superior parietal lobule. We observed one small cluster in left EBA (x = −50, y = −72, z = 10, 16 voxels), one small cluster in right EBA (x = 48, y = −66, z = 8, 11 voxels), and one cluster in right pSTS (x = 56, y = −42, z = 4, 20 voxels). The reverse contrast (left resection group vs controls) revealed no significant differences in activation.
Left versus right resection group.
Compared with the right resection group, left ATL resection patients show a stronger emotion effect (i.e., a two-sample t test comparing each of the four emotions vs neutral contrasts) in the left calcarine sulcus extending in the left occipital lobe and the precuneus (Fig. 5, violet outline). There were no significant differences in activation in the body-responsive ROIs. The reversed contrast (right vs left resection group) revealed no significant differences.
MVPA: within group results
Control subjects.
In the control group, all three ROIs reliably categorized between emotional and neutral stimuli, as was determined by permutation testing with 1000 permutations (right EBA: t(11) = 5.897, p < 0.001, d = 1.702; left EBA: t(11) = 3.939, p = 0.002, d = 1.137; right pSTS: t(11) = 7.966, p < 0.001, d = 2.3) (Fig. 6). For discriminating between the four emotions separately, the three ROIs performed significantly above chance level as well (right EBA: t(11) = 4.451, p = 0.001, d = 1.285; left EBA: t(11) = 5.946, p < 0.001, d = 1.716; right pSTS: t(11) = 3.362, p = 0.006, d = 0.970).
Right TLE.
In the right resection group, all three ROIs reliably categorize between emotional and neutral stimuli, as was determined by permutation testing with 1000 permutations (right EBA: t(9) = 5.514, p < 0.0001, d = 1.744; left EBA: t(9) = 7.856, p < 0.001, d = 2.484; right pSTS: t(9) = 3.354, p = 0.008, d = 1.06). For discriminating between the four emotions separately, bilateral EBA ROIs performed significantly above chance level as well (right EBA: t(9) = 3.475, p = 0.007, d = 1.099; left EBA: t(9) = 3.926, p = 0.003, d = 1.241). However, right pSTS does not reliably discriminate between the four emotions in this patient group (t(9) = 2.023, p = 0.222, uncorrected).
Left TLE.
In the left resection group, both right EBA and rpSTS reliably categorize emotional and neutral stimuli, as was determined by permutation testing with 1000 permutations (right EBA: t(8) = 3.578, p = 0.007, d = 1;192; right pSTS: t(8) = 3.5, p = 0.008, d = 1.167). In left EBA categorization, performance between emotional and neutral stimuli did not surpass chance level (t(8) = 2.412, p = 0.126, uncorrected). For discriminating between the four emotions, the left EBA performed significantly above chance level (t(8) = 4, p = 0.004, d = 1.333), unlike the other areas (right EBA: t(8) = 2.997, p = 0.051, uncorrected; right pSTS: t(8) = 2.667, p = 0.087, uncorrected).
MVPA: group differences
Linear mixed-model analysis on the accuracy minus chance values for the emotions versus neutral categorization showed no effect of group (F(2.30) = 1811, p = 0.181) nor ROI (F(2.37) = 1.632, p = 0.209) (Fig. 6). However, the interaction group × ROI was significant (F(4.37) = 4845, p = 0.003). Post hoc tests revealed a significant difference between the healthy controls and the right resection group in the right pSTS only (t(29.56) = 3.112, p = 0.012, 95% CI [3.456, 34.044]).
For the categorization of four emotions, linear mixed-model analysis on the accuracy minus chance values showed no significant effect of group (F(2.78) = 0.549, p = 0.580) nor a significant interaction between group and ROI (F(4.63) = 0.155, p = 0.960). The effect of ROI was significant (F(2.63) = 6.298, p = 0.003). Post hoc tests revealed a significant difference between left EBA and right pSTS (t(53.143) = 3.23, p = 0.005, 95% CI [2.176, 15.185]), with better performance in right STS.
Discussion
To investigate body emotion processing, we used artificial stimuli expressing emotions conveyed through motion (Sokolov et al., 2011; Krüger et al., 2013).
The first aim was to investigate whether the ATL and the amygdala in particular play a crucial role in the perception of emotional body expressions. The results reveal normal body emotion categorization performance in both patient groups. Hence, our data further challenge a key role for the amygdala in accurate interpretation of emotional body language. This is in line with neuropsychological evidence of preserved body expression recognition in patients with bilateral amygdalar atrophy due to Urbach-Wiethe disease (Atkinson et al., 2007; de Gelder et al., 2014). On the contrary, another syndrome characterized by amygdalar atrophy (i.e., bvFTD) is associated with deficits in recognizing body expressions (Van den Stock, De Winter, et al., 2015; Jastorff et al., 2016). However, those studies also revealed that the deficits were not related to structural integrity of the amygdala, but rather to the volume of a region in the left temporal pole (not overlapping the amygdala) and the left inferior frontal gyrus (Jastorff et al., 2016). Interestingly, we observed in the same sample of bvFTD a facial emotion recognition deficit that was associated with structural integrity of the amygdala (De Winter et al., 2016). These results suggest a dissociation in the structural neuroanatomy underlying body and face emotion recognition deficits. This notion is supported by the finding that the sensitivity of the amygdala is higher to facial than to body emotions (Kret et al., 2011) and suggests that, while the amygdala is necessary to recognize facial expressions, it is not necessary to recognize body expressions. This is in line with a functional hypothesis of the amygdala as a personal relevance detector and modulator of brain processes underlying adaptive behavior (Sander et al., 2003; Janak and Tye, 2015). The amygdala is functionally connected to primary visual areas but also to motor structures and frontal cortices (Bickart et al., 2012, 2014; Grèzes et al., 2014; Toschi et al., 2017). The disproportional effect of amygdalar lesions on processing facial compared with bodily expressions may thus relate to social stimulus proximity detection and regulating personal space (Kennedy et al., 2009). In particular, detailed perception of a facial expression typically implies closer proximity and hence higher personal relevance compared with perceiving a whole body. Similarly, our findings are in line with studies relating the amygdala to driving eye gaze and attention to the eyes of others (Adolphs et al., 2005; Gamer and Büchel, 2009).
The second aim of the study was to reveal whether amygdala lesions influence neural processes in distant areas, particularly body-sensitive regions.
Both patient groups showed decreased activation in left EBA and right fusiform gyrus. Additionally, the left resection group showed no emotional enhancement in right EBA and pSTS. The right resection group, however, shows an emotional response in the right EBA and STS. These findings support a central role for the amygdala in emotional enhancement of ventral (FBA) but not posterior lateral temporal cortex when perceiving body expressions. This is in line with facial expression recognition findings in patients with amygdalar lesions (Vuilleumier et al., 2004; De Winter et al., 2016) and structural connectivity studies in normal subjects reporting white matter tracts between amygdala and ventral posterior areas, but not STS (Smith et al., 2009; Gschwind et al., 2012).
On the other hand, the MVPA results in the right pSTS reveal a reduction in the right ATL patients in distinguishing emotional from neutral stimuli. This suggests that the ATL and amygdala directly or indirectly influence the function of the pSTS when processing body expressions. Reduced activation in pSTS in response to facial expressions has been observed in patients with ATL lesions (Vuilleumier et al., 2004; Ahs et al., 2014). While there is evidence that amygdala and pSTS are functionally connected (Pitcher et al., 2017; Sato et al., 2017), body emotion processing in posterior lateral temporal cortices, such as EBA and right pSTS, may rely primarily on nonamygdalar input, such as cortical visual areas and motor areas (Ewbank et al., 2011; Lahnakoski et al., 2012; Erickson et al., 2017; Zimmermann et al., 2018). One could speculate that altered input from, for example, FBA, due to absent amygdalar modulation, could explain the reduced emotion classification in the right pSTS.
The effects at distance we observe are not confined to the ipsilateral hemisphere. This is surprising because amygdala modulations have, up until now, been described predominantly ipsilateral to the lesion side (Vuilleumier et al., 2004). We speculate that these contralateral altered responses rely on altered input from the homotopic visual areas ipsilateral to the lesion side, which in turn receives diminished input from the lesioned amygdala. There is indeed evidence for corpus callosum deficits in TLE, which results in diminished interhemispheric transfer (Schneider et al., 2014).
Interestingly, we also observed strong distant effects of ATL resections beyond body-responsive visual areas (i.e., in frontal, parietal, occipital, and cingulate regions), which showed emotional enhancement in the control group. Widespread activation during body motion perception (Pavlova et al., 2017) and emotional body expression perception have consistently been reported (de Gelder et al., 2004; de Gelder, 2006; Grèzes et al., 2007; Pichon et al., 2008; van de Riet et al., 2009; Kret et al., 2011). The present results suggest that these activations may also be primarily amygdala-dependent.
The notion that specific emotions can be mapped to specific brain circuits has been increasingly challenged by psychological constructionist theories. These models hypothesize that processing of emotions relies on distributed brain networks and processing of specific emotions relies on coactivation patterns within and between those networks (Lindquist et al., 2012; Barrett, 2013; Jastorff et al., 2015; Huang et al., 2018). The amygdala is considered a “core affect” region in constructionist accounts, adding to the representation of emotional events in concert with other areas and networks. Indeed, the amygdala is strongly connected to other regions of the social brain. It is considered to be involved in multiple processes with a common ground of detecting arousal and salience in multiple types of stimuli (i.e., emotional, novel, rewarding stimuli) (Sander et al., 2003; Adolphs, 2010). As a “relevance detector,” it seems to modulate several of these social brain regions to facilitate adaptive behavior (Bickart et al., 2014). There is indeed evidence that the amygdala is both structurally and functionally linked to (pre)motor areas, such as the supplementary motor area and precentral and postcentral gyri (Pichon et al., 2008; Grèzes et al., 2014; Toschi et al., 2017).
Several limitations of the present study need to be addressed. The normal performance of body emotion recognition in the clinical groups may be related to insufficient sensitivity of our behavioral task, although the sensitivity of psychometric function modeling may surpass conventional accuracy measures (Biotti and Cook, 2016). Second, TLE is associated with ATL sclerosis and deficits in emotion recognition (Monti and Meletti, 2015), also before surgery. As we did not include presurgical data nor patients with active temporal lobe epilepsy, we cannot conclude that the group differences are due to the resection of the ATL.
A related issue to consider is the neural plasticity that may have taken place after the resection but also before, possibly compensating for the amygdala and ATL. While functional compensation may occur in contralateral amygdala, we did not find any evidence for increased responses to emotional stimuli in patients compared with healthy controls. Furthermore, in neither patient group did we observe an emotion effect in the intact amygdala, suggesting no compensatory activation from contralateral amygdala.
There is evidence that emotion recognition in TLE depends on intellectual capacity (Monti and Meletti, 2015), and we excluded patients with total IQ < 80. While this exclusion criterion benefits the validity of comparisons with the control group, it may also reduce the sensitivity to reveal IQ-dependent behavioral effects.
Our sample size is rather small and may negatively influence statistical power. To compensate for this, we applied a conservative statistical threshold (FWE correction) to maximize the reliability of the significance of the group differences. It should, however, be noted that this stringent control for Type I errors increases the probability of Type II errors, which we considered less relevant for the present study purpose.
Finally, antiepileptic medication in a part of the patient group, while absent in the control group, might also have influenced the results.
In conclusion, the findings reveal that unilateral ATL resection does not reduce recognition of body emotions despite the indications that it plays a central role in enhancing activation in the right fusiform gyrus during body emotion perception. Furthermore, the enhanced activity in a widely distributed set of regions during body emotion perception supports the role of the amygdala as a hub in the emotional brain.
Footnotes
L.V.d.V. was supported by Fonds Wetenschappelijk Onderzoek-Vlaanderen aspirant fellowship grant 11Z8917N. J.J. is a postdoctoral researcher supported by Fonds Wetenschappelijk Onderzoek-Vlaanderen. J.V.d.S. was supported by Fonds Wetenschappelijk Onderzoek-Vlaanderen and a KU Leuven starting grant. We thank all patients and healthy volunteers for cooperation; Dr. Kristof Vansteelandt for advice on statistics; and Jochen Weber for technical support.
The authors declare no competing financial interests.
- Correspondence should be addressed to Dr. Jan Van den Stock, Laboratory for Translational Neuropsychiatry, Department of Neurosciences, KU Leuven, O&N 2, bus 1027, Herestraat 49, 3000 Leuven, Belgium. jan.vandenstock{at}med.kuleuven.be