Abstract
Painful events in our environment are often accompanied by stimuli from other sensory modalities. These stimuli may influence the perception and processing of acute pain, in particular when they comprise emotional cues, like facial expressions of people surrounding us. In this whole-head magnetoencephalography (MEG) study, we examined the neuronal mechanisms underlying the influence of emotional (fearful, angry, or happy) compared to neutral facial expressions on the processing of pain in humans. Independent of their valence, subjective pain ratings for intracutaneous inputs were higher when pain stimuli were presented together with emotional facial expressions than when they were presented with a neutral facial expression. Source reconstruction using linear beamforming revealed pain-induced early (70–270 ms) oscillatory beta-band activity (BBA; 15–25 Hz) and gamma-band activity (GBA; 60–80 Hz) in the sensorimotor cortex. The presentation of faces with emotional expressions compared to faces with neutral expressions led to a stronger bilateral suppression of the pain-induced BBA, possibly reflecting enhanced response readiness of the sensorimotor system. Moreover, pain-induced GBA in the sensorimotor cortex was larger for faces expressing fear than for faces expressing anger, which might reflect the facilitation of avoidance-motivated behavior triggered by the concurrent presentation of faces with fearful expressions and painful stimuli. Thus, the presence of emotional cues, like facial expressions from people surrounding us, while receiving acute pain may facilitate neuronal processes involved in the preparation and execution of adequate protective motor responses.
Introduction
Facial expressions of people surrounding us provide the emotional context in most social situations and may therefore influence the processing of pain. The perception of emotional facial expressions modulates neuronal activity in the limbic system (Adolphs, 2002) and also elicits rapid and widely automatic responses in other brain areas (Batty and Taylor, 2003; Frith, 2009). Moreover, the cortical regions involved in the decoding of pain (Bushnell et al., 1999) and emotional facial expressions (Adolphs et al., 2000), like the somatosensory cortex, partly overlap. For these reasons, it is likely that emotional facial expressions modulate the processing of pain.
Emotional cues, such as pictures with affective content (Godinho et al., 2006) or pleasant (e.g., floral scents) or unpleasant odors (Villemure and Bushnell, 2009), influence pain processing. Compared to control stimuli, sensory inputs that induce positive emotion often lead to analgesic effects, i.e., reduction of pain perception (Meagher et al., 2001). In contrast, pain ratings of painful stimuli that are presented with inputs that induce negative emotions are frequently enhanced (Godinho et al., 2006). Interestingly, a recent study showed that both pleasant (i.e., lemon scent) and unpleasant odors lead to enhanced pain ratings (Martin, 2006), suggesting that the effect of positive emotional inputs on pain processing may depend on the type of stimulus (e.g., type of odor) used to induce positive emotions. To date, the impact of emotional facial expressions on pain perception and the neuronal mechanisms underlying the crossmodal influence of facial expression on the processing of pain are not understood.
A neuronal mechanism that may be crucial for the impact of facial expression on the processing of pain is synchronization of neuronal populations. Recent studies suggested that neuronal synchronization, in particular in higher-frequency beta-band activity (BBA; 15–30 Hz) and gamma-band activity (GBA; activity above 30 Hz), plays an important role for crossmodal processing (Senkowski et al., 2006, 2008; Lakatos et al., 2007; Kayser and Logothetis, 2009). Recently, neural synchronization mechanisms were also linked to pain processing and pain perception (Hauck et al., 2008). Specifically, the GBA in primary somatosensory (Gross et al., 2007; Tiemann et al., 2010) and sensorimotor cortex (Hauck et al., 2007) was positively correlated to pain perception. Furthermore, it has been shown that painful inputs suppress oscillatory activity in the alpha-band (7–15 Hz) and beta-band (Ploner et al., 2006a). This suggests that oscillatory responses, in particular in regions of the sensorimotor cortex, are an important neuronal mechanism for the encoding of pain.
In this magnetoencephalography (MEG) study, we investigated the influence of emotional facial expressions on pain processing, with a focus on higher-frequency oscillatory activity in the sensorimotor cortex. Using linear beamforming (Gross et al., 2001), pain-induced oscillatory responses were examined for faces with positive (happiness) and negative (fear, anger) valence and compared with the responses to neutral facial expressions. Independent of the valence of the facial expression, subjective pain ratings were higher for emotional than for neutral expressions, an effect that was paralleled by modulations in early pain-induced BBA in the sensorimotor cortex. Additionally, differences in the early pain-induced GBA were found between the facial expressions of fear and anger.
Materials and Methods
Participants.
Nineteen right-handed healthy volunteers participated in the experiment (mean age 24 years, range = 21–28 years, 8 female). One additional participant was excluded due to technical artifacts during the MEG recording. All participants had normal or corrected-to-normal vision and reported no history of neurological or psychiatric illness. The review board of the Hamburg Medical Association's Ethics Commission approved the study, and each subject provided written informed consent.
Stimuli.
The intracutaneous pain model (Bromm and Meier, 1984) was used to induce pain. Electrical stimulus pulses (16 ms duration) were delivered to the nearest proximity of nociceptors using a thin electrode that was fed through a hole inserted into the epidermal skin at the tip of the left middle finger. Before the experiment, each participant's individual pain threshold was determined by applying a staircase method with successive intensity increments and decrements of 0.02 mA. The individual threshold of a participant was computed as the average intensity at which the participant reported a given stimulus as painful. During the experiment, standard stimuli were applied with an intensity 1.5-fold of the individual pain threshold in 90% of all painful inputs (i.e., in P-only and bimodal PV trials; see below, Procedure). In addition, 10% of all painful stimuli were presented at twofold of the individual pain threshold. These stimuli, which served as catch trials, were presented to avoid a possible tendency of participants to rate all trials equally.
Visual stimuli comprised a set of pictures from the Ekman FEEST stimulus battery (Young et al., 2002). Faces of eight actors (four female) expressing three emotions (fear, anger, and happiness) and neutral control expressions were presented. The pictures of the faces were displayed at a visual angle of 10.8° in the horizontal plane and 15.2° in the vertical plane. Stimuli were projected from a calibrated LCD projector outside of the magnetically shielded chamber at a refresh rate of 60 Hz. Stimulus presentation was controlled using “Presentation” software (Neurobehavioural Systems).
Procedure.
Participants were presented with a continuous randomized stream of visual face stimuli (V-only), pain stimuli (P-only), and bimodal pain–visual (PV) stimuli (Fig. 1a). They were instructed to attend to the painful intracutaneous stimuli and to rate the perceived unpleasantness of these stimuli on a visual analog scale (VAS; range 1–100, where 1 indicates a very low and 100 a very high pain experience), independent of whether they were presented alone or in combination with a task-irrelevant face. Moreover, while maintaining central fixation, participants were asked not to pay particular attention to the faces. Unpleasantness ratings, which are a subjective measure of the affective-motivational dimension of the nociceptive experience (Price, 2000), have been previously shown to reflect emotional modulations of pain (Villemure and Bushnell, 2009). Visual and pain inputs had the same onsets in the PV trials. One thousand milliseconds after the onset of the P-only and PV stimuli, the VAS was presented for 1200 ms, during which the participants rated the unpleasantness of the painful inputs using a joystick in their right hand. A fixed interval of 400 ms followed the offset of V-only stimuli, which were presented for 600 ms. In addition, an interstimulus interval of 800–1200 ms (mean 1000 ms) followed the presentation of each stimulus. A central fixation cross was presented during periods in which no face or VAS was shown. In total (standard and catch trials), 300 P-only, 600 V-only (150 for each facial expression type), and 600 bimodal PV (150 for each facial expression type) inputs were presented. The order of stimuli was randomized with the limitation that not more than two faces with the same expression were shown consecutively. The experiment comprised 15 blocks, each of which lasted ∼4.5 min. In addition, one practice block was conducted to familiarize the participants with the experimental procedure.
MEG recordings and data preprocessing.
The MEG was recorded in a magnetically shielded chamber using a 275-channel whole-head system (Omega 2000, CTF Systems). Three defective MEG channels were removed from further analysis. Simultaneous to the MEG recording, eye movements and eye blinks were monitored using two electrooculogram electrodes for the offline artifact rejection. The head position of participants was measured before and after each experimental block. In the vast majority of blocks, the head displacements were <2 mm, and for all analyzed datasets, the head displacements were <10 mm. MEG data were low-pass filtered online (cutoff 300 Hz) and recorded at a sample rate of 1200 Hz. Analysis of MEG data was performed using Matlab (MathWorks) and FieldTrip (http://www.ru.nl/fcdonders/fieldtrip). For data reduction, the recorded signals were offline high-pass filtered at 2 Hz and low-pass filtered at 170 Hz. Moreover, data were resampled at 400 Hz. To remove line-noise artifacts, notch filters at 50, 100, and 150 Hz (bin width = 1 Hz) were used. Data were epoched from −800 ms before to 2400 ms after the onsets of V-only, P-only, and bimodal PV events. Trials containing eye blinks, eye movements, muscle artifacts, or signal jumps were rejected manually from further analysis. In addition, bad channels were offline interpolated by replacing them with the average of the neighboring channels. Specifically, two non-neighboring channels in one dataset and three non-neighboring channels in another dataset were interpolated.
Analysis of oscillatory responses.
Time–frequency representations (TFRs) were computed using the multitaper method applied to short sliding time windows. This method offers an optimal spectral concentration over the frequency range of interest (Mitra and Pesaran, 1999). The data in each time window were multiplied with a set of orthogonal Slepian tapers. The Fourier transforms of the tapered time windows were then calculated and the resulting power estimates were averaged across tapers. For the examination of low-frequency (2–30 Hz) and high-frequency (30–120 Hz) activities, fixed time windows (ΔT = 200 ms and ΔT = 150 ms, respectively) and fixed frequency smoothing (Δf = ±5 Hz and Δf = ±10 Hz, respectively) were applied, resulting in one taper for the low frequency range and two tapers for the high frequency range. For the calculation of total power (i.e., activity that comprises both evoked and induced oscillations), frequency transformation was performed on the single-trial level separately for each frequency before averaging for frequencies from 2 to 120 Hz (0.5 Hz frequency step size and 5 ms window step size). To reveal amplitude-normalized signal changes of total power in the poststimulus interval, the average power in the baseline interval (−300 to −100 ms before stimulus onset) was first subtracted and the resulting difference was divided by the baseline interval activity for each frequency as follows: Pow(t, f)normalized = 100 × (Pow(t, f)poststimulus − Pow(f)baseline)/Pow(f)baseline.
A linear beamforming approach was applied for the source reconstruction of total oscillatory responses (Van Veen et al., 1997; Gross et al., 2001). In short, this approach uses an adaptive spatial filter that passes activity from one specific location of interest with unit gain and maximally suppresses other activity. Since linear beamforming is based on the calculation of the cross-spectral density matrix over trials, this approach is particularly suitable for the analysis of total power in the human electroencephalogram (Schneider et al., 2008, 2011) and magnetoencephalogram (Bauer et al., 2006; Donner et al., 2009). The spatial filters for the beamforming analysis were computed based on pooled trials across the four facial expressions (separately for V-only and bimodal PV trials and separately for baseline and poststimulus intervals). The use of common filters ensures that differences in source space activity can be ascribed to power differences in the different facial expressions and not to differences between filters. Single trials were then projected through these filters and averaged in source space separately for each condition.
For the source reconstruction, T1-weighted structural MRIs were recorded on a 3 tesla MRT individually for all participants and single-shell models were derived from the segmentation of MRIs using the Statistical Parametric Mapping software (SPM2; http://www.fil.ion.ucl.ac.uk/spm). The lead field matrix was calculated on a boundary element model (BEM) for each grid point in the brain with a regular 7 mm grid. The source activity at each grid point was estimated by constructing a spatial filter using the lead field at this point and the cross-spectral density matrix. The cross-spectral density matrix was calculated between all MEG channels separately for a baseline time interval (centered at −200 ms) and for those poststimulus time points and frequencies in the beta-band and gamma-band for which robust amplitude changes were observed (Figs. 2, 3). To examine oscillatory responses in source space across participants, a BEM template grid was first created based on the template brain from the Montreal Neurological Institute (MNI; http://www.mni.mcgill.ca). Subsequently, the individual BEM head models were warped into the template grid and the inverse of that warp was applied to the individual grids. Due to the warping, a specific grid point (in MNI coordinates) is located in the same area of the brain across all participants. Since the most robust beta-band and gamma-band responses were observed in the ipsilateral and the contralateral sensorimotor hand region (Fig. 2), a symmetric hand region comprising sensorimotor cortex was used for the region of interest (ROI) analysis of total oscillatory activity (spatial coordinates for determining the ROI were taken from http://neuro.imm.dtu.dk/services/jerne/ninf/voi/hand_area.html).
For the study of effects on evoked power (i.e., oscillatory responses that are strictly phase locked across trials), the transformation to the frequency domain was performed on the average of all trials, i.e., the event-related response, separately for each frequency and condition. Evoked responses were calculated by subtracting the baseline interval activity from the poststimulus interval activity without further dividing these responses by the baseline (since evoked power is expected to be very low in the prestimulus interval). Moreover, the analysis of effects in evoked oscillatory responses was confined to the sensor level using two regions of interest that comprised six sensors each. The sensors were selected in accordance with the main activity patterns found in total oscillatory responses, which were localized to the sensorimotor hand region (Fig. 2a). Since there is no standard position for MEG sensors relative to the head position across participants, data of each participant were realigned to a standard helmet before the analysis of evoked responses. The procedure used has been described previously (Knosche, 2002).
Statistical analysis.
To examine whether pain ratings of non-catch trials were normally distributed, Shapiro–Wilk parametric tests were conducted within each participant and condition. If these tests indicated violations of normal distribution, median values of pain ratings were submitted to further analysis. In line with the activity pattern of pain-induced total oscillatory responses (Figs. 2, 3), which are to a large extent comparable with our previous study (Hauck et al., 2007), the focus of the statistical analysis of oscillatory responses was on the examination of BBA (13–23 Hz) and GBA (70–90 Hz) in the hand region of the sensorimotor cortex. An early time interval centered at 170 ms was subjected to the statistical analysis. The motivation for studying effects at this interval derives from electrophysiological studies on classical evoked brain responses (Ploner et al., 1999; Nakata et al., 2008). These studies suggested critical nociceptive processes in a cortical network including sensorimotor areas at this latency. Furthermore, neuronal activity around 170 ms has been closely related to the encoding of emotional facial expressions (Batty and Taylor, 2003; Eimer and Holmes, 2007). Therefore, it is possible that emotional facial expressions modulate pain-induced oscillatory responses around 170 ms. In the present study, robust modulations of the early BBA and GBA were observed. These modulations were paralleled by an increase in lower frequency total activity around 200 ms. For exploratory purposes, effects of emotional facial expressions on pain-induced lower frequency activity (3–9 Hz, centered at 170 ms) were examined (see Notes).
The statistical data analysis comprised two levels. The first level of analysis was conducted to investigate valence-specific effects of facial expressions on pain-induced oscillatory responses. For this analysis, the responses to V-only stimuli were subtracted from the responses to the respective visual inputs that were presented in the bimodal context. For instance, the response to a facial expression of fear (i.e., VFear) was subtracted from the response obtained for the same facial expression that was presented in combination with a painful stimulus (i.e., PVFear). This subtraction effectively eliminated response differences, which are solely driven by the visual inputs, leaving pain-induced oscillatory responses in the context of the crossmodal bias of the different facial expressions. For the analysis of evoked oscillatory responses, the response differences between PV and V-only stimuli were computed before the time–frequency transformation (Senkowski et al., 2007). The subtraction metric led to four conditions: PVFear minus VFear, PVAnger minus VAnger, PVHappiness minus VHappiness, and PVNeutral minus VNeutral. Data from these conditions were subjected to repeated-measures ANOVA using the factors of facial expression (Fear, Anger, Happiness, and Neutral) and ROI (left sensorimotor hand region and right sensorimotor hand region). If significant effects of the factor facial expression were found, planned follow-up comparisons between the three emotional facial expression conditions and the neutral expression condition were computed. Additionally, it was examined whether the responses differed between the three emotional facial expressions using ANOVAs with the factors of facial expression (Fear, Anger, and Happiness) and ROI. Follow-up pairwise comparisons between the different emotion expression conditions, which were computed if a significant effect of facial expression was found in the ANOVA, were corrected for multiple testing using Bonferroni correction. Finally, if parallel effects were observed for the impact of facial expressions on pain ratings and pain-induced oscillatory responses, Pearson correlation coefficients were computed for these effects across participants. Further details are provided in Results.
In the second level of analysis, the general effects of viewing faces, independent of their valence, on pain-induced responses were examined. To this end, oscillatory activity to P-only trials was compared with the activity to combined bimodal PV trials (averaged across the four conditions). The ANOVAs conducted comprised the factors of stimulation (P-only and averaged PV conditions) and ROI (left sensorimotor hand region and right sensorimotor hand region).
Results
Pain ratings
Since Shapiro–Wilk parametric tests, which were conducted within each participant and condition, revealed violations of normal distribution of non-catch trial pain ratings in the majority of tests (p < 0.01 in 46 of 72 tests), the median values of pain unpleasantness ratings were submitted to statistical analysis. Across participants and within conditions, no violations of normal distribution were found for the median values (none of the five conducted Shapiro–Wilk tests showed a p < 0.01). Therefore, parametric tests were used to analyze the median values of pain ratings. The ANOVA for pain ratings to non-catch trials using the factor facial expression (Fear, Anger, Happiness, and Neutral) revealed a significant main effect (F(3,17) = 3.31, p < 0.05). Follow-up contrasts showed significantly higher pain ratings for all three emotional expressions than for the neutral expression condition (Fear vs Neutral: F(1,17) = 6.14, p < 0.05; Anger vs Neutral: F(1,17) = 4.5, p < 0.05; Happiness vs Neutral: F(1,17) = 6.33, p < 0.05). Thus, the presentation of an emotional facial expression compared to the presentation of a neutral facial expression led to increased pain unpleasantness ratings (Fig. 1b). The comparisons between the three emotional facial expressions, however, did not reveal significant differences. Further analyses showed that the pain ratings for bimodal PV stimuli were for all conditions larger than the pain ratings for P-only stimuli (all p values < 0.0001). Finally, pain ratings in catch trials were significantly higher than pain ratings in non-catch trials (all p values < 0.0001; mean ratings for catch trials = 79; data not shown), demonstrating that participants reliably rated the painful stimuli.
Experimental setup and behavioral data. a, Participants were presented with a continuous stream of faces only, pain only, and bimodal PV stimuli. They were instructed to rate the painful events on a visual analog scale. Faces with three different emotions (fear, anger, and happiness) and neutral control expressions were presented. b, Pain ratings to standard trials were higher for the three emotional facial expression conditions than for the neutral control condition. Moreover, pain ratings were significantly higher when a face was presented simultaneously than in pain only trials.
Description of oscillatory response pattern
Time–frequency representations (2–120 Hz) of pain-induced oscillatory responses showed early modulations in total GBA and BBA (Figs. 2, 3). At ∼170 ms after stimulus onset, an enhancement of GBA was observed, which was most pronounced contralateral to the intracutaneous stimulation site and at central sensors (Pattern I). Moreover, at the same latency an early suppression of BBA was observed (Pattern II). Although the duration of electrical stimuli was just 16 ms, the very early (<100 ms) responses were contaminated by a broadband electrical stimulation artifact, which was strongest ipsilateral to the stimulation site. The artifact is smeared in time due to spectral smoothing in the time–frequency transformation but did not overlap with the time windows selected for the linear beamforming analysis. The observation that oscillatory GBA and BBA were predominantly observed in the sensorimotor hand region further suggests that the electromagnetic artifact did not contaminate the present findings (Fig. 2d).
Total oscillatory MEG responses to pain-only stimuli. a, Time–frequency representations for sensors located over sensorimotor areas (comprising 6 sensors each) and for central sensors (comprising 8 sensors). The selected sensors are highlighted in the head panels. Amplitude changes in early (∼170 ms) oscillatory BBA and GBA were observed (highlighted by rectangles). The electromagnetic artifact induced by the intracutaneous stimulation contaminated the very early (<100 ms) responses. b, Topographic maps for GBA (upper panel) and BBA (lower panel). c, Traces of GBA (upper panel) and BBA (lower panel). An increase in GBA was observed at ∼170 ms (Pattern I). At the same latency, the BBA showed a decrease that was most pronounced at central sensors (Pattern II). d, Results of the beamforming analysis of GBA (upper panel) and BBA (lower panel). Plotted are the highest and lowest quartiles of activity changes. A contralateral GBA increase was found that was paralleled by a bilateral BBA suppression.
Total oscillatory MEG responses for the contrasts of bimodal PV minus V-only stimuli. a, TFRs of total oscillatory responses for the different facial expression conditions (Bimodal PV minus V-only) at different sensor locations. The upper and the lower panels illustrate TFRs of combined sensors over sensorimotor regions ipsilateral and contralateral to the intracutaneous stimulation site, respectively. Comparable with the response patterns to pain-only stimuli (Fig. 2), short-latency (∼170 ms) modulations were observed for GBA (Pattern I) and BBA (Pattern II). The electromagnetic artifact induced by the intracutaneous stimulation contaminated the very early (<100 ms) responses. b, Traces of total GBA (left) and BBA (lower panel) at left and right sensor regions over the sensorimotor cortex and at central sensors.
Effects of facial expressions on total beta-band activity
A bilateral decrease of early BBA in the sensorimotor hand regions was found for all PV minus V conditions (Fig. 4). The ANOVA to examine valence-specific effects on this early pain-induced beta-band suppression (70–270 ms) using the factors of condition (PVFear minus VFear, PVAnger minus VAnger, PVHappiness minus VHappiness, and PVNeutral minus VNeutral) and ROI (left sensorimotor hand region and right sensorimotor hand region) revealed a significant main effect of facial expression (F(3,51) = 4.45, p < 0.01). Follow-up contrasts between the three emotion conditions (separately) and the neutral expression condition using the factors of condition and ROI revealed significant main effects of the factor facial expression (Fear vs Neutral: F(1,17) = 9.64, p < 0.01; Anger vs Neutral: F(1,17) = 8.54, p < 0.01; Happiness vs Neutral: F(1,17) = 8.9, p < 0.01). Thus, paralleling the findings of enhanced pain ratings for emotional compared to neutral facial expressions, the early BBA suppression was stronger in all three emotional expression conditions than in the neutral expression condition (Fig. 4b), while no significant differences were found between the three facial emotion conditions. No other main effects or interactions were found in these ANOVAs. Since the traces of sensors over contralateral sensorimotor regions indicated differences already at the onset of stimulation, oscillatory activity at −100 to 100 ms was analyzed in an additional ANOVA using the factors of condition and ROI. While this ANOVA revealed a significant main effect of ROI (F(1,17) = 20.66, p < 0.005)), due to a larger contralateral than ipsilateral BBA, no significant effects were found in relation to the factor condition.
Beamforming analysis of early beta-band responses for the contrasts of bimodal PV and the corresponding V-only stimuli. a, Robust pattern of BBA modulations at the sensorimotor hand region (highlighted by dotted circles) were found. Plotted are the highest and lowest quartiles of activity changes. b, The statistical analysis of BBA at the left and right sensorimotor hand region (bimodal PV minus visual-only) revealed significantly lower beta-band responses for the three emotional facial expression conditions [i.e., fear (F), anger (A), and happiness (H)] than for the neutral (N) expression condition. Horizontal lines are plotted to illustrate the mean values of BBA across the left and right sensorimotor hand regions.
Next, it was examined across participants whether the impact of emotional facial expression on pain ratings, reflected by higher ratings to emotional than to neutral facial expressions, correlated with the effects on the pain-induced BBA. Although no significant correlation between pain ratings (mean of emotional expression conditions minus neutral condition) and BBA suppression (mean of emotional expression conditions minus neutral condition) was found for the left sensorimotor hand region (Pearson's r = −0.23, p = 0.35), a significant positive correlation was found for the right sensorimotor hand region (Pearson's r = 0.51, p < 0.05). Participants with stronger effects of emotional facial expressions on pain ratings revealed larger differences in BBA between emotional and neutral facial expressions.
The general impact of faces on pain-induced BBA was examined in the second level of analysis in which responses to P-only trials were compared with averaged responses across PV conditions. The ANOVA for early BBA using the factors of stimulation (P-only and averaged PV conditions) and ROI (left sensorimotor hand region and right sensorimotor hand region) revealed a significant main effect of ROI, due to stronger amplitude suppression in the contralateral than in the ipsilateral region (F(1,17) = 5.23, p < 0.05). No other significant effects were found.
Effects of facial expressions on total gamma-band activity
The ANOVA to investigate valence-specific effects of facial expressions on early GBA using the factors of facial expression (Fear, Anger, Happiness, and Neutral) and ROI revealed a significant main effect of ROI (F(3,51) = 11.87, p < 0.005). The GBA was significantly larger in contralateral than in ipsilateral sensorimotor cortex (Fig. 5). No other significant effects were found. The ANOVA for the comparison between the GBA of the three emotional facial expression conditions using the factors of facial expression (Fear, Anger, and Happiness) and ROI revealed a significant main effect of facial expression (F(2,34), p < 0.05). Follow-up contrasts between all facial emotion expression pairs, revealed significantly larger pain-induced sensorimotor GBA for the facial expression of fear than for the expression of anger (F(1,17) = 7.39, p < 0.05, corrected for multiple testing). Since no differences in subjective ratings were found between the expressions of fear and anger, we did not compute a correlation analysis between pain ratings and the effects on GBA.
Beamforming analysis of total gamma-band responses for the contrasts of bimodal PV and the corresponding V-only stimuli. a, A robust increase in GBA was found at the right sensorimotor hand region at a short latency of 170 ms (left and right sensorimotor hand regions are highlighted by dotted circles). Plotted are the highest and lowest quartiles of activity changes. b, The statistical analysis of GBA (bimodal PV minus visual only) at the left and right sensorimotor hand region revealed significantly higher responses for the facial expression fear than for the facial expression anger but no significant differences between the three emotional facial expression conditions and the neutral expression condition. Horizontal lines are plotted to illustrate the mean values of GBA across the left and right sensorimotor hand regions.
Next, the general effects of faces on the pain-induced GBA were examined. The ANOVA using the factors of stimulation (P-only vs combined PV) and ROI revealed a significant main effect of ROI (F(1,17) = 24.78, p < 0.0001), indicating larger amplitudes in the contralateral than in the ipsilateral ROI. Furthermore, a significant main effect of stimulation was found (F(1,17) = 16.67, p < 0.001). The GBA was stronger in the combined PV conditions (average of left and right ROIs = 24%) than in the P-only condition (average of left and right ROIs = 14%). The stimulation × ROI interaction was not significant.
Effects of facial expressions on evoked oscillatory responses
The study of differences between the four emotional facial expression conditions on early evoked BBA and GBA at the sensor level using the factors of facial expression (Fear, Anger, Happiness, and Neutral) and ROI (sensors located over left vs right sensorimotor regions) did not reveal any significant main effects or interactions. Furthermore, the sensor level ANOVAs examining the effects of facial expressions on the early evoked BBA by comparing P-only with combined PV trials did not reveal significant main effects or interactions. For the early evoked GBA, however, a significant main effect of stimulation was observed (F(1,17) = 5.77, p < 0.05), indicating stronger responses for combined PV than for P-only trials.
Discussion
This MEG study examined the impact of emotional facial expressions on pain perception and pain processing in the human brain. We observed that the presentation of faces expressing fear, anger, and happiness compared to the presentation of neutral facial expressions led to an increase in subjective pain ratings. Emotional facial expression had a modulatory effect on early oscillatory beta- and gamma-band responses in the sensorimotor cortex.
Pain ratings
Our study revealed that facial expressions with negative and positive valence enhance subjective pain unpleasantness ratings compared to the presentation of faces with neutral expressions. The finding of enhanced pain ratings for faces with negative valence (i.e., fear and anger) is in line with previous reports of increased pain experience when painful stimuli are presented in the context of unpleasant pictures (Godinho et al., 2006; Rhudy et al., 2010). Surprisingly, facial expressions with positive valence (i.e., happiness) also led to enhanced pain ratings. This observation contradicts previous studies showing analgesic effects, i.e., reduced pain ratings or reduced pain perception thresholds, when painful stimuli are concurrently presented with pictures of positive valence (Meagher et al., 2001; Roy et al., 2009). The finding that emotional facial expressions lead to enhanced pain ratings independent of their valence may be linked to the important role of facial expression processing in the human brain, which involves a widespread network of cortical and subcortical structures (Adolphs, 2002; Calder and Young, 2005). Findings from patients with brain lesions (Adolphs et al., 2000), functional magnetic resonance imaging (Winston et al., 2003), and repetitive transcranial magnetic stimulation (Pitcher et al., 2008) demonstrated an involvement of right somatosensory cortex in the processing of emotional facial expression, which is widely independent of the valence of the expression. Thus, the presentation of emotional facial expressions in the present study may have facilitated the processing of painful stimuli in the somatosensory cortex, an assumption that is supported by our findings in oscillatory responses. Furthermore, it may be that painful inputs modulated the perception of happy faces. A recent study showed that painful inputs render pleasant pictures significantly less pleasant (Godinho et al., 2008). A biasing effect of painful stimuli on inputs with positive valence from other sensory modalities may have also contributed to the observation that a lemon scent odor, which was rated as pleasant when presented alone, enhanced pain perception (Martin, 2006). It can be speculated that a happy face presented in combination with a painful stimulus may be interpreted as if this person is laughing about oneself. While the present study focused on the impact of viewing faces with emotional expressions on pain processing, future studies may address how painful inputs modulate the perceived valence and arousal of facial expressions.
In the present study, participants fixated on the faces without an explicit task and rated the unpleasantness of concurrently presented painful inputs. Previous studies showed that in the absence of a demanding secondary task (Pessoa et al., 2002), the decoding of facial expressions occurs rapidly and widely automatically (Vuilleumier et al., 2001; Eimer and Holmes, 2007). Therefore, it is likely that the emotional facial expressions were automatically processed. Our study showed that viewing faces with both positive and negative valence enhances subjective pain perception compared to viewing faces with neutral expressions.
Pain-induced beta-band responses
In line with previous studies using tactile or painful stimuli (Salenius et al., 1997; Bauer et al., 2006; Hauck et al., 2007), the suppression of beta-band responses was localized to the sensorimotor cortex. Due to the widespread distribution of the BBA suppression, we cannot disentangle whether the reported effects mostly reflect processing in the motor cortex or processing in somatosensory areas. Previous studies showed that nociceptive stimuli cause a reduction of ongoing ∼20 Hz oscillations both in the primary motor cortex (Raij et al., 2004) and in somatosensory areas (Ploner et al., 2006a). Thus, it may be that the sensorimotor BBA reflect parallel processing in somatosensory and motor areas, which form a functional network through coupling of neuronal populations (Schnitzler et al., 2000; Brovelli et al., 2002). Notably, beta-band responses have been suggested to modulate the excitability of the sensorimotor cortex. Ploner et al. (2006b) showed that pain-induced suppression of BBA is positively correlated with the excitability of the sensorimotor system. In the present study, the enhanced early suppression of BBA for emotional compared to neutral facial expressions therefore suggests that facial expressions, which represent particularly powerful emotional cues, enhance the sensitivity of pain processing in the sensorimotor cortex.
The strongest pain-induced responses in our study were found in areas of the sensorimotor cortex. Pain processing in sensorimotor cortex primarily reflects the sensory-discriminative dimension of the nociceptive perception, which encompass the perceived intensity, location, and quality of painful events (Bushnell et al., 1999; Lorenz et al., 2005). Therefore, it is likely that participants incorporated the perceived intensity, which has not been separately monitored, in their pain ratings. The observation that pain-induced responses were reflected in early oscillatory responses in the beta- and gamma-band that were localized to the sensorimotor cortex supports this assumption. Together, our data suggest that emotional facial expressions, independent of their valence, increase the sensitivity of pain processing in the sensorimotor cortex through an enhanced suppression of the early pain-induced BBA. Since BBA suppression in the sensorimotor cortex has also been closely associated with the preparation of motor responses (Pfurtscheller and Lopes da Silva, 1999; Neuper et al., 2006), our data indicate that the concurrent presentation of an emotional facial expression while receiving a painful input may facilitate the preparation and execution of adequate protective motor responses.
Another interesting finding is that the effects of emotional facial expressions on subjective pain ratings were correlated in particular with the BBA modulations in the contralateral sensorimotor cortex. This suggests that the observed effects of emotional facial expression on the BBA most likely reflect pain-related processing and not the preparation of the motor response that was required for the rating of the painful stimulus. Since participants rated painful inputs with a joystick in their right hand, confounding effects of motor preparation on oscillatory responses would be expected to occur primarily in the left sensorimotor cortex.
Pain-induced gamma-band responses
In contrast to the BBA, the pain-induced GBA did not differ significantly between the neutral and emotional facial expression conditions. However, the direct comparison of the pain-induced GBA between the different emotion conditions revealed larger responses to faces expressing fear than to faces expressing anger. While both emotions have a negative valence, this finding is noteworthy when considering approach-related models for the processing of anger (Carver and Harmon-Jones, 2009). Recent studies demonstrated that the presentation of angry facial expressions facilitates approach-motivated motor behavior toward angry faces (Wilkowski and Meier, 2010). Angry facial expressions may communicate an intention to confront another person aggressively, thereby facilitating approach-motivated rather than avoidance-motivated behavior. By contrast, acute pain primarily triggers avoidance-motivated behavior (Bromm and Lorenz, 1998). Since the sensorimotor GBA has been, among other functions, linked to the initiation of a motor response (Crone et al., 1998), it can be speculated that the reduced GBA in the sensorimotor cortex may represent the conflict between approach-motivated behavior initiated by the angry facial expressions and avoidance-motivated behavior initiated by the painful inputs. Our observation that the strongest differences in pain-induced GBA were found between fearful and angry facial expressions fits with this interpretation. Similar to painful stimuli, fearful facial expressions have been associated with avoidance-motivated behavior (Davidson, 1998).
Factors that may have contributed to the reported effects of facial expressions on pain-induced oscillatory responses are arousal and attention (Anders et al., 2004; Wiech et al., 2008). Emotional arousal is reflected in modulations of neuronal synchronization across different frequency ranges (Keil et al., 2001; Steriade, 2004). Moreover, functional magnetic resonance studies have shown that attention is positively associated with pain perception and activity of nodes of the pain matrix (Villemure et al., 2009; Ploner et al., 2011). Since attention and arousal have not been explicitly monitored in the present study, it remains to be elucidated to what extent these factors may contribute to the impact of emotional facial expressions on pain processing.
Conclusion
Our study revealed several important findings. Compared to neutral facial expressions, the presentation of emotional facial expressions enhanced the perception of pain. Moreover, we found that both early BBA and GBA in the sensorimotor cortex were modulated by emotional facial expressions. Finally, the observation of a relationship between the effects on early BBA in the sensorimotor cortex and the effects on subjective pain ratings suggests an involvement of pain-induced oscillatory responses in the emotional modulation of pain processing and pain perception.
Notes
Supplemental material for this article is available at http://www.uke.de/institute/neurophysiologie/downloads/institut-neurophysiologie-pathophysiologie/Senkowski_2011_JN_Additional_Material.doc and shows effects of emotional facial expressions on pain-induced lower-frequency (3–9 Hz) activity. This material has not been peer reviewed.
Footnotes
-
This study was supported by grants from the Deutsche Forschungsgemeinschaft (SE 1859/1-2, D.S.; SFB TRR 58 B04, A.K.E.) and the European Union (ERC-2010-StG_20091209, D.S.; IST-2005-27268, NEST-PATH-043457, HEALTH-F2-2008-200728, ERC-2010-AdG-269716, A.K.E.). We thank K. Saha and S. Scheer for recruitment of participants and help with data recordings and Constanze Hipp and Christoph Kayser for their comments on the manuscript.
- Correspondence should be addressed to Dr. Daniel Senkowski, Department of Neurophysiology and Pathophysiology, University Medical Center Hamburg-Eppendorf, Martinistrasse 52, 20246 Hamburg, Germany. dsenkows{at}uke.uni-hamburg.de