Abstract
Atypical scan paths on emotional faces and reduced eye contact represent a prominent feature of autism symptomatology, yet the reason for these abnormalities remains a puzzle. Do individuals with autism spectrum disorders (ASDs) fail to orient toward the eyes or do they actively avoid direct eye contact? Here, we used a new task to investigate reflexive eye movements on fearful, happy, and neutral faces. Participants (ASDs: 12; controls: 11) initially fixated either on the eyes or on the mouth. By analyzing the frequency of participants' eye movements away from the eyes and toward the eyes, respectively, we explored both avoidance and orientation reactions. The ASD group showed a reduced preference for the eyes relative to the control group, primarily characterized by more frequent eye movements away from the eyes. Eye-tracking data revealed a pronounced influence of active avoidance of direct eye contact on atypical gaze in ASDs. The combination of avoidance and reduced orientation into an individual index predicted emotional recognition performance. Crucially, this result provides evidence for a direct link between individual gaze patterns and associated social symptomatology. These findings thereby give important insights into the social pathology of ASD, with implications for future research and interventions.
Introduction
Recent reports from the social-cognitive neurosciences have emphasized the importance of the eyes in carrying crucial information about emotional states of others. In autism spectrum disorders (ASDs), atypical eye contact is not only a diagnostic criterion (American Psychiatric Association, 1994), but also one of the earliest pathological signs, even for high-functioning individuals. Abnormalities in processing of information from the eyes (Leekam et al., 1998; Baron-Cohen et al., 2001a; Spezio et al., 2007a) are characterized by specific scan paths on emotional faces: individuals with ASD spend less time on faces in general (Pelphrey et al., 2002) and in particular focus less on the eye region (Klin et al., 2002). Despite some recent studies on gaze behavior in ASD, it still remains unclear why individuals with ASD show these abnormalities in face processing. A long-standing view suggests a general lack of social attention and specifically less attention toward the eyes as the main reason for the observed scan path on emotional faces in ASD. Some studies even found that individuals with ASD tend to focus more on the mouth region (Klin et al., 2002; Pelphrey et al., 2002). These findings may be taken as support for the missing orientation toward the eyes (Schultz, 2005). Thus, a failure to detect social saliency leads to a reduced orientation toward important sources of social information such as the eyes (Schultz, 2005; Neumann et al., 2006). Another emerging view, however, highlights the potential aversiveness of direct eye contact in ASD, resulting in an active avoidance reaction (Dalton et al., 2005; Kylliainen and Hietanen, 2006). Support for an aversion explanation of atypical gaze comes from studies showing increased skin conductance responses in children with autism when confronted with direct eye contact compared with averted gaze (for recent reviews on gaze and autism, see Nation and Penny, 2008) (Senju and Johnson, 2009). Additionally, another recent study (Joseph et al., 2008) showed that face-encoding skills of children and adolescents with ASD were mediated by emotional arousal in response to direct gaze.
Notably, those two processes do not have to be mutually exclusive, and the interplay of the two components may in fact account for the observed scan paths (Spezio et al., 2007b). Findings to date provide some evidence for each of these explanations; no study so far, however, addressed both processes specifically within the same research design. Here, we sought to further investigate the influence of (1) reduced social orientation to and (2) the active avoidance of the eyes on atypical gaze in ASD by analyzing participants' eye movements while they performed a new behavioral facial emotion classification task (Gamer and Büchel, 2009). We varied the initial fixation position on the face; thus, participants started processing a face either at the eyes or at the mouth and decided whether the face showed a happy, fearful, or neutral expression. Thereby, the task allows investigating both avoidance- and orientation-guided reflexive gaze behaviors, triggered by focusing the eyes or the mouth, respectively.
Materials and Methods
Nineteen neurotypical controls (NTs) (14 male; age range, 21–42 years; mean age, 30.37 years) with no known psychiatric or neurological disorder and 17 participants with ASD (12 male; age range, 22–48 years; mean age, 32.71 years) participated in the current study. Participants in the NT group were recruited through databases of the Max Planck Institute for Human Development (Berlin, Germany). Participants with ASD were recruited through psychiatrists who specialized in autism in the Berlin area and the outpatient clinic for autism in adulthood of the Charité University Medicine, Berlin, Germany. Diagnoses were made according to Diagnostic and Statistical Manual of Mental Disorders, Fourth Edition, criteria for Asperger syndrome and autism without mental retardation (American Psychiatric Association, 1994) using a videotaped semistructured interview, which included the assessment of childhood information. In 14 participants with available parental information, diagnoses were additionally confirmed with the Autism Diagnostic Interview-Revised (ADI-R) (Lord et al., 1994; Bölte and Poustka, 2001) and the Asperger Syndrome (and High-Functioning Autism) Diagnostical Interview (ASDI) (Gillberg et al., 2001). Restricting the eye movement data analysis to only the subsample of ASD participants with ADI-R data revealed the same pattern of results compared with the complete sample though very minor differences in p values for some analyses (most likely because of the reduced sample size). Groups were matched with respect to gender, age, and verbal intelligence quotient (IQ) (see Results). All participants had normal or corrected-to-normal vision, were native German speakers, received payment, and gave written informed consent in accordance with the requirements of the ethics committee of the Max Planck Institute for Human Development.
Visual stimuli were shown using Presentation (Version 12.4., Neurobehavioral Systems), and gaze behavior was measured via an integrated T120 60 HZ eye tracker (Tobii Systems). The eye-tracking device was built into the screen and did not require fixing participants' heads. The device tracks both eyes separately using corneal reflection.
One hundred twenty faces from a standardized dataset (Lundqvist et al., 1998) were chosen for the emotion classification task, based on a recent validation study (Goeleven et al., 2008). Twenty women and 20 men each displayed happy, fearful, and neutral expressions in a randomized order. Each image was rotated to ensure that, when displayed, the eyes were at the same vertical height. Additionally, an elliptic mask was applied resulting in images containing just the face (Fig. 1). All images were converted to grayscale, and the cumulative brightness was normalized across images.
In each trial, a fixation cross was presented initially (2 s), followed by the presentation of a face (150 ms) (Fig. 1). After a blank gray screen (2 s), participants were asked to indicate the emotional expression via button press on a standard keyboard. The emotional classification experiment followed a 2 × 3 within-subjects design with the factors “initial fixation” (eyes, mouth) and “emotion” (happy, fearful, neutral). To investigate the effect of initial fixation on gaze behavior, half of the faces within each emotion category were shifted either downward (Fig. 1, trial A) or upward (Fig. 1, trial B), so that the eyes or the mouth appeared at the location of the formerly presented fixation cross. Thereby, the task allowed the investigation of reflexive gaze behavior as a response to direct eye contact as well as eye movements, when initially fixating the mouth.
Eye movements analysis.
Eye movement behavior and pupil size were recorded at a rate of 60 data points per second (60 Hz), averaged over both eyes. Resulting time series were split into trials starting 1 s before face presentation up to 2 s thereafter. The gaze point coordinates were then convoluted with a Gaussian function (SD, 15 ms), and episodes were determined in which the first derivative of the resulting time series did not exceed 10 pixels (0.3 degree or 18°/s angular speed). Only those trials were included in which participants fixated within 50 pixels (1.5°) of the fixation cross before the onset of the face and that did not contain consecutive missing data points for >100 ms. Missing data points are typically caused by blinks, head motion, poor calibration, and, likely, by fast eye motion. To compare groups and conditions, the remaining valid trials were then categorized as trials containing a fixation change (upward or downward) or as trials not containing a fixation change within 1 s of face presentation. A fixation change was defined as occurring at least 50 pixels (1.5°) away from the original position of the fixation cross in vertical direction. We then calculated the proportion of fixation changes downward (when the eyes were presented at the location of the fixation cross) and upward (when the mouth was presented at the location of the fixation cross) as reflexive gaze responses triggered by the eyes or the mouth. Participants who had <20% valid trials for each of the six conditions were excluded from further analysis. Because of exclusion criteria (e.g., technical failure or participants' movements), the presented results are based on 12 individuals with ASD and 11 controls. Over all six conditions, included participants had, on average, 62.1% valid trials.
Additionally, we calculated an individual eye preference index for difference between the proportion of trials with a fixation change upward from the mouth and the proportion of trials with a fixation change downward from the eyes (Gamer and Büchel, 2009), as well as an individual eye preference control index (number of trials with a fixation change upward from the mouth divided by the number of trials with a fixation change upward from the mouth plus the number of trials with a fixation change downward from the eyes) separately for each emotion and on average over all emotions. Both indices indicate how often participants gazed away from the eyes compared with the mouth. The control index, however, did not take into account the number of trials with no fixation change and therefore validates the eye preference index by controlling for the effect of missing trials. Indeed, we did not expect that trials with a fixation change downward were missed with a different probability than trials with a fixation change upward, but it was possible that trials with fixation changes were missed with a different probability than trials with no fixation changes. If this would have been the case, we expected that those indices would differ considerably. Note that we cannot make any conclusions about the final target of the observed eye movements but only about the directionality attributable to methodological limitations in recording frequency.
Statistical analysis.
Since we were primarily interested in the main effects of the two factors initial fixation and emotion, as well as their interaction within and between groups, any data collected during the task were first analyzed applying a 2 × 3 [initial fixation (eyes vs mouth) × emotion (happy vs fear vs neutral)] repeated-measures ANOVA with between-subject factor group (ASD vs NT). Additionally, post hoc within-group ANOVAs, independent-samples t tests, as well as paired-samples t tests were conducted. All statistics used a significance level of p < 0.05, unless otherwise specified. Data were analyzed using SPSS (version 17.0 for Mac; SPSS Inc.).
Results
Demographics
Groups were matched with respect to gender and age (Table 1). To assure comparable verbal IQ levels between groups, we tested participants with a vocabulary IQ test [multiple choice vocabulary test (MWT)] (Lehrl et al., 1995) (Tables 1, 2). Additionally, we used the Autism Spectrum Quotient (AQ) (Baron-Cohen et al., 2001b) in both groups to control for clinically significant levels of autistic traits in the NT group. None of the controls scored above the cutoff score of 32; in fact, the highest score was 24, indicating a very low level of autistic traits in the NT group (ASD group: mean, 38.3; SD, 8.4; NT group: mean, 13.2; SD, 5.1).
We had valid eye movement data for a subsample of 12 individuals with ASD and 11 control participants (for subsample demographic information, see Table 1). Importantly, the subsample did not differ with respect to gender and verbal IQ, but ASD participants were significantly older than NT participants (Table 1). The factor age did, however, not influence eye movements (F(1,20) = 0.37, p = 0.57) and was therefore excluded from the model.
Behavioral results
Reaction times
We additionally analyzed participants' reaction times during the emotion classification task by condition with respect to the onset of the face stimuli. Individuals with ASD were generally slower in responding than NTs (ASD: 568 ms; NT: 432 ms; t(34) = 2.23; p = 0.03). The above-described 2 × 3 ANOVA with between-subject factor group yielded a significant three-way interaction (F(2,68) = 5.33; p = 0.007; partial η2 = 0.14), as well as a two-way interaction between emotion and initial fixation (F(2,68) = 6.86; p = 0.002; partial η2 = 0.17). Reaction times of the ASD group were influenced by an interaction of emotion and initial fixation (F(2,32) = 7.17; p = 0.003; partial η2 = 0.31), whereas the interaction did not reach significance for the NT group (p = 0.12). Both NT and ASD groups were slower in responding to fearful faces when they started their gaze on the mouth compared with the eyes (ASD: fear-eyes: 561.45; fear-mouth: 645.88; t(16) = −2.19; p = 0.04; NT: fear-eyes: 405.5; fear-mouth: 439.95; t(18) = −2.03; p = 0.057). For neutral faces, however, individuals with ASD were significantly slower in responding to faces when the eyes were presented initially (neutral-eyes: 630.14; neutral-mouth: 503.1; t(16) = 3.05; p = 0.008). Notably, individuals with ASD responded marginally faster to happy faces compared with neutral faces regardless of initial fixation (happy: 536.17; fearful: 603.67; t(16) = −2.02; p = 0.06).
Emotion classification performance
For the emotion classification task, individuals in the NT group showed on average a higher correct classification percentage than individuals in the ASD group (NT: 95.46; ASD: 91.14; t(34) = −2.13; p = 0.04). Performance over both groups was affected by a significant main effect of emotion (F(2,68) = 3.54; p = 0.035; partial η2 = 0.09) and a significant interaction of initial fixation and emotion (F(2,68) = 18.9; p = < 0.001; partial η2 = 0.36). The factor group marginally interacted with initial fixation (F(2,68) = 3.6; p = 0.07; partial η2 = 0.09). Especially in conditions when the mouth was fixated initially, the ASD group showed fewer correct classifications than NTs (across emotions: p = 0.014; happy-mouth: p = 0.055; fear-mouth: p = 0.03).
Both groups were separately influenced by an interaction of emotion and initial fixation (NT: F(2,36) = 6.89; p = 0.003; partial η2 = 0.28; ASD: F(2,32) = 12.14; p < 0.001; partial η2 = 0.43), whereas only individuals in the NT group showed a marginally significant main effect of the factor emotion (F(2,36) = 3.03; p = 0.061; partial η2 = 0.14). For happy compared with neutral faces, controls classified more emotional expressions correctly (happy: 96.58; neutral: 93.36; t(18) = 2.12; p = 0.048). Larger hit rates within the NT group were also observed when the eyes were presented at the fixation cross compared with the mouth across emotions (eyes: 93.46; mouth: 96.49; t(18) = −2.53; p = 0.02). Overall, individuals with ASD were able to discriminate a higher percentage of happy faces compared with fearful faces (happy: 94.37; fearful: 90.39; t(17) = 1.86; p = 0.08). For fearful faces, both participants in the ASD and NT groups could discriminate a higher percentage correctly when they initially fixated the eye region (ASD: p = 0.003; NT: p = 0.076), whereas for neutral faces the hit rates were larger when participants initially fixated the mouth (ASD: p = 0.05; NT: p = 0.02).
Eye Movements results
Analysis of eye movements revealed a three-way interaction of initial fixation, emotion, and group (F(2,42) = 2.89; p = 0.067; partial η2 = 0.12), which shortly fell from significance. The factor initial fixation additionally interacted with the factors group (F(1,21) = 8.03; p = 0.01; partial η2 = 0.28) and emotion (F(2,42) = 6.44; p = 0.004, partial η2 = 0.24). Averaged over emotions (ASD: 70.09%; NT: 30.85%; t(21) = 2.37; p = 0.005), as well as for all emotions separately (all p < 0.05), individuals with ASD gazed significantly more often downward than controls when initially fixating the eyes (Fig. 2A,C). There was no significant difference in the number of fixation changes upward from the mouth to the eyes between groups (all p > 0.1).
Within the NT group, eye movements were mediated by a significant main effect of initial fixation (F(1,10) = 5.06; p = 0.048; partial η2 = 0.34) and a significant two-way interaction of initial fixation and emotion (F(2,20) = 7.82; p = 0.003; partial η2 = 0.44). The ASD group showed only a trend for this interaction (p = 0.1). Generally, controls oriented their gaze to a greater proportion upward toward the eyes (when initially fixating the mouth) than downward away from the eyes (when initially fixating the eyes) (upward: 57.83%; downward: 30.85%; t(10) = −2.25; p = 0.048). The size of the effect was further modulated by the emotional expression, most prominently for fearful faces (upward: 60.74%; downward: 23.48%; t(10) = −2.97; p = 0.014) and neutral faces (upward: 60.81%; downward: 29.05%; t(10) = −3.02; p = 0.013), but not for happy faces (p = 0.43). In contrast to the controls, the ASD group showed a trend toward a greater proportion of fixation changes downward away from the eyes than upward from the mouth toward the eyes (upward: 46.78%; downward: 70.09%; t(11) = −1.8; p = 0.1), regardless of the emotion displayed.
To further investigate eye movements with respect to a possible interaction of the two initial fixation positions, we computed an index for eye preference for each participant (the difference between the proportion of trials with a fixation change upward from the mouth and downward from the eyes). Overall, the NT group's eye movements indicated a greater preference for the eyes than the ASD group's eye movements (averaged over emotions: p = 0.01; fearful: p = 0.002; neutral: p = 0.008; happy: p = 0.068). Again, only the NT group's gaze patterns were affected by the emotional expression: the index was significantly smaller for happy than for fearful faces (happy: 11.89; fear: 37.26; t(10) = −3.17; p = 0.01) and neutral faces (happy: 11.89; neutral: 31.76; t(10) = −2.74; p = 0.01) (Fig. 2B).
The number of valid trials was influenced by interactions of group with initial fixation (F(1,21) = 10.57; p = 0.004; partial η2 = 0.33) and marginally by emotion (F(2,42) = 3.15; p = 0.053; partial η2 = 0.13). The ASD group had more valid trials when initially fixating the eyes across emotions (p = 0.02), especially for happy faces (p = 0.009). To check whether the observed differences in number of valid trials may account for the effects on gaze behavior between groups described earlier, we calculated an additional individual index (eye preference control index), which accounts for differences in number of valid trials between trials with and without fixation changes (see Eye movements analysis). The individual values of the original eye preference index and the control index were highly correlated (Pearson's r = 0.96; p < 0.001, 2-tailed), thereby reflecting the same pattern of effects. Consequently, the reported between-group differences in fixation changes were not a result of differences in the number of valid trials but represent actual differences in gaze patterns on emotional faces between the ASD and NT groups.
To further approximate the differences between avoidance and orienting response, we further analyzed the latency (time between occurrence of facial stimuli and offset of first fixation change) of participants' fixation changes. Although not reaching statistical significance, the ASD group showed a tendency toward faster eye movements away from the eyes than the NT group (averaged over emotions: ASD mean, 449 ms; NT mean, 508 ms) (Table 2), whereas the ASD group gazed more slowly upward toward the eyes (when starting fixation on the mouth) compared with the NT group (averaged over emotions: ASD mean, 437 ms; NT mean, 404 ms). For both directions, the effects were most pronounced for fearful faces, but almost absent for happy faces (away from eyes, fearful: ASD: 437 ms; NT: 497 ms; happy: ASD: 438 ms, NT: 437 ms; toward eyes, fearful: ASD: 508 ms, NT: 402 ms; happy: ASD: 428 ms, NT: 438 ms), indicating a differential response to facial emotion. In line with previous reports, fearful faces, and especially wide-open eyes, seem more arousing than happy faces, which do not indicate a potential threat in the environment (Morris et al., 1996). Importantly, individuals with autism gazed away from the eyes faster than they gazed away from the mouth for fearful faces (eyes: mean, 437 ms; mouth: mean, 508 ms), while controls showed the opposite pattern (eyes: mean, 497 ms; mouth: mean, 402 ms). In sum, those data indicate that avoidance of the eyes is indeed more pronounced in individuals with ASD because they gaze away from the eyes faster than the NT group and do so relatively faster than gazing away from the mouth toward the eyes, especially for fearful faces.
Interestingly, and in line with an interpretation of aversion, the more often ASD participants gazed away from the eye region, the faster these eye movements occurred (averaged over emotions: Pearson's r = −0.93, p < 0.001, 2-tailed). Contrarily, there was no such correlation between latency and fixation changes upward from the mouth toward the eyes in ASD (correlations differed significantly; Fisher's z to p transformation: z = 2.89, p = 0.004). The NT group, in turn, showed no modulation of number of fixation changes by latency at all.
The individual eye preference index was furthermore associated with autistic symptomatology as well as emotion recognition performance. For the ASD group, the eye preference index correlated positively with individual overall performance in the emotion classification task (Pearson's r = 0.66; p = 0.019, 2-tailed). This effect was absent for the NT group. To further assess whether general illness severity rather than socioemotional impairments account for the observed ASD-specific gaze patterns, we investigated the influence of the ADI-R total score (as a measure of illness severity) versus the influence of emotion recognition performance in a stepwise regression analysis on the eye preference index as the dependent variable. The emotion recognition performance was a significant predictor of eye movement patterns (R2 change = 0.74; F change(1,7) = 20.2; β = 0.4, p = 0.003), whereas the ADI-R total score was excluded from the model. Additionally, the eye preference index correlated negatively with the ADI-R social score (Pearson's r = −0.7; p = 0.033, 2-tailed), whereas there was no correlation for the ADI-R communication score (p > 0.13) and ADI-R behavior score (p > 0.56). Thus, the greater the preference for the eye region in the ASD group, the less pronounced the social symptomatology, indicating a link between gaze pattern and social impairments but not general illness severity in autism. In addition, there was no relationship of verbal IQ and group specific gaze patterns for both groups (ASD, p = 0.12; NT, p = 0.89), nor for the AQ scores (ASD, p = 0.14, NT: p = 0.56). Taking both the performance and diagnostic correlative relationships on the gaze patterns in ASD together, the greater the preference for the eye region in the ASD group, the higher the percentage of correct emotion classification and the less severe the social symptomatology.
Analysis of additional eye movement parameters revealed that there were no global group differences in the average y-axis offset during presentation of the fixation cross (250 ms before up to 50 ms after face presentation; ASD: 1.1 pixels; NT: 4.6 pixels; t(21) = −1.55; p = 0.14), the average duration (ASD: 908 ms; NT: 1011 ms; t(16) = −1.27; p = 0.22), distance (ASD: 93.9 pixel; NT: 89.44 pixel; t(21) = 3.95; p = 0.7), and latency of fixation changes (ASD: 428 ms; NT: 459 ms; t(21) = −7.78; p = 0.44). Additionally, the number of trials with a fixation change downward from the mouth or upward from the eyes did not differ significantly from 0. Those very few trials (<0.5% of the trials per participant on average) were excluded from any further analysis. There were no group differences in x-axis offsets of fixation changes from the mouth upward toward the left or right half of the screen (i.e., toward the left or right eye) (ASD: −3.5 pixels; NT: −0.9 pixel; t(21) = −0.32; p = 0.75).
Discussion
The main goal of this study was to investigate the influence of diminished social orientation and active avoidance on the reduced focus to the eyes in ASD. The ASD group showed an overall reduced preference for the eye region compared with the NT group, which was specifically the result of an increase of fixation changes away from the eyes. Further analyses revealed group-specific directional effects when combining both components: the NT group shifted their gaze significantly more often toward the eyes than away from the eyes, replicating previous results in typically developed individuals (Gamer and Büchel, 2009). Conversely, individuals with ASD showed more fixation changes away from the eyes than toward the eyes (Fig. 2).
These results can be seen as support for the hypothesis of increased avoidance of eye contact in ASD. We are aware, however, that both avoidance and orienting responses likely coexist at any time, yet to varying degrees. For example, it is likely that an initial reflexive orientation is present in ASD (as indicated by the number of fixation changes toward the eyes), but that it is then attenuated by an avoidance of eye contact (e.g., fixations do not reach or the eyes). The combination of both forces might furthermore interact differentially in both groups, as assessed with the eye preference index. A clear disentangling of both processes can therefore not be performed based on the number of fixation changes alone. In fact, gaze away from the eyes might indicate avoidance, but it might also indicate an increased interest in the mouth, because, for example, it might be more informative.
We thus performed analyses on additional eye movement parameters, which further support the pronounced influence of active avoidance of eye contact in ASD. Though only on a descriptive level, the ASD group showed a tendency to gaze away from the eyes faster than the NT group and, importantly, did so relatively faster than gazing away from the mouth toward the eyes. This effect was most pronounced for fearful faces, indicating differential responses to facial emotion. In line with previous reports, fearful faces—and especially eyes—are more arousing than happy faces, which do not indicate a potential threat in the environment (Morris et al., 1996). Statistically reliable correlations underline these patterns, since the more often ASD participants gazed away from the eyes, the faster these eye movements occurred. Contrarily, there was no such correlation between latency and gaze away from the mouth in ASD and also no such relationship in the NT group.
Furthermore, if individuals with ASD gaze away from the eyes because the mouth is more informative, one would expect longer fixation times on the mouth to have a beneficial effect on emotion recognition performance. In contrast, we found the reverse to be true: the greater the eye preference index (i.e., the fewer fixations on the mouth), the more correct emotions were classified by the autistic individuals. Moreover, emotions differ distinctively with respect to their diagnostic features across the face (Smith et al., 2005). For fearful and neutral faces, the eye region may contain the most important information about the emotion, whereas for happy faces, the mouth might be more distinctive. Although the number of fixation changes was modulated by the emotional expression for the NT group, which replicates previous findings (Gamer and Büchel, 2009), the ASD group showed no such effect. Also, if the interest in the mouth was greater (than in the eyes), eye movements away from the mouth might occur less often and would be slower for certain emotions, where the mouth is most informative, such as for happy faces. However, the ASD group gazed away from the mouth faster for happy faces than for fearful faces, albeit only on a descriptive level. Thus, the lack of benefit from fixations toward the mouth, together with the missing modulatory effect of facial emotion on gaze can be interpreted as further support of the hypothesis of active avoidance of the eyes in ASD.
It has been suggested that avoidance of the eyes may be mediated by an increased emotional arousal (Kylliainen and Hietanen, 2006). Thus, it is important to mention that even typically developed participants seem to experience facial expressions as more intense when they focus on the eye region (Gamer and Büchel, 2009). With respect to the mechanisms of typically developed gaze behavior; however, the reported increase in intensity has been suggested to underscore the saliency of the eye region. The saliency, in turn, leads to an increased orientation toward this important region of the face (when starting to gaze on another prominent feature, such as the mouth). The increase in experienced intensity as a response to direct eye contact thus represents a rather adaptive arousal in neurotypical controls, enhancing social cognitive functioning. The results of the current study show that direct eye contact leads to differential reflexive gaze behavior within the framework of atypical gaze in ASD: instead of persisting gaze at the eyes (to collect important information about the emotional state of others), individuals with ASD gazed more frequently away from the eyes, independent of the displayed emotional expression. In line with previous findings (see above), these gaze patterns could be interpreted as maladaptive consequences of arousal, oppositional to the arousal effect on controls. Thus, direct eye contact has a rather positive effect on social cognition in controls, whereas this effect is reversed for individuals with ASD. The opposite directional group effects in overall eye preference underline this hypothesis: in the present study, ASD specific gaze was associated with performance and social symptomatology but not with impairments in repetitive behavior and communication. Thus, these relationships further illustrate how the combination of avoidance and orientation reflect specific socioemotional dysfunctioning within the heterogeneous pathophysiology of ASD, rather than general illness severity.
In sum, we used a novel task to further disentangle avoidance and orientation-related reflexive gaze behavior. We found that atypical gaze in ASD was prominently influenced by an increased number of reflexive eye movements away compared with eye movements toward the eye region in ASD. Within the context of previous literature (Hutt and Ounsted, 1966; Richer and Coss, 1976; Dalton et al., 2005; Kylliainen and Hietanen, 2006; Joseph et al., 2008), more and at the same time faster eye movements away from the eyes compared with away from the mouth in ASD favor the assumption of a pronounced avoidance of eye contact compared with reduced orientation. We acknowledge that future studies shall apply variations in task design such as using inverted faces, directing initial fixations to the nose, or showing only specific facial features instead of the whole face. Moreover, subtle latency characteristics have to be further investigated with higher temporal resolution of eye movement recordings. Possible limitations of the current study arise from repeated reports of basic visual-processing differences in autism in the literature (Minshew et al., 1999; Goldberg et al., 2002; Luna et al., 2007; Thakkar et al., 2008). We did not find such general group differences, suggesting that the observed gaze patterns in ASD represent socioemotional dysfunctioning in ASD, rather than global differences in oculomotor functioning. There is, however, a clear need for further investigations in particular in relation to brain abnormalities (for cerebellar pathology, see Nowinski et al., 2005).
The results of the current study complement research on the role of the amygdala in autism (Baron-Cohen et al., 2000). In typically developed controls, the amygdala seems to be not only involved in orienting toward socially salient stimuli (Gamer and Büchel, 2009; Gamer et al., 2010), but also in fear and aversion processing (LeDoux et al., 1988; Adolphs et al., 1998, 2005; Phelps et al., 2004; Pessoa et al., 2005a,b; Seymour et al., 2005; Anders et al., 2008; Asghar et al., 2008; Hadjikhani et al., 2008). A behavioral dissociation of aversion- and orientation-related gaze in ASD provides an opportunity to simultaneously identify and dissociate sociocognitive functions potentially subserved by different amygdalar nuclei (Ball et al., 2007; Hurlemann et al., 2008; Gamer et al., 2010). With respect to the goal to gain a comprehensive understanding of autism pathophysiology, the present results provide crucial information for future research investigating atypical gaze in relation to brain function. Finally, interventions and behavioral training studies may benefit from considering these results: although strategies motivating attention toward social salient cues seem generally justified, special attention should be directed to the reduction of the aversive effect of direct eye contact.
Footnotes
This work was supported by the German Federal Ministry of Education and Research and the Max Planck Society. We thank Jennifer Kirchner for assistance in establishing diagnoses and Hannah Bruehl for valuable comments on the manuscript.
- Correspondence should be addressed to Dorit Kliemann, Cluster of Excellence, “Languages of Emotion,” Freie Universität Berlin, Habelschwerdter Allee 45, 14195 Berlin, Germany. dorit.kliemann{at}fu-berlin.de