Original Research ARTICLE
Electrophysiological Correlates of Subliminal Perception of Facial Expressions in Individuals with Autistic Traits: A Backward Masking Study
- Centre for Human Psychopharmacology, Swinburne University of Technology, Melbourne, VIC, Australia
People with Autism spectrum disorder (ASD) show difficulty in social communication, especially in the rapid assessment of emotion in faces. This study examined the processing of emotional faces in typically developing adults with high and low levels of autistic traits (measured using the Autism Spectrum Quotient—AQ). Event-related potentials (ERPs) were recorded during viewing of backward-masked neutral, fearful and happy faces presented under two conditions: subliminal (16 ms, below the level of visual conscious awareness) and supraliminal (166 ms, above the time required for visual conscious awareness). Individuals with low and high AQ differed in the processing of subliminal faces, with the low AQ group showing an enhanced N2 amplitude for subliminal happy faces. Some group differences were found in the condition effects, with the Low AQ showing shorter frontal P3b and N4 latencies for subliminal vs. supraliminal condition. Although results did not show any group differences on the face-specific N170 component, there were shorter N170 latencies for supraliminal vs. subliminal conditions across groups. The results observed on the N2, showing group differences in subliminal emotion processing, suggest that decreased sensitivity to the reward value of social stimuli is a common feature both of people with ASD as well as people with high autistic traits from the normal population.
Autism spectrum disorder (ASD) is a neurodevelopmental disorder characterized by difficulties in reciprocal social interaction and a restricted range of interests and behaviors (American Psychiatric Association, 2000, 2013). It has been proposed that individuals with ASD have difficulties in the processing of facial expressions, particularly for complex or negative emotional expressions (Adolphs et al., 2001; Castelli, 2005; Golan et al., 2007). However, this finding is not uniform (e.g., Ozonoff et al., 1990; Piggot et al., 2004; Castelli, 2005). Several theoretical models have been proposed for explaining atypical socio-emotional behavior in autism (see reviews, Gaigg, 2012; Hamilton, 2013), and one of them, the Amygdala Theory of Autism, suggests that dysfunction or damage of the amygdala region leads to social impairments in autism (Baron-Cohen et al., 2000). Although there is substantial experimental support for atypical amygdala activation in individuals with autism (e.g., Critchley et al., 2000; Ashwin et al., 2006, 2007), some studies have questioned whether the amygdala plays a specific role of emotional recognition in autism (e.g., Paul et al., 2010; Birmingham et al., 2011; also see Zalla and Sperduti, 2013). Recently, impairments in functional networks as an explanation for socio-cognitive difficulties in autism have been receiving more support in the literature than dysfunction of a single brain region (Kana et al., 2011). For example, Di Martino et al. (2009) showed that autistic traits in neurotypical adults were related to the pregenual anterior cingulate cortex connectivity with insula, and more specifically, this connectivity being limited to the anterior mid-insula rather than anterior insula per se. Previous studies showed the importance of those regions for social function, particularly emotional processing, with the pregenual anterior cingulate cortex playing a crucial role in thinking about others’ thoughts and beliefs (Amodio and Frith, 2006; Gilbert et al., 2006), while insula is involved in the processing of sensations and emotions (Singer, 2006).
The amygdala has a central role in the processing of facial emotions in healthy populations. Although it is considered to respond primarily to threatening facial expressions, especially of fearful faces (Morris et al., 1996), it was also found to be involved in the processing of non-threatening facial expressions, such as happy and sad faces (Breiter et al., 1996; Wright et al., 2002; Yang et al., 2002). The amygdala is considered to have an essential role in a vigilance system for rapidly alerting other brain regions to the importance of social stimuli (LeDoux, 1996; Whalen, 1998; Schultz et al., 2000), suggesting that in autism, if amygdala function is disrupted, faces may be less socially salient, leading to reduced experience with emotional facial stimuli (Schultz et al., 2000; Hall et al., 2007). Based on those findings, several additional theoretical models of autism have been proposed in order to explain socio-emotional difficulties in autism. For example, the Relevance Detection Theory (Sander et al., 2003) suggests that the amygdala is part of an extended cortico-limbic system that is important for recognizing cues in environmental stimuli that immediately orient a person towards dangers or safety in the environment. In addition, the Social Motivation Theory of autism (Dawson et al., 2005) has proposed that abnormal social motivation prevents orienting to socially relevant stimuli including faces. This suggests that a lack of experience with faces during critical developmental periods in individuals with autism, leads to difficulties in expert and automatic processing of faces.
The amygdala can be engaged subconsciously by presenting images of facial emotions with very short duration and masking them (Morris et al., 1998; Whalen et al., 1998; Liddell et al., 2005), rendering them outside of conscious awareness or by presenting them under conditions of binocular suppression (Williams et al., 2004). One popular, although not universally accepted (see Pessoa and Adolphs, 2010), model proposes that fear-related responses are processed through a direct subcortical pathway comprising the amygdala, the superior colliculus and the pulvinar nucleus of the thalamus, for fast, but coarse analysis of potential threat that has adaptive survival value (LeDoux, 1996, 2000; Vuilleumier et al., 2003; Tamietto and de Gelder, 2010; Tamietto et al., 2012).
Research support for subliminal processing of emotional faces has been found in studies of affective subliminal priming (Murphy and Zajonc, 1993; Monahan et al., 2000; Nomura et al., 2004; Finkbeiner and Palermo, 2009; Jiang et al., 2013), and with cortically blind patients (de Gelder et al., 1999). For example, de Gelder et al. (1999) showed that a patient with right hemianopia, or blindness in the right visual field due to damage to his left occipital lobe, was able to guess the facial emotional expression at a level above chance, even if stimuli were not seen consciously.
An important research method that is often used to examine subliminal automatic responses is the backward masking paradigm (for a review of this paradigm, see Breitmeyer and Ogmen, 2000). Esteves and Öhman (1993) and Öhman and Soares (1994) were among the first researchers to use this paradigm in emotion research, with Öhman and Soares (1994) using common phobic objects, such as snakes, as stimuli, whereas Esteves and Öhman (1993) used emotional faces as stimuli. In the backward masking paradigm, there is a very brief presentation of the face stimuli followed by a mask that blocks the conscious recognition of a stimulus. The interval between the onset of the target and masking stimuli, the stimulus-onset-asynchrony (SOA), was the principal factor in correctly perceiving the target stimulus within this paradigm (Esteves and Öhman, 1993).
EEG backward-masking studies have demonstrated an ability to measure temporal processing of emotional stimuli presented below the level of visual awareness in healthy participants. Liddell et al. (2004) suggested that subliminal and supraliminal emotion processing could be distinguished with the N2/early P3 components representing orienting and N4/late P3 event integration, based on the Halgren and Marinkovic (1995) model of emotion processing. The results of their study supported this model, showing larger the N2 amplitude and faster P3a latencies for subliminally presented fearful faces compared to neutral faces, and larger P3b amplitude in response to supraliminal fearful faces (Liddell et al., 2004). This indicates that event-related potential (ERP) components can be used to show a double dissociation for subliminal vs. supraliminal processing of fearful facial expressions (Liddell et al., 2004). The finding of enhanced N2 for subliminal fearful faces was initially criticized by Pegna et al. (2008) for using a passive viewing task, citing Kiss and Eimer (2008) who suggested that the passive task may prevent participants from attending to masked stimuli. Contrary to Liddell et al. (2004), Pegna et al. (2008) found an increased N2 for fearful compared to non-fearful faces at longer durations of presentation (supraliminal rather than subliminal condition) by using an active emotion recognition task. However, a subsequent study (Pegna et al., 2011) indicated that a difference between subliminally presented fearful and non-fearful faces can be observed on both the early N170 component and also elicit posterior positivity and fronto-central negativity at around 200–250 ms, even when participants’ attention is engaged in an incidental task. Thus, the N2 peak is a component worth observing with regard to subliminal emotion processing. Following these results, the present study investigated both the N2/P3a and the N4/P3b while processing emotion.
However, several EEG backward-masking studies have reported earlier emotional differentiation in the subliminal condition. For example, Kiss and Eimer (2008) and Eimer et al. (2008) found an enhanced frontal positivity for both subliminal and supraliminal fearful faces compared to neutral faces between 140 ms and 180 ms post-stimulus. Pegna et al. (2008) examined processing of fearful and non-fearful (happy and neutral) faces at subliminal, intermediate and supraliminal stimulus durations of 16 ms, 33 ms, 66 ms, 133 ms and 266 ms. This study found emotion discrimination for subliminal faces in the N170 component, showing increased N170 responses to fearful compared to non-fearful faces for all three durations (subliminal, intermediate and supraliminal) over temporal electrodes.
The N170 is considered a face-specific component reflecting structural encoding of faces and represents the earliest stages of face processing (Bentin et al., 1996; Itier and Taylor, 2004). As there is evidence of atypical responses on the N170 ERP component in children and adults with autism compared to those without autism (e.g., McPartland et al., 2004; O’Connor et al., 2005, 2007; Hileman et al., 2011), component was also included in the design of the present study. There is mixed evidence about modulation of the N170 by facial expressions in both typically developing subjects (e.g., Batty and Taylor, 2003, 2006; Blau et al., 2007; Eimer et al., 2008; Luo et al., 2010) and subjects with autism (Dawson et al., 2004; O’Connor et al., 2005; Wong et al., 2008; Batty et al., 2011).
While there are several behavioral reports of impaired processing of briefly presented emotional faces in autism (e.g., Kamio et al., 2006; Hall et al., 2007), there is a relative lack of EEG studies on automatic face processing in the disorder. Recently, Fujita et al. (2013) measured visual evoked potentials (VEPs) elicited by subliminally presented fearful and neutral faces and objects in the upright and inverted position and found group differences between individuals with ASD and typically developing individual in the earliest VEP component (N1), indicating altered early visual processing of briefly presented emotional faces in the disorder.
Backward masking has been used in several fMRI studies investigating subliminal processing of emotional and social information in autism. While deficits were found in subliminal processing in ASD (Kamio et al., 2006; Hall et al., 2007; Kleinhans et al., 2011), other studies did not find differences between subjects with autism and healthy controls in the amygdala activation during sub-threshold presentation of facial expressions (Hall et al., 2010). However, additional evidence for atypical subliminal face processing in autism is found in psychophysiological studies using facial electromyography (EMG). Facial EMG studies with neurotypical subjects suggest that the observation of others’ emotional facial expressions automatically produces similar facial expressions or facial mimicry in the observer (Sato et al., 2013). However, atypical automatic facial mimicry to backwardly masked briefly presented happy and angry facial expressions was found in adults with ASD (Mathersul et al., 2013). Impaired recognition of briefly presented (but not backwardly masked) happy and angry faces was also found in young adults with ASD (Clark et al., 2008). Studying facial emotion processing in autism is important because they are crucial for social functioning and social communication (Grelotti et al., 2002; Dawson et al., 2005; Golan et al., 2007). In addition, understanding of automatic emotion processing in the disorder can provide better understanding and clarification of specificity of emotion processing in this group. The importance of understanding various aspects of emotion processing is particularly evident from a recent study (Tseng et al., 2016) that showed differences in neural activity for arousal but not valence on emotion processing between subjects with autism and neurotypically developing subjects.
The aim of the present study was to assess, using EEG, the processing of subliminal and supraliminal fearful and happy facial expressions by investigating differences in high and low autistic traits as measured by the Autism Spectrum Quotient (AQ; Baron-Cohen et al., 2001). Following the model (Liddell et al., 2004) that considers N2/early P3 components representing “orienting” and N4/late P3 “event integration”, our task was to investigate both sets of ERPs. Based on those findings, we predicted more prominent activity over N2 and P3a (early P3) components as a response to subliminal fearful and happy facial stimuli, and more prominent activity over N4 and P3b (late P3) ERP components as a response to supraliminal fearful and happy facial stimuli. We suggest that, based on previous fMRI findings of deficits in subliminal processing in autism (Kamio et al., 2006; Hall et al., 2007; Kleinhans et al., 2011), we would find group differences predominantly in subliminal condition (or orienting stages of face processing). This stage of emotion processing is important for creation of conscious emotional experience.
The investigation of individuals with high and low autistic traits can have some advantages compared to working with individuals with autism, particularly related to heterogeneity of autism and differing diagnostic criteria used. In research with facial emotional stimuli, this group of participants can exclude differences due to extensive training with face stimuli that individuals with autism can have through various intervention programs (e.g., Herbrecht et al., 2009), and that are rarely mentioned in research studies on face processing.
Materials and Methods
The participants were selected from a total population of 94 individuals who completed online surveys including the AQ and the Empathy Quotient (EQ). Based on their AQ scores, 26 participants (all right-handed, 13 females) were selected to participate in the EEG study (see Table 1). The selection procedure consisted of creating two groups of participants: those who achieved scores ≤10 (Low AQ) and ≥22 (High AQ) on the AQ questionnaire (total sample N = 94, median score = 15, SD = 7.8). The AQ cut off scores in this study are similar to some other studies (e.g., Sutherland and Crewther, 2010). Demographic information of participants was also collected online. All participants had a normal or corrected-to-normal vision, with no neurological impairment (including clinical autism). They gave informed signed consent to participate, and all experimental procedures were approved by the Swinburne University Human Research Ethics Committee (SUHREC). All participants volunteered to be part of the study. Participants consented to the study after carefully reading the Consent forms for this study, which were approved by the SUHREC. No vulnerable populations were involved.
All participants completed online questionnaires: the AQ and the EQ. Participants also completed the Advanced Raven’s Progressive Matrices before or after EEG testing.
The AQ is a self-administered questionnaire that consists of 50 questions, devised to quantitatively measure the degree to which a person with normal intelligence has autistic traits (Baron-Cohen et al., 2001; Woodbury-Smith et al., 2005). Participants respond using a 4-point rating scale, from “definitely agree” to “definitely disagree”.
The EQ is a self-administered questionnaire that consists of 40 questions assessing empathy (Baron-Cohen and Wheelwright, 2004) with high test-retest reliability (Lawrence et al., 2004). Lower scores on the EQ have been found in adults with autism (Baron-Cohen and Wheelwright, 2004) and in neurotypical men compared to women (Baron-Cohen and Wheelwright, 2004; Lawrence et al., 2004).
The Raven’s Advanced Progressive Matrices (RAPM) is a standardized nonverbal intelligence test, and is generally used as a test of general cognitive ability and intelligence (Raven, 2000). It consists of visually presented geometric figures where one part is missing and the missing part must be selected from a panel of suggested answers to complete the designs. In the present study, we used the RAPM with a time limit of 20 min, to eliminate general intelligence as a potential explanation of any differences found between AQ groups.
The stimuli consisted of grayscale photographs of the faces of 18 Caucasian models (nine male, nine female). The models’ faces depicted neutral, fearful and happy expressions (with both open and closed mouth exemplars) and were cropped with an oval shape, removing external features. The facial images were taken from the NimStim set (Tottenham et al., 2009), and masks were created by phase scrambling images of neutral faces using MatLab. The use of phase-scrambled images as a mask was based on past neuroimaging studies (e.g., Jacques and Rossion, 2007; Rousselet et al., 2007; Schultz and Pilz, 2009) that have used phase-scrambled stimuli because they make faces unrecognizable without altering their original power spectrum, luminance and contrast.
Subjects sat in an electrically shielded, dimly-lit and sound-attenuated room in front of a computer screen. The experiment was programmed with E-Prime 1.2 (Psychology Software Tools, Inc., Pittsburgh, PA, USA). Stimuli were presented in 8 blocks of 138 trials, each block consisting of a randomized presentation of both subliminal and supraliminal faces. Block order was counterbalanced across participants. Before starting the experimental procedure, participants were given a practice run. At the beginning of the experiment, a white fixation cross appeared in the middle of the screen, lasting for 700 ms. Shortly thereafter, a picture of a face stimulus was displayed for duration of 16 ms (subliminal condition) or 166 ms (supraliminal condition), immediately followed by the mask for 284 ms for subliminal presentation or 134 ms for supraliminal presentation, in order to keep the presentation time constant for 300 ms (see Figure 1). Stimuli were presented on a 24″ color LCD monitor driven at 100 Hz vertical refresh rate.
At the end of each trial, a question appeared on the screen asking for explicit emotion recognition, showing numbers 1–3 that were put next to the written label for each of the three facial expressions, with a new trial started only after participant response. Participants were allowed unlimited time to press an answer and were asked to always press the answer with the right hand. In the case of subliminal stimuli (stimuli showed below the threshold of visual awareness), participants were asked to guess the facial expression. The explicit recognition task was adopted because it gives equal importance to all facial expressions in both conditions. There was an equal number of trials in each condition for each facial expression (120 trials for each facial expression, for each condition).
EEG activity was recorded using a Neuroscan amplifier (Compumedics USA, Charlotte, NC, USA) from 64 electrodes, placed in accordance with the International 10–20 system. The amplification was set at 1000× EEG signals were band pass filtered 0.05–70 Hz with a sampling rate of 500 Hz. Electrode impedances were kept below 5 kΩ. The vertex (Cz) electrode was used as a reference. Recordings were re-referenced to the average reference as computed from all scalp electrodes (for the N170 component), and to the average of mastoids (for N2, P3a, P3b, N4). This re-referencing method of the average of mastoids used for the latter components was chosen because it is used most frequently for endogenous ERP components. However, mastoid re-reference can be problematic for the N170 because this ERP is usually largest over lateral posterior regions, close to mastoids (Luck, 2005). It was found that the common average reference yielded the largest N170 amplitude and the smallest amplitude at the vertex positive potential (VPP; Joyce and Rossion, 2005). On the other hand, it has been suggested that using a mastoid reference eliminates the specificity of the N170 response, yielding the smallest peaks at the N170 sites and the largest peaks at VPP sites (Schendan et al., 1998; Rossion et al., 2003; Joyce and Rossion, 2005). EOG was recorded from two electrodes placed at the external canthi of both eyes and from two electrodes on the infraorbital and supraorbital areas of the left eye to monitor for eye movements and blinks. The raw data were segmented into epochs of −200 to 800 ms around stimulus events. Trials in which the amplitude exceeded ±100 μV were automatically rejected, eliminating eye blinks and other movements. Trials with EOG artifacts (e.g., eye blinks or large eye movement) were also removed, with additional artifacts excluded upon visual inspection. Only participants with more than 50% of artifact-free trials in each condition were included in the final analysis. ERPs were averaged separately for each stimulus category (each emotion was averaged for subliminal and supraliminal threshold conditions), baseline corrected and low-pass filtered at 30 Hz (24 dB/octave).
Independent samples t-tests were conducted to compare groups on EQ scores and RAPM.
In the behavioral analysis we reported the accuracy rates and due to unlimited time given to participants for recognizing facial expressions, we did not report the reaction times (RTs).
The behavioral analysis consisted of a series of one tailed t-tests comparing accuracy to chance levels for the two conditions and for each of the emotions within the conditions. This was followed by the repeated measures ANOVA with emotion and condition as a within-subjects factors and the AQ group as a between-subjects factor.
Additional estimates of detection measure (d′) and response bias (c) (Macmillan and Creelman, 1991) were computed separately for each participant for each of the facial expressions in each condition. The signal detection is considered a particularly suitable behavioral analysis for the experimental procedure where participants need to guess as it shows inflated hit rates. Contrary to this method, the response accuracy percentage is known to be highly sensitive to response bias (Macmillan and Creelman, 1991). We performed repeated measures ANOVAs for the d′ values and the response bias c with emotion and condition as a within-subjects factors and the AQ group as a between-subjects factor.
The time windows for ERP components and grand averages were selected based on previous literature (Liddell et al., 2004; Kiss and Eimer, 2008; Pegna et al., 2008). The N170 ERP was examined at the lateral occipito-temporal sites P7 and P8, and N2, P3a, P3b and N4 components were examined at midline electrodes Fz, Cz, and Pz. The peak amplitudes and latencies were measured in the following latency windows: N170 (140–220 ms), N2 (180–300 ms), P3a (240–350 ms), P3b (400–700 ms) and N4 (300–500 ms). The amplitude and latency of each ERP component were quantified by the highest peak value within the chosen latency window.
ERP amplitude and latency were analyzed with repeated-measures ANOVA using the AQ Group as the between-subject factor, with emotion (neutral, happy, fearful), condition (subliminal, supraliminal), and hemisphere (left and right, only for N170)/electrode (Fz, Cz, Pz) as within-subject factors. Degrees of freedom were adjusted (Greenhouse-Geisser epsilon) for factors with greater than two levels. Paired-samples t-tests were performed to supplement the ERP findings. An alpha criterion level of 0.05 was used unless otherwise noted.
The independent samples test was used to compare groups on EQ scores. The study found significantly higher EQ scores for the Low AQ group (EQ = 53.50) compared to the High AQ group (EQ = 35.67; see Table 1), (t(24) = 3.995, p = 0.001). This result, indicating lower scores on the self-report empathy test in typically developing individuals with higher AQ, supplements previous evidence showing that the empathy difficulties, indicated by lower EQ scores, are also found in parents of children with autism, particularly fathers (Sucksmith et al., 2013). Sucksmith et al. (2013) suggested that lower scores on the EQ test may represent a reliable feature of the broad autism phenotype in fathers.
In addition, a Pearson’s correlation was run to determine the relationship between AQ and EQ scores. The results showed a very strong, negative correlation between the AQ and EQ (r = −0.81, N = 26, p < 0.0001; Figure 2). This finding again complements a previous research showing that the EQ is inversely correlated with the AQ (Baron-Cohen and Wheelwright, 2004).
Figure 2. A Pearson’s correlation between Autism Spectrum Quotient (AQ) and Empathy Quotient (EQ) scores. The results show a very strong, negative correlation between the AQ and EQ (r = −0.81, p < 0.0001).
There was no evidence for between-group differences on the RAPM (t(24) = 0.742, p = 0.47).
The analysis revealed an accuracy rate of 57.1% in the subliminal and 92.6% in the supraliminal conditions (above chance (33.3%) accuracy rate for both conditions, both p = 0.0001). Response accuracy was above chance (33.3%) for all facial expressions in supraliminal condition (all p < 0.000). For facial expressions in subliminal condition, accuracy rates were above chance for neutral and fearful faces (t(25) = 8.74, p = 0.0001 [neutral] and t(25) = 6.04, p = 0.0001 [fearful]), but it was below chance for happy faces (t(25) = 1.75, p = 0.09). Table 2 shows mean accuracy results as percentages across both groups and for each AQ group.
Table 2. Accuracy rates (%) across both groups (N = 26) and for Low AQ (N = 14) and High AQ (N = 12) groups.
An additional analysis applied the repeated measures ANOVA with emotion and condition as a within-subjects factors and the AQ group as a between-subjects factor. The results showed significant effect of Condition (F(1,24) = 255.5, p = 0.0001), indicating higher accuracy rates in the supraliminal condition. The results also found significant effects of Emotion (F(2,38) = 9.45, p = 0.001 and Emotion × Condition interaction (F(2,39) = 10.16, p = 0.001). The paired t-tests showed lower accuracy rates for happy compared to neutral and fearful faces across both conditions (all p = 0.001; 66.19% for happy, 77.6% for fearful and 88.44% for neutral faces. We also applied the repeated measures ANOVA on each condition and found a significant effect of Emotion (F(2,38) = 10.23, p = 0.001) in the subliminal condition, with happy faces showing lower accuracy rates compared to neutral (p = 0.002) and fearful (p = 0.0001) facial expressions (40.93% for happy, 62.82% for fearful and 67.18% for neutral faces). There was no main effect of AQ group for accuracy rates.
Across both AQ groups, the d′ values for trials in both conditions were significantly greater than zero (in subliminal condition d′ was 1.23 (SD = 0.66), t(25) = 9.55, p = 0.0001; in supraliminal condition the d′ values were 4.12, (SD = 0.9), t(25) = 23.39, p = 0.0001), suggesting a non-random response on both conditions.
The repeated measures ANOVA was applied for the d′ values. The analysis revealed the main effect of Condition (F(1,24) = 445.82, p = 0.0001), indicating larger d′ values for supraliminal than subliminal condition. In addition, significant effects of Emotion (F(2,37) = 25.18, p = 0.0001) and Emotion × Condition interaction (F(2,48) = 7.88, p = 0.001) were also found. The paired t-test indicated significant differences between all emotions, particularly showing larger d′ values for happy compared to neutral (p = 0.0001) and fearful faces (p = 0.01) and larger d′ values for fearful compared to neutral (p = 0.0001) faces. When each condition was analyzed separately, the results showed a main effect of Emotion (F(2,48) = 9.31, p = 0.0001) in the subliminal condition, showing lower d′ values for neutral faces compared to happy (p = 0.01) and fearful faces (p = 0.0001).
The repeated measures ANOVA was also applied for the response bias c. The results showed significant effects of Condition (F(1,24) = 62.72, p = 0.0001), with larger c for subliminal (mean = 0.391) than supraliminal condition (mean = 0.172). The main effects of Emotion (F(2,37) = 7.7, p = 0.003) and Emotion × Condition interaction (F(2,39) = 10.62, p = 0.001) were also found. The paired t-test revealed the higher c for happy compared to neutral (p = 0.001) and fearful faces (p = 0.03). An additional analysis for each condition revealed a main effect of Emotion (F(1,33) = 13.27, p = 0.0001) in the subliminal condition, with higher c for happy compared to neutral and fearful faces (all p = 0.0001). There were no significant Emotion effects in the supraliminal condition. These results indicate that participants were less likely to report happy faces than other faces in subliminal condition.
There was no main effect of AQ group either for the detection d′ values or response bias c.
The planned analysis for N2, P3a, P3b and N4 ERP components included analysis both across all regions and for each midline region. The final ERP analysis included both correct and incorrect behavioral responses.
N2 Amplitude and Latency
The analysis of the N2 amplitude data showed a significant effect of Region (F(2,38) = 30.35, p = 0.0001), indicating that the N2 was larger in the frontal (Fz) than central (Cz) and parietal (Pz) regions (both p = 0.0001) and also larger in the central (Cz) than parietal (Pz) (p = 0.02) region.
The results of N2 amplitude analysis showed a significant Emotion × Condition × AQ (F(2,47) = 3.54, p = 0.04) interaction. Separate ANOVAs were performed for each condition and a significant Emotion by AQ group interaction (F(2,48) = 5.2, p = 0.01) was found only in the subliminal condition. Further analysis showed a main effect of Emotion in the Low AQ (F(2,26) = 3.52, p = 0.05), with larger N2 amplitudes for happy compared to neutral faces (p = 0.03). No significant effects were found for N2 amplitudes for the High AQ group (see Figures 3A,B, 4).
Figure 3. Grand-average Event-related potential (ERP) waveforms of N2 amplitudes for Low AQ (A) and High AQ (B) groups (across all regions). The Low AQ group shows larger amplitudes for happy than neutral faces in the subliminal condition. The High AQ group does not show emotional differentiation.
Figure 4. Topographic maps for subliminal faces. Topographic maps show the cortical activities during 180–300 ms (N2) for subliminally presented fearful minus neutral and happy minus neutral faces in Low and High AQ groups. An increased negativity (blue color) is found for the Low AQ group for happy minus neutral faces predominantly in fronto-contral regions.
The N2 latency analysis showed a main effect of Condition (F(1,24) = 28.01, p = 0.0001), indicating that subliminal stimuli elicited shorter N2 latencies than supraliminal stimuli.
P3a Amplitude and Latency
No significant effects were found for P3a amplitudes.
The P3a latency analysis showed a main effect of Condition (F(1,24) = 18.15, p = 0.0001), indicating significantly shorter P3a latencies in the supraliminal than in the subliminal condition.
The results for P3a latencies also showed a significant Emotion × Condition × Region interaction (F(4,96) = 2.46, p = 0.05). This interaction was further examined by conducting an additional ANOVA for each region and results showed an Emotion × Condition interaction (F(1,48) = 3.62, p = 0.05) in the frontal (Fz) region. Further analysis in this region found a main effect of Emotion (F(2,38) = 3.76, p = 0.04) in the supraliminal condition, with a shorter P3a latency for happy compared to neutral faces (p = 0.05). In addition, there was a significant effect of Emotion (F(2,48) = 3.16, p = 0.05) in the central (Cz) region indicating shorter P3a latencies for happy compared to neutral faces (p = 0.05) across both conditions.
P3b Amplitude and Latency
The main effect of Condition was significant for both P3b amplitudes (F(1,24) = 17.2, p = 0.0001) and P3b latencies (F(1,24) = 28.63, p = 0.0001), indicating larger amplitudes and shorter latencies in the supraliminal than in the subliminal condition.
The results of P3b latencies showed a significant Condition × Region × AQ interaction (F(2,48) = 7.67, p = 0.001). Further analysis found the main effect of Condition (F(1,13) = 10.1, p = 0.007) in the frontal (Fz) region and in the Low AQ group only, indicating a significantly shorter P3b latency in the supraliminal than in the subliminal condition. There was no significant effect of Condition in the frontal (Fz) region in the High AQ group (F(11) = 1.62, p = 0.23).
N4 Amplitude and Latency
There were no significant main effects for N4 amplitudes.
The analysis for the N4 latency found a significant Condition × AQ interaction (F(1,24) = 4.83, p = 0.04). Separate ANOVAs for each group found a significantly shorter N4 latency under supraliminal than under subliminal condition only in the Low AQ group (Condition effect; F(1,13) = 7.53, p = 0.02). The Low AQ group also showed a significant Condition × Region interaction (F(2,26) = 5.42, p = 0.02), and further analysis indicated that the shorter N4 latency under the supraliminal condition is mostly observed in frontal (Fz; F(1,13) = 12.43, p = 0.004), and central (Cz; F(1,13) = 5.6, p = 0.03) regions.
Statistical significant differences between subliminal and supraliminal condition across all participants and for each ERP can be seen in the Table 3.
N170 Amplitude and Latency
The analysis on the N170 amplitude data showed a significant effect of Hemisphere (F(1,24) = 9.06, p = 0.01), which was due to a greater negativity over the right (−7.812 μV) than the left (−5.456 μV) hemisphere (Figure 5 shows hemisphere lateralization for each group). No such effect was observed for N170 latencies.
Figure 5. Grand-average ERP waveforms showing the N170 ERP for Low AQ (A) and High AQ (B) groups. There are larger amplitudes in the right hemisphere for both groups. However, the figures show reduced hemisphere lateralization in the High AQ.
The analysis of N170 peak latencies revealed a main effect of Condition, with shorter supraliminal than subliminal latencies (F(1,24) = 7.36, p = 0.01).
No other main effects were significant for the N170.
The main aim of this study was to examine the processing of subliminal facial expression reflecting difference emotional states in individuals with higher and lower autistic tendencies. Partial support was found for the main hypothesis that group differences would emerge between individuals with lower and higher autistic traits in emotion effect under subliminal viewing conditions. This effect was found only for subliminally presented happy faces in the frontal region on the N2 component, showing emotional differentiation in the Low AQ group, but not in the High AQ group. No group differences were found for processing of subliminal fearful faces, contrary to our expectations. However, recently Smith (2011) examined processing of a larger number of subliminal facial expressions within a backward-masking paradigm and confirmed processing of subliminal facial expressions at the N2 without finding evidence that the emotional modulation on this component in subliminal condition is specifically related to fearful faces as found in earlier studies (Liddell et al., 2004; Kiss and Eimer, 2008). Our findings of differential emotion processing on the N2 certainly can contribute to the present discussion on subliminal processing in this component.
There are discrepancies within the backward-masking literature regarding the N2 component and emotion modulation at subliminal and supraliminal levels. Liddell et al. (2004) and Kiss and Eimer (2008) found differences between facial expressions on N2 only in the subliminal condition, whereas Pegna et al. (2008) found such differences to occur in the supraliminal condition. Some other studies found modulations in the supraliminal condition on this component by using masked line drawings as face stimuli (e.g., Wilenius-Emet et al., 2004). However, although initially considering that the N2 represents early responding to conscious emotional information, Pegna et al. (2011) in a later study, confirmed the suggestion that the N2 represent an automatic, non-conscious attention-orienting response to emotionally relevant stimuli (Liddell et al., 2004). Although the present study provides support for the emotion modulation in subliminal condition on the N2, this was found only in the Low AQ group and for happy but not for fearful faces. However, the N2 component latency was shorter for subliminal vs. supraliminal stimuli across both AQ groups.
The N2 component has been linked to the activity of the amygdala (Halgren and Marinkovic, 1995; Krolak-Salmon et al., 2004) and of anterior cingulate cortex (van Veen and Carter, 2002), adding to the importance of this component when researching autism. Both of these regions are part of the brain’s limbic system and destructive lesions of any of these regions could lead to social and communicative impairments. An updated view of the amygdala theory of autism proposes that the brain circuit, in which the amygdala occupies a crucial place, is responsible for the detection of a larger category of biologically relevant stimuli, acting as a relevance detector and giving priority to salient signals, based on the motivation and contextual goals of the perceiver (Sander et al., 2003; Zalla and Sperduti, 2013). The social relevance detector account posits that although the amygdala is able to process social information under the unaware condition, its prime role is bringing to conscious awareness salient stimuli through emotional arousal (Vuilleumier and Schwartz, 2001). Hence, according to the relevance detection theory of autism, there is reduced top-down control and attentional modulation performed by the vMPFC in this group, leading to the inability of this prefrontal area to form salience maps for giving priority to specific environmental stimuli. As happy faces have been found to activate reward circuitry in neurotypical individuals (Phillips et al., 1998; O’Doherty et al., 2003), findings in the present study may indicate decreased sensitivity to reward value of social stimuli not only in people with autism compared to typically developing controls but also in individuals with higher autistic traits compared to individuals with lower autistic traits. In this way, our results based on processing of emotional faces presented below the level of visual awareness complement studies that showed specificity in responding to happy facial stimuli in people with autism (e.g., Sepeta et al., 2012) and neurotypical participants with high AQ (e.g., Gayle et al., 2012; Sims et al., 2012).
However, it should be added that given the evidence for magnocellular dysfunction in autism (McCleery et al., 2007; Sutherland and Crewther, 2010), an inability to process visual information correctly from much earlier in cortical processing should also be taken into consideration (Kveraga et al., 2007). The magnocellular pathway is more sensitive to stimuli of lower spatial frequencies (Merigan and Maunsell, 1993), has a faster conduction speed than the parvocellular pathway (Schroeder et al., 1989; Maunsell et al., 1999) and dominates input to the dorsal cortical stream. The parvocellular pathway, on the other hand, is more sensitive to stimuli of higher spatial frequencies (Merigan and Maunsell, 1993) and conveys higher spatial frequency information to the ventral cortical visual stream (Merigan and Maunsell, 1993). Activation of the pulvinar, amygdala and superior colliculus (making a subcortical visual pathway) is restricted to low spatial frequencies but enables a fast yet coarse processing of visual information, bypassing visual cortex (Morris et al., 1996, 1999). Kveraga et al. (2007) showed that fast magnocellular pathways connect early visual and object recognition regions with the orbitofrontal cortex via the dorsal stream and, in this way, facilitate object recognition by activating early predictions about objects. They found that the amygdala receives a substantially greater magnocellular input than parvocellular input and also suggested that their finding of increased right amygdala activation for M-biased stimuli is similar to findings of greater increase of right amygdala for masked fearful faces presented below awareness level to healthy subjects (Morris et al., 1999).
Understanding the role of sensory/afferent differences is of great importance for autism research. Sutherland and Crewther (2010) nonlinear flash VEP study showed atypical magnocellular processing in neurotypical adults with high AQ scores. This was based on the identification of short interaction time second order kernels as of magnocellular derivation (Kaplan and Shapley, 1986) through factors of high contrast gain, saturation at high contrast, and short latency (Klistorner et al., 1997; Jackson et al., 2013). In addition, McCleery et al. (2007) showed abnormal processing of luminance contrast in infants who were at a risk for autism, suggesting an abnormal magnocellular pathway at early stages of development which could have detrimental effects on neural areas that receive input from the magnocellular pathway.
We also found group differences for longer latencies, showing shorter N4 latencies under supraliminal than subliminal condition for the Low AQ group but not for the High AQ group. This component is considered to represent event integration, according to previous studies (Liddell et al., 2004). This finding indicates that in later stages of emotion processing AQ groups differ mostly on condition effect but not on emotion effect. It also brings into attention a question about possible differences in processing of subliminal or briefly-presented face stimuli based on autistic tendencies. We could suggest that it is needed a further study that would look more carefully into the threshold of briefly presented face stimuli in people along the spectrum of autistic traits before giving conclusive answer about emotion processing in subliminal and supraliminal condition.
The present study did not find any emotion modulation on later ERP component representing an event integration stage of emotion processing. Similarly, Pegna et al. (2008) did not find an effect of emotion on the N4 component, but only an effect of stimulus duration (condition), similar to the present study. In the present study, we found emotional effect on the P3a latency. Results for this component showed shorter P3a latencies for happy compared to neutral faces in the supraliminal condition over the frontal region, but shorter P3a latencies for happy compared to neutral faces across both conditions in the central condition. This result put into the question a possible role of the P3a in the orientation stage of emotion processing, particularly showing that the P3a showed shorter latencies in supraliminal compared to subliminal condition.
Although we adopted objective criteria for visual awareness, we did not find any group differences in behavioral analysis. In addition, personal reports of participants at the end of each block revealed that they were not able to see whole faces or recognize facial expressions when faces were presented in the subliminal condition, although some of them were able to notice eyes or mouths in the subliminal condition in some of the trials. However, behavioral results showed that both subliminal and supraliminal conditions in both groups had above chance accuracy rates, with significantly larger accuracy rates for the supraliminal condition. This brings to attention research on general anesthesia where, for example, there is an important unresolved question about the possibility of partial conscious perception (Ghoneim, 2000; Daunderer and Schwender, 2004; see review in Ghoneim et al., 2009). In addition, Pessoa et al. (2005) showed in a behavioral study that when backward masking fearful facial expressions there was no universal objective awareness threshold perception of fearful emotional expressions among subjects.
In the present study, the face-specific N170 component did not show emotion discrimination either in supraliminal or subliminal condition and no group differences were observed for this component. This is in line with previous findings of absent emotional modulation for this component in a backward-masking paradigm (Kiss and Eimer, 2008). However, several backward-masking studies (Pegna et al., 2008; Smith, 2011) found that the N170 is affected by facial expressions in both subliminal and supraliminal conditions. Pegna et al. (2008) suggested that conflicting results of N170 modulation by emotional faces might be due to different references that were used. For instance, Pegna et al. (2008) and Smith (2011) used average reference whereas Kiss and Eimer (2008) used linked earlobes as a reference. In the present study we used average reference, which is suggested to be more appropriate for investigating the N170 as linked earlobes are close to the N170 location. It is important to remember that the present study examined only subjects with high and low AQ, excluding the medium AQ group, which normally makes a great part of typical population.
The results of the present study also revealed a main effect of hemisphere (i.e., right hemisphere dominance) and faster N170 latency for stimuli presented under supraliminal vs. subliminal conditions. The N170 component is usually more prominent over the right hemisphere in typically developing individuals (Rossion et al., 2003; Jacques et al., 2007; Dalrymple et al., 2011). It is interesting that the present study shows enhanced right hemisphere activation for both supraliminal and subliminal condition, indicating that the right hemisphere maintains its prominence under very briefly presented facial stimuli. We did not find any interaction of emotion and hemisphere, but Pegna et al. (2008) found enhanced N170 for fearful compared to non-fearful faces in the right hemisphere in both subliminal and supraliminal conditions.
It is easy to notice big discrepancies in results among backward masking studies. We could mention several methodological factors that can affect modulation of ERP component in these studies such as, for example: the choice of a mask, such as using a pattern by scrambling neutral faces (our study; Kiss and Eimer, 2008; Smith, 2011) or a neutral face (e.g., Liddell et al., 2004; Eimer et al., 2008; Pegna et al., 2008); an active emotion recognition task (the present study, similar to Eimer et al., 2008; Kiss and Eimer, 2008; Pegna et al., 2008; Smith, 2011) or a passive viewing task (Liddell et al., 2004), etc. We believe that our study adds some important information and understanding into subliminal face processing, including understanding of emotional processing in subclinical autistic traits. However, it is also necessary to take into consideration that the study has some methodological limitations, particularly the number of participants. Although all our participants confirmed absence of formal autism diagnosis, we had two participants with AQ score above 32, which was previously found in adults identified with autism. For example, Baron-Cohen et al. (2001) found 80% of adults with autism scored above AQ score 32, whereas only 2% of controls did so whereas Woodbury-Smith et al. (2005) found that AQ score of 32 predictably identify 76% of people diagnosed with autism when using AQ in clinical sample. However, previous research also found that scientists and those in occupations/skills such as maths, physics and engineering had higher AQ scores than the mean (Baron-Cohen et al., 1998, 2001), and many of participants of our study were students in some science and computer oriented degree.
In addition, in the light of recent findings on the P3 component, the design of the present study could be improved to include computing reaction times, as an important factor in order to measure P3 as an index of decision process due to recent findings that showed the peak latency P3 coincides with response execution (Twomey et al., 2015).
In conclusion, the present study demonstrated that individuals with low and high AQ differ in the processing of subliminal happy faces, finding increased N2 amplitude for subliminally presented happy facial expressions only in the Low AQ group (although this finding is of moderate significance, we still consider it important), but not in the High AQ group. This suggests that these differences may be based on the reduced sensitivity to social salient stimuli in individuals with High AQ.
Each author contributed to the development of the final manuscript. DPC and JC supervised this research which was part of a research PhD dissertation (SV) funded by a NHMRC grant through DPC (NHMRC project ID 1004740). DPC designed the concept proposal and JC supervised and assisted with the EEG design and analysis components. The first draft was written by SV as part of her research dissertation.
This project was funded by the National Health and Medical Research Committee (NHMRC project ID 1004740) and the Swinburne University Postgraduate Support Scheme and the Brain and Psychological Science Research Centre support program.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Ashwin, C., Baron-Cohen, S., Wheelwright, S., O’Riordan, M., and Bullmore, E. T. (2007). Differential activation of the amygdala and the ‘social brain’ during fearful face-processing in Asperger Syndrome. Neuropsychologia 45, 2–14. doi: 10.1016/j.neuropsychologia.2006.04.014
Ashwin, C., Chapman, E., Colle, L., and Baron-Cohen, S. (2006). Impaired recognition of negative basic emotions in autism: a test of the amygdala theory. Soc. Neurosci. 1, 349–363. doi: 10.1080/17470910601040772
Baron-Cohen, S., Bolton, P., Wheelwright, S., Short, L., Mead, G., Smith, A., et al. (1998). Autism occurs more often in families of physicists, engineers, and mathematicians. Autism 2, 296–301. doi: 10.1177/1362361398023008
Baron-Cohen, S., Ring, H. A., Bullmore, E. T., Wheelwright, S., Ashwin, C., and Williams, S. (2000). The amygdala theory of autism. Neurosci. Biobehav. Rev. 24, 355–364. doi: 10.1016/S0149-7634(00)00011-7
Baron-Cohen, S., and Wheelwright, S. (2004). The empathy quotient: an investigation of adults with Asperger syndrome or high functioning autism and normal sex differences. J. Autism Dev. Disord. 34, 163–175. doi: 10.1023/b:jadd.0000022607.19833.00
Baron-Cohen, S., Wheelwright, S., Skinner, R., Martin, J., and Clubley, E. (2001). The autism spectrum quotient (AQ): evidence from Asperger syndrome/high-functioning autism, males and females, scientists and mathematicians. J. Autism Dev. Disord. 31, 5–17. doi: 10.1023/A:1005653411471
Batty, M., Meaux, E., Wittemeyer, K., Rogé, N., and Taylor, M. J. (2011). Early processing of emotional faces in children with autism: an even related potential study. J. Exp. Child Psychol. 109, 430–444. doi: 10.1016/j.jecp.2011.02.001
Birmingham, E., Cerf, M., and Adolphs, R. (2011). Comparing social attention autism and amygdala lesions: effects of stimulus and task condition. Soc. Neurosci. 6, 420–435. doi: 10.1080/17470919.2011.561547
Blau, V. C., Maurer, U., Tottenham, N., and McCandliss, B. D. (2007). The face-specific N170 component is modulated by emotional facial expression. Behav. Brain. Funct. 3, 1–13. doi: 10.1186/1744-9081-3-1
Breiter, H. C., Etcoff, N. L., Whalen, P. J., Kennedy, W. A., Rauch, S. L., Buckner, R. L., et al. (1996). Response and habituation of the human amygdala during visual processing of facial expression. Neuron 17, 875–887. doi: 10.1016/s0896-6273(00)80219-6
Clark, T. F., Winkielman, P., and McIntosh, D. N. (2008). Autism and the extraction of emotion from briefly presented facial expressions: stumbling at the first step of empathy. Emotion 8, 803–809. doi: 10.1037/a0014124
Critchley, H., Daly, E., Phillips, M., Brammer, M., Bullmore, E., Williams, S., et al. (2000). Explicit and implicit neural mechanisms for processing of social information from facial expressions: a functional magnetic resonance imaging study. Hum. Brain Mapp. 9, 93–105. doi: 10.1002/(SICI)1097-0193(200002)9:293::AID-HBM4>3.3.CO;2-Q
Dalrymple, K. A., Oruç, I., Duchaine, B., Pancaroglu, R., Fox, C. J., Iaria, G., et al. (2011). The anatomic basis of the right face-selective N170 IN acquired prosopagnosia: a combined ERP/fMRI study. Neuropsychologia 49, 2553–2563. doi: 10.1016/j.neuropsychologia.2011.05.003
Dawson, G., Webb, S. J., Carver, L., Panagiotides, H., and McPartland, J. (2004). Young children with autism show atypical brain responses to fearful versus neutral facial expressions of emotion. Dev. Sci. 7, 340–359. doi: 10.1111/j.1467-7687.2004.00352.x
Dawson, G., Webb, S., and McPartland, J. (2005). Understanding the nature of face processing impairment in autism: insights from behavioral and electrophysiological studies. Dev. Neuropsychol. 27, 403–424. doi: 10.1207/s15326942dn2703_6
de Gelder, B., Vroomen, J., Pourtois, G., and Weiskrantz, L. (1999). Non-Conscious recognition of affect in the absence of striate cortex. Neuroreport 10, 3759–3763. doi: 10.1097/00001756-199912160-00007
Di Martino, A., Shehzad, Z., Kelly, C. A. M., Roy, A. K., Gee, D. G., Uddin, L. Q., et al. (2009). Relationship between cingulo-insular functional connectivity and autistic traits in neurotypical adults. Am. J. Psychiatry 166, 891–899. doi: 10.1176/appi.ajp.2009.08121894
Esteves, F., and Öhman, A. (1993). Masking the face: recognition of emotional facial expressions as a function of the parameters of backward masking. Scand. J. Psychol. 34, 1–18. doi: 10.1111/j.1467-9450.1993.tb01096.x
Finkbeiner, M., and Palermo, R. (2009). The role of spatial attention in nonconscious processing: a comparison of face and non-face stimuli. Psychol. Sci. 20, 42–51. doi: 10.1111/j.1467-9280.2008.02256.x
Fujita, T., Kamio, Y., Yamasaki, T., Yasumoto, S., Hirose, S., and Tobimatsu, S. (2013). Altered automatic face processing in individuals with high-functioning autism spectrum disorders: evidence from visual evoked potentials. Res. Autism Spectr. Disord. 7, 710–720. doi: 10.1016/j.rasd.2013.03.001
Gayle, L. C., Gal, D., and Kieffaber, P. D. (2012). Measuring affective reactivity in individuals with autism spectrum personality traits using the visual mismatch negativity event-related brain potential. Front. Hum. Neurosci. 6:334. doi: 10.3389/fnhum.2012.00334
Ghoneim, M. M., Block, R. I., Haffarnan, M., and Mathews, M. J. (2009). Awareness during anesthesia: risk factors, causes and sequelae: a review of reported cases in the literature. Anesth Analg. 108, 527–535. doi: 10.1213/ane.0b013e318193c634
Gilbert, S. J., Spengler, S., Simons, J. S., Steele, J. D., Lawrie, S. M., Frith, C. D., et al. (2006). Functional specialization within rostral prefrontal cortex (area 10): a meta-analysis. J. Cogn. Neurosci. 18, 932–948. doi: 10.1162/jocn.2006.18.6.932
Golan, O., Baron-Cohen, S., Hill, J. J., and Rutherford, M. D. (2007). The ‘Reading the Mind in the Voice’ test-revised: a study of complex emotion recognition in adults with and without autism spectrum conditions. J. Autism Dev. Disord. 37, 1096–1106. doi: 10.1007/s10803-006-0252-5
Grelotti, D. J., Gauthier, I., and Schultz, R. T. (2002). Social interest and the development of cortical face specialization: what autism teaches us about face processing. Dev. Psychobiol. 40, 213–225. doi: 10.1002/dev.10028
Hall, G. B. C., Doyle, K. A. R., Goldberg, J., West, D., and Szatmari, P. (2010). Amygdala engagement in response to subthreshold presentations of anxious face stimuli in adults with autism spectrum disorders: preliminary insights. PLoS One 5:e10804. doi: 10.1371/journal.pone.0010804
Herbrecht, E., Poustka, F., Birnkammer, S., Duketis, E., Schlitt, S., Schmötzer, G., et al. (2009). Pilot evaluation of the Frankfurt Social Skills Training for children and adolescents with autism spectrum disorder. Eur. Child Adolesc. Psychiatry 18, 327–335. doi: 10.1007/s00787-008-0734-4
Hileman, C. M., Henderson, H. A., Mundy, P., Newell, L. C., and Jaime, M. (2011). ]Developmental and individual differences on the P1 and N170 ERP components in children with and without autism. Dev. Neuropsychol. 36, 214–236. doi: 10.1080/87565641.2010.549870
Jackson, B. L., Blackwood, E. M., Blum, J., Carruthers, S. P., Nemorin, S., Pryor, B. A., et al. (2013). Magno- and parvocellular contrast responses in varying degrees of autistic trait. PLoS One 8:e66797. doi: 10.1371/journal.pone.0066797
Jacques, C., and Rossion, B. (2007). Early electrophysiological responses to multiple face orientations correlate with individual discrimination performance in humans. Neuroimage 36, 863–876. doi: 10.1016/j.neuroimage.2007.04.016
Joyce, C., and Rossion, B. (2005). The face-sensitive N170 and VPP components manifest the same brain processes: the effect of reference electrode site. Clin. Neurophysiol. 116, 2613–2631. doi: 10.1016/j.clinph.2005.07.005
Kamio, Y., Wolf, J., and Fein, D. (2006). Automatic processing of emotional facial faces in high-functioning pervasive developmental disorders: an affective priming study. J. Autism Dev. Disord. 36, 155–165. doi: 10.1007/s10803-005-0056-z
Kana, R. K., Libero, L. E., and Moore, M. S. (2011). Disrupted cortical connectivity theory as an explanatory model for autism spectrum disorders. Phys. Life Rev. 8, 410–437. doi: 10.1016/j.plrev.2011.10.001
Kaplan, E., and Shapley, R. M. (1986). The primate retina contains two types of ganglion cells, with high and low contrast sensitivity. Proc. Natl. Acad. Sci. U S A 83, 2755–2757. doi: 10.1073/pnas.83.8.2755
Kleinhans, N. M., Richards, T., Johnson, L. C., Weaver, K. E., Greenson, J., Dawson, G., et al. (2011). fMRI evidence of neural abnormalities in the subcortical face processing system in ASD. Neuroimage 54, 697–704. doi: 10.1016/j.neuroimage.2010.07.037
Klistorner, A., Crewther, D. P., and Crewther, S. G. (1997). Separate magnocellular and parvocellular contributions from temporal analysis of the multifocal VEP. Vision Res. 37, 2161–2169. doi: 10.1016/s0042-6989(97)00003-5
Krolak-Salmon, P., Hénaff, M.-A., Vighetto, A., Bertrand, O., and Mauguiere, F. (2004). Early amygdala reaction to fear spreading in occipital, temporal, and frontal cortex: a depth electrode ERP study in human. Neuron 42, 665–676. doi: 10.1016/s0896-6273(04)00264-8
Lawrence, E. J., Shaw, P., Baker, D., Baron-Cohen, S., and David, A. S. (2004). Measuring Empathy—reliability and validity of the empathy quotient. Psychol. Med. 34, 911–919. doi: 10.1017/s0033291703001624
Liddell, B. J., Brown, K. J., Kemp, A. H., Barton, M. J., Das, P., Peduto, A., et al. (2005). A direct brainstem-amygdala-cortical ‘alarm’ system for subliminal signals of fear. Neuroimage 24, 235–243. doi: 10.1016/j.neuroimage.2004.08.016
Liddell, B. J., Williams, L. M., Rathjen, J., Shevrin, H., and Gordon, E. (2004). A temporal dissociation of subliminal versus supraliminal fear perception: an event-related potential study. J. Cogn. Neurosci. 16, 479–486. doi: 10.1162/089892904322926809
Luo, W., Feng, W., He, W., Wang, N. Y., and Luo, Y. J. (2010). Three stages of facial expression processing: ERP study with rapid serial visual presentation. Neuroimage 49, 1857–1867. doi: 10.1016/j.neuroimage.2009.09.018
Mathersul, D., McDonald, S., and Rushby, J. A. (2013). Automatic facial responses to briefly presented emotional stimuli in autism spectrum disorder. Biol. Psychol. 94, 397–407. doi: 10.1016/j.biopsycho.2013.08.004
Maunsell, J. H., Ghose, G. M., Assad, J. A., McAdams, C. J., Boudreau, C. E., and Noerager, B. D. (1999). Visual response latencies of magnocellular and parvocellular LGN neurons in macaque monkeys. Vis. Neurosci. 16, 1–14. doi: 10.1017/s0952523899156177
McCleery, J. P., Allman, E., Carver, L. J., and Dobkins, K. R. (2007). Abnormal magnocellular pathway visual processing in infants at risk for autism. Biol. Psychiatry 62, 1007–1014. doi: 10.1016/j.biopsych.2007.02.009
McPartland, J., Dawson, G., Webb, S. J., Panagiotides, H., and Carver, L. J. (2004). Event related brain potentials reveal anomalies in temporal processing of faces in autism spectrum disorder. J. Child Psychol. Psychiatry 45, 1235–1245. doi: 10.1111/j.1469-7610.2004.00318.x
Morris, J. S., Frith, C. D., Perrett, D. I., Rowland, D., Young, A. W., Calder, A. J., et al. (1996). A differential neural response in the human amygdala for fearful and happy facial expressions. Nature 383, 812–815. doi: 10.1038/383812a0
Murphy, S. T., and Zajonc, R. B. (1993). Affect, cognition and awareness: affective priming with optimal and suboptimal stimulus exposures. J. Pers. Soc. Psychol. 64, 723–739. doi: 10.1037/0022-3518.104.22.1683
Nomura, M., Ohira, H., Haneda, K., Iidaka, T., Sadato, N., Okada, T., et al. (2004). Functional association of the amygdala and ventral prefrontal cortex during cognitive evaluation of facial expressions primed by masked angry faces: an event-related fMRI study. Neuroimage 21, 352–363. doi: 10.1016/j.neuroimage.2003.09.021
O’Connor, K., Hamm, J. P., and Kirk, I. J. (2005). The neurophysiological correlates of face processing in adults and children with Asperger’s syndrome. Brain Cogn. 59, 82–95. doi: 10.1016/j.bandc.2005.05.004
O’Connor, K., Hamm, J. P., and Kirk, I. J. (2007). Neurophysiological responses to face, facial regions and objects in adults with Asperger’s syndrome: an ERP investigation. Int. J. Psychophysiol. 63, 283–293. doi: 10.1016/j.ijpsycho.2006.12.001
O’Doherty, J., Winston, J., Critchley, H., Perrett, D., Burt, D. M., and Dolan, R. J. (2003). Beauty in a smile: the role of medial orbitofrontal cortex in facial attractiveness. Neuropsychologia 41, 147–155. doi: 10.1016/s0028-3932(02)00145-8
Ozonoff, S., Pennington, B. F., and Rogers, S. J. (1990). Are there emotion perception deficits in young autistic children? J. Child Psychol. Psychiatry 31, 343–361. doi: 10.1111/j.1469-7610.1990.tb01574.x
Pegna, A. J., Landis, T., and Khateb, A. (2008). Electrophysiological evidence for early non-conscious processing of fearful facial expressions. Int. J. Psychophysiol. 70, 127–136. doi: 10.1016/j.ijpsycho.2008.08.007
Phillips, M. L., Bullmore, E. T., Howard, R., Woodruff, P. W., Wright, I. C., Williams, S. C., et al. (1998). Investigation of facial recognition memory and happy and sad facial expression perception: an fMRI study. Psychiatry Res. 83, 127–138. doi: 10.1016/s0925-4927(98)00036-5
Piggot, J., Kwon, H., Mobbs, D., Blasey, C., Lotspeich, L., Menon, V., et al. (2004). Emotional attribution in high-functioning individuals with autistic spectrum disorder: a functional imaging study. J. Am. Acad. Child Adolesc. Psychiatry 43, 473–480. doi: 10.1097/00004583-200404000-00014
Rossion, B., Joyce, C. A., Cottrell, G. W., and Tarr, M. J. (2003). Early lateralization and orientation tuning for face, word, and object processing in the visual cortex. Neuroimage 20, 1609–1624. doi: 10.1016/j.neuroimage.2003.07.010
Schendan, H. E., Ganis, G., and Kutas, M. (1998). Neurophysiological evidence for visual perceptual categorization of words and faces within 150 ms. Psychophysiology 35, 240–251. doi: 10.1111/1469-8986.3530240
Schroeder, C. E., Tenke, C. E., Arezzo, J. C., and Vaughan, H. G. Jr. (1989). Timing and distribution of flash-evoked activity in the lateral geniculate nucleus of the alert monkey. Brain Res. 477, 183–195. doi: 10.1016/0006-8993(89)91406-6
Schultz, R. T., Gauthier, I., Klin, A., Fulbright, R., Anderson, A., Volkmar, F., et al. (2000). Abnormal ventral temporal cortical activity during face discrimination among individuals with autism and Asperger syndrome. Arch. Gen. Psychiatry 57, 331–340. doi: 10.1001/archpsyc.57.4.331
Sepeta, L., Tsuchiya, N., Davies, M. S., Sigman, M., Bookheimer, S. Y., and Dapretto, M. (2012). Abnormal social reward processing in autism as indexed by pupillary responses to happy faces. J. Neurodev. Disord. 4:17. doi: 10.1186/1866-1955-4-17
Sims, T. B., Van Reekum, C. M., Johnstone, T., and Chakrabarti, B. (2012). How reward modulates mimicry: EMG evidence of greater facial mimicry of more rewarding happy faces. Psychophysiology 49, 998–1004. doi: 10.1111/j.1469-8986.2012.01377.x
Singer, T. (2006). The neuronal basis and ontogeny of empathy and mind reading: review of literature and implications for future research. Neurosci. Biobehav. Rev. 30, 855–863. doi: 10.1016/j.neubiorev.2006.06.011
Sucksmith, E., Allison, C., Baron-Cohen, S., Chakrabarti, B., and Hoekstra, R. A. (2013). Empathy and emotion recognition in people with autism, first-degree relatives, and controls. Neuropsychologia 51, 98–105. doi: 10.1016/j.neuropsychologia.2012.11.013
Sutherland, A., and Crewther, D. P. (2010). Magnocellular visual evoked potential delay with high autism spectrum quotient yields a neural mechanism for altered perception. Brain 133, 2089–2097. doi: 10.1093/brain/awq122
Tamietto, M., Pullens, P., de Gelder, B., Weiskrantz, L., and Goebel, R. (2012). Subcortical connections to human amygdala and their changes following destruction of the visual cortex. Curr. Biol. 22, 1449–1455. doi: 10.1016/j.cub.2012.06.006
Tottenham, N., Tanaka, J., Leon, A. C., McCarry, T., Nurse, M., Hare, T. A., et al. (2009). The NimStim set of facial expressions: judgments from untrained research participants. Psychiatry Res. 168, 242–249. doi: 10.1016/j.psychres.2008.05.006
Tseng, A., Wang, Z., Huo, Y., Goh, S., Russell, J. A., and Peterson, B. S. (2016). Differences in neural activity when processing emotional arousal and valence in autism spectrum disorders. Hum. Brain Mapp. 37, 443–461. doi: 10.1002/hbm.23041
Vuilleumier, P., Armony, J. L., Driver, J., and Dolan, R. J. (2003). Distinct spatial frequency sensitivities for processing faces and emotional expressions. Nat. Neurosci. 6, 624–631. doi: 10.1038/nn1057
Whalen, P. J., Rauch, S. L., Etcoff, N. L., McInerney, S. C., Lee, M. B., and Jenike, M. A. (1998). Masked presentations of emotional facial expressions modulate amygdala activity without explicit knowledge. J. Neurosci. 18, 411–418.
Williams, M. A., Morris, A. P., McGlone, F., Abbott, D. F., and Mattingley, J. B. (2004). Amygdala responses to fearful and happy facial expressions under conditions of binocular suppression. J. Neurosci. 24, 2898–2904. doi: 10.1523/jneurosci.4977-03.2004
Wong, T. K. W., Fung, P. C. W., Chua, S. E., and McAlonan, G. M. (2008). Abnormal spatiotemporal processing of emotional facial expressions in childhood autism: dipole source analysis of event-related potentials. Eur. J. Neurosci. 28, 407–416. doi: 10.1111/j.1460-9568.2008.06328.x
Woodbury-Smith, M. R., Robinson, J., Wheelwright, S., and Baron-Cohen, S. (2005). Screening adults for Asperger syndrome using the AQ: a preliminary study of its diagnostic validity in clinical practice. J. Autism Dev. Disord. 35, 331–335. doi: 10.1007/s10803-005-3300-7
Wright, C. I., Martis, B., Shin, L. M., Fischer, H., and Rauch, S. L. (2002). Enhanced amygdala responses to emotional versus neutral schematic facial expressions. Neuroreport 13, 785–790. doi: 10.1097/00001756-200205070-00010
Yang, T. T., Menon, V., Eliez, S., Blasey, C., White, C. D., Reid, A. J., et al. (2002). Amygdalar activation associated with positive and negative facial expressions. Neuroreport 13, 1737–1741. doi: 10.1097/00001756-200210070-00009
Keywords: autism, backward masking, subliminal facial emotions, subconscious, EEG
Citation: Vukusic S, Ciorciari J and Crewther DP (2017) Electrophysiological Correlates of Subliminal Perception of Facial Expressions in Individuals with Autistic Traits: A Backward Masking Study. Front. Hum. Neurosci. 11:256. doi: 10.3389/fnhum.2017.00256
Received: 29 July 2016; Accepted: 28 April 2017;
Published: 23 May 2017.
Edited by:Edmund C. Lalor, University of Rochester, USA
Reviewed by:Ian J. Kirk, University of Auckland, New Zealand
Michael J. Crosse, Albert Einstein College of Medicine, USA
Copyright © 2017 Vukusic, Ciorciari and Crewther. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: David P. Crewther, email@example.com