Abstract
The human brain has evolved specialised mechanisms to enable the rapid detection of threat cues, including emotional face expressions (e.g., fear and anger). However, contextual cues – such as gaze direction – influence the ability to recognise emotional expressions. For instance, anger paired with direct gaze, and fear paired with averted gaze are more accurately recognised compared to alternate conjunctions of these features. It is argued that this is because gaze direction conveys the relevance and locus of the threat to the observer. Here, we used continuous flash suppression (CFS) to assess whether the modulatory effect of gaze direction on emotional face processing occurs outside of conscious awareness. Previous research using CFS has demonstrated that fearful facial expressions are prioritised by the visual system and gain privileged access to awareness over other expressed emotions. We hypothesised that if the modulatory effects of gaze on emotional face processing occur also at this level, then the gaze-emotion conjunctions signalling self-relevant threat will reach awareness faster than those that do not. We report that fearful faces gain privileged access to awareness over angry faces, but that gaze direction does not modulate this effect. Thus, our findings suggest that previously reported effects of gaze direction on emotional face processing are likely to occur once the face is detected, where the self-relevance and locus of the threat can be consciously appraised.
Similar content being viewed by others
Introduction
Faces provide a rich source of social information (e.g., identity, gender, age, mental perspectives and emotional states), the processing of which is supported by specialised brain regions1,2,3,4,5. The accurate perception of social information in faces enables us to effectively understand the perspectives of conspecifics. Emotional expression is a particularly important source of information as it provides rapid insights into the internal states of others and guides how we approach, interact with, or avoid our conspecifics. Thus, accurately evaluating and responding to such cues is critical for maintaining social relationships, but arguably also serves an adaptive function for survival6,7,8.
Emotional expressions conveyed by others, if accurately perceived and categorised, can provide important warning signals of threat and danger to the observer. But while perception of emotional expressions depends primarily on facial configuration, it is also influenced by other face-bound contextual cues. For instance, gaze direction is an important contextual cue known to influence judgments of emotional expression, particularly those that signal threat to the observer (i.e., anger and fear). This is arguably because gaze direction contextualises the relevance and locus of the threat to the observer9,10,11. Specifically, angry faces with direct gaze signal threat directed at the observer, by a conspecific. In contrast, fearful faces with averted gaze indicate an external source of threat in the immediate environment (e.g., a predator)10. Indeed, these gaze direction and emotion combinations have been found to result in the subjective perception of the expressed emotion as being more intense than alternate combinations9,10. Further, fearful faces have been found to be more frequently detected during an attentional blink paradigm when they display averted rather than direct gaze, whereas angry faces are more frequently detected when they display direct gaze12. It has also been shown that certain gaze direction and emotion conjunctions are more accurately and readily recognised than others13,14,15,16,17. As such, it has thus been suggested that gaze direction acts to provide additional contextual cues that boost the perceptual salience and valence of the emotional threat signal – because the locus of threat conveyed is of greater relevance to the observer (hereafter referred to as “Threat+” faces) – than that signalled by alternate gaze and emotion conjunctions (i.e., “Threat−” faces)18.
Consistent with findings of a heightened perceptual sensitivity to Threat+ faces (anger paired with direct gaze, fear paired with averted gaze) than Threat− faces (anger paired with averted gaze, fear paired with direct gaze) is the idea that evolutionary pressure has led to the hardwiring of specialized brain mechanisms that prioritize the processing and response to threatening stimuli6. The ‘low-road’ hypothesis theorises that the rapid responding to threat-related visual cues is supported by a specialized subcortical pathway from the retina to the amygdala, which bypasses the visual cortex and operates outside of visual awareness19,20,21,22,23,24,25. Ancient brain structures like the superior colliculus and pulvinar are thought to operate in the absence of conscious awareness to rapidly orient attention to the threatening stimulus and facilitate a response26. Although neuroanatomical support for this pathway has been limited27, recent research shows evidence for an afferent subcortical pulvinar connection to the amygdala that facilitates fear recognition28. There is also consistent evidence to suggest that threat is processed outside of visual awareness29,30. Continuous flash suppression (CFS) is a technique that suppresses stimuli from visual awareness and is thought to target early subcortical visual processing31,32,33,34. Experiments using this technique consistently show that faces with fearful expressions gain privileged access to visual awareness over faces with neutral or happy expressions35,36,37,38,39,40,41,42,43,44. Further, fMRI studies using CFS have also shown that amygdala responses elicited by fearful faces – that are suppressed from awareness – are similar to those elicited by consciously perceived fearful faces36,45. These findings support suggestions of a ‘low road’ pathway which operates without awareness to prioritize the detection of threatening stimuli.
Although there is evidence for the processing of fearful faces outside of visual awareness, it is currently unclear whether the modulatory effects of gaze on emotion perception occur at the same preconscious stages of processing. For example, would a fearful face with direct gaze (i.e., a Threat− face) gain access to awareness earlier than an angry face with direct gaze (i.e., a Threat + face), despite the angry face presenting a more relevant threat signal to the observer? Some behavioural evidence from studies using the Garner Interference task show that when expressed emotions are unambiguous, the emotion is encoded before gaze interferes15, suggesting that the effect of gaze may occur at later stages of cognitive processing. However, evidence from electroencephalography (EEG) studies suggest that neural responses to human gaze direction and emotional expression interact as early as 200 milliseconds after stimulus presentation11,17,46,47,48. In recent studies, researchers have created stimuli with different levels of perceived threat using different combinations of identical gaze directions (direct and averted) and expressed emotions (angry and fear)18,48. By matching the gaze directions and emotion expression across conditions, they were able to identify brain responses caused by the conjunction of gaze direction and emotion, and thus the perceptual integration of these cues, while controlling out activation produced by either gaze direction or emotion expression alone. They found support for perceptual integration in ventral face-selective and dorsal motor cortices 170 ms after the stimulus presentation, as evidenced by enhanced neural responses to Threat + conjunctions compared to Threat− conjunctions. This suggests that the detection and behavioural response to self-relevant threat signals are supported by fast-acting neural circuits. However, no study has yet assessed whether the fast integration of gaze and emotion cues occurs outside of awareness leading to the faster detection of Threat + stimuli.
In the current study, we aimed to use the breaking-CFS (b-CFS) paradigm (Fig. 1) to test whether the modulatory effects of gaze direction on emotional face processing occur outside of awareness. We hypothesised that if gaze direction and emotion expression is integrated at the earliest preconscious stages of face processing – and acts to boost the early detection of self-relevant threat signals – then Threat+ faces should gain access to awareness faster (i.e., exhibit shorter suppression times) than Threat− faces.
Results
Using a b-CFS paradigm, we assessed whether the integration of emotional expression (fear versus anger) and gaze direction (direct versus averted) occurs at early stages of face processing, leading to the prioritised awareness of emotion and gaze conjunctions that signal self-relevant threat (Fig. 1). Mean suppression times for each stimulus condition are summarised in Fig. 2. A two-way repeated measures ANOVA with factors: emotion (angry, fear) and gaze direction (direct, averted) revealed a main effect of emotion, with participants being significantly faster to detect fearful faces than angry faces (F(23) = 21.89, p < 0.001, η² = 0.488; mean difference = 158.77 ms, SD = 168.88). Our analyses did not reveal evidence for a significant main effect of gaze direction (F(23) = 2.09, p = 0.162, η² = 0.083). Moreover, we did not detect the hypothesised interaction between emotion and gaze direction on suppression times (F(23) = 0.241, p = 0.628, η² = 0.010). This suggests that Threat+ and Threat− conjunctions of expressed emotion and gaze direction did not differentially affect suppression times.
To confirm that there were no differences in suppression times for Threat+ and Threat− face conjunctions, we averaged suppression times across both Threat+ (i.e., anger-direct, fear-averted) and Threat− (i.e., anger-averted, fear-direct) face conditions and compared them using a paired t-test. This analysis revealed no evidence for a significant effect of self-relevant gaze-emotion conjunctions on suppression times (t(23) = 0.491, p = 0.628, d = 0.100).
Discussion
Using a b-CFS paradigm, we assessed whether the integration of emotional expression and gaze direction occurs at early stages of face processing, leading to the prioritised awareness of emotion and gaze conjunctions that signal self-relevant threat (i.e., Threat+). Consistent with previous research38,41,42,44,49, our study found that faces with fearful expressions gain privileged access to conscious awareness over angry expressions. However, we found no evidence that gaze direction modulates this effect. These findings suggest that, whilst the effects of gaze direction on emotion processing may act to emphasise the locus and self-relevance of threat14,17,18,46,47,48,50,51, they are likely to manifest after the face is detected and is assigned conscious appraisal.
In a recent study using a similar set of face stimuli to those used here, participants were required to discriminate between fear and anger18. When participants were asked to consciously evaluate the face, gaze direction was shown to boost emotion categorization performance. This suggests that the modulatory effects of gaze direction may act to lower discrimination thresholds and sharpen the discriminability of threat-signalling emotions, rather than lower the threshold for detecting the face. This would be particularly advantageous under conditions of uncertainty, where the face is detected but is poor in signal quality (e.g., in the dark)52. In such cases, information about gaze direction would be critical for disambiguating whether the threat is relevant to the observer and thus requiring additional neural and attentional resources. In support of this idea, the effects of gaze15 on emotion discrimination are strongest when the face morphology weakly exhibits the expressed emotion17,53, compared to when the expressed emotion is more intensely conveyed18,52.
Our findings suggest the modulatory effects of gaze direction on emotional face processing requires conscious awareness. This is consistent with previous evidence from speeded emotion categorization tasks, showing that a conscious and strategic use of gaze information is required to disambiguate angry and fearful faces16. In addition, the current data implicates a cortical locus for the integration of emotion and gaze information54,55,56,57. This is consistent with neuroanatomical and neuroimaging studies showing an overlap of cortical systems that process these features. For instance, representations of conjunctions of gaze and emotion have been identified in temporal3,58,59 and premotor cortices18. Neuroimaging and patient studies have also consistently shown the amygdala to be involved in integrating gaze and emotion expression47,50,53,60,61. Importantly, it has been suggested that two parallel routes to the amygdala facilitate threat perception22,23,24,25,62,63. The evolutionary old subcortical route is thought to elicit early responses to unambiguous threat, particularly fearful faces28. Alternatively, the cortical route is thought to processes ambiguous threat signals (e.g., fearful faces with direct gaze)11. Our data suggest that while a subcortical pathway may facilitate the rapid detection of fearful faces, cortical mechanisms must be involved in linking contextual information that signals their relevance to the observer.
The results from our study failed to show a main effect of gaze direction on suppression times despite previous reports of a direct gaze advantage under b-CFS conditions35,64,65,66,67. This finding is again consistent with a heightened sensitivity to emotional expression and dedicated brain mechanisms for the rapid detection of fearful and angry faces irrespective of gaze68. In support of this, previous research using Garner Interference tasks show that when emotion can be easily recognised, it is processed before gaze interferes15. Other CFS studies also suggest the effects of gaze on early face processing are likely to occur after the emotion cue is processed. For instance, studies examining emotional face processing report much shorter mean suppression times (i.e., in the range of 1000–4000 ms)38,42,69 compared to studies examining gaze direction (~6000 ms)35,64,65,66. Previous studies reporting a direct gaze advantage using b-CFS have typically used faces with neutral expressions35,64,65,66,67. Thus, gaze direction might only affect suppression times under CFS when the face expresses a subtle emotion11,62,63. In such a case, the positioning of the sclera and iris might be more visible compared to stimuli used in the current study11.
Finally, it should be noted that there is currently a debate about whether CFS indeed measures unconscious processing32,33,34,49. In this study, the reported differences in suppression time between fearful and angry faces are assumed to reflect differences in the unconscious processing these stimuli receive while being suppressed. However, it is possible that these differences reflect a bias for fearful faces that manifests at very early post-conscious stages of processing32,33. Importantly, however, this interpretation does not detract from our main conclusion that eye gaze direction does not influence emotional face processing in the absence of awareness. Also, it is argued that many previous reports of differences in the preconscious processing of social stimuli may be explained by differences in other stimulus dimensions such as low-level properties like luminance or contrast30,70,71,72. Indeed, it is possible that the main effect of emotion reported in our study was influenced by differences in visibility of sclera between fearful faces compared to angry faces16. However, such differences cannot account for our lack of evidence for a modulatory effect of gaze on emotional face processing in the absence of awareness. Specifically, the elegance of our study lies in the design of conjunction stimuli: Threat+ and Threat− categories both comprised fearful and angry faces with direct and averted gaze. The use of such stimuli allowed us to examine the integration of emotion expression and gaze direction at the earliest stages of face processing while keeping physical low-level properties of the stimuli matched, overall, across conjunction conditions.
Method
Ethics statement
All participants provided written and informed consent before participating in this study. They also consented to the sharing of all their de-identified research data on online research repositories (e.g., Open Science Framework). All experimental protocol used in this study – including the sharing of de-identified research data – were approved by the Macquarie University Human Research Ethics Committee (Caruana #5201200021) and were carried out in accordance with the enforced guidelines.
Participants
Twenty-four healthy subjects (9 females; mean age, 47.13 ± 13.76 years) participated in the study. All participants had normal vision and no psychiatric history or neurological impairment. Participants were either (1) members of the general public who were recruited via online advertisements or our department’s research participation register and were paid for their time, or (2) students from undergraduate psychology units at Macquarie University who received course credit. Our sample size was selected to be comparable to previous CFS studies investigating the effect of gaze direction and emotional expression70,73. A power analysis also revealed that the current sample size was approximately double that required to detect a gaze direction by emotion interaction of the same magnitude reported in previous emotion recognition studies using the same stimuli as those used here18,48,74.
CFS stimulus and apparatus
During CFS, a target stimulus (e.g., a face) is presented to an observer’s non-dominant eye. This stimulus is rendered invisible by simultaneously presenting a high-contrast, dynamic mask, to the dominant eye. The mask dominates conscious perception, and ‘suppresses’ perception of the target stimulus, until it ‘breaks’ into conscious awareness. The time taken for observers to detect the stimulus is an index of the target stimulus’ potency in accessing conscious awareness33,34,75. In the current study, participants viewed dichoptic displays on a Samsung SynchMaster SA950 HD LED monitor (60 × 34 cm, 120 Hz), through a mirror stereoscope. Participants were seated 70 cm from the screen with their head stabilized in a chin rest. Two red frames (10.4° X 10.4°) were displayed side-by-side on the screen, separated by 21.6o of visual angle. This ensured that the left and right frames were only visible to the left and right eye, respectively. Fusion contours (random noise pixels; width 0.5o) were presented within each of the frames to facilitate binocular fusion of the two images presented to each eye. A black fixation point was continuously presented in the centre of each frame (0.2o). Participants were asked to maintain fixation on this point for the duration of the experiment. Prior to testing, participants completed the near convergence test to establish eye dominance76. Upon setting-up the stereoscope, we ensured that participants perceived a single (i.e., fused) frame when viewing the screen binocularly. We also confirmed that participants could only see one frame when viewing the screen monocularly with each eye.
We used a selection of face stimuli employed by El Zein et al.18, which were originally taken from the Radboud Faces Database77 but adapted such that the subjective intensity of expressed emotions were equated across anger and fear emotion categories. This intensity calibration was informed by subjective ratings obtained from a sample of 19 individuals, and the adaptation was subsequently verified by another sample of 10 naïve individuals (for more extensive details on stimulus adaptation and validation see El Zein et al., 2015, p.15–16)18. Faces appeared as greyscale images (2o × 3o) within an oval to obscure hairlines (see Fig. 1 for example stimuli). The original stimulus set developed by El Zein et al., comprised 24 identities varying in emotion intensity across 7 levels (for both fear and anger) across three gaze directions (directed at participant or averted 45° to the left or right). For use in the current study, we selected only the most extreme intensities to elicit reliable discrimination of Threat+ and Threat− based on previous findings from El Zein et al.18. This resulted in 4 conditions for each identity: 2 emotions (angry versus fearful) * 2 gaze directions (averted versus direct). Importantly, the two Threat+ and two Threat− conditions contained the same two gaze directions and two emotional expressions but differed exclusively in their unique feature conjunctions. Thus, any prioritized preconscious processing of Threat+ stimuli over Threat− stimuli by the visual system (as evidenced by shorter suppression times), could not be explained by the separate features alone. All images were presented against a grey background. In all conditions, face stimuli were presented within the frame visible to the non-dominant eye, whilst high-contrast contour-rich coloured masks (9.2o) were presented in the frame visible to the dominant eye to induce interocular suppression and thus render the stimuli temporarily invisible (see Fig. 1. for a detailed description of stimulus development and presentation).
Procedure
The experiment comprised 384 trials (96 trials per condition), with each stimulus being repeated four times. At the beginning of each trial, both red frames, with fusion contours and the central fixation point were presented against a grey background (see Fig. 1). After 100 milliseconds (ms), the masks were presented within the frame visible to the dominant eye, flashing at a frequency of 10 Hz. At the same time, a face was gradually presented within the frame visible to the non-dominant eye by linearly increasing the stimulus contrast from 0–100% over a one second period. Face stimuli were presented to either the top-left, top-right, bottom-left or bottom-right quadrant of the fusion contour, centred at an eccentricity of 3 degrees of visual angle from fixation. The location of stimulus presentation was counterbalanced such that stimuli from each condition were presented to all quadrants equally. After eight seconds, the contrast of the mask linearly decreased from 100–0% over a two-second period.
Participants were explicitly instructed to report the location of the face stimulus as soon as any part of it became visible (i.e., broke suppression) and were not required to make judgments about the face itself. Participants used their right hand only to indicate whether the stimulus was presented left or right of fixation by pressing the left arrow key with their index finger or the right arrow key with their middle finger. Suppression times, the time taken to locate each stimulus, were recorded. Only correct responses were analysed (mean accuracy was above 89% for all conditions: Angry-Direct, M = 89.00%, SD = 6.07%; Angry-Averted M = 90.17%, SD = 6.14%; Fear-Direct, M = 91.21%, SD = 6.14%; Fear-Averted M = 91.04%, SD = 5.92%). Individual trials in which the participant reported prematurely (<200 ms) or responded after the mask was ramped down were also excluded from the analyses (retained trials: mean = 94.12%, SD = 6.12%). Participants were given a short break half-way through the experiment.
Analyses
Statistical analyses were conducted using the free software JASP (JASP Team, 2017). The effects of our four conditions on mean suppression times were examined using two-way repeated measures ANOVA.
Suppression times were used as an index of stimulus potency in reaching conscious awareness33,34,75. Each individual’s mean suppression times were log-transformed (log10) to account for the positive skew of the reaction time data78,79. All statistical analyses were conducted on log transformed data, but Fig. 2 summarises untransformed data to facilitate the intuitive interpretation of the results.
Data availability
Data used in the reported analyses are available to download from the Open Science Framework: https://osf.io/2dkrq/.
References
Harries, M. & Perrett, D. Visual processing of faces in temporal cortex: Physiological evidence for a modular organization and possible anatomical correlates. Journal of Cognitive Neuroscience 3, 9–24, https://doi.org/10.1162/jocn.1991.3.1.9 (1991).
Haxby, J. V. et al. The Effect of Face Inversion on Activity in Human Neural Systems for Face and Object Perception. Neuron 22, 189–199, https://doi.org/10.1016/S0896-6273(00)80690-X (1999).
Haxby, J. V., Hoffman, E. A. & Gobbini, M. I. Human Neural Systems for Face Recognition and Social Communication. Society of Biological Psychiatry 22, 189–199 (2002).
Kanwisher, N., McDermott, J. & Chun, M. M. The fusiform face area: A module in human extrastriate cortex specialized for face perception. Journal of Neuroscience 17, 4302–4311 (1997).
Perrett, D. I. et al. Visual cells in the temporal cortex sensitive to face view and gaze direction. Proceedings of the Royal Society of London. Series B, Biological Sciences 223 (1985).
Öhman, A. Of snakes and faces: An evolutionary perspective on the psychology of fear. Scandinavian Journal of Psychology 50, 543–552, https://doi.org/10.1111/j.1467-9450.2009.00784.x (2009).
Öhman, A., Lundqvist, D. & Esteves, F. The face in the crowd revisited: A threat advantage with schematic stimuli. Journal of Personality and Social Psychology 80, 381–396 (2001).
Öhman, A. & Mineka, S. Fears, phobias, and preparedness: Toward an evolved module of fear and fear learning. Psychological Review 108, 483–522 (2001).
Sander, D., Grafman, J. & Zalla, T. In Reviews in the Neurosciences Vol. 14 303 (2003).
Sander, D., Grandjean, D., Kaiser, S., Wehrle, T. & Scherer, K. R. Interaction effects of perceived gaze direction and dynamic facial expression: Evidence for appraisal theories of emotion. European Journal of Cognitive Psychology 19, 470–480, https://doi.org/10.1080/09541440600757426 (2007).
Adams, R. B. et al. Amygdala responses to averted vs direct gaze fear vary as a function of presentation speed. Social Cognitive and Affective Neuroscience 7, 568–577, https://doi.org/10.1093/scan/nsr038 (2012).
Milders, M., Hietanen, J. K., Leppänen, J. M. & Braun, M. Detection of emotional faces is modulated by the direction of eye gaze. Emotion 11, 1456–1461, https://doi.org/10.1037/a0022901 (2011).
Adams, R. B. & Kleck, R. E. Effects of Direct and Averted Gaze on the Perception of Facially Communicated. Emotion 5, 3–11, https://doi.org/10.1037/1528-3542.5.1.3 (2005).
Adams, R. B. & Kleck, R. E. Perceived Gaze Direction and the Processing of Facial Displays of Emotion. Psychological Science 14, 644–647, https://doi.org/10.1046/j.0956-7976.2003.psci_1479.x (2003).
Graham, R. & LaBar, K. S. Garner interference reveals dependencies between emotional expression and gaze in face perception. Emotion 7, 296–313, https://doi.org/10.1037/1528-3542.7.2.296 (2007).
Bindemann, M., Mike Burton, A. & Langton, S. R. H. How do eye gaze and facial expression interact? Visual Cognition 16, 708–733, https://doi.org/10.1080/13506280701269318 (2008).
N’Diaye, K., Sander, D. & Vuilleumier, P. Self-relevance processing in the human amygdala: Gaze direction, facial expression, and emotion intensity. Emotion 9, 798–806, https://doi.org/10.1037/a0017845 (2009).
El Zein, M., Wyart, V. & Grèzes, J. Anxiety dissociates the adaptive functions of sensory and motor response enhancements to social threats. eLife 4, e10274, https://doi.org/10.7554/eLife.10274 (2015).
Öhman, A. The role of the amygdala in human fear: Automatic detection of threat. Psychoneuroendocrinology 30, 953–958, https://doi.org/10.1016/j.psyneuen.2005.03.019 (2005).
Skuse, D. Genetic influences on the neural basis of social cognition. Philosophical Transactions of the Royal Society B: Biological Sciences 361, 2129–2141, https://doi.org/10.1098/rstb.2006.1935 (2006).
Tamietto, M. & de Gelder, B. Neural bases of the non-conscious perception of emotional signals. Nature Reviews Neuroscience 11, 697–709, https://doi.org/10.1038/nrn2889 (2010).
Gelder, B., Snyder, J., Greve, D., Gerard, G. & Hadjikhani, N. Fear fosters flight: A mechanism for fear contagion when perceiving emotion expressed by a whole body. Vol. 101 (2004).
Morris, J. S. et al. A neuromodulatory role for the human amygdala in processing emotional facial expressions. Vol. 121 (Pt 1) (1998).
Rudrauf, D., Mehta, S. & Grabowski, T. J. Disconnection’s renaissance takes shape: Formal incorporation in group-level lesion studies. Cortex 44, 1084–1096, https://doi.org/10.1016/j.cortex.2008.05.005 (2008).
Vuilleumier, P. L. et al. Distinct spatial frequency sensitivities for processing faces and emotional expressions. Nature Neurosci. 6, 624–631 (2003).
LeDoux, J. E. The Emotional Brain: The Mysterious Underpinnings of Emotional Life. (Simon and Schuster, 1996).
Pessoa, L. & Adolphs, R. Emotion processing and the amygdala: From a ‘low road’ to ‘many roads’ of evaluating biological significance. Nature Reviews Neuroscience 11, 773–782, https://doi.org/10.1038/nrn2920 (2010).
McFadyen, J., Mattingley, J. B. & Garrido, M. I. An afferent subcortical white matter pathway to the amygdala facilitates fear recognition. bioRxiv, 277368, https://doi.org/10.1101/277368 (2018).
Gayet, S., Paffen, C. L. E., Belopolsky, A. V., Theeuwes, J. & Van der Stigchel, S. Visual input signaling threat gains preferential access to awareness in a breaking continuous flash suppression paradigm. Cognition 149, 77–83, https://doi.org/10.1016/j.cognition.2016.01.009 (2016).
Hedger, N., Gray, K., Garner, M. & Adams, W. Are Visual Threats Prioritized Without Awareness? A Critical Review and Meta-Analysis Involving 3 Behavioral Paradigms and 2696 Observers. Vol. 142 (2016).
Tsuchiya, N. & Koch, C. Continuous flash suppression reduces negative afterimages. Nature Neuroscience 8, 1096–1101, https://doi.org/10.1038/nn1500 (2005).
Yang, E., Brascamp, J., Kang, M.-S. & Blake, R. On the use of continuous flash suppression for the study of visual processing outside of awareness. Frontiers in Psychology 5, https://doi.org/10.3389/fpsyg.2014.00724 (2014).
Gayet, S., Van der Stigchel, S. & Paffen, C. L. E. Breaking continuous flash suppression: competing for consciousness on the pre-semantic battlefield. Frontiers in Psychology 5, https://doi.org/10.3389/fpsyg.2014.00460 (2014).
Stein, T., Hebart, M. N. & Sterzer, P. Breaking Continuous Flash Suppression: A New Measure of Unconscious Processing during Interocular Suppression? Frontiers in Human Neuroscience 5, https://doi.org/10.3389/fnhum.2011.00167 (2011).
Stein, T., Senju, A., Peelen, M. V. & Sterzer, P. Eye contact facilitates awareness of faces during interocular suppression. Cognition 119, 307–311, https://doi.org/10.1016/j.cognition.2011.01.008 (2011).
Jiang, Y. & He, S. Cortical Responses to Invisible Faces: Dissociating Subsystems for Facial-Information Processing. Current Biology 16, 2023–2029, https://doi.org/10.1016/j.cub.2006.08.084 (2006).
Jiang, Y. et al. Dynamics of processing invisible faces in the brain: Automatic neural encoding of facial expression information. NeuroImage 44, 1171–1177, https://doi.org/10.1016/j.neuroimage.2008.09.038 (2009).
Yang, E., Zald, D. H. & Blake, R. Fearful expressions gain preferential access to awareness during continuous flash suppression. Emotion 7, 882–886, https://doi.org/10.1037/1528-3542.7.4.882 (2007).
Troiani, V. & Schultz, R. T. Amygdala, pulvinar, and inferior parietal cortex contribute to early processing of faces without awareness. Frontiers in Human Neuroscience 7, https://doi.org/10.3389/fnhum.2013.00241 (2013).
Sterzer, P., Hilgenfeldt, T., Freudenberg, P., Bermpohl, F. & Adli, M. Access of emotional information to visual awareness in patients with major depressive disorder. Psychological Medicine 41, 1615–1624, https://doi.org/10.1017/S0033291710002540 (2011).
Oliver, L. D., Mao, A. & Mitchell, D. G. V. “Blindsight” and subjective awareness of fearful faces: Inversion reverses the deficits in fear perception associated with core psychopathic traits. Cognition and Emotion 29, 1256–1277, https://doi.org/10.1080/02699931.2014.976182 (2015).
Zhan, M., Hortensius, R. & de Gelder, B. The Body as a Tool for Anger Awareness–Differential Effects of Angry Facial and Bodily Expressions on Suppression from Awareness. PloS one 10, e0139768, http://europepmc.org/abstract/MED/26469878, http://europepmc.org/articles/PMC4607361?pdf=render, http://europepmc.org/articles/PMC4607361, https://doi.org/10.1371/journal.pone.0139768 (2015).
Gray, K., Adams, W., Hedger, N., E Newton, K. & Garner, M. Faces and Awareness: Low-Level, Not Emotional Factors Determine Perceptual Dominance. Vol. 13 (2013).
Capitao, L. et al. Anxiety Increases Breakthrough of Threat Stimuli in Continuous Flash Suppression. Vol. 14 (2014).
Vizueta, N., Patrick, C. J., Jiang, Y., Thomas, K. M. & He, S. Dispositional fear, negative affectivity, and neuroimaging response to visually suppressed emotional faces. NeuroImage 59, 761–771, https://doi.org/10.1016/j.neuroimage.2011.07.015 (2012).
Sato, W., Yoshikawa, S., Kochiyama, T. & Matsumura, M. The amygdala processes the emotional significance of facial expressions: an fMRI investigation using the interaction between expression and face direction. NeuroImage 22, 1006–1013, https://doi.org/10.1016/j.neuroimage.2004.02.030 (2004).
Conty, L., Dezecache, G., Hugueville, L. & Grezes, J. Early Binding of Gaze, Gesture, and Emotion: Neural Time Course and Correlates. Journal of Neuroscience 32, 4531–4539, https://doi.org/10.1523/jneurosci.5636-11.2012 (2012).
El Zein, M., Gamond, L., Conty, L. & Grèzes, J. Selective attention effects on early integration of social signals: Same timing, modulated neural sources. NeuroImage 106, 182–188, https://doi.org/10.1016/j.neuroimage.2014.10.063 (2015).
Stein, T., Seymour, K., Hebart, M. N. & Sterzer, P. Rapid Fear Detection Relies on High Spatial Frequencies. Psychological Science 25, 566–574, https://doi.org/10.1177/0956797613512509 (2014).
Sato, W., Kochiyama, T., Uono, S. & Yoshikawa, S. Amygdala integrates emotional expression and gaze direction in response to dynamic facial expressions. NeuroImage 50, 1658–1665, https://doi.org/10.1016/j.neuroimage.2010.01.049 (2010).
Hadjikhani, N., Hoge, R., Snyder, J. & de Gelder, B. Pointing with the eyes: The role of gaze in communicating danger. Brain and Cognition 68, 1–8, https://doi.org/10.1016/j.bandc.2008.01.008 (2008).
Graham, R. & LaBar, K. S. Neurocognitive mechanisms of gaze-expression interactions in face processing and social attention. Neuropsychologia 50, 553–566, https://doi.org/10.1016/j.neuropsychologia.2012.01.019 (2012).
Cristinzio, C., N’Diaye, K., Seeck, M., Vuilleumier, P. & Sander, D. Integration of gaze direction and facial expression in patients with unilateral amygdala damage. Brain 133, 248–261, https://doi.org/10.1093/brain/awp255 (2009).
Eccles, J. C. A unitary hypothesis of mind-brain interaction in the cerebral cortex. Proceedings of the Royal Society of London. B. Biological Sciences 240, 433–451 (1990).
Panagiotaropoulos Theofanis, I., Deco, G., Kapoor, V. & Logothetis Nikos, K. Neuronal Discharges and Gamma Oscillations Explicitly Reflect Visual Consciousness in the Lateral Prefrontal Cortex. Neuron 74, 924–935, https://doi.org/10.1016/j.neuron.2012.04.013 (2012).
Crick, F. & Koch, C. Are we aware of neural activity in primary visual cortex? Nature 375, 121–123 (1995).
Crick, F. & Koch, C. Why neuroscience may be able to explain consciousness. Scientific American 273 (1995).
Engell, A. D. & Haxby, J. V. Facial expression and gaze-direction in human superior temporal sulcus. Neuropsychologia 45, 3234–3241, https://doi.org/10.1016/j.neuropsychologia.2007.06.022 (2007).
Wicker, B., Perrett, D. I., Baron-Cohen, S. & Decety, J. Being the target of another’s emotion: a PET study. Neuropsychologia 41, 139–146, https://doi.org/10.1016/S0028-3932(02)00144-6 (2003).
Boll, S., Gamer, M., Kalisch, R. & Büchel, C. Processing of facial expressions and their significance for the observer in subregions of the human amygdala. NeuroImage 56, 299–306, https://doi.org/10.1016/j.neuroimage.2011.02.021 (2011).
Ewbank, M., Fox, E. & Calder, A. The interaction between gaze and facial expression in the amygdala and extended amygdala is modulated by anxiety. Frontiers in Human Neuroscience, https://doi.org/10.3389/fnhum.2010.00056 (2010).
Garvert, M. M., Friston, K. J., Dolan, R. J. & Garrido, M. I. Subcortical amygdala pathways enable rapid face processing. NeuroImage 102, 309–316, https://doi.org/10.1016/j.neuroimage.2014.07.047 (2014).
Garrido Marta, I., Barnes Gareth, R., Sahani, M. & Dolan Raymond, J. Functional Evidence for a Dual Route to Amygdala. Current Biology 22, 129–134, https://doi.org/10.1016/j.cub.2011.11.056 (2012).
Chen, Y.-C. & Yeh, S.-L. Look into my eyes and I will see you: Unconscious processing of human gaze. Consciousness and Cognition 21, 1703–1710, https://doi.org/10.1016/j.concog.2012.10.001 (2012).
Madipakkam, A. R., Rothkirch, M., Guggenmos, M., Heinz, A. & Sterzer, P. Gaze Direction Modulates the Relation between Neural Responses to Faces and Visual Awareness. Journal of Neuroscience 35, 13287–13299, https://doi.org/10.1523/jneurosci.0815-15.2015 (2015).
Seymour, K., Rhodes, G., Stein, T. & Langdon, R. Intact unconscious processing of eye contact in schizophrenia. Schizophrenia Research: Cognition 3, 15–19, https://doi.org/10.1016/j.scog.2015.11.001 (2016).
Yokoyama, T., Noguchi, Y. & Kita, S. Unconscious processing of direct gaze: Evidence from an ERP study. Neuropsychologia 51, 1161–1168, https://doi.org/10.1016/j.neuropsychologia.2013.04.002 (2013).
Darwin, C. The Expression of the Emotions In Man And Animals. (John Murray, 1872).
Stein, T. & Sterzer, P. Not just another face in the crowd: Detecting emotional schematic faces during continuous flash suppression. Emotion 12, 988–996, https://doi.org/10.1037/a0026944 (2012).
Adams, W., Gray, K., Garner, M. & Graf, E. On the “special” status of emotional faces… Comment on Yang, Hong, and Blake (2010). Vol. 11 (2011).
Stein, T., Gayet, S. & Peelen, M. Unconscious processing of facial dominance: The role of low-level factors in access to awareness (2018).
Gayet, S., Stein, T. V. & Peelen, M. The danger of interpreting detection differences between image categories: A brief comment on “Mind the snake: Fear detection relies on low spatial frequencies” (Gomes, Soares, Silva, & Silva, 2018). (2019).
Akechi, H. et al. Absence of Preferential Unconscious Processing of Eye Contact in Adolescents With Autism Spectrum Disorder. Autism Research 7, 590–597, https://doi.org/10.1002/aur.1397 (2014).
Ioannou, C. et al. Shared mechanism for emotion processing in adolescents with and without autism. Scientific Reports 7, https://doi.org/10.1038/srep42696 (2017).
Jiang, Y., Costello, P. & He, S. Processing of invisible stimuli: Advantage of upright faces and recognizable words in overcoming interocular suppression. Psychological Science 18, 349–355 (2007).
Rice, M. L., Leske, D. A., Smestad, C. E. & Holmes, J. M. Results of ocular dominance testing depend on assessment method. Journal of American Association for Pediatric Ophthalmology and Strabismus 12, 365–369, https://doi.org/10.1016/j.jaapos.2008.01.017 (2008).
Langner, O. et al. Presentation and validation of the Radboud Faces Database. Cognition & Emotion 24, 1377–1388, https://doi.org/10.1080/02699930903485076 (2010).
Gayet, S. & Stein, T. Between-Subject Variability in the Breaking Continuous Flash Suppression Paradigm: Potential Causes, Consequences, and Solutions. Frontiers in Psychology 8, https://doi.org/10.3389/fpsyg.2017.00437 (2017).
Heyman, T. & Moors, P. Frequent Words Do Not Break Continuous Flash Suppression Differently from Infrequent or Nonexistent Words: Implications for Semantic Processing of Words in the Absence of Awareness. PLOS ONE 9, e104719, https://doi.org/10.1371/journal.pone.0104719 (2014).
Acknowledgements
This work was supported by the Australian Research Council (ARC) and the ARC Centre of Excellence for Cognition and its Disorders (CCD; www.ccd.edu.au) [CE110001021]. Drs Caruana and Seymour were recipients of two ARC CCD Cross Program Support Scheme grants which directly supported this work. Dr Seymour also received financial support from the Society for Mental Health Research.
Author information
Authors and Affiliations
Contributions
N.C. and K.S. designed the experiment. C.I. collected the data. N.C. and K.S. analysed the data. M.E.Z. provided experimental stimuli and consulted on data analysis. N.C. prepared the initial draft of the manuscript and all authors contributed to the manuscript’s revision.
Corresponding author
Ethics declarations
Competing interests
The authors declare no competing interests.
Additional information
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Caruana, N., Inkley, C., Zein, M.E. et al. No influence of eye gaze on emotional face processing in the absence of conscious awareness. Sci Rep 9, 16198 (2019). https://doi.org/10.1038/s41598-019-52728-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1038/s41598-019-52728-y
- Springer Nature Limited
This article is cited by
-
Holistic processing of gaze cues during interocular suppression
Scientific Reports (2022)