Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 22 May 2019
Sec. Perception Science
This article is part of the Research Topic Typical and Atypical Processing of Gaze View all 16 articles

Socially Communicative Eye Contact and Gender Affect Memory

  • 1Brain, Attention, and Reality Laboratory, Department of Psychology, University of British Columbia, Vancouver, BC, Canada
  • 2Atypical Perception Laboratory, Department of Psychology, MacEwan University, Edmonton, AB, Canada
  • 3Cognition and Natural Behaviour Laboratory, Department of Psychology, University of Waterloo, Waterloo, ON, Canada

Because of their value as a socially communicative cue, researchers have strived to understand how the gaze of other people influences a variety of cognitive processes. Recent work in social attention suggests that the use of images of people in laboratory studies, as a substitute for real people, may not effectively test socially communicative aspects of eye gaze. As attention affects many other cognitive processes, it is likely that social attention between real individuals could also affect other cognitive processes, such as memory. However, from previous work alone, it is unclear whether, and if so how, socially communicative eye gaze affects memory. The present studies test the assumption that socially communicative aspects of eye gaze may impact memory by manipulating the eye gaze of a live speaker in the context of a traditional recognition paradigm used frequently in the laboratory. A female (Experiment 1) or male (Experiment 2) investigator read words aloud and varied whether eye contact was, or was not, made with a participant. With both female and male investigators, eye contact improved word recognition only for female participants and hindered word recognition in male participants. When a female investigator prolonged their eye contact (Experiment 3) to provide a longer opportunity to both observe and process the investigator’s eye gaze, the results replicated the findings from Experiments 1 and 2. The findings from Experiments 1–3 suggest that females interpret and use the investigator’s eye gaze differently than males. When key aspects from the previous experiments were replicated in a noncommunicative situation (i.e., when a video of a speaker is used instead of a live speaker; Experiment 4), the memory effects observed previously in response to eye gaze were eliminated. Together, these studies suggest that it is the socially communicative aspects of eye gaze from a real person that influence memory. The findings reveal the importance of using social cues that are communicative in nature (e.g., real people) when studying the relationship between social attention and memory.

Introduction

From the moment a child is born, she/he begins to engage with her/his mother to communicate her/his needs and to have those needs met. This is the earliest example of the importance of social interactions, and intuitively, the importance of interpreting social cues from others remains essential throughout one’s life. Our eyes are central to social interaction, as they convey a wealth of information about our emotional and mental states which people use to decode our behaviors and intentions (Emery, 2000). During a social interaction, people tend to look at other peoples’ eyes to gauge whether they are interested (Argyle et al., 1974; Ellsworth and Ross, 1975), paying attention (Kleinke et al., 1975), and what their intentions may be (Kleinke, 1986; Baron-Cohen, 1995; Emery, 2000; Shimojo et al., 2003; Ristic et al., 2005; Frischen and Tipper, 2006). Accordingly, it has been argued that one’s ability to attend to the eyes of others plays a critical role in understanding and facilitating social interaction (Kleinke et al., 1975; Cary, 1978; Campbell et al., 1990; Perrett and Emery, 1994; Emery, 2000; Vertegaal et al., 2001; Tomasello et al., 2005). On the other hand, failing to properly attend to the eyes of others has been linked to deficits in social functioning in autism spectrum disorder (see Senju and Johnson, 2009a for a review) as well as social anxiety disorder (Wieser et al., 2009; Schneier et al., 2011). Indeed, researchers have theorized that eye gaze represents a special social attentional cue (Baron-Cohen, 1995) that may be processed by dedicated neural mechanisms (such as that revealed by activity in the superior temporal sulcus, Campbell et al., 1990; Itier and Batty, 2009).

Researchers have attempted to study the eyes’ importance as a social attentional cue by using variants of classic visual attention paradigms in conjunction with socially relevant stimuli (e.g., an image of a face looking at you). In these laboratory-based tasks, such a stimulus is presented on a computer screen and a person’s eye movements – and other attentional behaviors in response to the stimulus – are recorded. Using different tasks (e.g., free viewing: Yarbus, 1967; Walker-Smith et al., 1977; Mojzisch et al., 2006; Birmingham et al., 2008, 2009; Kuhn et al., 2009; Schrammel et al., 2009; Foulsham et al., 2010; Foulsham and Sanderson, 2013; attentional cueing: Friesen and Kingstone, 1998, 2003; Driver et al., 1999; Langton and Bruce, 1999; Ristic et al., 2002; Vuilleumier, 2002; Senju and Hasegawa, 2005; Zwickel and Võ, 2010; Palanica and Itier, 2011; Wiese et al., 2012; Kuhn et al., 2014; Wykowska et al., 2014; Rensink and Kuhn, 2015; visual search: von Grunau and Anston, 1995; Senju et al., 2005; Doi and Ueda, 2007; Senju and Csibra, 2008; Doi et al., 2009; Palanica and Itier, 2011; and face detection: Macrae et al., 2002; Pageler et al., 2003; Vuilleumier et al., 2005; Conty et al., 2006, 2007; Itier et al., 2007, 2011) and a variety of stimuli (e.g., images of faces: Laidlaw et al., 2012; complex scenes: Vuilleumier et al., 2005; Birmingham et al., 2008, 2009; and dynamic videos: Foulsham et al., 2010; Foulsham and Sanderson, 2013; Kuhn et al., 2014; Wykowska et al., 2014), these studies show that people prefer to look at the eyes over any other feature on the face (Birmingham et al., 2008, 2009; Laidlaw et al., 2012; Levy et al., 2012) and that individuals are extremely sensitive to the signals they convey (e.g., people attend to where other people look, especially when they look at them, von Grunau and Anston, 1995; Senju et al., 2005; Vuilleumier et al., 2005; Conty et al., 2007; Doi and Ueda, 2007; Senju and Johnson, 2009b; Freeth et al., 2013).

While these laboratory tasks have made use of a variety of different social stimuli that vary in complexity and approximation to real-life social interactions, the stimuli in these tasks are seldom real people. Recently, researchers have asked similar questions about how we attend to the eyes in more natural settings, where the stimuli are live people instead of static images (Zuckerman et al., 1983; Patterson et al., 2002, 2007; Kuhn and Tatler, 2005; Tatler and Kuhn, 2007; Kuhn et al., 2008, 2016; Foulsham et al., 2011; Laidlaw et al., 2011, 2016; Risko et al., 2012; Wesselmann et al., 2012; Gallup et al., 2012a,b, 2014; Freeth et al., 2013; Wu et al., 2013, 2014; Gobel et al., 2015; Kompatsiari et al., 2018). These studies reveal that the way people respond, both behaviorally and neurologically, to a real person that they can interact with (or a robot that makes head and eye movements that resemble those of a human, Kompatsiari et al., 2018) is often different than the way they attend to an image of a person. For instance, in socially communicative settings where interactions between live people can occur (i.e., people involved in the interaction are aware that they can both send signals to and receive signals from each other; Risko et al., 2012; Gobel et al., 2015; Jarick and Kingstone, 2015; Myllyneva and Hietanen, 2015, 2016; Nasiopoulos et al., 2015; Risko and Kingstone, 2015; Conty et al., 2016; Risko et al., 2016), people will only look at one another if it is socially acceptable to do so.

While these same concerns regarding the ecological validity of attention to social stimuli should also apply to memory for social stimuli, images of people are used as stimuli in most of the work investigating how eye gaze affects memory. Some of these studies report that memory for an image of face (Hood et al., 2003; Mason et al., 2004; Smith et al., 2006) and for words (Fry and Smith, 1975; Kelley and Gorham, 1988; Macrae et al., 2002; Falck-Ytter et al., 2014) is improved when these stimuli are associated with direct gaze (though see Beattie, 1981; Conty et al., 2010; Nemeth et al., 2013). In the investigations that have used live people as stimuli, a speaker’s eye contact has been associated with improved memory for what the speaker has said (Otteson and Otteson, 1980; Sherwood, 1987; Fullwood and Doherty-Sneddon, 2006; Helminen et al., 2016). Some of the earliest studies examined how children’s academic performance could be affected by a teacher’s eye contact (Otteson and Otteson, 1980; Sherwood, 1987). For instance, Sherwood (1987) suggested that learning could be enhanced in a classroom when instructors made eye contact with members in the audience. In a more recent study, male participants remembered more details from a story told by a male storyteller who looked at them relative to a storyteller who looked away, but female participants did not (Helminen et al., 2016). The research using images and live people as stimuli seems to indicate that eye contact enhances the processing and retention of information. However, other laboratory research suggests eye contact may actually hinder performance (Beattie, 1981; Conty et al., 2010; Nemeth et al., 2013) consistent with the notion that eye contact draws attention and other cognitive resources away from the task at hand (Nemeth et al., 2013).

An important limitation of all past studies using live speakers is that researchers have generally not measured and/or systematically manipulated when a given participant actually experiences eye contact with the investigator. For example, in past studies, listeners were normally exposed to either a speaker who never made eye contact with the listener(s) in an audience or a speaker who periodically made eye contact with some undefined subset of listeners (Otteson and Otteson, 1980; Sherwood, 1987; Fullwood and Doherty-Sneddon, 2006; Helminen et al., 2016). As such, it is unclear how much eye contact a listener actually made with the speaker (if they experienced any eye contact at all). More importantly, it is also unclear whether the specific information that a listener recalled was actually the information that was presented when the speaker made eye contact, as the temporal synchrony between the speaker’s eye contact and the spoken information was not controlled (Otteson and Otteson, 1980; Sherwood, 1987; Fullwood and Doherty-Sneddon, 2006; Helminen et al., 2016). As a result, these studies cannot determine whether memory effects related to gaze reflect, for example, an enhancement or a decline from direct gaze. At best, the mixed results from the research that has used images of faces suggest that both factors may be in play. As such, it remains unclear whether mutual eye contact actually enhances or hinders memory for verbal information. What is needed is a paradigm that is controlled enough to study the effect of eye gaze, without compromising the signal that eye contact provides in a natural setting (Jarick and Kingstone, 2015; Myllyneva and Hietanen, 2015; Nasiopoulos et al., 2015; Risko and Kingstone, 2015; Conty et al., 2016; Helminen et al., 2016; Risko et al., 2016). The goal of the present work is to develop a rigorous paradigm that would avoid this limitation and enable us to investigate whether eye contact enhances or hinders memory for spoken information.

As previous laboratory work has successfully measured other gaze-related memory effects using recognition tests (e.g., gaze cuing to visual word stimuli presented on a computer screen, Fry and Smith, 1975; Kelley and Gorham, 1988; Macrae et al., 2002; Hood et al., 2003; Mason et al., 2004; Smith et al., 2006; Dodd et al., 2012; Falck-Ytter et al., 2014), the studies presented in this paper will use a variant of these classic recognition tasks. The basic methodology is as follows. In an initial study phase, a participant will be seated across from an investigator (or a video of an investigator) who reads words out loud. Critically, before the investigator reads each word s/he will either look up to make eye contact with the participant or keep gaze down at the computer screen to avoid eye contact. Afterward, the participant will perform a recognition test containing the words studied with eye contact, the words studied without eye contact, and new words. The key dependent measure will be recognition accuracy. During the study, and in order to systematically control what information is presented with eye contact, a laptop computer screen, that is only visible to the investigator, will indicate the word to be read aloud and instructions on whether or not to make eye contact with the participant on a given trial. Participants will also be instructed to make eye contact with the investigator during the experiment and to look at the investigator’s eyes if making eye contact is not possible (i.e., the investigator was looking down at the screen rather than at the participant). The investigator will monitor whether the participant makes eye contact, and participants who fail to make eye contact throughout the experiment will be excluded. In previous work, direct eye gaze has enhanced and hindered memory performance, so the effect that gaze could have in the present studies was very much an open question.

Gender has been suggested as a modulating factor in the effect of eye gaze (e.g., Helminen et al., 2016). Often eye contact helps all participants recognize a face, regardless of the their gender (Macrae et al., 2002; Hood et al., 2003; Mason et al., 2004; Smith et al., 2006). However, in some contexts, one gender will benefit from eye contact, but the other will not (e.g., Otteson and Otteson, 1980; Goodman et al., 2012; Helminen et al., 2016). Finally, researchers have observed gender differences in how attentive participants are to the eyes (Connellan et al., 2000; Lutchmaya et al., 2002) and the nonverbal signals of others (Hall, 1978; Rosenthal et al., 1979; McClure, 2000), as well as how responsive participants are to these signals (e.g., females maintain more distance between themselves and a virtual agent that makes eye contact than males; Bailenson et al., 2001; Bayliss et al., 2005). Because of this, gender should be systematically controlled since it is a factor that could influence how eye gaze affects performance. The experiments reported here use either a female (Experiments 1, 3, and 4) or a male (Experiments 2 and 4) investigator who looks at male and female participants. This experimental setup has the added benefit of permitting an examination of whether gender will influence any observed eye gaze-induced memory effects.

Experiment 1

It is currently unclear whether socially communicative eye contact helps or hinders memory. To determine this, we manipulated whether an investigator reading words aloud made eye contact with a participant or not and determined how this manipulation affected participants’ word recognition. If the investigator’s eye contact is helpful when a participant encodes information, then recognition performance would be best for words spoken while the investigator made eye contact. Alternatively, if the investigator’s eye contact interferes with encoding, recognition performance would be worse for words spoken while the investigator made eye contact with the participant. Since the gender of both participants’ and the gaze cue (e.g., investigator) have been reported to modulate the effect of gaze on memory, this factor was systematically manipulated across the studies reported in this paper. In Experiment 1, the investigator was female.

Method

Participants

Eighty-four undergraduate students from the University of British Columbia (42 males, 42 females) received course credit for participating. All reported speaking English as their first language. All had normal or corrected to normal vision and were naive about the purpose of the experiment. All participants gave informed consent before participating and all associated methods were approved by the University of British Columbia’s Research Ethics Board [Towards a More Natural Approach to Attention Research 1-200, certificate #H10-00527, & Research in Cognitive Ethology, #H04-80767].

Design

A 2 (Investigator gaze: eye contact and no eye contact) by 2 (Participant gender: male and female) mixed design was used, where investigator gaze was manipulated within participant and participant gender was a between-participant variable.

Apparatus

E-Prime 2.01 controlled the timing and presentation of stimuli read aloud by the investigator to the participant and logged response accuracy and response times (RTs) in the recognition test. The stimuli were presented on a 17-in. monitor with a 1920 × 1,080 pixel resolution.

Stimuli

The stimulus pool consisted of the 120 words from Macdonald and Macleod (1998). The words were nouns 5–10 letters long, with frequencies greater than 30 per million (Thorndike and Lorge, 1944). From the 120 words, 3 lists containing 40 words each were randomly generated. For a given participant, two lists were selected for study; one list was presented with eye contact and the other list without. The third list was reserved for a recognition test. List selection was counterbalanced across participants such that each word was presented in each of the different conditions (i.e., with eye contact, without eye contact, new words for recognition) an even number of times across participants.

Procedure

Participants heard words for a later memory test. Participants were not informed that they would complete a memory test after hearing the words. During the initial encoding phase, participants were seated ~40 in. across from a female investigator who read aloud words individually. Critically, while the investigator read the words, she either looked up to make eye contact briefly (less than a second) with the participant or kept gaze down at the computer screen to avoid eye contact. Eighty words in total were read aloud in random order to the participants, half of which were presented with eye contact and the other half without.

A laptop screen (only visible to the investigator) indicated when a word was to be read aloud and provided instructions on whether or not to make eye contact with the participant on a given trial. To begin each trial, a blank screen appeared for 1,500 ms. Next, the instruction to look up at the participant or look down at the laptop was presented to the investigator. After 1,000 ms, a word also appeared and remained on screen for 3,000 ms. As soon as the word appeared on screen, the investigator would then look as instructed either toward the participant or down at the computer screen while she read the word aloud. On trials where the investigator made eye contact, the investigator would look back down at the computer screen as she finished saying the word (~1 s). Next, a blank white screen would appear for 500 ms to alert the investigator of the end of the trial. The words and eye contact instructions were randomly intermixed. Since neither the investigator nor participant had knowledge of the trial sequence, this would prevent any systematic change in a participant’s eye gaze throughout the experiment. The investigators (who were authors MZ and CB) were trained to maintain a neutral facial expression and a consistent tone of voice irrespective of the gaze condition, though some natural non-systematic variation within and across participants was expected. The investigators memorized the instruction script for each phase, followed the instruction prompt on the computer screen (i.e., “look at participant or look at screen”), and subsequently read the words to the participant. Additionally, investigators rehearsed before testing to help maintain consistency.

To further ensure that eye contact between the investigator and participant was controlled, participants were instructed to make eye contact with the investigator during the experiment, and if making eye contact was not possible (i.e., the investigator was looking down at the screen) to look at the investigator’s face. This way, eye contact could be made easily when the investigator looked up at the participant. Thus, eye contact was monitored by the investigator on a trial-by-trial basis. In instances when a participant failed to make eye contact with the investigator consistently (e.g., on one or more trials), the participant was excluded. Any failures to make eye contact were extremely rare, and of the 2 (out of 84) participants who were excluded, there was no ambiguity, i.e., they consistently failed to make eye contact. It is important to note that that people are very sensitive at judging where people are looking and when eye contact is being made, even when judging that behavior via a static image or video (e.g., Anderson et al., 2011). The instructional sequence visible to the investigator during the encoding phase is presented in Figure 1A.

FIGURE 1
www.frontiersin.org

Figure 1. The depiction of the experimental setup and procedure used in Experiment 1. (A) The instructional sequence that was visible to the investigator for different trials during the encoding phase. When instructed, the investigator would lift her eyes to make eye contact with the participant as the word appeared on screen and was read aloud. The investigator is depicted from the participant’s perspective on each trial type. (B) The trial sequence that was presented to the participants during the recognition phase of the experiment.

Once the encoding phase was complete, the investigator would open a recognition test on the laptop and turn the laptop to face the participant. After reading the recognition test instructions to the participant, the investigator would monitor the participant’s performance as they completed four practice trials (which were excluded from the analysis). Next, the investigator left the participant alone in the room to complete the recognition test. The recognition test contained the words studied with eye contact, the words studied without eye contact, and 40 new words. The test words appeared on a computer screen in white font against a black background and were presented in random order. A fixation cross was presented for 500 ms before each word. When a word appeared, the subjects were instructed to make a “new” or “old” response for each test word by pressing buttons labeled “New” and “Old” on the keyboard. There was a 500-ms blank interval before each word appeared on screen, and the word offset with the subject’s key response. The response accuracy and response times were recorded. The trial sequence used during the recognition phase trial is presented in Figure 1B. Once the recognition task was complete, the participant remained seated until the investigator came back to the room.

Results

A two-way mixed ANOVA was conducted on response time (RT) and response accuracy (percentage correct), with investigator gaze (two levels: with eye contact and without eye contact) as the within-participant factor and participant gender (two levels: male and female) as the between-participant factor.

Response Time

Mean RTs are presented in Figure 2. There were no main effects of investigator gaze (F(1,82) = 0.04, MSE = 20,386.56, p = 0.84) or participant gender (F(1,82) = 1.08, MSE = 229,174.88, p = 0.30). Nor was there an interaction between investigator gaze and participant gender (F(1,82) = 0.36, MSE = 20,386.56, p = 0.55).

FIGURE 2
www.frontiersin.org

Figure 2. RT as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Percentage Correct

Analysis of the accuracy data (see Figure 3) revealed no main effect of investigator gaze (F(1,82) = 0.37, MSE = 36.69, p = 0.55) or participant gender (F(1,82) = 0.88, MSE = 321.28, p = 0.35). Critically, there was an interaction between investigator gaze and participant gender (F(1,82) = 15.84, MSE = 36.69, p < 0.001), such that female participants recognized more words that were spoken while the investigator made eye contact (79%) than when they did not (75%; t(41) = 3.27, SEM = 1.31, p < 0.005). However, male participants recognized fewer words read while the investigator made eye contact (73%) than when they did not (76%; t(41) = 2.37, SEM = 1.33, p < 0.05).

FIGURE 3
www.frontiersin.org

Figure 3. Percentage correct as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Discussion

The results from this initial experiment demonstrate that memory is improved for words that were encoded with eye contact, but this effect was only observed in female participants. These findings suggest that eye contact has differential effects on memory for verbal information in males and females. Given that females are more attentive and responsive to nonverbal behavior than males (Hall, 1978; Connellan et al., 2000; Bailenson et al., 2001; Lutchmaya et al., 2002; Bayliss et al., 2005; Yee et al., 2007; Marschner et al., 2015), it is possible that, in the context of the present study, females dedicated more attention to information delivered during eye contact than males, which may have resulted in deeper processing and better retention of words presented with than without eye contact (Craik and Tulving, 1975).

Alternatively, it could be the case that making eye contact with the opposite sex produces higher levels of arousal compared to making eye contact with the same sex (Argyle and Dean, 1965; Donovan and Leavitt, 1980). Research has demonstrated that eye contact elevates physiological arousal (Kleinke and Pohlen, 1971; Nichols and Champness, 1971; Gale et al., 1978; Wieser et al., 2009; Helminen et al., 2011, 2016), and that high levels of arousal can interfere with performance on similar tasks (Jelicic et al., 2004; Smeets et al., 2007). While it is possible that eye contact holds one’s attention by increasing arousal (i.e., affective arousal theory; Kelley and Gorham, 1988; Senju and Johnson, 2009b; Mather and Sutherland, 2011), eye contact between genders could produce excess arousal and anxiety, and thus, interfere with memory. Accordingly, male participants could have experienced more arousal than the female participants while making eye contact with the female investigator and were more distracted while words spoken with eye contact as a result (Nemeth et al., 2013). Since both the task and processing the eye contact are competing for cognitive resources, performance on the cognitive task suffers. We examined these possibilities in Experiment 2.

Experiment 2

In the previous experiment, only female participants benefited from the investigator’s gaze. This finding could be attributed to the female investigator’s eye contact distracting male participants from processing what is said (Nemeth et al., 2013). It could be the case that female participants were simply more attentive to the female investigator’s eye contact than males irrespective of the investigator’s gender (Hall, 1978; Connellan et al., 2000; Bailenson et al., 2001; Lutchmaya et al., 2002; Bayliss et al., 2005; Yee et al., 2007; Marschner et al., 2015). We seek to distinguish between these possibilities in Experiment 2 by using a male investigator. If males now benefit from eye contact (and females are possibly hindered by eye contact), this would support the idea that the investigator’s gender contributes to the memory effect vis-à-vis its relation to the participant. However, if the results replicate Experiment 1, then a participant’s gender is a contributing factor to how eye gaze influences memory, a finding that would also be consistent with the notion that females interpret nonverbal social cues differently than males.

Method

Participants

Eighty-four undergraduate students from the University of British Columbia (42 males, 42 females) who had not previously participated in Experiment 1 received course credit for participating. All had normal or corrected to normal vision and were naive about the purpose of the experiment.

Design, Apparatus, Stimuli, and Procedure

The design, apparatus, stimuli, and procedure were identical to those used in the previous study, with the exception that now a male investigator read the words aloud to the participants instead of a female investigator.

Results

Data analysis followed the same procedure that was used in Experiment 1.

Response Time

Mean RTs are presented in Figure 4. There was a marginally significant main effect of participant gender (F(1,82) = 3.35, MSE = 140,584.08, p = 0.07), such that females (967 ms) were faster to respond than males (1,055 ms). No other main effects or interaction were significant (all other F’s < 1).

FIGURE 4
www.frontiersin.org

Figure 4. RT as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Percentage Correct

Analysis of the accuracy data (see Figure 5) revealed no main effect of investigator gaze (F(1,82) = 0.15, MSE = 41.95, p = 0.70) or participant gender (F(1,82) = 0.03, MSE = 463.44, p = 0.87). Critically, there was an interaction between investigator gaze and participant gender (F(1,82) = 15.22, MSE = 41.95, p < 0.001), such that female participants recognized more words that were spoken while the investigator made eye contact (77%) than when they did not (72%; t(41) = 3.68, SEM = 1.17, p < 0.001). However, male participants recognized fewer words read while the investigator made eye contact (73%) than when they did not (77%; t(41) = 2.16, SEM = 1.62, p < 0.05).

FIGURE 5
www.frontiersin.org

Figure 5. Percentage correct as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Discussion

Results from Experiment 2 replicate the findings reported in Experiment 1, wherein females benefited from eye contact on recognition tests and males did not. Furthermore, the present experiment rules out a same-gender explanation of Experiment 1, where memory was improved only when making eye contact with a person of the same gender (i.e., female investigator and female participants). Here, female participants showed a memory benefit from eye contact with a male investigator as well. These results are consistent with the notion that females are generally more attentive to gaze cues than males (Hall, 1978; Connellan et al., 2000; Bailenson et al., 2001; Lutchmaya et al., 2002; Bayliss et al., 2005; Yee et al., 2007; Marschner et al., 2015). It is possible that female participants decode the various signals that could be embedded within the investigator’s eye contact (e.g., a signal to pay attention, a signal that the investigator is watching you, information about the investigator’s mental state, etc.) more readily than males. Thus, females quickly interpret this eye contact as a signal to pay attention, and dedicate more cognitive resources to words presented while the investigator looks at them (e.g., Otteson and Otteson, 1980; Sherwood, 1987). On the other hand, males may find interpreting the investigator’s eye contact distracting since it is uninformative to the task, or perhaps they require more cognitive resources than females to process the different social signals. As a result, their performance on the task at hand is impaired.

Experiment 3

In the two previous experiments, the eye contact that was initiated by the investigator was quite brief (i.e., a quick glance (less than 1 s) up at the participant as the investigator said the word aloud). Although this brief glance may have provided enough time for females to decode the eye contact, it is possible that male participants needed longer periods of eye contact in order to decode this social cue. Indeed, in more natural settings, people tend to engage in eye contact with others between 1.7 and 3.6 s (Argyle and Dean, 1965; Helminen et al., 2011). To determine whether more eye contact can help males decode social cues more effectively, in the present study, the investigator made eye contact for a longer period of time (approximately 3 s). In order to maximize any effect of prolonged eye contact on the male participant’s arousal, whether it is beneficial arousal or distracting arousal, we used a female investigator. If prolonged durations enable men to use eye contact as a social cue to enhance verbal information processing, then the present study should reduce or eliminate the effects of participant gender on word recognition.

Method

Participants

Eighty-four undergraduate students from the University of British Columbia (42 males, 42 females) who had not previously participated in Experiment 1 or 2 received course credit for participating. All had normal or corrected to normal vision and were naive about the purpose of the experiment.

Design, Apparatus, Stimuli, and Procedure

The design, apparatus, stimuli, and procedure were identical to those used in Experiments 1 and 2, with the exception that the female investigator made prolonged eye contact when instructed to look at the participant instead of brief eye contact to further accentuate any effect of eye gaze. As in the previous experiments, a laptop that only the investigator could see provided the same instructions on what word to say and whether to make eye contact or not as the word was spoken. On trials where the investigator made eye contact with the participant, the investigator lifted her eyes and read the word when the laptop displayed the word and made continuous eye contact with the participant until the screen flashed white after 3,000 ms, at which point the investigator would return her gaze down to the computer screen.

Results

Data analysis followed the same procedure that was used in Experiments 1 and 2.

Response Time

Mean RTs are presented in Figure 6. There was a main effect of participant gender (F(1,82) = 15.71, MSE = 200,664.95, p < 0.001), such that females (1,019 ms) were faster to respond than males (1,341 ms). No other main effects or interactions were significant (all other F’s < 1).

FIGURE 6
www.frontiersin.org

Figure 6. RT as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Percentage Correct

Analysis of the accuracy data (see Figure 7) revealed a main effect of investigator gaze (F(1,82) = 5.03, MSE = 47.31, p < 0.05), such that words presented with investigator eye contact (75%) were accurately recognized more than those presented without (73%). There was no main effect of participant gender (F(1,82) = 1.77, MSE = 389.94, p = 0.19). Critically, there was an interaction between investigator gaze and participant gender (F(1,82) = 6.37, MSE = 47.31, p < 0.02), such that female participants recognized more words that were spoken while the investigator made eye contact (79%) than when they did not (74%; t(41) = 3.45, SEM = 1.47, p < 0.001). However, male participants were no more likely to recognize words read with (72%) or without investigator eye contact (72%; t(41) = 0.19, SEM = 1.53, p = 0.54).

FIGURE 7
www.frontiersin.org

Figure 7. Percentage correct as a function of Participant Gender (Female versus Male) and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Comparison Between Experiments 1, 2, and 3

An additional follow-up analysis comparing all three experiments was run to reveal any differences (or similarities) in the effects eye contact had on memory in each experiment. A three-way mixed ANOVA was conducted on mean response accuracy with investigator gaze (two levels: with eye contact and without eye contact) as the within-participant factor and experiment (three levels: female investigator with brief glance, male investigator with brief glance, and female investigator with prolonged gaze) and participant gender (two levels: male and female) as between-participant factors. The analysis of the response accuracy data revealed a marginal main effect of investigator gaze (F(1,246) = 3.67, MSE = 42.00, p = 0.06), such that participants recognized more words that the investigator said while making eye contact than when they did not. Critically, this was qualified by a reliable interaction between investigator gaze and participant gender (F(1,246) = 35.16, MSE = 42.00, p < 0.001), such that female participants recognized more words that were spoken while the investigator made eye contact (78%) than when they did not (74%; t(125) = 6.00, SEM = 0.76, p < 0.001). However, male participants recognized fewer words read while the investigator made eye contact (73%) than when they did not (75%; t(125) = 2.67, SEM = 0.87, p < 0.01). No other main effects or interactions were significant (all F’s < 1.27). Note that the lack of a three-way interaction (p = 0.65) is supported by a Bayes factor estimated using Monte Carlo sampling via the BayesFactor package in R (Morey and Rouder, 2018). Specifically, the Bayes factor was 0.125:1 when comparing a model with the three-way interaction to a model without (i.e., only two-way interactions and main effects). In other words, there is about eight times more evidence against including a three-way interaction in the model.

Discussion

The results of Experiment 3 replicated the finding in both Experiments 1 and 2 that females recognized more words and were more sensitive to words presented with eye contact than without eye contact. Unlike the previous studies, male participants’ recognition performance was not significantly different in the eye contact and no eye contact conditions. Further, the follow-up analysis comparing all three experiments revealed that the length of eye contact did not modify the effect eye contact had on memory in female participants.

The failure to observe any interaction between Experiment and Investigator gaze or Experiment, Investigator gaze, and Participant gender suggests that exaggerating eye contact does not help the males encode the nonverbal eye contact cues, nor did it enhance the eye contact benefit observed in female participants (as compared to female participants in Experiments 1 and 2)2. These results also suggest that the memory benefits that arise from eye contact in Experiments 1–3 are unlikely to be mediated by an arousal response, since prolonging the eye contact in the present experiment would have, if anything, increased arousal, which could modify how eye contact affects performance3. Instead, we suggest that these results are more consistent with the idea that eye contact provides a signal to pay attention, and that interpreting this social signal is responsible for enhancing information processing in the previous studies. If anything, it seems that lengthening the investigator’s eye contact reduced the interference eye contact caused the males in Experiments 1 and 2. By providing the signal longer, males may have had enough resources to both process the eye contact and perform the memory task without these two tasks competing for cognitive resources. Since they did not perform better when the investigator made eye contact than when they did not, it seems that this cue may not have had social relevance for the males.

The interpretations discussed here all assume that the investigator’s eye gaze is being interpreted (at least by female participants) as a socially communicative cue. While the live interaction between the investigator and the participant ensures that social communication can occur, it is also possible that a nonsocial cue associated with the investigator’s eye gaze could also be driving the reported effects. Before concluding that socially communicative aspects of gaze produce these effects, it is important to exclude a nonsocial interpretation that could possibly account for the facilitatory effect of gaze. This will be addressed in Experiment 4 where the social communicative aspects of eye gaze will be dissociated from the purely perceptual cues by using a video of the investigator instead of a live investigator.

Experiment 4

The previous studies demonstrated that females benefited from an investigator’s gaze on a subsequent memory test, whereas males did not. These gender-specific memory effects could be driven by a socially communicative cue that is embedded in the investigator’s eye contact (i.e., when someone looks at you, it is a signal to pay attention). According to this idea, females might have been sensitive to the social cue embedded in the investigator’s eye contact, and could have used it to facilitate their performance on the recognition test. However, male participants might have failed to interpret and apply the investigator’s eye contact as a signal to pay attention, and as a result their performance at test could have been hindered by the investigator’s eye gaze.

A different possibility altogether is that there was nothing socially communicative about the investigator’s gaze that drove the memory effects observed in the previous studies. For example, these effects could have arisen by observing the investigators shifting their gaze up from the computer monitor. In the previous studies, the investigators either kept their eyes on the computer screen while they read a word, or they lifted them to make eye contact just before saying a word. Observing just the movement of the eyes up from the computer screen could be an indicator that a word is about to be spoken, much in the same way the onset of a flashing light at a crosswalk indicates that one should pay attention for pedestrians. There is nothing inherently “social” about either of these cues, but they both serve the purpose of a warning cue that informs a participant to increase attention to an upcoming stimulus (i.e., a word or a pedestrian in the latter case). In fact, a variety of perceptual cues (i.e., arrows, flashes in the periphery, etc.) are known to generate changes in attention (e.g., Posner, 1980; Friesen et al., 2004; Bayliss et al., 2005; Hietanen et al., 2006; Ristic et al., 2007; Mulckhuyse and Theeuwes, 2010; Shin et al., 2011; Hayward and Ristic, 2015). Given that both perceptual and socially communicative cues were embedded in the live investigator’s eye contact in the previous experiments, it is unclear which cue was actually driving the memory effects observed in the previous experiments.

The aim of Experiment 4 is to clarify whether socially communicative cues are responsible for the eye gaze-related effects observed in the previous experiments. One way to isolate the social aspects of eye gaze from the perceptual ones is to have observers watch a video of the investigator instead of interacting with a live investigator. Numerous studies have demonstrated that people respond differently, both behaviorally and neurologically, when looking at the eye gaze of people presented in images versus actual, physically present people (Hietanen et al., 2008; Itier and Batty, 2009; Teufel et al., 2010; Laidlaw et al., 2011; Pönkänen et al., 2011a,b; Risko et al., 2012, 2016; Schilbach et al., 2013; Schilbach, 2015). Furthermore, the eye gaze and gestures of people depicted in images and videos have less influence on the communication (Heath and Luff, 1993; Gullberg and Holmqvist, 2006) and attention (Varao-Sousa and Kingstone, 2015; Wammes and Smilek, 2017) of an observer than they typically would during an encounter with a live person. Presumably, this is because the people depicted in the images and videos cannot see the observer and therefore their gaze behavior is not actively communicating with the observer, and vice versa (De Jaegher et al., 2010; Schilbach, 2010; Risko et al., 2012).

By using a video recording of the investigator in the present study, the socially interactive context that was produced by using a live investigator in Experiments 1, 2, and 3 is removed. If the previous findings are replicated, it would suggest that perceptual cues derived from the eye gaze of someone in a video are enough to generate the memory benefits and deficits associated with eye contact, and that a socially communicative context is not required to generate these memory effects. However, eliminating eye-gaze related memory effects would be evidence for the idea that perceptual cues are not driving these previously observed effects. Instead, it would suggest that the socially communicative eye gaze from an individual that an observer could potentially interact with is required to produce these memory effects.

Method

Participants

To examine investigator gender as a factor in one experiment rather than in two separate experiments, as was the case in Experiments 1 and 2, the sample size was doubled to 168 undergraduate students from the University of British Columbia (84 males, 84 females) who had not participated in any of the previous experiments. All received course credit for participating and had normal or corrected to normal vision and were naive about the purpose of the experiment.

Design

A 2 (Investigator gaze: eye contact and no eye contact) by 2 (Participant gender: male and female) by 2 (Investigator gender: male and female) mixed design was used, where investigator gaze was manipulated within participant and participant gender and investigator gender were a between-participant variables.

Apparatus and Stimuli

The apparatus and stimuli were identical to those reported in the Experiments 1 and 2; however, the participants now watched a video of the investigator from either Experiment 1 (female investigator) or Experiment 2 (male investigator). The videos shown to each participant were recorded by a camera that was placed in front of the investigator, on a tripod that was adjusted so that the camera was positioned at the investigator’s eye level. This position was chosen to simulate the distance, height, and eye level of a participant who would have sat across from the live investigator in the previous experiments. A confederate also sat directly behind the camera and looked at the investigator’s eyes. The intention was to simulate the same live interaction to ensure that any variation in expressiveness due to a live context (Experiments 1–3) was also generated when the investigator was videotaped. During the recordings, the investigator read the words aloud as in the previous experiments, i.e., when prompted by the laptop, the investigator either looked toward the computer screen or, to simulate eye contact for the viewer, briefly toward the camera lens. A total of six different videos were made to ensure that across participants, each word would appear in each condition evenly.

Videos were presented full screen at the recorded resolution (1,920 × 1,080 pixels), on a 17-in. monitor. Participants were seated approximately 60 cm from the screen. Sound from the videos was also played through speakers built into the computer.

Procedure

The procedure was identical to those used in Experiments 1 and 2, with the exception that a participant was first assigned to watch a video of either a male or female investigator saying the words out loud. Participants were also instructed to look at the investigators’ eyes throughout the experiment. Based on a wealth of past work indicating that there is a preferential bias to look at the eyes of people when they are shown in photos or in videos (Laidlaw et al., 2011; Risko et al., 2012, 2016), we expected participants to readily comply with the instructions. Their performance during practice and self-report after testing support this prediction.

Results

A three-way mixed ANOVA was conducted on response time (RT) and response accuracy, with investigator gaze (two levels: eye contact and no eye contact) as the within-participant factor and participant gender (two levels: male and female) and investigator gender (two levels: male and female) as the between-participant factors.

Response Time

Mean RTs are presented in Figure 8. There was a main effect of investigator gender (F(1,166) = 13.47, MSE = 214,475.60, p < 0.001), such that participants were faster to respond with the male investigator (1,018 ms) than the female investigator (1,203 ms). No other main effects or interactions were significant (all other F’s < 1).

FIGURE 8
www.frontiersin.org

Figure 8. RT as a function of Investigator Gender (Female versus Male), Participant Gender (Female versus Male), and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Percentage Correct

Analysis of the accuracy data (Figure 9) revealed a main effect of participant gender (F(1,166) = 4.43, MSE = 542.79, p < 0.05), such that female participants were more accurate (72%) than male participants (66%). No other main effects or interactions were significant (all other F’s < 1.4).

FIGURE 9
www.frontiersin.org

Figure 9. Percentage correct as a function of Investigator Gender (Female versus Male), Participant Gender (Female versus Male), and Investigator gaze (With eye contact versus Without eye contact). Note that new words have been plotted in this figure as a reference point, but were not included in the analysis. Error bars represent the 95% confidence interval as defined by Masson and Loftus (2003).

Comparison Between Live and Videotaped Investigators

To reveal any difference in the memory effects generated by the eye gaze of a live investigator (in Experiments 1 and 2) and a videotaped investigator, a four-way mixed ANOVA was conducted on response accuracy with Investigator gaze (two levels: with eye contact and without eye contact) as the within-participant factor and Investigator presence (two levels: live and videotaped), Investigator gender (two levels: male and female) and Participant gender (two levels: male and female) as between-participant factors.

The analysis of mean RTs revealed that there was a main effect of investigator presence (F(1,328) = 12.19, MSE = 199,677.54, p < 0.001), such that participants were faster to recognize words that were said by the in-person investigator (987 ms) than the videotaped investigator (1,111 ms). There was also a main effect of investigator gender (F(1,328) = 11.77, MSE = 199,677.54, p < 0.001), such that participants were faster to recognize words that were said by the male investigator (991 ms) than the female investigator (1,109 ms). There was a marginal interaction between investigator presence and investigator gender (F(1,328) = 3.80, MSE = 199,677.54, p = 0.05), such that when the investigator appeared over video, participants recognized words that were said by a male investigator (1,018 ms) faster than a female investigator (1,203 ms). However, when the investigator was in-person, participants recognized words that were said by a male investigator (965 ms) as fast as words said by a female investigator (965 ms). No other main effects or interactions were significant (all F’s < 2.6).

The analysis of the accuracy data revealed a main effect of investigator presence (F(1,328)=13.43, MSE = 467.57, p < 0.001), such that participants recognized more words said by an in-person investigator (75%) than a videotaped investigator (69%). Critically, there was a three-way interaction between investigator gaze, investigator presence, and participant gender (F(1,328) = 15.88, MSE = 44.75, p < 0.001). When the investigator was in-person, there was an interaction between investigator gaze and participant gender (F(1,166) = 31.38, MSE = 38.85, p < 0.001) such that female participants recognized more words that were spoken while the investigator made eye contact (78%) than when they did not (74%; t(83) = 4.91, SEM = 0.87, p < 0.001). However, male participants recognized fewer words read while the investigator made eye contact (73%) than when they did not (76%; t(83) = 3.19, SEM = 1.04, p < 0.005). In contrast, the same analysis for the videotaped investigator, presented in Section “Percentage correct” (Experiment 4), yielded no effect of investigator gaze, nor an interaction between investigator gaze and participant gender. These findings indicate that the failure to observe any memory effects in Experiment 4 is because eye gaze from a live investigator is fundamentally different from the eye gaze of a videotaped investigator.

Discussion

In Experiment 4, there was no evidence that eye gaze displayed over video influenced memory. This was true regardless of the investigator or participant’s gender. This stands in direct contrast with the previous experiments that demonstrated that eye contact from a live investigator improved memory in female participants, and reduced memory in male participants. Taken together, the results demonstrate that people attend to the eye gaze of those they interact with in real life differently than the eye gaze of people depicted in images.

The current data dovetail with a growing body of research suggesting that viewing a live person elicits different neurological (Hietanen et al., 2006; Pönkänen et al., 2011a,b) and behavioral responses (Heath and Luff, 1992) than viewing an image of a person. Indeed, the finding that participants who listened to a live investigator recognized more words, faster than those who listened to a video of the same words, supports this notion. By using images, the present study stripped away the social cues that would typically be present during a live encounter, leaving only the perceptual cues associated with eye gaze. Since the effect of eye gaze was eliminated when presented over video, the implication is that the effects observed previously in Experiments 1–3 were driven by communicative cues associated with eye contact instead of noncommunicative perceptual cues.

General Discussion

In the present studies, female participants benefited from the investigator’s gaze more than male participants on a subsequent memory test. Specifically, female participants recognized more words in a subsequent memory test when the words were previously associated with eye contact from an investigator than when they were not. This was true regardless of the investigator’s gender (female in Experiments 1 and 3 and male in Experiment 2), and whether the investigator’s gaze was a quick glance (as in Experiments 1 and 2) or a prolonged stare (in Experiment 3). In contrast, male participants showed no benefit of the investigator’s gaze on subsequent memory tests when the investigator’s gaze was held longer (Experiment 3), and actually recognized fewer words on a subsequent memory test when they were associated with brief eye contact from the investigator (Experiments 1 and 2) relative to when they were not. While these findings suggest that eye gaze could provide a useful social cue to pay attention for females, this is not the case for males. Importantly, the data suggest it was something socially communicative about eye gaze (rather than something nonsocially communicative), since eye gaze effects disappeared in the absence of a communicative setting when the investigator was presented over video (Experiment 4).

Our finding that females and not males showed a memory enhancement to gaze contradicts a recent study by Helminen et al. (2016) who showed the opposite. Our methods were very different, such that Helminen et al. read aloud narratives, manipulated eye gaze spontaneously, and tested memory recall. Because of this, it is unclear whether the specific information that listeners recalled in Helminen et al. was spoken while the speaker made eye contact or not because the temporal synchrony between the speaker’s eye contact and spoken information was not controlled. As a result, the data from Helminen et al. (2016) are equivocal as to whether memory effects related to gaze reflect an enhancement from direct gaze or a decline resulting from gaze aversion. Our methods, however, can determine whether memory for words is enhanced or diminished by directly pairing each word with direct or averted gaze, and comparing memory in those conditions to a neutral (no gaze) condition. Our design allowed us to discover that females benefited from eye contact by remembering those words more compared to the control (no gaze) condition, while males showed a reduction in memory for words paired with eye contact compared to the control. Thus, we believe our methods can more accurately depict the how eye gaze modulates memory in males and females.

Gender Differences in Eye Gaze-Related Memory Effects

These findings converge with a body of literature that suggests males and females attend to social information differently. In comparison to males, females dedicate more attention to social stimuli, such as faces and eyes (Connellan et al., 2000; Lutchmaya et al., 2002) and more easily decode the nonverbal signals exhibited by others (Hall, 1978; Rosenthal et al., 1979; McClure, 2000). More importantly, previous studies have demonstrated that females are more responsive to eye gaze (Bailenson et al., 2001; Bayliss et al., 2005). These data suggest that females are more sensitive to social signals in general and eye gaze in particular, and thus the investigator’s eye contact engaged their attention in the first three experiments (despite being irrelevant to the task).

The general interpretation of the memory effect observed in females is that when eye contact accompanies information, it is interpreted as signaling the intent to communicate information that warrants attention (Duncan, 1972; Niederehe and Duncan, 1974; Kampe et al., 2003; Csibra and Gergely, 2009; Senju and Johnson, 2009a). This is particularly noteworthy since the investigator’s eye contact is actually irrelevant to the task at hand. As such, when people communicate with each other and their message is preceded or accompanied with eye contact, their eye contact serves to highlight the most important parts of their message. It would follow that in the present studies, when the female participant was with the speaker, information spoken with eye contact would be attended more and recognized better than information presented without eye contact.

While female participants noticed, and decoded the social signals associated with the investigator’s gaze with apparent ease, male participants did not. It is possible that male participants were simply insensitive to the investigator’s eye contact and/or did not think the eye contact was an important cue, and dedicated the same amount of attention to words regardless of whether the investigator looked at them or not. However, this explanation seems unlikely given that in male participants, the investigator’s eye contact, if anything, had a negative effect on their performance in Experiments 1 and 2. Moreover, participants were instructed to look at the investigator’s eyes throughout the study, so it seems inconceivable that eye contact with the investigator went unnoticed.

Another intriguing possibility is that the males were sensitive to the investigator’s eye gaze, but were unable to decode which signal the investigator intended to convey since eye contact provides a variety of different social signals. While eye contact may signal to pay attention, as previously discussed, eye contact also signals that one is being monitored (Guerin, 1986; Risko and Kingstone, 2011; Pönkänen et al., 2011b; Freeth et al., 2013; Baltazar et al., 2014; Marschner et al., 2015; Nasiopoulos et al., 2015; Hazem et al., 2017), and facilitates decoding of the emotional and intentional messages of others. Discerning which message is most important and/or appropriate (and requires the most attention) in a given situation may be more challenging for males than females. In the present context, males may have struggled to dissociate which signal embedded in the investigator’s eye contact was most relevant (or if they did make this distinction, they appear not to have acted on it). As such, processing or actively ignoring the investigator’s eye contact may have interfered with the males’ ability to pay attention to information that was being spoken. This idea converges with the finding in previous work that the presence of eyes interfered with performance on a Stroop task (Beattie, 1981; Conty et al., 2010; Nemeth et al., 2013), presumably because processing the eyes required the same cognitive resources (e.g., selective attention) used to perform the task. According to this idea, the brief eye contact provided in Experiments 1 and 2 was too difficult for the males to decode while simultaneously completing another task. As a result, they were unable to dedicate enough attention to the task of attending to what the investigator said when they were looked at, and their performance suffered. However, the less subtle signal provided in Experiment 3 did not alter the previous data pattern, undermining the interpretation that males just need a more salient gaze signal for it to yield a performance benefit.

Another similar, though slightly different, explanation of how a participant’s gender modified whether eye contact had a positive or negative impact on memory comes from a proposal put forward by Conty et al. (2016). The authors proposed that direct gaze first captures one’s attention and then triggers self-referential processing, i.e., a heightened processing of contextual information in relation with the self (Northoff et al., 2006). According to this account, direct gaze can have both positive and negative effects on performance since the tendency to pay attention to the direct gaze of others either facilitates or interferes with performance on a task (e.g., direct gaze may facilitate processing a face, but hinder processing information that is not related to the face). However, once direct gaze has triggered self-referential processing, any information associated with it would be prioritized. Indeed, a large body of research suggests that memory is improved for information processed in relation to oneself (i.e., the self-referential memory effect; Macrae et al., 2002; Northoff et al., 2006; Kim, 2012).

In Experiments 1–3, it is possible that the extent to which the speaker’s eye contact triggered self-referential processing differed between males and females. Females may have processed both the speaker’s eye contact and the self-referential cue it provides simultaneously, or simply processed these two signals more efficiently and sequentially. Thus, any interference (if any was experienced) in hearing what the speaker said, caused by simply processing the speaker’s eye contact, was overridden by the self-referential processing benefit triggered through the speaker’s eye contact. Males on the other hand may notice and process the speaker’s eye contact, but not the self-referential cue it provides. As a result, the speaker’s eye contact only interferes with processing what the speaker says. This could be due to interference caused by processing any self-referential cue in the context of the task (i.e., any self-referential cue could be distracting since it is irrelevant to the task of listening to everything the speaker says), or only self-referential cues conveyed through eye contact. Moving forward, it is still unclear which social signals communicated by the live investigator are being interpreted by participants. The findings are consistent with the idea that eye contact provides a social signal to pay attention, which results in memory benefits for information communicated with eye contact. However, there are many nonverbal social cues conveyed during a live encounter that can influence the way an observer pays attention. In fact, some research might suggest that head movements, rather than eye movements, are more important in eliciting attentional shifts in more natural contexts (Tomasello et al., 1998, 2005; Emery, 2000). Even though we have attributed the effects in Experiments 1, 2, and 3 to a social signal conveyed by eye gaze, it is possible that a different social signal that is also associated with the investigators’ gaze could be driving these effects. Future work could clarify whether it was the investigator’s eye contact or a general social cue that was associated with the eye contact that generated the memory benefits observed in the female participants in the previous studies.

Furthermore, although we have considered the no eye contact condition as the control condition, strictly speaking, it is theoretically possible that our observed memory effects reflect a performance decline without eye contact rather than an enhancement with eye contact. Recall that in previous research, listeners watched a speaker who never made eye contact with any listener in an audience or one who periodically made eye contact with some undefined listeners (Otteson and Otteson, 1980; Sherwood, 1987; Fullwood and Doherty-Sneddon, 2006). Even in instances where a single listener is present (in Experiments 1–3 and in Helminen et al., 2016), it is possible that a listener’s memory was improved for the information spoken while making eye contact with the speaker, and it is also possible that a listener had worse memory for information presented while the speaker avoided eye contact. This reduction in memory could be due to the observer feeling excluded by the speaker (a possibility considered but not addressed by Fullwood and Doherty-Sneddon, 2006) or because the speaker’s gaze directs the observer’s attention elsewhere. That said, the fact that participants in our study performed better in the presence of a real investigator than a videotaped investigator, even in the no eye contact condition, suggests that performance is being enhanced in the eye contact condition rather than diminished in the no contact condition. It remains for future studies, however, to confirm this interpretation by testing the adequacy of our baseline, for instance, comparing it to a situation where the speaker could make eye contact with someone other than the participant.

Contributions of Live and Non-live Settings to Eye Gaze-Related Memory Effects

By presenting a video of an investigator (Experiment 4) instead of a live investigator (Experiments 1–3), the socially communicative function of eye gaze was removed. A live investigator generates an interactive context in which both the investigator and the participants can convey and observe signals with their eyes (Risko et al., 2012, 2016; Baltazar et al., 2014; Gobel et al., 2015; Jarick and Kingstone, 2015; Myllyneva and Hietanen, 2015; Nasiopoulos et al., 2015; Risko and Kingstone, 2015; Conty et al., 2016; Hietanen, 2016; Hazem et al., 2017). This is not the case when the investigator is presented over video, since the investigator cannot observe any signals that the participants convey through their eye gaze. This notion is also supported by research showing that social centers in the brain are more activated when observing live people than when viewing images of people (Hietanen et al., 2006; Schilbach et al., 2010; Pönkänen et al., 2011a,b; Schilbach, 2015). Viewing a pre-recorded investigator enabled a strong test of whether nonsocial signals embedded in eye gaze could generate memory effects that were previously observed in response to a live investigator’s eye gaze in Experiments 1–3. The results were unequivocal. The memory effects previously observed in response to a live investigator’s eye gaze disappeared. Without a socially communicative context, eye gaze had no effect on memory. This finding provides strong support for the idea that socially communicative signals conveyed through eye gaze influence memory. This also stands in contrast with a noncommunicative explanation of how eye contact could affect memory. The notion above presents a challenge to researchers who have generally assumed that using images enables them to study social aspects of eye gaze present in real life with real people. However, if this assumption is misplaced, and indeed more and more research is suggesting this assumption may be, then there could be broad-reaching implications as researchers have been using images to study the social effects of eye gaze for decades.

Memory Mechanisms Affected by Eye Contact

While the present work has demonstrated that manipulating a speaker’s eye contact during encoding/consolidation can influence recognition memory for semantic (word) information, it has not explored or manipulated memory for other types of materials (e.g., faces) or different memory processes (e.g., retrieval). An important question related to memory retrieval is whether a speaker’s eye contact affects recall as well recognition. While studies from natural settings suggest that viewers generally recall more information when speakers periodically make eye contact than when they do not (Otteson and Otteson, 1980; Sherwood, 1987; Fullwood and Doherty-Sneddon, 2006), this notion remains to be tested in a rigorous paradigm that permits one to assess who is, and is not, receiving eye contact, and what information specifically is being delivered in those moments.

Another interesting question relevant to memory retrieval is whether making eye contact during retrieval will help or hinder this process. Some research suggests that direct gaze during the retrieval process can enhance memory for a face (Hood et al., 2003; Smith et al., 2006). However, this question has yet to be tested in a paradigm that systematically manipulates a speaker’s eye gaze during the retrieval process.

In sum, there are a number of different aspects of memory that could be influenced by eye contact. Future studies could extend the present work by exploring the effects of eye contact on all of the different components of memory mentioned at the outset of this section. They could also, for example, examine whether the effects are eye contact specific, or general to other visual (e.g., pointing) or nonvisual (e.g., verbal) cues.

Conclusion

The significance of the eyes in human communication has fascinated scientists for centuries. While the present findings only begin to scratch the surface of this broad area of investigation, this work does highlight the importance of systematically examining gender and conducting studies in contexts where eye contact can be communicative. Indeed, in the absence of a communicative context, eye gaze did not modulate recognition performance. This conclusion has tremendous implications for social theories of human communication, memory, and cognition more broadly, as images of the eyes have been used to manipulate and measure social behavior and social neural mechanisms of various cognitive processes. Using real people in future studies will enable the assessment of the social effects of eye gaze in particular, and social signals in general, thereby enhancing our understanding of the cognitive and neural bases of human communication and social interaction.

Ethics Statement

This study was carried out in accordance with the recommendations of the University of British Columbia’s Research Ethics Board [Towards a More Natural Approach to Attention Research 1-200, certificate #H10-00527, & Research in Cognitive Ethology, #H04-80767], with written and oral informed consent from all subjects. All subjects gave written informed consent in accordance with the Declaration of Helsinki. The protocol was approved by the University of British Columbia’s Research Ethics Board in accordance with the guiding ethical principles of the Tri-Council Policy Statement (TCPS2 2014), the International Conference on Harmonization Good Clinical Practice Guidelines (ICH-GCP) and the requirements of the US Department of Health and Human Services, as set out in the Federal Policy for the Protection of Human Subjects, 45CFR Part 46, sub-part A.

Author Contributions

SL was the lead investigator for all of the projects reported in this dissertation and was primarily responsible for design conception, data management and analysis, and report composition. AK and MJ acted in a supervisory capacity during project conception and report composition. MZ and CB were involved in data collection, management, and organization.

Funding

This project was funded by grants to AK from the Natural Sciences and Engineering Research Council of Canada, and the Social Sciences and Humanities Research Council of Canada. Financial support was also received by postgraduate scholarship to SL from the Natural Sciences and Engineering Research Council of Canada, and a fellowship from the University of British Columbia.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Footnotes

1. www.pstnet.com

2. We thank an anonymous reviewer for raising the possibility that the same effect might not be obtained with a male investigator and prolonged gaze. Based on the current data indicating that there is no effect of investigator gender or gaze duration, we think a 3-way interaction is unlikely, but we acknowledge that it is possible.

3. Note that nervous system arousal was not directly measured or analyzed in this study and therefore we cannot be certain that arousal was mediating memory performance. Future studies should incorporate measures of arousal using galvanic skin response and/or pupil dilation to investigate this further.

References

Anderson, N., Risko, E. F., and Kingstone, A. (2011). Exploiting human sensitivity to gaze for tracking the eyes. Behav. Res. Methods 43, 843–852. doi: 10.3758/s13428-011-0078-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Argyle, M., and Dean, J. (1965). Eye-contact, distance and affiliation. Sociometry 28, 289–304. doi: 10.2307/2786027

PubMed Abstract | CrossRef Full Text | Google Scholar

Argyle, M., Lefebvre, L., and Cook, M. (1974). The meaning of five patterns of gaze. Eur. J. Soc. Psychol. 4, 125–136. doi: 10.1002/ejsp.2420040202

CrossRef Full Text | Google Scholar

Bailenson, J. N., Blascovich, J., Beall, A. C., and Loomis, J. M. (2001). Equilibrium revisited: mutual gaze and personal space in virtual environments. Presence Teleop. Virt. 10, 583–598. doi: 10.1162/105474601753272844

CrossRef Full Text | Google Scholar

Baltazar, M., Hazem, N., Vilarem, E., Beaucousin, V., Picq, J.-L., and Conty, L. (2014). Eye contact elicits bodily self-awareness in human adults. Cognition 133, 120–127. doi: 10.1016/j.cognition.2014.06.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Baron-Cohen, S. (1995). Mindblindness an essay on autism and theory of mind. (Cambridge, MA, US: The MIT Press).

Google Scholar

Bayliss, A. P., di Pellegrino, G., and Tipper, S. P. (2005). Sex differences in eye gaze and symbolic cueing of attention. Q. J. Exp. Psychol. A 58, 631–650. doi: 10.1080/02724980443000124

CrossRef Full Text | Google Scholar

Beattie, G. W. (1981). A further investigation of the cognitive interference hypothesis of gaze patterns during conversation. Br. J. Soc. Psychol. 20, 243–248. doi: 10.1111/j.2044-8309.1981.tb00493.x

CrossRef Full Text | Google Scholar

Birmingham, E., Bischof, W. F., and Kingstone, A. (2008). Gaze selection in complex social scenes. Vis. Cogn. 16, 341–355. doi: 10.1080/13506280701434532

CrossRef Full Text | Google Scholar

Birmingham, E., Bischof, W. F., and Kingstone, A. (2009). Get real! Resolving the debate about equivalent social stimuli. Vis. Cogn. 17, 904–924. doi: 10.1080/13506280902758044

CrossRef Full Text | Google Scholar

Campbell, R., Heywood, C. A., Cowey, A., Regard, M., and Landis, T. (1990). Sensitivity to eye gaze in prosopagnosic patients and monkeys with superior temporal sulcus ablation. Neuropsychologia 28, 1123–1142. doi: 10.1016/0028-3932(90)90050-X

PubMed Abstract | CrossRef Full Text | Google Scholar

Cary, M. S. (1978). The role of gaze in the initiation of conversation. Soc. Psychol. 41, 269–271. doi: 10.2307/3033565

CrossRef Full Text | Google Scholar

Connellan, J., Baron-Cohen, S., Wheelwright, S., Batki, A., and Ahluwalia, J. (2000). Sex differences in human neonatal social perception. Infant Behav. Dev. 23, 113–118. doi: 10.1016/S0163-6383(00)00032-1

CrossRef Full Text | Google Scholar

Conty, L., George, N., and Hietanen, J. K. (2016). Watching eyes effects: when others meet the self. Conscious. Cogn. 45, 184–197. doi: 10.1016/j.concog.2016.08.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Conty, L., Gimmig, D., Belletier, C., George, N., and Huguet, P. (2010). The cost of being watched: stroop interference increases under concomitant eye contact. Cognition 115, 133–139. doi: 10.1016/j.cognition.2009.12.005

CrossRef Full Text | Google Scholar

Conty, L., N’Diaye, K., Tijus, C., and George, N. (2007). When eye creates the contact! ERP evidence for early dissociation between direct and averted gaze motion processing. Neuropsychologia 45, 3024–3037. doi: 10.1016/j.neuropsychologia.2007.05.017

PubMed Abstract | CrossRef Full Text | Google Scholar

Conty, L., Tijus, C., Hugueville, L., Coelho, E., and George, N. (2006). Searching for asymmetries in the detection of gaze contact versus averted gaze under different head views: a behavioural study. Spat. Vis. 19, 529–545. doi: 10.1163/156856806779194026

PubMed Abstract | CrossRef Full Text | Google Scholar

Craik, F. I. M., and Tulving, E. (1975). Depth of processing and the retention of words in episodic memory. J. Exp. Psychol. Gen. 104, 268–294. doi: 10.1037/0096-3445.104.3.268

CrossRef Full Text | Google Scholar

Csibra, G., and Gergely, G. (2009). Natural pedagogy. Trends Cogn. Sci. 13, 148–153. doi: 10.1016/j.tics.2009.01.005

PubMed Abstract | CrossRef Full Text | Google Scholar

De Jaegher, H., Di Paolo, E., and Gallagher, S. (2010). Can social interaction constitute social cognition? Trends Cogn. Sci. 14, 441–447. doi: 10.1016/j.tics.2010.06.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Dodd, M. D., Weiss, N., McDonnell, G. P., Sarwal, A., and Kingstone, A. (2012). Gaze cues influence memory…but not for long. Acta Psychol. 141, 270–275. doi: 10.1016/j.actpsy.2012.06.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Doi, H., and Ueda, K. (2007). Searching for a perceived stare in the crowd. Perception 36, 773–780. doi: 10.1068/p5614

CrossRef Full Text | Google Scholar

Doi, H., Ueda, K., and Shinohara, K. (2009). Neural correlates of the stare-in-the-crowd effect. Neuropsychologia 47, 1053–1060. doi: 10.1016/j.neuropsychologia.2008.11.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Donovan, W. L., and Leavitt, L. A. (1980). Physiologic correlates of direct and averted gaze. Biol. Psychol. 10, 189–199. doi: 10.1016/0301-0511(80)90014-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Driver, J., Davis, G., Ricciardelli, P., Kidd, P., Maxwell, E., and Baron-Cohen, S. (1999). Gaze perception triggers reflexive visuospatial orienting. Vis. Cogn. 6, 509–540. doi: 10.1080/135062899394920

CrossRef Full Text | Google Scholar

Duncan, S. (1972). Some signals and rules for taking speaking turns in conversations. J. Pers. Soc. Psychol. 23, 283–292. doi: 10.1037/h0033031

CrossRef Full Text | Google Scholar

Ellsworth, P., and Ross, L. (1975). Intimacy in response to direct gaze. J. Exp. Soc. Psychol. 11, 592–613. doi: 10.1016/0022-1031(75)90010-4

CrossRef Full Text | Google Scholar

Emery, N. J. (2000). The eyes have it: the neuroethology, function and evolution of social gaze. Neurosci. Biobehav. Rev. 24, 581–604. doi: 10.1016/S0149-7634(00)00025-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Falck-Ytter, T., Carlström, C., and Johansson, M. (2014). Eye contact modulates cognitive processing differently in children with autism. Child Dev. 1–11. doi: 10.1111/cdev.12273

CrossRef Full Text | Google Scholar

Foulsham, T., Cheng, J. T., Tracy, J. L., Henrich, J., and Kingstone, A. (2010). Gaze allocation in a dynamic situation: effects of social status and speaking. Cognition 117, 319–331. doi: 10.1016/j.cognition.2010.09.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Foulsham, T., and Sanderson, L. A. (2013). Look who’s talking? Sound changes gaze behaviour in a dynamic social scene. Vis. Cogn. 21, 922–944. doi: 10.1080/13506285.2013.849785

CrossRef Full Text | Google Scholar

Foulsham, T., Walker, E., and Kingstone, A. (2011). The where, what and when of gaze allocation in the lab and the natural environment. Vis. Res. 51, 1920–1931. doi: 10.1016/j.visres.2011.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Freeth, M., Foulsham, T., and Kingstone, A. (2013). What affects social attention? Social presence, eye contact and autistic traits. PLoS One 8:e53286. doi: 10.1371/journal.pone.0053286

PubMed Abstract | CrossRef Full Text | Google Scholar

Friesen, C. K., and Kingstone, A. (1998). The eyes have it! Reflexive orienting is triggered by nonpredictive gaze. Psychon. Bull. Rev. 5, 490–495. doi: 10.3758/BF03208827

CrossRef Full Text | Google Scholar

Friesen, C. K., and Kingstone, A. (2003). Covert and overt orienting to gaze direction cues and the effects of fixation offset. NeuroReport 14, 489–493. doi: 10.1097/01.wnr.0000058776.36017.5d

CrossRef Full Text | Google Scholar

Friesen, C. K., Ristic, J., and Kingstone, A. (2004). Attentional effects of counterpredictive gaze and arrow cues. J. Exp. Psychol. Hum. Percept. Perform. 30, 319–329. doi: 10.1037/0096-1523.30.2.319

PubMed Abstract | CrossRef Full Text | Google Scholar

Frischen, A., and Tipper, S. P. (2006). Long-term gaze cueing effects: evidence for retrieval of prior states of attention from memory. Vis. Cogn. 14, 351–364. doi: 10.1080/13506280544000192

CrossRef Full Text | Google Scholar

Fry, R., and Smith, G. F. (1975). The effects of feedback and eye contact on performance of a digit-coding task. J. Soc. Psychol. 96, 145–146. doi: 10.1080/00224545.1975.9923275

CrossRef Full Text | Google Scholar

Fullwood, C., and Doherty-Sneddon, G. (2006). Effect of gazing at the camera during a video link on recall. Appl. Ergon. 37, 167–175. doi: 10.1016/j.apergo.2005.05.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Gale, A., Kingsley, E., Brookes, S., and Smith, D. (1978). Cortical arousal and social intimacy in the human female under different conditions of eye contact. Behav. Process. 3, 271–275. doi: 10.1016/0376-6357(78)90019-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Gallup, A. C., Chong, A., and Couzin, I. D. (2012a). The directional flow of visual information transfer between pedestrians. Biol. Lett. 8, 520–522. doi: 10.1098/rsbl.2012.0160

CrossRef Full Text | Google Scholar

Gallup, A. C., Chong, A., Kacelnik, A., Krebs, J. R., and Couzin, I. D. (2014). The influence of emotional facial expressions on gaze-following in grouped and solitary pedestrians. Sci. Rep. 4:5794. doi: 10.1038/srep05794

PubMed Abstract | CrossRef Full Text | Google Scholar

Gallup, A. C., Hale, J. J., Sumpter, D. J. T., Garnier, S., Kacelnik, A., Krebs, J. R., et al. (2012b). Visual attention and the acquisition of information in human crowds. Proc. Natl. Acad. Sci. USA 109, 7245–7250. doi: 10.1073/pnas.1116141109

CrossRef Full Text | Google Scholar

Gobel, M. S., Kim, H. S., and Richardson, D. C. (2015). The dual function of social gaze. Cognition 136, 359–364. doi: 10.1016/j.cognition.2014.11.040

PubMed Abstract | CrossRef Full Text | Google Scholar

Goodman, L. R., Phelan, H. L., and Johnson, S. A. (2012). Sex differences for the recognition of direct versus averted gaze faces. Memory 20, 199–209. doi: 10.1080/09658211.2011.651089

PubMed Abstract | CrossRef Full Text | Google Scholar

Guerin, B. (1986). Mere presence effects in humans: a review. J. Exp. Soc. Psychol. 22, 38–77. doi: 10.1016/0022-1031(86)90040-5

CrossRef Full Text | Google Scholar

Gullberg, M., and Holmqvist, K. (2006). Visual attention towards gestures in face-to-face interaction vs. on screen visual attention towards gestures in face-to-face interaction vs. on screen. Pragmatics and Cognition. Available at: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.16.2066&rep=rep1&type=pdf

Google Scholar

Hall, J. A. (1978). Gender effects in decoding nonverbal cues. Psychol. Bull. 85, 845–857. doi: 10.1037/0033-2909.85.4.845

CrossRef Full Text | Google Scholar

Hayward, D. A., and Ristic, J. (2015). Exposing the cuing task: the case of gaze and arrow cues. Atten. Percept. Psychophys. 77, 1088–1104. doi: 10.3758/s13414-015-0877-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Hazem, N., George, N., Baltazar, M., and Conty, L. (2017). I know you can see me: social attention influences bodily self-awareness. Biol. Psychol. 124, 21–29. doi: 10.1016/j.biopsycho.2017.01.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Heath, C., and Luff, P. (1992). Media space and communicative asymmetries: preliminary observations of video-mediated interaction. Hum.-Comput. Interact. 7, 315–346. doi: 10.1207/s15327051hci0703_3

CrossRef Full Text | Google Scholar

Heath, C., and Luff, P. (1993). “Disembodied conduct: interactional asymmetries in video-mediated communication” in Technology in working order: Studies of work, interaction, and technology. ed. G. Button (New York, NY: Routledge), 35–54. Available at: https://pdfs.semanticscholar.org/bb2b/742c5deaf8364ce3f863ac708f042d19e82c.pdf

Google Scholar

Helminen, T. M., Kaasinen, S. M., and Hietanen, J. K. (2011). Eye contact and arousal: the effects of stimulus duration. Biol. Psychol. 88, 124–130. doi: 10.1016/j.biopsycho.2011.07.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Helminen, T. M., Pasanen, P., and Hietanen, J. K. (2016). Learning under your gaze: the mediating role of affective arousal between perceived direct gaze and memory performance. Psychol. Res. 80, 159–171. doi: 10.1007/s00426-015-0649-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Hietanen, J. K. (2016). Eye contact: to see and to be seen. Int. J. Psychol. 51:167. Available at: https://insights.ovid.com/internationalpsychology/ijpsy/2016/07/001/eye-contact-seen/992/00011205

Google Scholar

Hietanen, J. K., Leppänen, J. M., Peltola, M. J., Linna-Aho, K., and Ruuhiala, H. J. (2008). Seeing direct and averted gaze activates the approach-avoidance motivational brain systems. Neuropsychologia 46, 2423–2430. doi: 10.1016/j.neuropsychologia.2008.02.029

PubMed Abstract | CrossRef Full Text | Google Scholar

Hietanen, J. K., Nummenmaa, L., Nyman, M. J., Parkkola, R., and Hämäläinen, H. (2006). Automatic attention orienting by social and symbolic cues activates different neural networks: an fMRI study. NeuroImage 33, 406–413. doi: 10.1016/j.neuroimage.2006.06.048

PubMed Abstract | CrossRef Full Text | Google Scholar

Hood, B. M., Macrae, C. N., Cole-Davies, V., and Dias, M. (2003). Eye remember you: the effects of gaze direction on face recognition in children and adults. Dev. Sci. 6, 67–71. doi: 10.1111/1467-7687.00256

CrossRef Full Text | Google Scholar

Itier, R. J., and Batty, M. (2009). Neural bases of eye and gaze processing: the core of social cognition. Neurosci. Biobehav. Rev. 33, 843–863. doi: 10.1016/j.neubiorev.2009.02.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Itier, R. J., Van Roon, P., and Alain, C. (2011). Species sensitivity of early face and eye processing. NeuroImage 54, 705–713. doi: 10.1016/j.neuroimage.2010.07.031

PubMed Abstract | CrossRef Full Text | Google Scholar

Itier, R. J., Villate, C., and Ryan, J. D. (2007). Eyes always attract attention but gaze-orienting is task dependent: evidence from eye movement monitoring. Neuropsychologia 45, 1019–1028. doi: 10.1016/j.neuropsychologia.2006.09.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Jarick, M., and Kingstone, A. (2015). The duality of gaze: eyes extract and signal social information during sustained cooperative and competitive dyadic gaze. Front. Psychol. 6:1423. doi: 10.3389/fpsyg.2015.01423

PubMed Abstract | CrossRef Full Text | Google Scholar

Jelicic, M., Geraerts, E. G., Merckelbach, H. L. G. J., and Guerrieri, R. (2004). Acute stress enhances memory for emotional words, but impairs memory for neutral words. Int. J. Neurosci. 114, 1343–1351. doi: 10.1080/00207450490476101

CrossRef Full Text | Google Scholar

Kampe, K. K. W., Frith, C. D., and Frith, U. (2003). “Hey John”: signals conveying communicative intention toward the self activate brain regions associated with “Mentalizing” regardless of modality. J. Neurosci. 23, 5258–5263. doi: 10.1523/JNEUROSCI.23-12-05258.2003

PubMed Abstract | CrossRef Full Text | Google Scholar

Kelley, D. H., and Gorham, J. (1988). Effects of immediacy on recall of information. Commun. Educ. 37, 198–207. doi: 10.1080/03634528809378719

CrossRef Full Text | Google Scholar

Kleinke, C. L. (1986). Gaze and eye contact: a research review. Psychol. Bull. 100, 78–100. doi: 10.1037/0033-2909.100.1.78

PubMed Abstract | CrossRef Full Text | Google Scholar

Kleinke, C. L., and Pohlen, P. D. (1971). Affective and emotional responses as a function of other person's gaze and cooperativeness in a two-person game. J. Pers. Soc. Psychol. 17, 308–313. doi: 10.1037/h0030600

PubMed Abstract | CrossRef Full Text | Google Scholar

Kleinke, C. L., Staneski, R. A., and Berger, D. E. (1975). Evaluation of an interviewer as a function of interviewer gaze, reinforcement of subject gaze, and interviewer attractiveness. J. Pers. Soc. Psychol. 31, 115–122. doi: 10.1037/h0076244

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, H. (2012). A dual-subsystem model of the brain’s default network: self-referential processing, memory retrieval processes, and autobiographical memory retrieval. NeuroImage 61, 966–977. doi: 10.1016/j.neuroimage.2012.03.025

CrossRef Full Text | Google Scholar

Kompatsiari, K., Ciardo, F., Tikhanoff, V., Metta, A., and Wykowska, A. (2018). On the role of eye contact in gaze cueing. Sci. Rep. 8:17842. doi: 10.1038/s41598-018-36136-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Kuhn, G., Caffaratti, H. A., Teszka, R., and Rensink, R. A. (2014). A psychologically-based taxonomy of misdirection. Front. Psychol. 5, 1–14. doi: 10.3389/fpsyg.2014.01392

CrossRef Full Text | Google Scholar

Kuhn, G., and Tatler, B. W. (2005). Magic and fixation: now you don’t see it, now you do. Perception 34, 1155–1161. doi: 10.1068/p3409bn1

CrossRef Full Text | Google Scholar

Kuhn, G., Tatler, B. W., and Cole, G. G. (2009). You look where I look! Effect of gaze cues on overt and covert attention in misdirection. Vis. Cogn. 17, 925–944. doi: 10.1080/13506280902826775

CrossRef Full Text | Google Scholar

Kuhn, G., Tatler, B. W., Findlay, J. M., and Cole, G. G. (2008). Misdirection in magic: implications for the relationship between eye gaze and attention. Vis. Cogn. 16, 391–405. doi: 10.1080/13506280701479750

CrossRef Full Text | Google Scholar

Kuhn, G., Teszka, R., Tenaw, N., and Kingstone, A. (2016). Don’t be fooled! Attentional responses to social cues in a face-to-face and video magic trick reveals greater top-down control for overt than covert attention. Cognition 146, 136–142. doi: 10.1016/j.cognition.2015.08.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Laidlaw, K. E. W., Foulsham, T., Kuhn, G., and Kingstone, A. (2011). Potential social interactions are important to social attention. Proc. Natl. Acad. Sci. USA 108, 5548–5553. doi: 10.1073/pnas.1017022108

PubMed Abstract | CrossRef Full Text | Google Scholar

Laidlaw, K. E. W., Risko, E. F., and Kingstone, A. (2012). A new look at social attention: orienting to the eyes is not (entirely) under volitional control. J. Exp. Psychol. Hum. Percept. Perform. 38, 1132–1143. doi: 10.1037/a0027075

PubMed Abstract | CrossRef Full Text | Google Scholar

Laidlaw, K. E. W., Rothwell, A., and Kingstone, A. (2016). Camouflaged attention: covert attention is critical to social communication in natural settings. Evol. Hum. Behav. 37, 449–455. doi: 10.1016/j.evolhumbehav.2016.04.004

CrossRef Full Text | Google Scholar

Langton, S. R. H., and Bruce, V. (1999). Reflexive visual orientation in response to the social attention of others. Vis. Cogn. 6, 541–567. doi: 10.1080/135062899394939

CrossRef Full Text | Google Scholar

Levy, J., Foulsham, T., and Kingstone, A. (2012). Monsters are people too. Biol. Lett. 9:20120850. doi: 10.1098/rsbl.2012.0850

PubMed Abstract | CrossRef Full Text | Google Scholar

Lutchmaya, S., Baron-Cohen, S., and Raggatt, P. (2002). Foetal testosterone and eye contact in 12-month-old human infants. Infant Behav. Dev. 25, 327–335. doi: 10.1016/S0163-6383(02)00094-2

CrossRef Full Text | Google Scholar

Macdonald, P. A., and Macleod, C. M. (1998). The influence of attention at encoding on direct and indirect remembering. Acta Psychol. 98, 291–310. doi: 10.1016/S00016918(97)00047-4

CrossRef Full Text | Google Scholar

Macrae, C. N., Hood, B. M., Milne, A. B., Rowe, A. C., and Mason, M. F. (2002). Are you looking at me? Eye gaze and person perception. Psychol. Sci. 13, 460–464. doi: 10.1111/1467-9280.00481

CrossRef Full Text | Google Scholar

Marschner, L., Pannasch, S., Schulz, J., and Graupner, S. T. (2015). Social communication with virtual agents: the effects of body and gaze direction on attention and emotional responding in human observers. Int. J. Psychophysiol. 97, 85–92. doi: 10.1016/j.ijpsycho.2015.05.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Mason, M. F., Hood, B. M., and Macrae, C. N. (2004). Look into my eyes: gaze direction and person memory. Memory 12, 637–643. doi: 10.1080/09658210344000152

PubMed Abstract | CrossRef Full Text | Google Scholar

Masson, M. E. J., and Loftus, G. R. (2003). Using confidence intervals for graphically based data interpretation. Can. J. Exp. Psychol. 57, 203–220. doi: 10.1037/h0087426

PubMed Abstract | CrossRef Full Text | Google Scholar

Mather, M., and Sutherland, M. R. (2011). Arousal-biased competition in perception and memory. Perspect. Psychol. Sci. 6, 114–133. doi: 10.1177/1745691611400234

PubMed Abstract | CrossRef Full Text | Google Scholar

McClure, E. B. (2000). A meta-analytic review of sex differences in facial expression processing and their development in infants, children, and adolescents. Psychol. Bull. 126, 424–453. doi: 10.1037/0033-2909.126.3.424

PubMed Abstract | CrossRef Full Text | Google Scholar

Mojzisch, A., Schilbach, L., Helmert, J. R., Pannasch, S., Velichkovsky, B. M., and Vogeley, K. (2006). The effects of self-involvement on attention, arousal, and facial expression during social interaction with virtual others: a psychophysiological study. Soc. Neurosci. 1, 184–195. doi: 10.1080/17470910600985621

PubMed Abstract | CrossRef Full Text | Google Scholar

Morey, R. D., and Rouder, J. N. (2018). BayesFactor: Computation of Bayes Factors for Common Designs. R Package Version 0.9.12–4.2. Available at: https://CRAN.R-project.org/package=BayesFactor

Google Scholar

Mulckhuyse, M., and Theeuwes, J. (2010). Unconscious attentional orienting to exogenous cues: a review of the literature. Acta Psychol. 134, 299–309. doi: 10.1016/j.actpsy.2010.03.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Myllyneva, A., and Hietanen, J. K. (2015). There is more to eye contact than meets the eye. Cognition 134, 100–109. doi: 10.1016/j.cognition.2014.09.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Myllyneva, A., and Hietanen, J. K. (2016). The dual nature of eye contact: To see and to be seen. Soc. Cogn. Affect. Neurosci. 11, 1089–1095. doi: 10.1093/scan/nsv075

PubMed Abstract | CrossRef Full Text | Google Scholar

Nasiopoulos, E., Risko, E. F., and Kingstone, A. (2015). “Social attention, social presence, and the dual function of gaze” in The many faces of social attention eds. A. Puce and B. Bertenthal (Cham: Springer), 129–155.

Google Scholar

Nemeth, D., Turcsik, A. B., Farkas, G., and Janacsek, K. (2013). Social communication impairs working-memory performance. Appl. Neuropsychol. Adult 20, 211–214. doi: 10.1080/09084282.2012.685134

CrossRef Full Text | Google Scholar

Nichols, K. A., and Champness, B. G. (1971). Eye gaze and the GSR. J. Exp. Soc. Psychol. 7, 623–626. doi: 10.1016/0022-1031(71)90024-2

CrossRef Full Text | Google Scholar

Niederehe, G., and Duncan, S. (1974). On signalling it’s your turn to speak. J. Exp. Soc. Psychol. 10, 234–247. doi: 10.1016/00221031(74)90070-5

CrossRef Full Text | Google Scholar

Northoff, G., Heinzel, A., de Greck, M., Bermpohl, F., Dobrowolny, H., and Panksepp, J. (2006). Self-referential processing in our brain – A meta-analysis of imaging studies on the self. NeuroImage 31, 440–457. doi: 10.1016/j.neuroimage.2005.12.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Otteson, J. P., and Otteson, C. R. (1980). Effect of teacher’s gaze on children’s story recall. Percept. Mot. Skills 50, 35–42. doi: 10.2466/pms.1980.50.1.35

CrossRef Full Text | Google Scholar

Pageler, N. M., Menon, V., Merin, N. M., Eliez, S., Brown, W. E., and Reiss, A. L. (2003). Effect of head orientation on gaze processing in fusiform gyrus and superior temporal sulcus. NeuroImage 20, 318–329. doi: 10.1016/S1053-8119(03)00229-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Palanica, A., and Itier, R. J. (2011). Searching for a perceived gaze direction using eye tracking. J. Vis. 11:19. doi: 10.1167/11.2.19

PubMed Abstract | CrossRef Full Text | Google Scholar

Patterson, M. L., Iizuka, Y., Tubbs, M. E., Ansel, J., Tsutsumi, M., and Anson, J. (2007). Passing encounters east and west: comparing Japanese and American pedestrian interactions. J. Nonverbal Behav. 31, 155–166. doi: 10.1007/s10919-007-0028-4

CrossRef Full Text | Google Scholar

Patterson, M. L., Webb, A., and Schwartz, W. (2002). Passing encounters: patterns of recognition and avoidance in pedestrians. Basic Appl. Soc. Psychol. 24, 57–66. doi: 10.1207/S15324834BASP2401

CrossRef Full Text | Google Scholar

Perrett, D., and Emery, N. (1994). Understanding the intentions of others from visual signals: neurophysiological evidence. Curr. Psychol. Cogn. 13, 683–694. Available at: http://psycnet.apa.org/record/1995-24608-001

Google Scholar

Pönkänen, L. M., Alhoniemi, A., Leppänen, J. M., and Hietanen, J. K. (2011a). Does it make a difference if I have an eye contact with you or with your picture? An ERP study. Soc. Cogn. Affect. Neurosci. 6, 486–494. doi: 10.1093/scan/nsq068

CrossRef Full Text | Google Scholar

Pönkänen, L. M., Peltola, M. J., and Hietanen, J. K. (2011b). The observer observed: frontal EEG asymmetry and autonomic responses differentiate between another person’s direct and averted gaze when the face is seen live. Int. J. Psychophysiol. 82, 180–187. doi: 10.1016/j.ijpsycho.2011.08.006

CrossRef Full Text | Google Scholar

Posner, M. I. (1980). Orienting of attention. Q. J. Exp. Psychol. 32, 3–25. doi: 10.1080/00335558008248231

CrossRef Full Text | Google Scholar

Rensink, R. A., and Kuhn, G. (2015). A framework for using magic to study the mind. Front. Psychol. 6, 1–14. doi: 10.3389/fpsyg.2014.01508

CrossRef Full Text | Google Scholar

Risko, E. F., and Kingstone, A. (2011). Eyes wide shut: implied social presence, eye tracking and attention. Atten. Percept. Psychophys. 73, 291–296. doi: 10.3758/s13414-010-0042-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Risko, E. F., and Kingstone, A. (2015). “Attention in the wild: visual attention in complex, dynamic and social environments” in Cambridge handbook of applied perception research. eds. R. R. Hoffman, P. A. Hancock, M. W. Scerbo, R. Prasuraman, and J. L. Szalma (Cambridge, UK: Cambridge University Press), 466–487.

Google Scholar

Risko, E. F., Laidlaw, K. E. W., Freeth, M., Foulsham, T., and Kingstone, A. (2012). Social attention with real versus reel stimuli: toward an empirical approach to concerns about ecological validity. Front. Hum. Neurosci. 6:143. doi: 10.3389/fnhum.2012.00143

PubMed Abstract | CrossRef Full Text | Google Scholar

Risko, E. F., Richardson, D. C., and Kingstone, A. (2016). Breaking the fourth wall of cognitive science: real-world social attention and the dual function of gaze. Curr. Dir. Psychol. Sci. 25, 70–74. doi: 10.1177/0963721415617806

CrossRef Full Text | Google Scholar

Ristic, J., Friesen, C. K., and Kingstone, A. (2002). Are eyes special? It depends on how you look at it. Psychon. Bull. Rev. 9, 507–513. doi: 10.3758/BF03196306

PubMed Abstract | CrossRef Full Text | Google Scholar

Ristic, J., Mottron, L., Friesen, C. K., Iarocci, G., Burack, J. a., and Kingstone, A. (2005). Eyes are special but not for everyone: the case of autism. Brain Res. Cogn. Brain Res. 24, 715–718. doi: 10.1016/j.cogbrainres.2005.02.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Ristic, J., Wright, A., and Kingstone, A. (2007). Attentional control and reflexive orienting to gaze and arrow cues. Psychon. Bull. Rev. 14, 964–969. doi: 10.3758/BF03194129

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenthal, R., Hall, J. A., DiMatteo, M. R., Rogers, P. L., and Archer, D. (1979). Sensitivity to nonverbal communication: The PONS test. (Baltimore: John Hopkins University Press).

Google Scholar

Schilbach, L. (2010). A second-person approach to other minds. Nat. Rev. Neurosci. 11:449. doi: 10.1038/nrn2805-c1

PubMed Abstract | CrossRef Full Text | Google Scholar

Schilbach, L. (2015). Eye to eye, face to face and brain to brain: novel approaches to study the behavioral dynamics and neural mechanisms of social interactions. Curr. Opin. Behav. Sci. 3, 130–135. doi: 10.1016/j.cobeha.2015.03.006

CrossRef Full Text | Google Scholar

Schilbach, L., Timmermans, B., Vasudevi, R., Costall, A., Bente, G., Schlicht, T., et al. (2013). Toward a second-person neuroscience. Behav. Brain Sci. 36, 1–77. Available at: http://journals.cambridge.org/abstract_S0140525X12000660

Google Scholar

Schilbach, L., Wilms, M., Eickhoff, S. B., Romanzetti, S., Tepest, R., Bente, G., et al. (2010). Minds made for sharing: initiating joint attention recruits reward-related neurocircuitry. J. Cogn. Neurosci. 22, 2702–2715. doi: 10.1162/jocn.2009.21401

PubMed Abstract | CrossRef Full Text | Google Scholar

Schneier, F. R., Rodebaugh, T. L., Blanco, C., Lewin, H., and Liebowitz, M. R. (2011). Fear and avoidance of eye contact in social anxiety disorder. Compr. Psychiatry 52, 81–87. doi: 10.1016/j.comppsych.2010.04.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Schrammel, F., Pannasch, S., Graupner, S. T., Mojzisch, A., and Velichkovsky, B. M. (2009). Virtual friend or threat? the effects of facial expression and gaze interaction on psychophysiological responses and emotional experience. Psychophysiology 46, 922–931. doi: 10.1111/j.1469-8986.2009.00831.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Senju, A., and Csibra, G. (2008). Gaze following in human infants depends on communicative signals. Curr. Biol. 18, 668–671. doi: 10.1016/j.cub.2008.03.059

PubMed Abstract | CrossRef Full Text | Google Scholar

Senju, A., and Hasegawa, T. (2005). Direct gaze captures visuospatial attention. Vis. Cogn. 12, 127–144. doi: 10.1080/13506280444000157

CrossRef Full Text | Google Scholar

Senju, A., Hasegawa, T., and Tojo, Y. (2005). Does perceived direct gaze boost detection in adults and children with and without autism? The stare-in-the-crowd effect revisited. Vis. Cogn. 12, 1474–1496. doi: 10.1080/13506280444000797

CrossRef Full Text | Google Scholar

Senju, A., and Johnson, M. H. (2009a). Atypical eye contact in autism: models, mechanisms and development. Neurosci. Biobehav. Rev. 33, 1204–1214. doi: 10.1016/j.neubiorev.2009.06.001

CrossRef Full Text | Google Scholar

Senju, A., and Johnson, M. H. (2009b). The eye contact effect: mechanisms and development. Trends Cogn. Sci. 13, 127–134. doi: 10.1016/j.tics.2008.11.009

CrossRef Full Text | Google Scholar

Sherwood, J. V. (1987). Facilitative effects of gaze upon learning. Percept. Mot. Skills 64, 1275–1278. doi: 10.2466/pms.1987.64.3c.1275

CrossRef Full Text | Google Scholar

Shimojo, S., Simion, C., Shimojo, E., and Scheier, C. (2003). Gaze bias both reflects and influences preference. Nat. Neurosci. 6, 1317–1322. doi: 10.1038/nn1150

PubMed Abstract | CrossRef Full Text | Google Scholar

Shin, M.-J., Marrett, N., and Lambert, A. J. (2011). Visual orienting in response to attentional cues: spatial correspondence is critical, conscious awareness is not. Vis. Cogn. 19, 730–761. doi: 10.1080/13506285.2011.582053

CrossRef Full Text | Google Scholar

Smeets, T., Giesbrecht, T. M., Jelicic, M., and Merckelbach, H. L. G. J. (2007). Context-dependent enhancement of declarative memory performance following acute psychosocial stress. Biol. Psychol. 76, 116–123. doi: 10.1016/j.biopsycho.2007.07.001

PubMed Abstract | CrossRef Full Text | Google Scholar

Smith, A. D., Hood, B. M., and Hector, K. (2006). Eye remember you two: gaze direction modulates face recognition in a developmental study. Dev. Sci. 9, 465–472. doi: 10.1111/j.1467-7687.2006.00513.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Tatler, B. W., and Kuhn, G. (2007). “Don’t look now: the magic of misdirection” in Eye movements: A window on mind and brain. eds. R. P. G. Van Gompel, M. H. Fischer, W. S. Murray, and R. L. Hill (Elsevier Science), 697–714.

Google Scholar

Teufel, C., Fletcher, P. C., and Davis, G. (2010). Seeing other minds: attributed mental astates influence perception. Trends Cogn. Sci. 14, 376–382. doi: 10.1016/j.tics.2010.05.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Thorndike, E. L., and Lorge, I. (1944). The teacher’s word book of 30,000 words. (Oxford, England: Bureau of Publications, Teachers Co.).

Google Scholar

Tomasello, M., Call, J., and Hare, B. (1998). Five primate species follow the visual gaze of conspecifics. Anim. Behav. 55, 1063–1069. doi: 10.1006/anbe.1997.0636

PubMed Abstract | CrossRef Full Text | Google Scholar

Tomasello, M., Carpenter, M., Call, J., Behna, T., and Moll, H. (2005). Understanding and sharing intentions: the origins of cultural cognition. Behav. Brain Sci. 28, 675–735. doi: 10.1017/S0140525X05000129

PubMed Abstract | CrossRef Full Text | Google Scholar

Varao-Sousa, T. L., and Kingstone, A. (2015). Memory for lectures: how lecture format impacts the learning experience. PLoS One 10:e0141587. doi: 10.1371/journal.pone.0141587

PubMed Abstract | CrossRef Full Text | Google Scholar

Vertegaal, R., Slagter, R., Van Der Veer, G., and Nijholt, A. (2001). “Eye gaze patterns in conversations: there is more to conversational agents than meets the eyes” in Proceedings of the SIGCHI conference on human factors in computing systems (AMC). 3, 301–308.

Google Scholar

von Grunau, M., and Anston, C. (1995). The detection of gaze direction: a stare-in-the-crowd effect. Perception 24, 1297–1313. doi: 10.1068/p241297

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuilleumier, P. (2002). Perceived gaze direction in faces and spatial attention: a study in patients with parietal damage and unilateral neglect. Neuropsychologia 40, 1013–1026. doi: 10.1016/S0028-3932(01)00153-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Vuilleumier, P., George, N., Lister, V., Armony, J., and Driver, J. (2005). Effects of perceived mutual gaze and gender on face processing and recognition memory. Vis. Cogn. 12, 85–101. doi: 10.1080/13506280444000120

CrossRef Full Text | Google Scholar

Walker-Smith, G. J., Gale, A. G., and Findlay, J. M. (1977). Eye movement strategies involved in face perception. Perception 6, 313–326. doi: 10.1068/p060313

PubMed Abstract | CrossRef Full Text | Google Scholar

Wammes, J. D., and Smilek, D. (2017). Examining the influence of lecture format on degree of mind wandering. J. Appl. Res. Mem. Cogn. 6, 174–184. doi: 10.1016/j.jarmac.2017.01.015

CrossRef Full Text | Google Scholar

Wesselmann, E. D., Cardoso, F. D., Slater, S., and Williams, K. D. (2012). To be looked at as though air: civil attention matters. Psychol. Sci. 23, 166–168. doi: 10.1177/0956797611427921

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, E., Wykowska, A., Zwickel, J., and Müller, H. J. (2012). I see what you mean: how attentional selection is shaped by ascribing intentions to others. PLoS One 7:e45391. doi: 10.1371/journal.pone.0045391

PubMed Abstract | CrossRef Full Text | Google Scholar

Wieser, M. J., Pauli, P., Alpers, G. W., and Mühlberger, A. (2009). Is eye to eye contact really threatening and avoided in social anxiety?-An eye-tracking and psychophysiology study. J. Anxiety Disord. 23, 93–103. doi: 10.1016/j.janxdis.2008.04.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, D. W.-L., Bischof, W. F., and Kingstone, A. (2013). Looking while eating: the importance of social context to social attention. Sci. Rep. 3:2356. doi: 10.1038/srep02356

PubMed Abstract | CrossRef Full Text | Google Scholar

Wu, D. W.-L., Bischof, W. F., and Kingstone, A. (2014). Natural gaze signaling in a social context. Evol. Hum. Behav. 35, 211–218. doi: 10.1016/j.evolhumbehav.2014.01.005

CrossRef Full Text | Google Scholar

Wykowska, A., Wiese, E., Prosser, A., and Müller, H. J. (2014). Beliefs about the minds of others influence how we process sensory information. PLoS One 9:e94399. doi: 10.1371/journal.pone.0094339

PubMed Abstract | CrossRef Full Text | Google Scholar

Yarbus, A. L. (ed.) (1967). “Eye movements during perception of complex objects” in Eye movements and vision. US: Springer, 171–211.

Google Scholar

Yee, N., Bailenson, J. N., Urbanek, M., Chang, F., and Merget, D. (2007). The unbearable likeness of being digital: the persistence of nonverbal social norms in online virtual environments. Cyberpsychol. Behav. 10, 115–121. doi: 10.1089/cpb.2006.9984

PubMed Abstract | CrossRef Full Text | Google Scholar

Zuckerman, M., Miserandino, M., and Bernieri, F. (1983). Civil inattention exists—in elevators. Personal. Soc. Psychol. Bull. 9, 578–586. doi: 10.1177/0146167283094007

CrossRef Full Text | Google Scholar

Zwickel, J., and Võ, M. L.-H. (2010). How the presence of persons biases eye movements. Psychon. Bull. Rev. 17, 257–262. doi: 10.3758/PBR.17.2.257

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: gaze, eye contact, attention, memory, gender

Citation: Lanthier SN, Jarick M, Zhu MJH, Byun CSJ and Kingstone A (2019) Socially Communicative Eye Contact and Gender Affect Memory. Front. Psychol. 10:1128. doi: 10.3389/fpsyg.2019.01128

Received: 06 November 2018; Accepted: 29 April 2019;
Published: 22 May 2019.

Edited by:

Paola Ricciardelli, University of Milano-Bicocca, Italy

Reviewed by:

Agnieszka Wykowska, Istituto Italiano di Tecnologia, Italy
Juan Lupiáñez, University of Granada, Spain

Copyright © 2019 Lanthier, Jarick, Zhu, Byun and Kingstone. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sophie N. Lanthier, c25sYW50aGllckBnbWFpbC5jb20=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.