Skip to main content

BRIEF RESEARCH REPORT article

Front. Hum. Neurosci., 25 October 2023
Sec. Cognitive Neuroscience

Face processing and early event-related potentials: replications and novel findings

  • Department of Psychology, California State University of San Bernardino, San Bernardino, CA, United States

This research explores early Event-Related Potentials (ERPs) sensitivity to facial stimuli, investigating various facial features aimed to unveil underlying neural mechanisms. Two experiments, each involving 15 undergraduate students, utilized a multidimensional stimulus set incorporating race, gender, age, emotional expression, face masks, and stimulus orientation. Findings highlight significant modulations in N170 and P200 amplitudes and latencies for specific attributes, replicating prior research and revealing novel insights. Notably, age-related facial feature variations, facial inversion, and the presence of face masks significantly impact neural responses. Several speculative explanations are proposed to elucidate these results: First, the findings lend support to the idea that the increased N170 amplitude observed with facial inversion is closely tied to the activation of object-sensitive neurons. This is further bolstered by a similar amplitude increase noted when masks (effective objects) are added to faces. Second, the absence of an additional amplitude increase, when inverting face images with face masks suggests that neural populations may have reached a saturation point, limiting further enhancement. Third, the study reveals that the latency deficit in N170 induced by facial inversion is even more pronounced in the subsequent ERP component, the P200, indicating that face inversion may impact multiple stages of face processing. Lastly, the significant increase in P200 amplitude, typically associated with face typicality, for masked faces in this study aligns with previous research that demonstrated elevated P200 amplitudes for scrambled faces. This suggests that obscured faces may be processed as typical, potentially representing a default state in face processing.

1. Introduction

Electroencephalography (EEG) plays a crucial role in cognitive neuroscience, particularly in studying facial processing (Rossion, 2014). The favorable location of the Fusiform Face Area (FFA) in the inferior temporal cortex, specialized for facial stimuli, enhances EEG data collection through scalp sensors. Researchers utilize EEG to investigate FFA activity, exploring various factors and cognitive processes. Participants encounter a range of facial and non-facial stimuli, engaging in tasks like memory and identification. This approach yields event-related potentials (ERPs) obtained through EEG data averaging, revealing positive and negative voltage deflections (ERP components) (Rossion and Jacques, 2008). These components correspond to distinct neural sources, encapsulating different facial processing stages.

Key ERP components commonly addressed in early vision studies include the P1 (P100), a positive deflection occurring approximately 100 ms post-stimulus onset. It primarily reflects occipital lobe activity and exhibits heightened responses to facial stimuli (Itier and Taylor, 2004; Herrmann et al., 2005; Kaltwasser et al., 2014; Moradi et al., 2017; Gantiva et al., 2020). The N170, a negative deflection within the 130–200 ms timeframe, originates from FFA, showing larger amplitudes for faces and a pronounced sensitivity to facial features and configurations (Johnson, 2005; Rossion and Jacques, 2011; Hinojosa et al., 2015; Schindler and Bublatzky, 2020). Lastly, the P2 (P200) component peaks around 150–275 ms post-stimulus onset, characterized by its sensitivity to attentional processes (Carretié et al., 2013) and facial prototypicality (Schweinberger and Neumann, 2016).

Studying the Fusiform Face Area (FFA) is of utmost importance because it sheds light on a fundamental aspect of brain processing: the brain’s specialization in processing specific information categories like faces, places, tools, and body parts, which likely extends to other sensory modalities and cognitive processes (Pascual-Leone and Hamilton, 2001).

For a visual stimulus to elicit a face-sensitive N170 response, it must contain enough information in terms of local elements and their arrangement to create the perception of a face (Rossion and Jacques, 2011). Due to the cyclical resurgence of respiratory viruses, face masks are expected to persist. A recent study (Freud et al., 2022) reports that adding masks significantly impairs face recognition, contradicting the notion of easy adaptation to a masked world. This raises the question of how mask-wearing affects early face processing and the retrieval of vital facial information like age, gender, race/ethnicity, and facial expressions, characteristics, usually easily discernible in unmasked faces.

To that extent, a set of facial stimuli was carefully curated, varying across five binary dimensions: gender (male/female), race (white/black), facial expression (happy/angry), age (young/old), and the presence or absence of a face mask. This approach not only facilitates the exploration of interactions among these variables but also offers the opportunity to reexamine prior research with divergent findings regarding the influence of these factors on P100 and N170 processes. Some studies, for instance, reported no discrimination by N170 based on emotional expression (Eimer and Holmes, 2002; Herrmann et al., 2002; Eimer et al., 2003), while others noted larger amplitudes in response to fearful faces (Batty and Taylor, 2003; Williams et al., 2006; Blau et al., 2007; Luo et al., 2010), some even at earlier latencies (Walker et al., 2008). Social category modulation of N170 has also produced mixed results, with some studies showing no effect (Caldara et al., 2004; He et al., 2009; Wiese et al., 2009) or increased N170 responses to other-race faces (Walker et al., 2008). Conversely, no gender effects have been reported on P100 or N170 components (Mouchetant-Rostaing et al., 2000; Mouchetant-Rostaing and Giard, 2003). Regarding facial age, both young and older participants exhibit heightened N170 amplitudes when presented with older faces compared to younger ones (Wiese et al., 2008; Wiese, 2012). However, the N170’s sensitivity to age-related factors diminishes when age and race/ethnicity factors are presented concurrently, suggesting potential modulation by contextual or task-related variables (Wiese, 2012). These findings indicate that N170 responsiveness to age differs from its reactivity to race/ethnicity. These discrepancies across studies likely stem from variations in stimulus characteristics, task demands, experimental design, and stimulus presentation, posing challenges for direct comparisons.

The findings from 15 participants in the study revealed a significant and substantial increase in both N170 and P200 activity when comparing masked and unmasked faces, as elaborated in the Results and Discussion sections. This heightened N170 response bears a resemblance to the pattern observed with inverted faces (Rossion et al., 2000), which is thought to involve the additional engagement of object-sensitive neurons, supported by evidence from fMRI studies. These studies suggest that inverted faces become more akin to objects, eliciting stronger responses in object-sensitive brain regions (Yovel and Kanwisher, 2005; Epstein et al., 2006) while reducing activity in face-selective areas (Yovel and Kanwisher, 2005; Mazard et al., 2006). Given that face masks, perceptually, introduce an element of both object addition and reduced facial visibility, it is plausible that the hypothesis regarding inverted faces may extend to masked faces. Consequently, it can be hypothesized that inverted and masked faces elicit comparable ERP patterns. To investigate this hypothesis, the original experiment was modified by introducing a new binary dimension: stimulus orientation (upright/inverted). To maintain an equivalent number of trials as the original experiment, the binary dimension of emotional expression was simplified to include only “happy” faces (see Figure 1A). In this second experiment, an additional 15 participants were recruited.

FIGURE 1
www.frontiersin.org

Figure 1. Experimental procedures and stimulus selection. (A) Illustration of experimental stimuli. This panel presents representative samples of the stimuli employed in two distinct experiments, each comprising a total of 416 stimuli. These stimuli were strategically chosen to investigate alterations in neural responses to facial stimuli, specifically examining the impact of face masks. The stimulus set was meticulously designed to encompass a balanced representation of various facial attributes, including an equitable distribution of both Caucasian and African American faces (dimension 1), male and female faces (dimension 2), as well as happy and angry facial expressions (dimension 3), drawn from the openly accessible RADIATE face database. To introduce an additional dimension, an “older” rendition of the faces was created using the photo editing application FaceApp. Furthermore, a fifth dimension (presence or absence of a mask) was introduced by digitally incorporating masks onto every image using Adobe Photoshop. For enhanced interpretability of the study results, a secondary experimental study was devised, which also featured inverted face images. Notably, the “emotion” dimension was excluded from the stimulus set in this second study to prevent an unwieldy number of trials. (B) Stimulus presentation and response task. All 416 stimuli are presented in random order. Participants are instructed to use a button box to indicate whether the displayed face is male or female. Faces are displayed for a minimum of 1 s and remain on screen until a participant response is recorded or up to 4 s if no response is detected. The intertrial interval lasts 1 s, during which a cross is displayed in the center of the monitor. (C) Topographic map with sensor locations. This panel illustrates the topographic map displaying the sensor locations utilized in both experiments. The reported study results are based on the averaged signal recorded from the sensors highlighted in yellow. However, topographic maps based on all sensors can be found in the Supplementary material section (see Supplementary Figures S1–S5).

To tackle the issue of the multiple comparison problem, a concern that frequently impacts ERP-related research and has been thoroughly discussed by Luck and Gaspelin (2017), the analysis focused on a single electrode combination for all assessments and a robust non-parametric randomization test to evaluate variations in amplitude and latencies among the various conditions (see Analysis in the Materials and methods section for more information).

2. Materials and methods

2.1. Subjects

This study involved two experiments, each with 15 undergraduate students. In experiment 1, there were 10 female participants and 5 male participants, all with a mean age of 24 years. Experiment 2 consisted of 12 female participants and 3 male participants, with a mean age of 25 years. All participants were psychology majors at the California State University of San Bernardino, and they received class credit in appreciation of their participation. Prior to their involvement, each student provided informed consent, and no student participated in more than one experiment.

2.2. Stimuli

To generate the image dataset used for Experiment 1, original face images from 26 white and 26 black models (equally distributed across gender) were selected from the RADIATE database (Conley et al., 2018). This database provides open-access face stimuli, featuring racially and ethnically diverse models displaying various emotional expressions. Two emotional expressions, namely “happy” and “angry,” were chosen for each model, resulting in a set of 104 unique face images. To expand the dataset, an AI aging filter (FaceApp) was employed to create an “older version” of each model, effectively doubling the number of stimuli. Subsequently, the dataset was doubled again by digitally adding a facemask to each face image, employing Adobe Photoshop. As a result, the dataset was expanded to include a total of 416 face images. For Experiment 2, the initial image dataset from Experiment 1 was halved by removing all “angry” faces. Then, the dataset was brought back to a total of 416 images by adding an inversed version of the remaining faces. Throughout both experiments, all face images were thus equally divided based on several attributes, including race (black/white), gender (male/female), age (young/old), use of face mask (mask/no mask), and emotional expression (happy/angry). In Experiment 2, an additional attribute, face orientation (upright/inverted), was considered, in place of emotional expression. Figure 1A illustrates examples of the stimuli used for experiments 1 and 2, featuring one black and one white male and female model and the stimuli derived from it.

2.3. Experimental procedure and EEG equipment

The experiment involved participants wearing a 64-channel EEG-cap (BrainVision), with only 32 channels effectively utilized (sampling rate of 500 Hz) and the reference electrode placed at the FCz location, following the standard 10–20 EEG system. The EEG electrodes were connected to a BrainVision actiCHamp active channel amplifier (BrainVision) and checked for proper conductivity (impedance below 5kΩ for each electrode) before starting the recording. During the recording session, participants were seated in a dimly lit, quiet room, in front of a 19-inch Dell monitor, positioned 50 cm away from their heads. The experiment presented 416 face images (see “Stimuli”) sequentially, each displayed at the center of the screen with a visual angle of 17° × 23°. The order of the images was randomized to minimize any potential biases. To ensure participants’ attention, they were instructed to indicate the gender of each face using a button box (see Figure 1B). Each face was displayed for at least 1 s, and disappeared from the screen as soon as it was evaluated, followed by a 1-s inter-trial interval displaying a cross at the center of the screen before the next face appeared (see Figure 1B). Participants were encouraged to respond as quickly and accurately as possible. Any images not evaluated within 4 s disappeared from the screen to maintain the experimental flow. The entire experimental paradigm was created using Experiment Builder by SR Research.

2.4. Analysis

2.4.1. Stimulus presentation and data acquisition

Each stimulus used in both Experiment 1 and 2 contained a small black square engineered in the bottom right corner, allowing precise timing information through a screen-positioned photodiode. This setup ensured accurate timestamps for the onset and offset of the 416 face stimuli used in the experiments.

2.4.2. Data segmentation and preprocessing

The EEG data were analyzed offline using the FieldTrip Matlab software toolbox (Oostenveld et al., 2011). Noisy trials were removed using the Fieldtrip data browser function, and the timestamps were utilized to generate 416 data segments, each lasting 2 s, spanning 0.5 s before stimulus onset to 1.5 s after onset. This broader range than needed was selected to prevent edge effects caused by preprocessing. Subsequently, the raw data was filtered (3–45 Hz) and demeaned. The unusual high-pass filter (3 Hz) was selected because it effectively removed slow drifts observed in the EEG signal (Pinto et al., 2019). DFT notch filters were applied at 60 and 120 Hz.

2.4.3. Electrode selection and grand average ERP waveforms

In alignment with established practices in the field (e.g., Gao et al., 2009), ERP waveforms and their associated components were computed by averaging across specific occipito-temporal electrodes, namely P7, P3, O1, O2, P4, and P8 (as illustrated in Figure 1C, highlighted in yellow). A consistent electrode selection was maintained across all ERP components to prevent multiple implicit comparisons (Luck and Gaspelin, 2017). However, data from all recorded sites were utilized to generate topographic maps, which can be found in the Supplementary material. Following the averaging of data for each pair of experimental conditions and each participant, grand average ERP waveforms were computed by averaging across the results from all 15 participants for each experiment. The shaded regions in Figures 2, 3 represent the standard error of the grand average, denoted as ±SE. To ensure smooth curves and error bands, a 5-point moving average with a window of 0.010 s was applied.

FIGURE 2
www.frontiersin.org

Figure 2. ERP waveforms for Experiment 1. This figure presents the grand mean ERP waveforms, computed from a total of 15 datasets with 416 trials. The waveforms are grouped according to specific conditions: “young/old” (A), “white/black” (B), “male/female” (C), “mask/no mask” (D), the first and second half of the session (E), and “angry/happy” (F). To generate the conditions depicted in each panel, the electrode data from the left-side cluster (P7, P3, and O1) and the right-side cluster (P4, P8, and O2) were first averaged across all participants. The shaded areas in the graphs represent the standard error (+/−) with N = 15 participants. Significance testing along the waveforms for differences between the two conditions was conducted using paired t-tests (see Materials and methods) with an alpha level of 0.01 to account for multiple comparisons. Statistically significant differences are indicated by pink horizontal lines beneath the waveforms. For amplitude comparisons between the P100, N170, and P200 components of both conditions, a non-parametric randomization test was employed (see Materials and methods and Table 1). Statistically significant differences are represented by one (p < 0.05), two (p < 0.01), or three (p < 0.001) asterisks at the relevant locations.

FIGURE 3
www.frontiersin.org

Figure 3. ERP waveforms for Experiment 2. This figure illustrates the grand mean ERP waveforms, derived from 15 datasets with 416 trials. Like Figure 2, the waveforms are grouped according to specific conditions: “young/old” (A), “white/black” (B), “male/female” (C), “mask/no mask” (D), and the first and second half of the session (E). However, in Experiment 2, the analysis was performed separately for upright (left subpanels) and inverted (right subpanels) stimuli. (F) Displays the grand mean ERP waveforms for the “mask vs. no mask” conditions, both for upright stimuli (solid curves) and inverted stimuli (dotted curves). The shaded areas in the graphs represent the standard error (+/−) with N = 15 participants. Significance testing along the waveforms for differences between the two conditions (A–E) was conducted using paired t-tests (see Materials and methods) with an alpha level of 0.01 to account for multiple comparisons. Statistically significant differences are indicated by pink horizontal lines beneath the waveforms. For amplitude comparisons between the P100, N170, and P200 components of each pair of conditions (A–E) and for the two “mask” conditions or the two “no mask” conditions (F), a non-parametric randomization test was employed (see Materials and methods and Table 1). Statistically significant differences are represented by one (p < 0.05), two (p < 0.01), or three (p < 0.001) asterisks at the relevant locations. The asterisks in panel F are color-coded to distinguish between comparisons of the two “mask” conditions (blue) and the two “no mask” conditions (red).

2.4.4. Statistical analysis

To evaluate statistical distinctions along the ERP waveforms across experimental conditions, the study implemented a “running p-value” approach utilizing a paired t-test with a sample size of N = 15. This method involved assessing variations in intervals of 0.010 s (equivalent to 5 samples) throughout the ERP waveforms. Pink horizontal line segments were incorporated to indicate locations where p-values < 0.01 were observed beneath the curves for enhanced visual representation.

Although the study is structured around a factorial design featuring 5 independent variables, only three-way ANOVA (using the Matlab function anovan.m) were utilized to investigate interactions between different pairs of independent variables concerning the amplitudes of P100, N170, and P200 (see Supplementary material for results). This choice was made because higher-dimensional ANOVAs, would of have required averaging smaller trial quantities and potentially led to unstable ERP waveforms and associated components.

For the presentation of topographic maps illustrating the amplitudes of the various ERP components, the Matlab function plottopography.m was employed (available through the Mathworks file exchange). Detailed explanations regarding the derivation of ERP component peak values, utilized for both the ANOVA and the creation of topographic maps, are provided in the subsequent section.

2.4.5. Non-parametric approach for amplitude comparison

Amplitude differences in the EEG components (P100, N170, P200) between two experimental conditions were explored using a non-parametric approach. This approach involved 100,000 iterations, during which—for each iteration—data from 416 trials for each participant were randomly divided into two pseudo conditions, yielding grand average ERP waveforms for each condition by averaging across occipito-temporal electrodes and participants (see above). Amplitude differences for each pair of pseudo-ERP waveforms were calculated within specific time windows: 80–120 ms (P100), 150–300 ms (P200), and 100–200 ms (N170) for each iteration. The resulting 100,000 amplitude differences were sorted from high to low, allowing computation of the ranking of the experimentally observed differences for each pair of experimental conditions and each ERP component of interest among the randomizations. For instance, a rank of 5 indicated that only 4 out of 100,000 randomizations yielded a greater amplitude difference, resulting in a probability of 5/100,000 (thus, p = 0.00005). In Experiment 2, this non-parametric test was performed separately for trials featuring upright and inverse face images, as well as for trials featuring face images with and without masks (see Table 1). An identical approach, using the same time windows for each ERP component, was also used to compute differences in latency between masked and unmasked faces, and upright and inverted faces.

TABLE 1
www.frontiersin.org

Table 1. Non-parametric randomization test results.

3. Results

3.1. Experiment 1

Participants viewed 416 images, which were presented one by one in random order. Their sole task was to identify whether each face image depicted a female or a male by using a button box (see Figure 1B).

The 416-image stimulus set was purposely designed to be split into two sets of 208 images each, based on five distinct face attributes (see Figure 1A for examples). Consequently, this division resulted in five pairs of grand averaged ERP waveforms. These pairs were categorized as follows: “young” and “old” (see Figure 2A), “white Caucasian” and “Black African American” (see Figure 2B), “male” and “female” (see Figure 2C), “mask” and “no mask” (see Figure 2D), and “angry” and “happy” (see Figure 2F). Additionally, for each participant, the trials were divided into two groups: those presented first (trials 1 to 208) and those presented last (trials 209–416) (see Figure 2E).

To visualize and analyze the differences in the measured EEG signals between the selected pairs of conditions (see Materials and methods), the following features were included in each panel: a shaded error band representing ± standard error for each waveform, short horizontal lines under each pair of waveforms to indicate the timepoints at which the waveforms statistically differed from each other (with alpha = 0.01), and asterisks (see caption Figure 2) to indicate differences in amplitudes between the waveforms for the key ERP components (P100, N170, and P200) that are relevant to this study, based upon a non-parametric test (see Materials and methods).

Results, obtained through a 3-way ANOVA, aimed at exploring interactions among the binary dimensions of “age,” “emotion,” and “mask,” as well as “gender,” “race,” and “mask,” and their impact on the amplitudes of the P100, N170, and P200 components, are presented in Supplementary Tables S1, S2. Complementing these findings, Supplementary Figures S1, S2 show topographic maps depicting amplitude distributions for these components under various conditions.

3.2. Experiment 2

Following experiment 1, a group of 15 new participants was recruited to replicate the study, incorporating notable modifications. Specifically, the “emotional expression” factor was omitted, and a new “stimulus orientation dimension” was introduced. These adjustments were informed by the findings of experiment 1, which highlighted substantial differences, particularly in N170 and P200 amplitudes, between masked and unmasked faces. In contrast, no discernible distinctions were observed between “angry” and “happy” faces. This alignment with the concept that emotional effects on early vision may not necessarily signify an influence of cognitive processes (Raftopoulos, 2023). Because similar amplitude differences have been reported for inverted faces (Rossion et al., 2000), the introduction of face stimulus orientation provided an opportunity to compare and study the effect of both modulations, and hence learn more about the underlaying neural mechanisms. For examples of stimuli used for Experiment 2, and how they differ compared with Experiment 1, see Figure 1A.

Similar to the results shown for Experiment 1 (see Figure 2), the panels in Figures 3AE illustrates the results obtained from splitting the data based on one pair of experimental conditions. To examine the effect of inverting the stimuli, upright (left subpanels) and inverted (right subpanels) stimuli were analyzed separately. Additionally, to compare the effects of both masks and stimulus orientation, the data was split into four groups: no masks and masks, either upright or inverted (see Figure 3F).

The results, derived from a 3-way ANOVA, were directed toward investigating interactions among the binary dimensions of “stimulus orientation” and “mask,” along with either “age” (Supplementary Table S3), “gender” (Supplementary Table S4), or “race” (Supplementary Table S5), and their influence on the amplitudes of the P100 and N170 components (see Supplementary material). To complement these findings, Supplementary Figures S3–S5 present topographic maps illustrating amplitude distributions for these components under conditions involving masked vs. unmasked stimuli and upright vs. inverted faces.

3.3. Amplitude and latencies of the P100, N170, and P200

Table 1 provides a comprehensive summary of amplitude differences for all ERP components considered in both experiment 1 and experiment 2. A non-parametric randomization test was employed to compare amplitudes between any two experimental conditions (see Materials and methods). Additionally, this test helped identify latency shifts (not shown in Table 1) between different conditions for different ERP components. When analyzing the EEG response separately for inverted or upright face image stimuli, no latency shift was observed between masked and unmasked images. However, the response to inverted unmasked faces showed a significant delay compared to upright unmasked faces for the N170 (6 ms, p = 0.017) and the P200 (20 ms, p < 0.00001). Inverting masked faces also exhibited a delay, albeit not statistically significant (4 ms, p = 0.08), but a significant delay for the P200 (20 ms, p < 0.00001) compared to upright masked faces. For a more detailed interpretation of these findings, please see the “Discussion” section below.

4. Discussion

4.1. Factors influencing the N170 amplitude and latencies

This study found that N170 amplitudes increased with appearent complexity of processing faces. Older faces, for instance, known to be more challenging in terms of emotional expression recognition (Grondhuis et al., 2021), elicited larger N170 responses, noted in both Experiment 1 and 2, consistent with prior research (Wiese et al., 2008). Interestingly, the inversion of faces abolished this age-related difference (see Figure 3), although a separate study observed larger inversion effects for young faces compared to old faces (Wiese et al., 2008). The presence of a face mask, which also imposes additional processing demands, was also found to increase N170 amplitudes and P200 responses in both experiments, aligning with recent studies reporting similar N170 increments due to face masks (Prete et al., 2022; Proverbio and Cerri, 2023). It’s noteworthy, however, that another study (Żochowska et al., 2022) did not observe changes in N170 amplitudes with face masks. Remarkably, none of these studies noted a corresponding increase in P200 amplitude, possibly indicative of cognitive task variations.

Another critical aspect impacting face processing is facial inversion, a phenomenon evident across all conditions in our study except for the “masked face condition” (see below). This effect resulting in larger N170s is well-documented (Rossion et al., 2000; Rousselet et al., 2004; Sadeh and Yovel, 2010), been observed to impact the N250 (Hashemi et al., 2019; Abreu et al., 2023), an ERP component that falls outside the scope of this study. The absence of familiarity with faces (Ito and Urland, 2003), giving rise to the “own race bias effect,” has also been linked to heightened challenges in processing faces, resulting in larger N170 responses to other-race faces (Sun et al., 2014; Yao and Zhao, 2019). However, it’s important to note that the current study did not identify any sensitivity of early ERP components to race or skin color. This absence of sensitivity can be attributed to the predominant representation of Latino participants in the sample, which may not offer the requisite diversity to thoroughly investigate own-race effects. Furthermore, gender-based differences may affect N170 amplitudes, as evidenced by larger N170s for male faces, primarily due to the gender imbalance in our participant pool. Nevertheless, this finding warrants further investigation with a more balanced participant pool.

4.2. Potential neural mechanisms for increased N170 response to inverted faces

One plausible explanation for the enhanced N170 response to inverted faces lies in the early recruitment of additional neural mechanisms, rather than a simple increase in activity within existing neural populations during the N170 time-window. These findings align with Rossion’s hypothesis, which posits the involvement of object-sensitive neurons in augmenting the N170 amplitude observed for inverted faces (Rossion et al., 2000). This hypothesis may also be extended to elucidate the increased N170 response observed in our study for masked faces, as masks themselves can be considered objects. Intriguingly, the study demonstrates that the addition of face masks has a comparable effect on N170 amplitude as inverting the maskless face stimulus. However, combining a mask and inversion did not lead to an additive increase, suggesting a potential neural saturation point.

4.3. Delayed N170 response for inverted faces

The delay in the N170 response to inverted faces is often linked to alterations in the spatial relationships among facial features. Additionally, an amplification of the N170 component and a corresponding shift in latency have been associated with a reduced ability to recognize faces. For instance, a study demonstrated that gradually rotating facial images from an upright to an upside-down position resulted in a declining ability to identify faces (Jacques and Rossion, 2007). However, it’s important to emphasize that the reduced face recognition ability alone may not completely explain the observed delay in both the N170 and P200 components. This becomes evident in the present study, where no such delay was observed for faces with face masks compared to unmasked faces, despite face masks also significantly impeding face identification (Freud et al., 2022).

4.4. Factors influencing the P200 amplitude and latencies

The P200 component is well-established as being linked to configural face encoding, with more typical faces consistently yielding larger P200 amplitudes (Wuttke and Schweinberger, 2019). Conversely, as deviations from the norm increase, relatively smaller P200 amplitudes are typically observed (Halit et al., 2000; Latinus and Taylor, 2006; Schweinberger and Neumann, 2016). An intriguing revelation from our study is that both masked faces, as observed in our current investigation, and scrambled faces (Latinus and Taylor, 2006), elicit substantial P200 amplitudes. In contrast, a similar robust increase in the P200 component was not observed in response to facial inversion. This intriguing finding suggests that faces lacking spatial and configurational information may be processed as highly typical or in a default state, with the amplitude modulation being influenced by the presence of spatial face information.

Moreover, the inversion of faces was found to induce delays in both N170 and P200 latencies. This phenomenon has been noted previously (Latinus and Taylor, 2006), as well as in the case of Mooney faces compared to other face types. However, it is crucial to highlight that, unlike in our current study, previous research primarily emphasized the delay in N170 latency, as indicated by N170-to-P200 peak analyses. In contrast, this study revealed that the observed latency for inverted faces increased from 4–6 ms (N170) to approximately 20 ms (P200), demonstrating that this delay is not limited to one specific processing stage.

4.5. Effects of neural adaptation on the N170 and P200 components

The investigation of diminished ERP components resulting from repetition effects typically involves comparing responses to two identical faces or faces sharing common attributes (e.g., identity) vs. responses to two distinct faces. For an in-depth review, (see Schweinberger and Neumann, 2016). Previous studies (Amihai et al., 2011; Walther et al., 2013) have demonstrated that the N170 component is influenced when preceded by another face, regardless of whether the sequentially presented faces represent the same individual or different individuals. The present study provides further insight by showing that the categorical face adaptation effect accumulates over the duration of a session, leading to a gradual decline in N170 amplitude throughout the experimental session, as depicted in Figures 2E, 3E. Even when upright and inverted face stimuli were interleaved and separately analyzed, the apparent “neural adaptation effect” remained consistent across sessions. It’s noteworthy that this phenomenon has also been observed in data collected from the occipital lobe of rhesus macaque monkeys using non-face stimuli (Brunet et al., 2014). Therefore, researchers utilizing block designs to compare experimental conditions should consider this effect to ensure the reliability and validity of their interpretations.

4.6. Limitations of the study

It is essential to acknowledge certain limitations in our study. The homogeneity of our participant pool, primarily consisting of college students identifying as Hispanic/Latine, may influence our findings. Furthermore, the gender distribution in the sample was skewed toward females, suggesting the need for a more balanced participant pool in future investigations.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Institutional Review Board of the California State University San Bernardino. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

NB: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Supervision, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author declares financial support was received for the research, authorship, and/or publication of this article. NB received grant support from the National Institutes of Health (NIH)/National Institutes of Neurological Disorders and Stroke (NINDS), R15NS121788.

Acknowledgments

NB would like to express appreciation to Krisha Orgo, Camryn Amundsen, Diana Guevara, Britney Aguirre, Emily Mendiola, Candace Perez, Marshina Brown, and Stephanie Villalpando for their valuable assistance with data collection.

Conflict of interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fnhum.2023.1268972/full#supplementary-material

References

Abreu, A. L., Fernández-Aguilar, L., Ferreira-Santos, F., and Fernandes, C. (2023). Increased N250 elicited by facial familiarity: an ERP study including the face inversion effect and facial emotion processing. Neuropsychologia 188:108623. doi: 10.1016/j.neuropsychologia.2023.108623

CrossRef Full Text | Google Scholar

Amihai, I., Deouell, L. Y., and Bentin, S. (2011). Neural adaptation is related to face repetition irrespective of identity: a reappraisal of the N170 effect. Exp. Brain Res. 209, 193–204. doi: 10.1007/s00221-011-2546-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Batty, M., and Taylor, M. J. (2003). Early processing of the six basic facial emotional expressions. Cogn. Brain Res. 17, 613–620. doi: 10.1016/S0926-6410(03)00174-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Blau, V. C., Maurer, U., Tottenham, N., and McCandliss, B. D. (2007). The face-specific N170 component is modulated by emotional facial expression. Behav. Brain Funct. 3, 1–13. doi: 10.1186/1744-9081-3-7

CrossRef Full Text | Google Scholar

Brunet, N. M., Bosman, C. A., Vinck, M., Roberts, M., Oostenveld, R., Desimone, R., et al. (2014). Stimulus repetition modulates gamma-band synchronization in primate visual cortex. Proc. Natl. Acad. Sci. 111, 3626–3631. doi: 10.1073/pnas.1309714111

PubMed Abstract | CrossRef Full Text | Google Scholar

Caldara, R., Rossion, B., Bovet, P., and Hauert, C.-A. (2004). Event-related potentials and time course of the ‘other-race’ face classification advantage. Neuroreport 15, 905–910. doi: 10.1097/00001756-200404090-00034

PubMed Abstract | CrossRef Full Text | Google Scholar

Carretié, L., Kessel, D., Carboni, A., López-Martín, S., Albert, J., Tapia, M., et al. (2013). Exogenous attention to facial vs non-facial emotional visual stimuli. Soc. Cogn. Affect. Neurosci. 8, 764–773. doi: 10.1093/scan/nss068

PubMed Abstract | CrossRef Full Text | Google Scholar

Conley, M. I., Dellarco, D. V., Rubien-Thomas, E., Cohen, A. O., Cervera, A., Tottenham, N., et al. (2018). The racially diverse affective expression (RADIATE) face stimulus set. Psychiatry Res. 270, 1059–1067. doi: 10.1016/j.psychres.2018.04.066

PubMed Abstract | CrossRef Full Text | Google Scholar

Eimer, M., and Holmes, A. (2002). An ERP study on the time course of emotional face processing. Neuroreport 13, 427–431. doi: 10.1097/00001756-200203250-00013

PubMed Abstract | CrossRef Full Text | Google Scholar

Eimer, M., Holmes, A., and McGlone, F. P. (2003). The role of spatial attention in the processing of facial expression: an ERP study of rapid brain responses to six basic emotions. Cogn. Affect. Behav. Neurosci. 3, 97–110. doi: 10.3758/CABN.3.2.97

PubMed Abstract | CrossRef Full Text | Google Scholar

Epstein, R. A., Higgins, J. S., Parker, W., Aguirre, G. K., and Cooperman, S. (2006). Cortical correlates of face and scene inversion: a comparison. Neuropsychologia 44, 1145–1158. doi: 10.1016/j.neuropsychologia.2005.10.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Freud, E., Di Giammarino, D., Stajduhar, A., Rosenbaum, R. S., Avidan, G., and Ganel, T. (2022). Recognition of masked faces in the era of the pandemic: no improvement despite extensive natural exposure. Psychol. Sci. 33, 1635–1650. doi: 10.1177/09567976221105459

CrossRef Full Text | Google Scholar

Gantiva, C., Sotaquirá, M., Araujo, A., and Cuervo, P. (2020). Cortical processing of human and emoji faces: an ERP analysis. Behav. Inf. Technol. 39, 935–943. doi: 10.1080/0144929X.2019.1632933

CrossRef Full Text | Google Scholar

Gao, L., Xu, J., Zhang, B., Zhao, L., Harel, A., and Bentin, S. (2009). Aging effects on early-stage face perception: an ERP study. Psychophysiology 46, 970–983. doi: 10.1111/j.1469-8986.2009.00853.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Grondhuis, S. N., Jimmy, A., Teague, C., and Brunet, N. M. (2021). Having difficulties reading the facial expression of older individuals? Blame it on the facial muscles, not the wrinkles. Front. Psychol. 12:620768. doi: 10.3389/fpsyg.2021.620768

PubMed Abstract | CrossRef Full Text | Google Scholar

Halit, H., de Haan, M., and Johnson, M. H. (2000). Modulation of event-related potentials by prototypical and atypical faces. Neuroreport 11, 1871–1875. doi: 10.1097/00001756-200006260-00014

PubMed Abstract | CrossRef Full Text | Google Scholar

Hashemi, A., Pachai, M. V., Bennett, P. J., and Sekuler, A. B. (2019). The role of horizontal facial structure on the N170 and N250. Vis. Res. 157, 12–23. doi: 10.1016/j.visres.2018.02.006

PubMed Abstract | CrossRef Full Text | Google Scholar

He, Y., Johnson, M. K., Dovidio, J. F., and McCarthy, G. (2009). The relation between race-related implicit associations and scalp-recorded neural activity evoked by faces from different races. Soc. Neurosci. 4, 426–442. doi: 10.1080/17470910902949184

PubMed Abstract | CrossRef Full Text | Google Scholar

Herrmann, M. J., Aranda, D., Ellgring, H., Mueller, T. J., Strik, W. K., Heidrich, A., et al. (2002). Face-specific event-related potential in humans is independent from facial expression. Int. J. Psychophysiol. 45, 241–244. doi: 10.1016/S0167-8760(02)00033-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Herrmann, M. J., Ehlis, A.-C., Ellgring, H., and Fallgatter, A. J. (2005). Early stages (P100) of face perception in humans as measured with event-related potentials (ERPs). J. Neural Transm. 112, 1073–1081. doi: 10.1007/s00702-004-0250-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Hinojosa, J. A., Mercado, F., and Carretié, L. (2015). N170 sensitivity to facial expression: a meta-analysis. Neurosci. Biobehav. Rev. 55, 498–509. doi: 10.1016/j.neubiorev.2015.06.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Itier, R. J., and Taylor, M. J. (2004). N170 or N1? Spatiotemporal differences between object and face processing using ERPs. Cereb. Cortex 14, 132–142. doi: 10.1093/cercor/bhg111

CrossRef Full Text | Google Scholar

Ito, T. A., and Urland, G. R. (2003). Race and gender on the brain: electrocortical measures of attention to the race and gender of multiply categorizable individuals. J. Pers. Soc. Psychol. 85, 616–626. doi: 10.1037/0022-3514.85.4.616

PubMed Abstract | CrossRef Full Text | Google Scholar

Jacques, C., and Rossion, B. (2007). Early electrophysiological responses to multiple face orientations correlate with individual discrimination performance in humans. NeuroImage 36, 863–876. doi: 10.1016/j.neuroimage.2007.04.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Johnson, M. H. (2005). Subcortical face processing. Nat. Rev. Neurosci. 6, 766–774. doi: 10.1038/nrn1766

CrossRef Full Text | Google Scholar

Kaltwasser, L., Hildebrandt, A., Recio, G., Wilhelm, O., and Sommer, W. (2014). Neurocognitive mechanisms of individual differences in face cognition: a replication and extension. Cogn. Affect. Behav. Neurosci. 14, 861–878. doi: 10.3758/s13415-013-0234-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Latinus, M., and Taylor, M. J. (2006). Face processing stages: impact of difficulty and the separation of effects. Brain Res. 1123, 179–187. doi: 10.1016/j.brainres.2006.09.031

PubMed Abstract | CrossRef Full Text | Google Scholar

Luck, S. J., and Gaspelin, N. (2017). How to get statistically significant effects in any ERP experiment (and why you shouldn’t). Psychophysiology 54, 146–157. doi: 10.1111/psyp.12639

PubMed Abstract | CrossRef Full Text | Google Scholar

Luo, W., Feng, W., He, W., Wang, N.-Y., and Luo, Y.-J. (2010). Three stages of facial expression processing: ERP study with rapid serial visual presentation. NeuroImage 49, 1857–1867. doi: 10.1016/j.neuroimage.2009.09.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Mazard, A., Schiltz, C., and Rossion, B. (2006). Recovery from adaptation to facial identity is larger for upright than inverted faces in the human occipito-temporal cortex. Neuropsychologia 44, 912–922. doi: 10.1016/j.neuropsychologia.2005.08.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Moradi, A., Mehrinejad, S. A., Ghadiri, M., and Rezaei, F. (2017). Event-related potentials of bottom-up and top-down processing of emotional faces. Basic Clin. Neurosci. 8, 27–36. doi: 10.15412/J.BCN.03080104

PubMed Abstract | CrossRef Full Text | Google Scholar

Mouchetant-Rostaing, Y., and Giard, M.-H. (2003). Electrophysiological correlates of age and gender perception on human faces. J. Cogn. Neurosci. 15, 900–910. doi: 10.1162/089892903322370816

PubMed Abstract | CrossRef Full Text | Google Scholar

Mouchetant-Rostaing, Y., Giard, M., Bentin, S., Aguera, P., and Pernier, J. (2000). Neurophysiological correlates of face gender processing in humans. Eur. J. Neurosci. 12, 303–310. doi: 10.1046/j.1460-9568.2000.00888.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Oostenveld, R., Fries, P., Maris, E., and Schoffelen, J.-M. (2011). FieldTrip: open source software for advanced analysis of MEG, EEG, and invasive electrophysiological data. Comput. Intell. Neurosci. 2011, 1–9. doi: 10.1155/2011/156869

PubMed Abstract | CrossRef Full Text | Google Scholar

Pascual-Leone, A., and Hamilton, R. (2001). The metamodal organization of the brain. Prog. Brain Res. 134, 427–445. doi: 10.1016/S0079-6123(01)34028-1

CrossRef Full Text | Google Scholar

Pinto, S., Tremblay, P., Basirat, A., and Sato, M. (2019). The impact of when, what and how predictions on auditory speech perception. Exp. Brain Res. 237, 3143–3153. doi: 10.1007/s00221-019-05661-5

CrossRef Full Text | Google Scholar

Prete, G., D’Anselmo, A., and Tommasi, L. (2022). A neural signature of exposure to masked faces after 18 months of COVID-19. Neuropsychologia 174:108334. doi: 10.1016/j.neuropsychologia.2022.108334

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., and Cerri, A., Gallotta C. Facemasks selectively impair the recognition of facial expressions that stimulate empathy. An ERP study. Psychophysiology. (2023);e14280, 60, doi: 10.1111/psyp.14280

CrossRef Full Text | Google Scholar

Raftopoulos, A. (2023). Does the emotional modulation of visual experience entail the cognitive penetrability of early vision? Rev. Philos. Psychol., 1–24. doi: 10.1007/s13164-023-00695-9

CrossRef Full Text | Google Scholar

Rossion, B. (2014). Understanding face perception by means of human electrophysiology. Trends Cogn. Sci. 18, 310–318. doi: 10.1016/j.tics.2014.02.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Rossion, B., Gauthier, I., Tarr, M. J., Despland, P., Bruyer, R., Linotte, S., et al. (2000). The N170 occipito-temporal component is delayed and enhanced to inverted faces but not to inverted objects: an electrophysiological account of face-specific processes in the human brain. Neuroreport 11, 69–72. doi: 10.1097/00001756-200001170-00014

PubMed Abstract | CrossRef Full Text | Google Scholar

Rossion, B., and Jacques, C. (2008). Does physical interstimulus variance account for early electrophysiological face sensitive responses in the human brain? Ten lessons on the N170. NeuroImage 39, 1959–1979. doi: 10.1016/j.neuroimage.2007.10.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Rossion, B., and Jacques, C. (2011). “The N170: understanding the time course of face perception in the human brain” in The Oxford Handbook of Event-Related Potential Components, 115–142.

Google Scholar

Rousselet, G. A., Macé, M. J.-M., and Fabre-Thorpe, M. (2004). Animal and human faces in natural scenes: how specific to human faces is the N170 ERP component? J. Vis. 4, 13–21. doi: 10.1167/4.1.2

CrossRef Full Text | Google Scholar

Sadeh, B., and Yovel, G. (2010). Why is the N170 enhanced for inverted faces? An ERP competition experiment. Neuroimage 53, 782–789. doi: 10.1016/j.neuroimage.2010.06.029

PubMed Abstract | CrossRef Full Text | Google Scholar

Schindler, S., and Bublatzky, F. (2020). Attention and emotion: an integrative review of emotional face processing as a function of attention. Cortex 130, 362–386. doi: 10.1016/j.cortex.2020.06.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Schweinberger, S. R., and Neumann, M. F. (2016). Repetition effects in human ERPs to faces. Cortex 80, 141–153. doi: 10.1016/j.cortex.2015.11.001

CrossRef Full Text | Google Scholar

Sun, G., Zhang, G., Yang, Y., Bentin, S., and Zhao, L. (2014). Mapping the time course of other-race face classification advantage: a cross-race ERP study. Brain Topogr. 27, 663–671. doi: 10.1007/s10548-013-0348-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Walker, P. M., Silvert, L., Hewstone, M., and Nobre, A. C. (2008). Social contact and other-race face processing in the human brain. Soc. Cogn. Affect. Neurosci. 3, 16–25. doi: 10.1093/scan/nsm035

PubMed Abstract | CrossRef Full Text | Google Scholar

Walther, C., Schweinberger, S. R., Kaiser, D., and Kovács, G. (2013). Neural correlates of priming and adaptation in familiar face perception. Cortex 49, 1963–1977. doi: 10.1016/j.cortex.2012.08.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H. (2012). The role of age and ethnic group in face recognition memory: ERP evidence from a combined own-age and own-race bias study. Biol. Psychol. 89, 137–147. doi: 10.1016/j.biopsycho.2011.10.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H., Schweinberger, S. R., and Hansen, K. (2008). The age of the beholder: ERP evidence of an own-age bias in face memory. Neuropsychologia 46, 2973–2985. doi: 10.1016/j.neuropsychologia.2008.06.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Wiese, H., Stahl, J., and Schweinberger, S. R. (2009). Configural processing of other-race faces is delayed but not decreased. Biol. Psychol. 81, 103–109. doi: 10.1016/j.biopsycho.2009.03.002

CrossRef Full Text | Google Scholar

Williams, L. M., Palmer, D., Liddell, B. J., Song, L., and Gordon, E. (2006). The ‘when’and ‘where’of perceiving signals of threat versus non-threat. NeuroImage 31, 458–467. doi: 10.1016/j.neuroimage.2005.12.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Wuttke, S. J., and Schweinberger, S. R. (2019). The P200 predominantly reflects distance-to-norm in face space whereas the N250 reflects activation of identity-specific representations of known faces. Biol. Psychol. 140, 86–95. doi: 10.1016/j.biopsycho.2018.11.011

PubMed Abstract | CrossRef Full Text | Google Scholar

Yao, Q., and Zhao, L. (2019). Using spatial frequency scales for processing own-race and other-race faces: an ERP analysis. Neurosci. Lett. 705, 167–171. doi: 10.1016/j.neulet.2019.04.059

PubMed Abstract | CrossRef Full Text | Google Scholar

Yovel, G., and Kanwisher, N. (2005). The neural basis of the behavioral face-inversion effect. Curr. Biol. 15, 2256–2262. doi: 10.1016/j.cub.2005.10.072

PubMed Abstract | CrossRef Full Text | Google Scholar

Żochowska, A., Jakuszyk, P., Nowicka, M. M., and Nowicka, A. (2022). Are covered faces eye-catching for us? The impact of masks on attentional processing of self and other faces during the COVID-19 pandemic. Cortex. 149, 173–87.

Google Scholar

Keywords: ERP, EEG, N170, P200, age, face inversion, face mask, amplitude

Citation: Brunet NM (2023) Face processing and early event-related potentials: replications and novel findings. Front. Hum. Neurosci. 17:1268972. doi: 10.3389/fnhum.2023.1268972

Received: 28 July 2023; Accepted: 05 October 2023;
Published: 25 October 2023.

Edited by:

Vilfredo De Pascalis, Sapienza University of Rome, Italy

Reviewed by:

Carlo Lai, Sapienza University of Rome, Italy
Joseph Ciorciari, Swinburne University of Technology, Australia

Copyright © 2023 Brunet. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Nicolas M. Brunet, nicolas.brunet@csusb.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.