Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 16 June 2022
Sec. Perception Science
This article is part of the Research Topic Impact of Face Covering on Social Cognition and Interaction View all 19 articles

The Recognition of Facial Expressions Under Surgical Masks: The Primacy of Anger

  • Cognitive Electrophysiology Laboratory, Department of Psychology, University of Milano-Bicocca, Milan, Italy

Background: The need to wear surgical masks in everyday life has drawn the attention of psychologists to the negative effects of face covering on social processing. A recent but not homogeneous literature has highlighted large costs in the ability to recognize emotions.

Methods: Here it was investigated how mask covering impaired the recognition of facial mimicry in a large group of 220 undergraduate students. Sex differences in emotion recognition were also analyzed in two subgroups of 94 age-matched participants. Subjects were presented with 112 pictures displaying the faces of eight actors (4 women and 4 men) wearing or not wearing real facemasks, and expressing seven emotional states (neutrality, surprise, happiness, sadness, disgust, anger and fear). The task consisted in categorizing facial expressions while indicating the emotion recognizability with a 3-point Likert scale. Scores underwent repeated measures ANOVAs.

Results: Overall, face masking reduced emotion recognition by 31%. All emotions were affected by mask covering except for anger. Face covering was most detrimental to sadness and disgust, both relying on mouth and nose expressiveness. Women showed a better performance for subtle expressions such as surprise and sadness, both in masked and natural conditions, and men for fear recognition (in natural but especially masked conditions).

Conclusion: Anger display was unaffected by masking, also because corrugated forehead and frowning eyebrows were clearly exposed. Overall, facial masking seems to polarize non-verbal communication toward the happiness/anger dimension, while minimizing emotions that stimulate an empathic response in the observer.

Introduction

It is known that surgical masks (used pervasively to counter the transmission of coronavirus) might negatively affect and impair social processing. Impairment might concern the recognition of face identity (Carragher and Hancock, 2010; Noyes et al., 2021), emotion reading (Roberson et al., 2012; Carbon, 2020; Ruba and Pollak, 2020; Calbi et al., 2021; Grundmann et al., 2021; Marini et al., 2021; Carbon et al., 2022), trustworthiness judgment (Biermann et al., 2021), face likability and closeness impression (Grundmann et al., 2021), as well as speech comprehension (Singh et al., 2021). Relatedly, previous literature showed that face blurring impairs the understanding of emotional signals including body language (Proverbio et al., 2018). Although emotions conveyed by bodily expressions are quite easily recognizable (de Gelder et al., 2015), face obscuration reduces pantomime comprehension in healthy subjects, as opposed to patients with bilateral amygdala damage (Adolphs et al., 2003). This indicates how facial mimicry is crucial in nonverbal communication. For example, when facial expressions are incongruent with bodily expressions (of anger, for instance) response times are much slower during a matching-to-sample task in controls (Kret and de Gelder, 2013), thus suggesting that bodily expressions are better recognized when accompanied by a face that expresses the same emotion (Meeren et al., 2005).

To investigate at which extent face covering impaired social communication Grundmann et al. (2021) performed a large study on 191 individuals of different ages and sexes and found that facemasks diminished people’s ability to accurately categorize facial expressions and affected the perceptions of person trustworthiness, likability, and closeness.

Generally, the mouth region is thought to be most informative for happy, surprised and disgusted expressions, while the eyes area is considered more informative for fearful and angry expressions. For example, the white sclera expansion, typical of fear display, is especially at the basis of its innate recognition (e.g., Jessen and Grossmann, 2014; Barrett, 2018). Both the mouth and eyes areas are informative for neutral and sad expressions (Smith et al., 2005;Wegrzyn et al., 2017). In addition, joy can very well detected through the smiling mouth, but especially the “smiling” eyes. Years of psychological (e.g., Ekman et al., 1998) and lately engineering research on pattern recognition (e.g., Ugail and Al-dahoud, 2019) have shown that the more reliable indicators of a genuine happy facial expression are indeed the eyes. Angry facial expressions are associated with a strong activation of the corrugator supercilii (i.e., the muscle involved in frowning), whereas happy facial expressions are associated with a strong activation of the zygomaticus major, (the muscle involved in smiling) (Rymarczyk et al., 2019). Based on the above findings we expected lower costs in accuracy due to the covering of the lower part of the face (face masking) during recognition of anger and happiness expressions, but the available literature was not completely homogeneous at this regard.

Noyes et al. (2021) recently explored the effect of masks and sunglasses wearing on familiar and unfamiliar face matching and emotion categorization in 100 participants. They found that, while masks did not reduce the recognition of angry faces, facial expressions of disgust, happy, fear and surprise were most affected by it. A large reduction in categorization accuracy for disgust expressions was found in particular. Sadness detection was difficult both mask less and with mask covering, so that the performance was not significantly impaired by masking. Results are not fully consistent across studies. In a recent study by Marini et al. (2021) investigating the impact of facemasks on emotion recognition (but only with three emotions) they showed an impaired recognition of sad and fearful expressions in the masked condition, with no effect on neutral expressions. Among the three expressions, sadness was the most affected and happiness the least affected. In this study, sadness was more hardly detected with mask covering the mouth area.

One of the problems with the available studies is that many of them digitally applied a mask or a foulard on the face picture in order to create identically expressive faces, across the masked vs. non-masked category (e.g., Carragher and Hancock, 2010; Carbon, 2020; Calbi et al., 2021; Grundmann et al., 2021; Marini et al., 2021; Carbon et al., 2022). While this procedure might assure an optimal matching between masked and unmasked expressions, however it lacks likelihood and ecological value. Indeed, digitally applied masks are not stretched by the facial expression thus reducing the verisimilitude. Furthermore, they deprive the visual image of details that are present in the real masked face, such as mask sucking or folding. In reality, surgical masks are, for example, deformed by the vertical opening of the mouth in expressions like surprise or laughing, or during verbal speech; likewise, they are stretched horizontally for smiling. Indeed the masks adapt to, and reveal, the underneath muscular movements, which can be picked up by an observer. In order to maintain the visibility of mask bending and stretching due to underneath facial mimicry, in this study, actors wore real surgical masks during shooting. Several repetitions and much effort was devoted to the perfect matching between expressions produced with or without masks.

The aim of the study was to gain clear knowledge on the effects of face masking on the comprehensibility of a large variety of facial expressions (i.e., the six basic Ekman emotions: fear, anger, joy, sadness, disgust, and surprise plus neutrality) by using real and non-digital facemasks, unlike many of the previous studies quoted above. In fact, it is possible that digital masks further limited the possibility of recognizing facial mimicry because they are fixed and do not show dynamic deformations of the fabric, made possible by its elasticity. For example, real masks can show inhalation-related sucking associated with startle reaction in the surprise or fear expressions. Again, they can also show vertical and horizontal stretching of the tissue due to smiling or nose wrinkling. Therefore, it is possible that emotion recognition under digital facemasks was currently under-estimated.

In addition, we wished to investigate if face masking affected the two sexes differently. According to the available psychological and neuroscientific literature, overall, females would be more accurate in identifying emotional facial expressions then males (e.g., McClure, 2000; Montagne et al., 2005; Proverbio et al., 2006; Proverbio, 2021). Indeed, a recent study involving perception of masked faces (Grundmann et al., 2021) showed that being a man was associated with a reduced accuracy in emotion recognition than being a woman, without specific interactions with face masking conditions.

Materials and Methods

Participants

220 undergraduate students of local University self-recruited through online advertisement posted on the student’s web site. Six of them were excluded because older than 35 years. They aged between 18 and 35 years (mean = 21.617, SD = 2.91) and 47 of them were males). Experiments were conducted with the understanding and written consent of each participant according to the Declaration of Helsinki (BMJ 1991; 302: 1194) with approval of the Ethical Committee of the Psychology department of local University approved the study (protocol number: RM-2021-401). It was conducted online from June 25 until July 8 2021 and programmed in Google forms https://www.google.com/forms. Participation was free and not rewarded.

Stimuli

10 actors (master psychology students) of Caucasian ethnicity were recruited (five females and five males) aging 23 years on average (SD = 1.333) for photos taking. High-resolution pictures of their faces were self- taken with a cell phone at about 40 cm of distance in light controlled conditions, while standing up against a white wall. Actors were required to avoid wearing earrings, glasses, make up, hairpins, pliers, any type of hair embellishments, mustaches, beard. They were also instructed to wear a black t-shirt and gather the hair behind the head. The pictures of two actors were discarded in that showing a different mimicry in the natural vs. masked condition; their pictures were therefore used only as stimuli for the training phase, to accustom the subjects to the task, without showing them the faces selected for the experimental phase. For each of the seven emotions, actors were instructed to imagine a vivid emotional state, while concentrating on a specific autobiographic scenario through the Stanislavsky method, and express it spontaneously while ignoring the presence/lack of surgical masks. For “surprise” emotion, they were instructed to think of a positive surprise. They trained repeatedly in order to reach the same degree of intensity across subjects and emotions (see Figure 1 for some examples). Each of the 10 actors provided written consent and filled in the privacy release form.

FIGURE 1
www.frontiersin.org

Figure 1. Example of stimuli (facial expressions) in the natural and masked condition. Overall, stimuli were created by taking photographs of natural expressions of eight actors (four male and four females) concentrated on their inner imaginary emotional state, through the Stanislavsky method. Masks were worn in reality and not digitally recreated. This revealed, for example the mouth/lips contraction associated with anger, the large mouth opening associated with disgust, the air intake (inhalation) characteristic of surprise or fear resulting in evident mask sucking.

Stimulus set was validated on a group of 50 students (25 females, 24 males and 1 gender fluid) aging on average 23.7 years (min = 17, max = 34 years). All participants had normal vision, no neurological or psychiatric deficits and possessed diploma, BA or Master degrees. Participants were shown, randomly mixed and once at a time, the 56 pictures relative to the seven facial expressions acted by the eight female and male actors. Subjects were required to rapidly observe the picture and decide which one of the seven emotions typed below was more appropriate to describe the viewed facial expression, by clicking a check mark within a few seconds. Pictures were displayed at the center of the screen and the experimental session lasted 10 min.

Overall performance for correctly identifying facial emotions in unmasked faces was remarkably high = 87.35% (with a chance rate of 16.7%). No participant performed below an overall rate of 75.0%. In more details, accuracy was 98.47% for joy, 86.73% for surprise, 80.1% for sadness, 89.29% for anger, 72.70% for fear, 85.97% for disgust and 98.21% for neutrality. These recognition rates (in line with the data reported by Carbon, 2020; Carbon et al., 2022) outperform the accuracy of recognizing facial expressions reported by other studies in the literature (e.g., 57.85% for anger and disgust in Aviezer et al. (2008) and 57.85 for negative emotions in Derntl et al. (2009) thus supporting the qualitative validity of the stimuli.

Stimulus set was also evaluated for facial attractiveness by a further group of 12 students (seven females and five males) aged between 18 and 25 years. Judges were requested to evaluate the attractiveness of neutral unmasked pictures of all identities, by using a 3-point Likert scale, where 1 stood for “not attractive,” 2 for “average” and 3 for “attractive.” The results showed a perfect balance across the two sexes and indicated an “average” degree of attractiveness for the facial stimuli (Females = 1.83; SD = 0.78; Males = 1.82; SD = 0.76). This characteristic of stimuli promotes the generalizability of results to the normally looking population

Procedure

After giving written and informed consent participants were administered a questionnaire about demographic information (such as age, sex, manual dexterity, educational qualification and e-mail address). This section was followed by the emotion-recognition task, consisting in 112 experimental trials, in which participants were first shown a portrait photograph of an adult face to be inspected for about 2 s. The images were equiluminant as assessed by subjecting their luminance values to an analysis of variance (F = 0.099, p = 0.992). Photos were in color, had the same size (3.37 cm × 5 cm; 199 × 295 pixels; 3° 22′ × 5°) and were displayed at the center of the screen, on a white background.

Immediately below the face, there was a list of words (neutrality, happiness, surprise, fear, anger, sadness, disgust), from which they had to select the emotion that they deemed the most appropriate to characterize the face. Next, participants judged how clearly they considered the expression recognizable on a 3-point Likert scale (ranging from “1 = not much” to “3 = very much”). The original wording was in Italian. The emotion was scored 0 if a different incorrect expression was selected. 5 s were allowed for perceiving and responding to the two queries. Participants were instructed to observe one face at a time and to respond within 5 s, not missing any answer. Only one choice per face was allowed. The task lasted about 15 min.

Data Analysis

The individual scores obtained from each individual, for each of the 7 facial expressions and condition, underwent a 3-ways repeated-measures ANOVA whose factors of variabilities were: one between-groups named “sex” (with 2 levels, female and male), and two within-groups named “condition” (with 2 levels, natural and masked) and “emotion” (with 7 levels, happiness, neutrality, surprise, anger, sadness, fear, disgust).

In order to properly assess the statistical effect of the sex of participants (who were females in majority) in a balanced population, two subgroups of participants were created: the group of males comprised all male participants recruited (N = 47) and a blind selection of females (N = 47) chose on the basis of their date of birth (by paring each of the male with a same-age female). The statistical power achieved by the current sample size (N = 94) was computed using the program G*Power 3.1 (Faul et al., 2009) for comparing 2 independent groups.

As a result of this blind procedure, the age of the two sub-groups was identical (males: 23.042, fameless: 23.042). A 3-ways repeated-measures ANOVA was also performed on the data relative to this sample. Factors of variabilities were: one between-groups named “sex” (with 2 levels, female and male), and two within-groups named “condition” (with 2 levels, natural and masked) and “emotion” (with 7 levels, happiness, neutrality, surprise, anger, sadness, fear, disgust). Multiple post hoc comparisons were performed using Tukey’s test. Greenhouse-Geisser correction was applied in case of epsilon < 1 and epsilon corrected p value were computed.

Results

The factor condition was statistically significant [F(1,212) = 212;p < 0.001, ε = 1], with emotion recognizability being higher in the natural [2.31, standard error (SE) = 0.02] than masked (1.59, SE = 0.02) condition. The factor emotion was also significant [F(6,1272) = 191; p < 0.001, ε = 0.79, ε-corrected p value = 0.001]. Post hoc comparisons showed that overall positive emotions were recognized more easily than negative emotions (p < 0.001), except for anger, as shown in Figure 2 (neutral = 2.422, SE = 0.029; happy = 2.3, SE = 0.03; surprise = 2.02; SE = 0.03; anger = 2.22, SE = 0.03; sadness = 1.788, SE = 0.02; fear = 1.48; SE = 0.04; disgust = 1.42, SE = 0.02.). Happiness was recognized more easily (p < 0.001), the recognizability of fear and disgust was equally poor, while that of neutral and angry expressions was equally high.

FIGURE 2
www.frontiersin.org

Figure 2. Mean scores of recognizability (along with SE values) attributed by participants (N = 214) to the various facial expressions regardless of masking condition. Scale ranged from 0 = “not much recognizable” to 3 = “very well recognizable”.

Surgical masks (covering the nose and mouth area) strongly reduced recognizability of all emotions, as shown by the statistical significance of condition × emotion [F(6,1272) = 160; p < 0.001, ε = 0.911, ε-corrected p value = 0.001], except for anger. Post hoc comparisons showed that neutral and happy expressions were equally well recognizable under the mask, but worse than angry expressions. Again, negative emotions such as disgust, sadness and fear were much poorly recognized than positive emotions in masked conditions. Figure 3 shows the mean scores for each facial expression as a function of the masking condition. Negative emotions such as sadness and disgust, more relying on the nose and mouth area expressivity, were most penalized by mask covering.

FIGURE 3
www.frontiersin.org

Figure 3. Mean scores of recognizability (along with SE values) attributed by participants (N = 214) to the various facial expressions as a function of masking condition. Scale ranged from 0 = “not much recognizable” to 3 = “very well recognizable”.

The sex of viewer affected the ability to recognize the emotions regardless of face covering, as shown by the significance of emotion × sex interaction [F(6,1272) = 4.14; p < 0.001, ε = 0.776, ε-corrected p value = 0.001]. The ANOVA performed on the two subgroups of 47 males and 47 females yielded the same significances as the main ANOVA, i.e.: condition (p < 0.001), emotion (p < 0.001), emotion × condition (p < 0.001) and emotion × sex interaction [F(6,552) = 4.138; p < 0.001, ε = 0.778, ε-corrected p value = 0.001].

As for the last interaction and similarly to ANOVA applied to the whole population (see Figure 4 for mean values and SEs), post hoc showed that while women were better at recognizing surprise (p < 0.004) and sadness (p < 0.05), males were better at recognizing fear expressions (p < 0.005). Simple effect analysis showed that this male advantage in recognizing fear was even stronger (see Figure 5) in the masked conditions (p < 0.004).

FIGURE 4
www.frontiersin.org

Figure 4. Mean scores of recognizability (along with SE values) attributed by female and male participants (N = 94) to the various facial expressions as a function of masking condition.

FIGURE 5
www.frontiersin.org

Figure 5. Mean scores (along with SE values) of recognizability attributed by participants (N = 94) to the various facial expressions as a function of sex of viewers and masking condition. Scale ranged from 0 = “not much recognizable” to 3 = “very well recognizable”.

Discussion

In the natural (mask less) conditions, positive emotions (happiness, neutrality, positive surprise) were recognized more accurately than negative emotions such as fear, sadness or disgust. This positive/negative valence distinction is based on the dichotomy on approach/avoidance attitude to emotions supported by previous neuroimaging and electrophysiological literature (Davidson, 1995; Balconi et al., 2017). Overall, masking heavily affected emotion comprehension with a 31% decay in recognizability (namely, going from 2.31 in the natural condition to 1.59 in the masked condition, on a scale where 0 indicated “not much recognizable” and 3 stood for “very well recognizable”). Overall, these findings fit with previous recent literature showing how facemasks reduce emotion recognition accuracy (Roberson et al., 2012; Carbon, 2020; Ruba and Pollak, 2020; Calbi et al., 2021; Marini et al., 2021; Grundmann et al., 2021; Carbon et al., 2022). In our study, face masking was most detrimental for sadness and especially disgust detection, than positive emotions such as happiness. This pattern of results agrees with previous studies, for example Marini et al. (2021), finding that sadness was the most affected and happiness the least affected expression by face masking.

However, we found that mask covering did not affect the recognition of angry faces, which replicates some findings obtained with non-digital masks by Noyes et al. (2021) (see their Figure 7), who also found that the mask and sunglasses conditions did not significantly differ in the angry expressions. The primacy of anger among the biologically relevant emotions has been shown by several studies (e.g., Mancini et al., 2020).

Conversely, the emotional display whose recognition was most affected by mask covering was disgust (also in Noyes et al.’s, 2021 study). Indeed, disgust’s more evident markers (nasiolabial lifting and grimacing and nose wrinkling) are hidden by surgical masks in the masking condition. At this regard it is known that successful recognition of anger versus disgust requires one to process information located in the eye/brow region (which was disclosed) as opposed to the mouth/nose region (which was covered by masks), respectively (Yitzhak et al., 2020). Again, in a study by Ponari et al. (2012) where emotion recognition was hampered by stimuli in which an upper or lower half-face showing an emotional expression was combined with a neutral half-face it was shown that neutral lower half-face interfered with recognition of disgust, whereas the neutral upper half (i.e., the eyes area) impaired the recognition of anger. This difference may probably explain the supremacy of anger and the poor recognition of disgust in the present study.

Women Better at Recognizing Sadness and Surprise

In our study, females outperformed males in the recognition of sadness and surprise. Several evidences in the literature consistently reported a similar pattern of results for both sadness (Montagne et al., 2005; Deng et al., 2016; Li et al., 2020) and surprise (Montagne et al., 2005). In addition, according to some investigations, women seem to be more sensitive to sadness whereas men seem to be more sensitive to anger (Brody et al., 1995; Deng et al., 2016; He et al., 2018). In another study by Montagne et al. (2005) women were reported to be significantly more accurate than men at identifying sadness and surprise. Furthermore, Li et al. (2020)’s study, performed in 1,063 participants varying in sex and age, reported that women performed significantly better at recognizing facial expressions of sadness and disgust.

As for the specific effect of masking, Grundmann et al. (2021) tested 191 participants (52.9% female) aging from 19 to 79 years and found that emotion-recognition accuracy declined for masked (vs. unmasked) faces. More interestingly, they showed lower accuracy to being male vs. female, being old (vs. young), and to seeing an old (vs. young) target face. In a study by Calbi et al. (2021) involving only three affective displays (neutrality, happiness and anger) it was found that female participants gave more negative ratings than male ones when evaluating angry and neutral facial expressions, and more positive ratings when evaluating happy facial expressions. This was discussed in terms of women’ stronger sensibility to face expressivity and better decoding of emotions through facial expressions (e.g., McClure, 2000; Proverbio et al., 2007; Hall et al., 2010; Proverbio, 2021). Consistently, Hoffmann et al. (2010) found that women were better at identifying subtle, less intense emotions (such as sadness), but equally good at identifying clearly expressed emotions (such as fear). Apart from that, it is generally believed that women are more sensitive to emotional facial cues (Proverbio, 2017).

Men Better at Detecting Fear

In this study, males outperformed women in recognizing fearful expressions (especially masked ones). The increased male ability to recognize fear (relying mostly on the processing of the eyes area, with the typical sclera enlargement) when faces were covered by surgical masks, might depend on the fact the eyes were even more focally attended in the masked condition, being the only uncovered face area. However, Sullivan et al. (2017), investigating the percentage of time young women and men spent fixating the eyes and mouth areas of facial expressions (including fear), found that both sexes spent 63.6% of their time looking at the eyes (and 36.4% of the time at the mouth) with no difference across sexes.

In the literature, a male advantage in the processing of fearful expressions is not commonly found, except for an fMRI study, observing regional brain responses to face versus shape identification, in which men showed more significant modulations by both fear and anger affective traits than women (Li et al., 2020).

On a different verge, Riva et al. (2011) have instead found that the observers’ ability to detect pain in a female face was lower than their ability to detect pain in male faces, i.e., that male pain faces are more easily processed at the reflexive level. Relatedly, Simon et al. (2006) in an fMRI study found that observing male (vs. female) individuals expressing pain activated in the observers a much greater threat-related response, including the activation of the ventromedial prefrontal cortex, posterior and anterior insula, somatosensory areas, and amygdala. In another study, where healthy subjects were provoked by money taken by an opponent and given the opportunity to retaliate, men showed a higher amygdala activation during provocation, and the amygdala activation correlated with trait anger scores in men, but not in women (Repple et al., 2018). As well-known amygdala nuclei are the brain structures most involved in fear and threat processing (Adolphs et al., 1995).

Summary

Overall, while face masking reduced the comprehension of all facial expressions but anger (conveying an aggressive display), it was most detrimental for sadness and especially disgust detection (conveying a second person, more passive negative state). The larger impairment for the recognition of the above expressions might depend on their mainly relying on the expressivity of mouth (especially sadness: Smith et al., 2005; Wegrzyn et al., 2017) and nose areas (especially disgust: Yitzhak et al., 2020; Noyes et al., 2021), which were covered by masks. Instead, the angry expression was totally unaffected by face masking. This effect, different from previous studies, might be related to the ecological use of real and non-digital masks, allowing a more complex analysis of facial patterns.

In general, women showed a better performance for positive emotions, both in masked and natural conditions, and men for fear recognition (in natural but especially masked conditions). At this regard, it might be interesting to consider that sex differences in the hemispheric activation for emotion processing were reported. Cahill et al. (2001) found that enhanced memory for emotional video clips was associated with activity of the right amygdala in men, and of the left amygdala in women. In addition, an fMRI study investigating the emotional response to odors by Royet et al. (2003) found a sex difference in the activation of the left orbitofrontal cortex, which was greater in women compared to men. On the other side, Bourne and Watling (2015) found that for males, but not females, greater reported use of negative emotion strategies was associated with stronger right hemisphere lateralization for processing negative emotions. In the light of the well know right/left asymmetry for negative/positive emotions (Canli et al., 1998; Davidson et al., 1999) these studies might provide the neural underpinnings for the higher male accuracy in fear recognition (right amygdala), and of the higher female accuracy for detecting subtle positive emotional cues (e.g., Calbi et al., 2021), but further investigations are certainly needed to reach a definitive conclusion.

More in general, our study suggests the opportunity of studying the effect of face masking with really worn facemasks (instead of digitally applied ones) because there might be a difference in the way masks elastically respond to underneath facial muscles contractions, by deforming and stretching differently as a function of the facial expression. Furthermore, the typical inhalation associated, for example, to the surprised or fearful reaction (startle response), which results in mask sucking, will not be observable with digitally applied masks.

In general, wearing masks hampers facial affect recognition, and it might be particularly challenging for individuals with neuropsychiatric or neurodevelopmental conditions (Pavlova and Sokolov, 2021).

In this study, face masking was strongly detrimental to the comprehension of emotional markers, especially of non-aggressive negative states (such as sadness, disgust and fear). The only expression, whose recognition was not impaired by masking was indeed anger (associated with angry eyes, forehead wrinkling and contraction of mouth and lip muscles).

The primacy of anger among other more subtle emotions (such as sadness) has been reported in previous other studies (Öhman, 1993; Esteves et al., 1994; Fox et al., 2000), who found increased psychophysiological responding to masked angry faces relative to masked happy faces. The present data showed how face masking was able to polarize emotion comprehension toward the negative/positive opposite dimensions (happiness/anger or approach/withdrawal), while causing a deficit in social interaction and communication of softer emotions that usually trigger an empathic resonance in the observer (sadness, fear, disgust). The limited recognition of distressed people’s emotions might supposedly bring to a reduction of personal concern and empathic response (Israelashvili et al., 2020), within the population. This hypothesis strongly agrees with the recent findings by Rymarczyk et al. (2019), which, in a study using simultaneously recorded electromyography (EMG) and fMRI signals, showed that the perception of fear and disgust strongly activated brain regions involved in simulative processes and in empathy, such as mirror neurons (the fronto/parietal MNS) and limbic regions (e.g., the Anterior Insula (AI). Furthermore, the more empathic were the observers, the stronger was the reaction to these facial expressions. This seriously raises the question of a possible reduction in the observers’ empathic capacity in the absence of subtle, lower facial cues covered by facemasks. In fact, the present pattern of results indicates a selective decrease in the ability to recognize emotions that normally stimulate an empathic response (e.g., sadness, disgust, and fear) in face masking conditions.

Study Limits

One possible limitation of this study is that static faces were used instead of dynamic videos for conveying affective information, since, naturally, the emotional valence of such stimuli is enhanced in naturalistic conditions (e.g., Ambadar et al., 2005; Rymarczyk et al., 2019). However, this study, and its novel pattern of results, should be compared with the pre-existing literature where masked static faces were used (Carbon, 2020; Ruba and Pollak, 2020; Calbi et al., 2021; Grundmann et al., 2021; Marini et al., 2021). It would be very interesting, in the near future, to investigate if this sparing of anger from the detrimental effects of masking can also be observed in dynamic conditions.

Data Availability Statement

The original contributions presented in this study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Ethics Statement

The studies involving human participants were reviewed and approved by Ethical Committee of University of Milano-Bicocca. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

AP conceived and planned the experiment, performed statistical analyses and data illustration, and wrote the manuscript. AC prepared the stimuli and carried out the data collection. AP and AC interpreted the data. Both authors provided critical feedback and helped shape the research, analysis and manuscript.

Funding

This study was supported by the Università degli studi di Milano-Bicocca 2018-ATE-0003 grant number 28882.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We are extremely grateful to all subjects of the study for their free participations and particularly to the 10 actors involved in photo shooting, for their incredible patience and commitment.

References

Adolphs, R., Tranel, D., and Damasio, A. R. (2003). Dissociable neural systems for recognizing emotions. Brain Cogn. 52, 61–69. doi: 10.1016/s0278-2626(03)00009-5

CrossRef Full Text | Google Scholar

Adolphs, R., Tranel, D., Damasio, H., and Damasio, A. R. (1995). Fear and the human amygdala. J. Neurosci. 15, 5879–5891. doi: 10.1523/JNEUROSCI.15-09-05879.1995

PubMed Abstract | CrossRef Full Text | Google Scholar

Ambadar, Z., Schooler, J. W., and Cohn, J. F. (2005). Deciphering the enigmatic face: the importance of facial dynamics in interpreting subtle facial expressions. Psychol. Sci. 16, 403–410. doi: 10.1111/j.0956-7976.2005.01548.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Aviezer, H., Hassin, R. R., Ryan, J., Grady, C., Susskind, J., Anderson, A., et al. (2008). Angry, disgusted, or afraid? Studies on the malleability of emotion perception. Psychol. Sci. 19, 724–732. doi: 10.1111/j.1467-9280.2008.02148.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Balconi, M., Vanutelli, M. E., and Grippa, E. (2017). Resting state and personality component (BIS/BAS) predict the brain activity (EEG and fNIRS measure) in response to emotional cues. Brain Behav. 7:e00686. doi: 10.1002/brb3.686

PubMed Abstract | CrossRef Full Text | Google Scholar

Barrett, L. F. (2018). Seeing fear: it’s all in the eyes? Trends Neurosci. 41, 559–563. doi: 10.1016/j.tins.2018.06.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Biermann, M., Schulze, A., Unterseher, F., Atanasova, K., Watermann, P., Krause-Utz, A., et al. (2021). Trustworthiness appraisals of faces wearing a surgical mask during the Covid-19 pandemic in Germany: an experimental study. PLoS One 16:e0251393. doi: 10.1371/journal.pone.0251393

PubMed Abstract | CrossRef Full Text | Google Scholar

Bourne, V. J., and Watling, D. (2015). Individual differences in emotion lateralisation and the processing of emotional information arising from social interactions. Laterality 20, 95–111. doi: 10.1080/1357650X.2014.925910

PubMed Abstract | CrossRef Full Text | Google Scholar

Brody, L. R., Lovas, G. S., and Hay, D. H. (1995). Gender differences in anger and fear as a function of situational context. Sex Roles 32, 47–78.

Google Scholar

Cahill, L., Haier, R. J., White, N. S., Fallon, J., Kilpatrick, L., Lawrence, C., et al. (2001). Sex-related difference in amygdala activity during emotionally influenced memory storage. Neurobiol. Learn. Mem. 75, 1–9. doi: 10.1006/nlme.2000.3999

PubMed Abstract | CrossRef Full Text | Google Scholar

Calbi, M., Langiulli, N., Ferroni, F., Montalti, M., Kolesnikov, A., Gallese, V., et al. (2021). The consequences of COVID-19 on social interactions: an online study on face covering. Sci. Rep. 11:2601. doi: 10.1038/s41598-021-81780-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Canli, T., Desmond, J. E., Zhao, Z., Glover, G., and Gabrieli, J. D. E. (1998). Hemispheric asymmetry for emotional stimuli detected with fMRI. Neuroreport 9, 3233–3239. doi: 10.1097/00001756-199810050-00019

PubMed Abstract | CrossRef Full Text | Google Scholar

Carbon, C. C. (2020). Wearing face masks strongly confuses counterparts in reading emotions. Front. Psychol. 11:566886. doi: 10.3389/fpsyg.2020.566886

PubMed Abstract | CrossRef Full Text | Google Scholar

Carbon, C. C., Held, M. J., and Schütz, A. (2022). Reading emotions in faces with and without masks is relatively independent of extended exposure and individual difference variables. Front. Psychol. 17:856971. doi: 10.3389/fpsyg.2022.856971

PubMed Abstract | CrossRef Full Text | Google Scholar

Carragher, D. J., and Hancock, P. J. B. (2010). Surgical face masks impair human face matching performance for familiar and unfamiliar faces. Cogn. Res. Princ. Implic. 5:59. doi: 10.1186/s41235-020-00258-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Davidson, R. J. (1995). “Cerebral asymmetry, emotion and affective style,” in Brain Asymmetry, eds R. J. Davidson and K. Hughdahl (Cambridge, MA: MIT Press), 361–387. doi: 10.1016/j.biopsycho.2005.06.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Davidson, R. J., Abercrombie, H., Nitschke, J. B., and Putnam, K. (1999). Regional brain function, emotion and disorders of emotion. Curr. Opin. Neurobiol. 9, 228–234.

Google Scholar

de Gelder, B., de Borst, A. W., and Watson, R. (2015). The perception of emotion in body expressions. Wiley Interdiscip. Rev. Cogn. Sci. 6, 149–158. doi: 10.1002/wcs.1335

PubMed Abstract | CrossRef Full Text | Google Scholar

Deng, Y., Chang, L., Yang, M., Huo, M., and Zhou, R. (2016). Gender differences in emotional response: inconsistency between experience and expressivity. PLoS One 11:e0158666. doi: 10.1371/journal.pone.0158666

PubMed Abstract | CrossRef Full Text | Google Scholar

Derntl, B., Seidel, E. M., Kainz, E., and Carbon, C. C. (2009). Recognition of emotional expressions is affected by inversion and presentation time. Perception 38, 1849–1862. doi: 10.1068/p6448

PubMed Abstract | CrossRef Full Text | Google Scholar

Esteves, F., Dimberg, U., and Öhman, A. (1994). Automatically elicited fear: conditioned skin conductance responses to masked facial expressions. Cogn. Emot. 8, 393–413.

Google Scholar

Ekman, P., Friesen, W. V., and O’Sullivan, M. (1988). Smiles when lying. J. Pers. Soc. Psychol. 54, 414–420.

Google Scholar

Faul, F., Erdfelder, E., Buchner, A., and Lang, A. G. (2009). Statistical power analyses using G*Power 3.1: tests for correlation and regression analyses. Behav. Res. Methods 41, 1149–1160. doi: 10.3758/BRM.41.4.1149

PubMed Abstract | CrossRef Full Text | Google Scholar

Fox, E., Lester, V., Russo, R., Bowles, R. J., Pichler, A., and Dutton, K. (2000). Facial expressions of emotion: are angry faces detected more efficiently? Cogn. Emot. 14, 61–92. doi: 10.1080/026999300378996

PubMed Abstract | CrossRef Full Text | Google Scholar

Grundmann, F., Epstude, K., and Scheibe, S. (2021). Face masks reduce emotion-recognition accuracy and perceived closeness. PLoS One 16:e0249792. doi: 10.1371/journal.pone.0249792

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, J. K., Hutton, S. B., and Morgan, M. J. (2010). Sex differences in scanning faces: Does attention to the eyes explain female superiority in facial expression recognition? Cogn. Emot. 24, 629–637.

Google Scholar

He, Z., Liu, Z., Wang, J., and Zhang, D. (2018). Gender differences in processing fearful and angry body expressions. Front. Behav. Neurosci. 12:164. doi: 10.3389/fnbeh.2018.00164

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoffmann, H., Kessler, H., Eppel, T., Rukavina, S., and Traue, H. C. (2010). Expression intensity, gender and facial emotion recognition: women recognize only subtle facial emotions better than men. Acta Psychol. 135, 278–283. doi: 10.1016/j.actpsy.2010.07.012

PubMed Abstract | CrossRef Full Text | Google Scholar

Israelashvili, J., Sauter, D., and Fischer, A. (2020). Two facets of affective empathy: concern and distress have opposite relationships to emotion recognition. Cogn. Emot. 34, 1112–1122. doi: 10.1080/02699931.2020.1724893

PubMed Abstract | CrossRef Full Text | Google Scholar

Jessen, S., and Grossmann, T. (2014). Unconscious discrimination of social cues from eye whites in infants. Proc. Natl. Acad. Sci. U. S. A. 111, 16208–16213. doi: 10.1073/pnas.1411333111

PubMed Abstract | CrossRef Full Text | Google Scholar

Kret, M. E., and de Gelder, B. (2013). When a smile becomes a fist: the perception of facial and bodily expressions of emotion in violent offenders. Exp. Brain Res. 228, 399–410. doi: 10.1007/s00221-013-3557-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, G., Zhang, S., Le, T. M., Tang, X., and Li, C. R. (2020). Neural responses to negative facial emotions: sex differences in the correlates of individual anger and fear traits. Neuroimage 221:117171. doi: 10.1016/j.neuroimage.2020.117171

PubMed Abstract | CrossRef Full Text | Google Scholar

Mancini, C., Falciati, L., Maioli, C., and Mirabella, G. (2020). Threatening facial expressions impact goal-directed actions only if task-relevant. Brain Sci. 10:794. doi: 10.3390/brainsci10110794

PubMed Abstract | CrossRef Full Text | Google Scholar

Marini, M., Ansani, A., Paglieri, F., Caruana, F., and Viola, M. (2021). The impact of facemasks on emotion recognition, trust attribution and re-identification. Sci. Rep. 11:5577. doi: 10.1038/s41598-021-84806-5

PubMed Abstract | CrossRef Full Text | Google Scholar

McClure, E. B. (2000). A meta-analytic review of sex differences in facial expression processing and their development in infants, children, and adolescents. Psychol. Bull. 126, 424–453. doi: 10.1037/0033-2909.126.3.424

PubMed Abstract | CrossRef Full Text | Google Scholar

Meeren, H. K., van Heijnsbergen, C. C., and de Gelder, B. (2005). Rapid perceptual integration of facial expression and emotional body language. Proc. Natl. Acad. Sci. U. S. A. 102, 16518–16523. doi: 10.1073/pnas.0507650102

PubMed Abstract | CrossRef Full Text | Google Scholar

Montagne, B., Kessels, R. P. C., Frigerio, E., de Haan, E. H., and Perrett, D. I (2005). Sex differences in the perception of affective facial expressions: Do men really lack emotional sensitivity? Cogn. Process. 6, 136–141. doi: 10.1007/s10339-005-0050-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Noyes, E., Davis, J. P., Petrov, N., Gray, K., and Ritchie, K. L. (2021). The effect of face masks and sunglasses on identity and expression recognition with super-recognizers and typical observers. R. Soc. Open Sci. 8:201169. doi: 10.1098/rsos.201169

PubMed Abstract | CrossRef Full Text | Google Scholar

Öhman, A. (1993). “Fear and anxiety as emotional phenomenon: clinical phenomenology, evolutionary perspectives, and information-processing mechanisms,” in Handbook of Emotions, eds M. Lewis and J. M. Haviland (New York, NY: Guildford Press), 511–536.

Google Scholar

Pavlova, M. A., and Sokolov, A. A. (2021). Reading covered faces. Cereb. Cortex 14:bhab311. doi: 10.1093/cercor/bhab311

PubMed Abstract | CrossRef Full Text | Google Scholar

Ponari, M., Conson, M., D’Amico, N. P., Grossi, D., and Trojano, L. (2012). Mapping correspondence between facial mimicry and emotion recognition in healthy subjects. Emotion 12, 1398–1403. doi: 10.1037/a0028588

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M. (2017). Sex differences in social cognition: the case of face processing. J. Neurosci. Res. 95, 222-234. doi: 10.1002/jnr.23817

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M. (2021). Sex differences in the social brain and in social cognition. J. Neurosci. Res. [Online ahead of print]. doi: 10.1002/jnr.24787

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., Brignone, V., Matarazzo, S., Del Zotto, M., and Zani, A. (2006). Gender and parental status affect the visual cortical response to infant facial expression. Neuropsychologia 44, 2987–2999. doi: 10.1016/j.neuropsychologia.2006.06.015

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., Matarazzo, S., Brignone, V., Del Zotto, M., and Zani, A. (2007). Processing valence and intensity of infant expressions: the roles of expertise and gender. Scand. J. Psychol. 48, 477–485. doi: 10.1111/j.1467-9450.2007.00616.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Proverbio, A. M., Ornaghi, L., and Gabaro, V. (2018). How face blurring affects body language processing of static gestures in women and men. Soc. Cogn. Affect. Neurosci. 13, 590–603. doi: 10.1093/scan/nsy033

PubMed Abstract | CrossRef Full Text | Google Scholar

Repple, J., Habel, U., Wagels, L., Pawliczek, C. M., Schneider, F., and Kohn, N. (2018). Sex differences in the neural correlates of aggression. Brain Struct. Funct. 223, 4115–4124. doi: 10.1007/s00429-018-1739-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Riva, P., Sacchi, S., Montali, L., and Frigerio, A. (2011). Gender effects in pain detection: speed and accuracy in decoding female and male pain expressions. Eur. J. Pain 15, 985.e1–985.e11. doi: 10.1016/j.ejpain.2011.02.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Roberson, D., Kikutani, M., Döge, P., Whitaker, L., and Majid, A. (2012). Shades of emotion: what the addition of sunglasses or masks to faces reveals about the development of facial expression processing. Cognition 125, 195–206.

Google Scholar

Royet, J., Plailly, J., Delon-Martin, C., Kareken, D. A., and Segebarth, C. (2003). fMRI of emotional responses to odors: influence of hedonic valence and judgment, handedness, and gender. Neuroimage 20, 713–728. doi: 10.1016/S1053-8119(03)00388-4

CrossRef Full Text | Google Scholar

Ruba, A. L., and Pollak, S. D. (2020). Children’s emotion inferences from masked faces: implications for social interactions during COVID-19. PLoS One 15:e0243708. doi: 10.1371/journal.pone.0243708

PubMed Abstract | CrossRef Full Text | Google Scholar

Rymarczyk, K., Żurawski, Ł, Jankowiak-Siuda, K., and Szatkowska, I. (2019). Empathy in facial mimicry of fear and disgust: simultaneous EMG-fMRI recordings during observation of static and dynamic facial expressions. Front. Psychol. 27:701. doi: 10.3389/fpsyg.2019.00701

PubMed Abstract | CrossRef Full Text | Google Scholar

Simon, D., Craig, K. D., Miltner, W. H., and Rainville, P. (2006). Brain responses to dynamic facial expressions of pain. Pain 126, 309–318. doi: 10.1016/j.pain.2006.08.033

PubMed Abstract | CrossRef Full Text | Google Scholar

Singh, L., Tan, A., and Quinn, P. C. (2021). Infants recognize words spoken through opaque masks but not through clear masks. Dev. Sci. [Epub ahead of print]. doi: 10.1111/desc.13117.

PubMed Abstract | CrossRef Full Text | Google Scholar

Smith, M. L., Cottrell, G. W., Gosselin, F., and Schyns, P. G. (2005). Transmitting and decoding facial expressions. Psychol. Sci. 16, 184–189.

Google Scholar

Sullivan, S., Campbell, A., Hutton, S. B., and Ruffman, T. (2017). What’s good for the goose is not good for the gander: age and gender differences in scanning emotion faces. J. Gerontol. B Psychol. Sci. Soc. Sci. 72, 441–447.

Google Scholar

Ugail, H., and Al-dahoud, A. A. (2019). Genuine smile is indeed in the eyes - the computer aided non-invasive analysis of the exact weight distribution of human smiles across the face. Adv. Eng. Inf. 42:100967.

Google Scholar

Wegrzyn, M., Vogt, M., Kireclioglu, B., Schneider, J., and Kissler, J. (2017). Mapping the emotional face. How individual face parts contribute to successful emotion recognition. PLoS One 12:e0177239. doi: 10.1371/journal.pone.0177239

PubMed Abstract | CrossRef Full Text | Google Scholar

Yitzhak, N., Pertzov, Y., Guy, N., and Aviezer, H. (2020). Many ways to see your feelings: successful facial expression recognition occurs with diverse patterns of fixation distributions. Emotion [Epub ahead of print]. doi: 10.1037/emo0000812

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: emotions, face masking, facial expression, face processing, sex difference, empathy

Citation: Proverbio AM and Cerri A (2022) The Recognition of Facial Expressions Under Surgical Masks: The Primacy of Anger. Front. Neurosci. 16:864490. doi: 10.3389/fnins.2022.864490

Received: 28 January 2022; Accepted: 31 May 2022;
Published: 16 June 2022.

Edited by:

Rufin VanRullen, Centre National de la Recherche Scientifique (CNRS), France

Reviewed by:

Krystyna Rymarczyk, University of Social Sciences and Humanities, Poland
Zahra Zargol Moradi, King’s College London, United Kingdom

Copyright © 2022 Proverbio and Cerri. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Alice M. Proverbio, mado.proverbio@unimib.it

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.