Skip to main content

BRIEF RESEARCH REPORT article

Front. Psychol., 28 April 2020
Sec. Emotion Science

Attentional Bias to Facial Expressions of Different Emotions – A Cross-Cultural Comparison of ≠Akhoe Hai||om and German Children and Adolescents

  • 1Department of Psychology, Brandenburg Medical School Theodor Fontane, Neuruppin, Germany
  • 2Comparative Developmental Psychology, Department of Education and Psychology, Freie Universität Berlin, Berlin, Germany
  • 3Graduate School “Languages of Emotion”, Freie Universität Berlin, Berlin, Germany
  • 4Department of Linguistics, Cognitive Sciences, University of Potsdam, Potsdam, Germany

The attentional bias to negative information enables humans to quickly identify and to respond appropriately to potentially threatening situations. Because of its adaptive function, the enhanced sensitivity to negative information is expected to represent a universal trait, shared by all humans regardless of their cultural background. However, existing research focuses almost exclusively on humans from Western industrialized societies, who are not representative for the human species. Therefore, we compare humans from two distinct cultural contexts: adolescents and children from Germany, a Western industrialized society, and from the ≠Akhoe Hai||om, semi-nomadic hunter-gatherers in Namibia. We predicted that both groups show an attentional bias toward negative facial expressions as compared to neutral or positive faces. We used eye-tracking to measure their fixation duration on facial expressions depicting different emotions, including negative (fear, anger), positive (happy), and neutral faces. Both Germans and the ≠Akhoe Hai||om gazed longer at fearful faces, but shorter on angry faces, challenging the notion of a general bias toward negative emotions. For happy faces, fixation durations varied between the two groups, suggesting more flexibility in the response to positive emotions. Our findings emphasize the need for placing research on emotion perception into an evolutionary, cross-cultural comparative framework that considers the adaptive significance of specific emotions, rather than differentiating between positive and negative information, and enables systematic comparisons across participants from diverse cultural backgrounds.

Introduction

A large body of research suggests that humans pay more attention to negative than positive information (e.g., Baumeister et al., 2001; Rozin and Royzman, 2001). The enhanced sensitivity to negative information – resulting in increased alertness and the mobilization of attentional resources – is most likely an evolutionary adaptive behavior, as the ability to successfully detect and appropriately respond to threatening and potentially harmful situations increases the probability of survival (Öhman and Mineka, 2001). Positive information, on the other hand, promotes flexible and explorative behavior, and as a result, supports social bonding and positive interactions (Fredrickson, 1998). Unlike in case of neglecting negative information, the consequences of a missed opportunity to react appropriately to positive information seem much less severe (Baumeister et al., 2001).

Given the evolutionary significance of the fast detection of and appropriate reaction to potential threats (Öhman and Mineka, 2001), it seems likely that the attentional bias to negative emotions is a universal trait shared by all humans regardless of their cultural background. However, researchers have largely focused on humans from industrialized, Western societies, who are not representative for the human species (Henrich et al., 2010). Although some studies showed that emotional facial expressions are not universally recognized (Jack et al., 2012; Gendron et al., 2014; Crivelli et al., 2016) and that humans attend to positive or negative information differently depending on their cultural backgrounds (Grossmann et al., 2012), such cross-cultural comparisons including several human populations from diverse cultural, social, and ecological backgrounds remain scarce.

Furthermore, the notion of an exclusive, biologically prepared bias toward negative information has been challenged, as meta-analyses also confirmed a bias toward positive information (Pool et al., 2016), or demonstrated that the negativity bias is influenced by, for example, anxious concerns, type of negative information, or age (Bar-Haim et al., 2007; van Rooijen et al., 2017; Lisk et al., 2019). Thus, to better understand these apparently inconsistent findings, it is crucial to understand developmental pathways of attentional biases. However, unlike for adults, there is substantially less research with children and adolescents (Vaish et al., 2008). The bias toward negative information emerges early in ontogeny (Leppänen and Nelson, 2012), since after an initial positivity bias (Vaish et al., 2008), infants between 5 and 7 months of age pay more attention to negative information, such as fearful faces (Grossmann and Jessen, 2017). From about 4 years of age, children seem to prefer both negative and positive emotional stimuli over neutral information (Elam et al., 2010; Burris et al., 2017), while adolescents show a bias toward negative emotions (Grose-Fifer et al., 2013). In adulthood, there is substantial evidence for a bias toward negative information across different domains, such as social interactions and relationships, learning or emotion processing (Baumeister et al., 2001). Several studies show, however, that a negativity bias in children, adolescents and adults is specifically found in more anxious individuals (Bar-Haim et al., 2007; Shechner et al., 2013). In older adults, there is a shift toward a bias for positive information, although this seems to differ across cultural contexts (Fung et al., 2008). Taken together, at different times in ontogeny, humans preferentially pay attention to negative and/or positive information, with the overall pattern that – at least in Western societies – the bias toward negative emotions emerges in the second half of infants’ first year of life, and, although also positive information is preferred over neutral stimuli (Pool et al., 2016), a pronounced positive bias only emerges again in older adults.

To investigate attentional biases, researchers often focus on emotional information and use facial expressions of basic emotions to compare humans’ responses to negative, positive, and neutral faces. Unlike the proposed general bias to negative information (Rozin and Royzman, 2001), there is increasing evidence that humans’ responses to faces vary across types of negative emotions (for reviews, see Frischen et al., 2008; Vaish et al., 2008; Yiend, 2010). However, studies vary in their conclusions with regard to which negative emotion attracts most attention. For example, while Williams et al. (2005) found that angry faces are detected faster than fearful faces, many studies report a fear bias (Vaish et al., 2008), supported by neurobiological evidence showing a stronger activation of the amygdala in response to fearful as compared to angry faces (Whalen et al., 2001). Interestingly, some studies suggest that threat-related stimuli, such as angry faces, are even avoided by children and adolescents, particularly by anxious youths (Lisk et al., 2019). To compare findings across studies and to identify general patterns, however, is difficult, as most studies focus on one type of negative emotion (either anger or fear) and therefore do not investigate if responses differ depending on the type of negative emotion (e.g., Hansen and Hansen, 1988; Rossignol et al., 2013).

We aimed at filling two of these several gaps in research on attentional biases: the lack of research with humans from non-Western societies, and the limited knowledge about the processing of different types of negative emotion in children and adolescents. We studied two very distinct cultural groups: the ≠Akhoe Hai||om, who are semi-nomadic egalitarian hunter-gatherers in northern Namibia; and Germans from two large cities as representatives of a Western industrialized society. We used eye-tracking in a free viewing task to investigate if the groups’ gazing patterns on facial expressions varied depending on the type of the depicted emotion, and compared two negative emotions (fear, anger) with a positive emotion (happy) and neutral faces. We hypothesized that given the evolutionary significance of negative, potentially threatening information, participants from both cultural backgrounds should look longer at negative facial expressions compared to neutral or positive faces. Although we cannot claim to identify universal behaviors based on two samples, we suggest that if we find an attentional bias toward negative information in both groups, which differ with regard to several factors, such as dwelling, subsistence, and social organization, it seems at least likely that this trait is shared by many humans regardless of their cultural backgrounds.

Materials and Methods

Participants

We focused on adolescents and children as we were not able to recruit sufficient numbers of adults from the ≠Akhoe Hai||om. Using opportunity sampling, we first collected the data in Namibia and tested adolescents and children who were available and willing to participate. It is important to note that determining the exact ages of the ≠Akhoe Hai||om was difficult, since many children are not officially registered when they are born, and therefore exact dates of birth days are often not known. Second, we tested adolescents and children in Germany and tried to match participants in terms of age and gender to the ≠Akhoe Hai||om sample.

The ≠Akhoe Hai||om in northern Namibia are semi-nomadic hunter-gatherers, characterized by egalitarian social structures and the common practice of sharing of resources (Widlok, 1999). Their traditional lifestyle is changing, as they have become increasingly sedentary and have taken up alternative subsistence strategies, like gardening or animal husbandry, and formalized schooling has been introduced. Participants were recruited from the Khomxa Khoeda Primary School at Farm 6 and comprised 30 participants (15 males, 15 females; mean age = 11.93 years, SD = 2.87, age range 7–19 years).

Germany is an industrialized Western European nation that values individual independence. The German sample consisted of 21 participants (9 males, 12 females, mean age = 13.76 years, SD = 1.48, age range 12–18 years), who were recruited from two schools in two large cities (Marion-Dönhoff-Gymnasium, Hamburg; Bertolt-Brecht High School, Berlin).

Although it has been shown that affectivity or depressed mood may affect attention for positive facial expressions (e.g., Isaac et al., 2014), we were not able to assess this in our participants, as there are no standardized, culture-fair questionnaires for measuring affectivity in the ≠Akhoe Hai||om.

The study was conducted in accordance with the Declaration of Helsinki and the ethical guidelines of the German Psychological Society. The study did not require approval by an Institutional Review Board, as it did not involve any invasive techniques, ethically problematic procedures, or deception [see the regulations on freedom of research in the German Constitution, §5 (3)]. Permission to conduct this study with the ≠Akhoe Hai||om was obtained from the “Working Group of Indigenous Minorities in Southern Africa” (WIMSA) in Windhoek, and the local school’s principal, Efraim Kavetuna. Prior to testing, each participant was informed about the background and procedure of the study by a video recording in their native language, and gave their informed consent verbally. In Germany, parents gave their written informed consent.

Stimuli

We selected pictures (440 × 550 pixels) of 14 German adults (7 females) showing negative (fearful, angry), positive (happy), and neutral faces from the FACES database (Ebner et al., 2010)1. Thus, unlike German participants, the ≠Akhoe Hai||om looked at stimuli from a different ethnical group. Although there are databases with stimuli representing a greater ethnical diversity (e.g., NimStim; Tottenham et al., 2009), there is none with facial expressions of the ≠Akhoe Hai||om. Therefore, we used stimuli from the FACES database, since it depicts naturalistic facial expressions of amateur actors rather than professionals, and the photographs, standardized in size, color and background, have been extensively rated with regard to the type of depicted emotional expression (Ebner et al., 2010).

To enable the comparison of fixations between two emotions, each stimulus consisted of a pair of pictures, with each picture showing a different facial expression of the same person (e.g., person A with a happy face and person A with a fearful face). The picture pairs were presented on gray background, separated by a gap of 10 pixels. Each facial expression was combined with every other facial expression, resulting in six possible picture combinations (fear-angry, fear-happy, fear-neutral, angry-happy, angry-neutral, happy-neutral). The position of each facial expression (left or right) was counterbalanced and the order of presenting these picture pairs was randomized. Each participant saw an equal number of pictures of each facial expression and their combinations with the corresponding other facial expressions.

Procedure

All participants were tested in a separate room in their schools. A table-mounted, monitor-integrated eye-tracker (Tobii T60) was used to measure the participants’ fixations with an infrared corneal reflection system. Participants were seated in front of the monitor (distance: 60–70 cm), on which the stimuli were presented. As we used a free-viewing task, the participants were only informed that a series of facial expressions will be shown on a screen we asked them to watch (Germans: verbal information; ≠Akhoe Hai||om: video-recorded information in their native language). Before the experiment started, we conducted an automated five-point calibration (with a moving red dot on white background) and only started testing after successful calibration. To start the experiment, the participant’s gaze was centered on the screen by presenting a fixation cross for 500 ms. Then a picture pair appeared for 3,000 ms, followed by a fixation cross shown for 500 ms to again center the participant’s gaze, followed by the next picture pair. In total, 168 different trials were presented (28 trials per condition of each picture pair) to each participant in one session, which lasted ~15–20 min.

Data Analysis

Raw data were extracted (see Supplementary Data Sheet 1) and then collapsed for analysis using the software Matlab® (R2014b). For the analysis, we defined one of the two facial expressions as the target emotion and the other as the distractor emotion. Depending on the focus of the analysis, each emotion was considered either as target or distractor. We defined each of the two facial expressions in a trial as a separate area of interest (AoI), and measured the fixation duration separately for each of the two AoIs. A fixation was scored if the gaze remained stationary within a radius of 50 pixels for at least 100 ms. The measurement of first fixations was not suitable, as this would have required a more controlled testing situation (e.g., restriction of head movements, control of pre-trial fixation position, and no distractions, such as noise), which was not possible to achieve. Therefore, we focused on fixation duration over the course of stimulus presentation, which enabled us to derive maintained attention but also avoidance over the time of stimulus presentation (Lisk et al., 2019).

We conducted two sets of analyses: first, we compared the average cumulative fixation duration on each target emotion in comparison to the average cumulative fixation duration on all corresponding distractors combined (e.g., target = happy vs. all distractors = neutral + fear + anger). Second, we compared the mean fixation duration on each target emotion with the mean fixation duration on a specific distractor (e.g., target = happy vs. distractor = anger). While we conducted the first analysis to control for the possibility that the fixation duration on a specific target emotion was influenced by the simultaneously presented distractor emotion, the second analysis enabled us to directly compare fixation durations between two different emotions. Because mean ages varied between our samples, we first tested if age had an influence on fixation duration. We found no effect and therefore excluded this variable from further analyses.

To test if fixation durations differed across emotions, we compared a particular emotion combination consisting of the target emotion and distractor(s) by fitting a linear mixed model in R (3.1.1, R Core Team, 2013) using the function lmer of the R-package lme4 (Baayen, 2008; Bates et al., 2014). We ran a linear mixed model with no intercept and no slopes, with fixation duration as dependent variable, and the different levels of the predictor emotion combination as fixed effects, as well as subject as random effect to account for individual differences. Since the levels of the variable emotion combination were not independent, we allowed for all possible correlations. P-values were derived by comparing the full model with all levels of emotion combination included to a reduced model with the specific combination excluded. The significance of this comparison was tested using a likelihood ratio test (R function anova with argument test set to “Chisq”).

Results

First Analysis: Target Emotion vs. All Distractors Combined

Germans fixated on both fearful and happy faces significantly longer than on all other corresponding distractors combined (fear vs. distractors: χ2 = 5.37, df = 1, p = 0.021; happy vs. distractors: χ2 = 6.54, df = 1, p = 0.011), while they fixated on neutral and angry faces significantly shorter compared to their respective distractors (neutral vs. distractors: χ2 = 10.97, df = 1, p = 0.001; anger vs. distractors: χ2 = 5.63, df = 1, p = 0.018) (Figure 1 and Table 1). Similarly, the ≠Akhoe Hai||om fixated on fearful faces longer than on their distractors (fear vs. distractors: χ2 = 13.30, df = 1, p < 0.001), while they fixated on both neutral and angry faces significantly shorter than on their distractors (neutral vs. distractors: χ2 = 5.38, df = 1, p = 0.020; anger vs. distractors: χ2 = 15.76, df = 1, p < 0.001). While both groups exhibited similar fixation patterns for fearful, angry, and neutral faces, they responded differently to happy faces, since unlike German participants, the ≠Akhoe Hai||om did not fixate on them longer compared to all other emotions (happy vs. distractors:χ2 = 0.11, df = 1, p = 0.738).

FIGURE 1
www.frontiersin.org

Figure 1. Mean cumulative fixation duration for the comparison of the target emotion (e.g., Fear) vs. all distractors combined (e.g., All.Fear) for Germans and ≠Akhoe Hai||om. Scales show the mean fixation duration in milliseconds. Significant differences are indicated by asterisks (**p < 0.01, ***p < 0.001).

TABLE 1
www.frontiersin.org

Table 1. Estimates for the mean cumulative fixation duration (in ms) on the target emotion vs. all distractors (see also Figure 1) calculated for both groups.

Second Analysis: Target Emotion vs. Specific Distractor

Both populations fixated on fearful expressions longer than on angry and neutral faces (≠Akhoe Hai||om: fear vs. anger χ2 = 31.49, df = 1, p < 0.001; fear vs. neutral χ2 = 4.98, df = 1, p = 0.026; Germans: fear vs. anger χ2 = 13.69, df = 1, p < 0.001; fear vs. neutral χ2 = 3.35, df = 1, p = 0.067), and on angry faces shorter than happy faces, although this was only a trend for Germans (≠Akhoe Hai||om: χ2 = 7.99, df = 1, p = 0.005; Germans: χ2 = 3.36, df = 1, p = 0.067). Unlike Germans, the ≠Akhoe Hai||om fixated on fearful faces significantly longer compared to happy faces (≠Akhoe Hai||om: χ2 = 6.28, df = 1, p = 0.012, Germans: χ2 = 0.10, df = 1, p = 0.740). Unlike the ≠Akhoe Hai||om, Germans fixated on happy faces longer than on neutral faces (Germans: χ2 = 17.84, df = 1, p < 0.001, ≠Akhoe Hai||om: χ2 = 0.14, df = 1, p = 0.723). In sum, resembling the pattern found in the first analysis, both groups fixated on fearful faces longer than on angry or neutral faces, and on angry faces shorter than on happy faces (Figure 2 and Table 2), while they varied to some extent in their responses to happy faces.

FIGURE 2
www.frontiersin.org

Figure 2. Mean fixation duration for the comparison of a target emotion (in capital letters) vs. a distractor emotion (in lower case letters) for Germans and the ≠Akhoe Hai||om. fe, fear; an, anger; ha, happy; ne, neutral. Scales show the mean fixation duration in milliseconds. Significant differences are indicated by asterisks (*p < 0.05, **p < 0.01, ***p < 0.001).

TABLE 2
www.frontiersin.org

Table 2. Estimates for the mean fixation duration (in ms) on a target emotion vs. a specific distractor emotion (see also Figure 2) for both groups.

Discussion

We expected that participants from both cultural backgrounds would look longer at negative facial expressions compared to neutral or positive faces. In line with this expectation, both the ≠Akhoe Hai||om and Germans fixated on fearful expressions longer, regardless of the emotion they were combined with. Contrary to our prediction, both groups fixated on angry faces significantly shorter than on all other emotions. Thus, our hypothesis was only partially confirmed, since we did not find a general attentional bias toward negative emotions, as participants of both groups attended more to fearful than angry facial expressions.

To better understand this finding, it is important to place it into an evolutionary framework, and to consider the adaptive function of reactions to specific emotional expressions (Schmidt and Cohn, 2001; Öhman et al., 2001). Both angry and fearful expression may be perceived as potential threatening information, but they emerge for different reasons, and therefore require different responses. Fearful expressions are shown in response to a threat in the environment, and the individual perceiving this expression on someone’s face needs to detect the source of the threat, and is therefore orienting attention to this face (Öhman, 2008). The white sclera of the human eye (Kobayashi and Kohshima, 1997) attracts this unconscious orienting response, mediated by the amygdala, from an early age on (Whalen et al., 2004; Jessen and Grossmann, 2014; for a review see Emery, 2000). Angry expressions, on the other hand, may represent a direct threat from a social partner and may therefore result in avoiding the other’s gaze (Marsh et al., 2005). In support of these predictions, we found that participants looked longer at fearful, but shorter on angry faces, regardless of their cultural background.

Other approaches, which categorize emotions based on the different reaction tendencies they elicit, come to a different conclusion. Williams et al. (2005) propose that not fearful, but angry faces signal an immediate threat to recipients and should therefore attract their attention, while fearful faces divert recipients’ attention, as the potential threat comes from a source in the environment, but not from the signaler itself. Similarly, Adams and Kleck (2005), who studied the impact of gaze direction of emotional expressions on approach and avoidance responses, described anger as an approach-oriented emotion, as direct gaze increased the perceived intensity of angry faces, but fear as avoidance-oriented emotion, since averted gaze increased the perceived intensity of fearful expressions. These predictions resulting from different reaction tendencies – longer looks toward angry than fearful facial expressions – are not supported by our findings. However, unlike Adams and Kleck (2005), we did not manipulate gaze direction, but only used direct gaze. Furthermore, since our stimuli depicted adults, who could be perceived as physically or mentally superior by our participants, we cannot exclude an impact on children’s and adolescents’ gaze patterns. Thus, an alternative explanation for our findings of shorter fixation durations on angry faces is that they perceived anger in adult faces as more threatening than adult participant would do. To address this issue, future studies should use stimuli resembling young participants’ ages, and should vary gaze direction across angry and fearful faces (see van Rooijen et al., 2017).

A possible explanation for these inconsistent findings across studies regarding the attentional bias to either fearful or angry faces comes from Mogg et al. (2007). They differentiated between reflexive and reflective attention, and suggest that initially, both angry and fearful faces automatically attract attention. At a later stage, attention is averted from angry faces, while for fearful faces, attention is maintained to identify the most appropriate response. This suggests that different mechanisms may underlie the processing of negative faces, depending on how long such stimuli are presented. Indeed, several studies with children and adolescents between 3 and 18 years draw different conclusions with regards to attentional biases toward anger or fear, depending on whether early fixations (<120 ms after stimulus onset) or maintained attention (dwell time across stimulus presentation) are measured (Pool et al., 2016; Lisk et al., 2019). An attentional bias toward angry faces is evident during initial fixations (Shechner et al., 2013), while maintained gazes at angry faces are avoided (Lisk et al., 2019). Thus, unlike initial orienting, maintained attention – as measured in our study – suggests some volitional “top-down control,” which is expected to increase with children’s age (Lagattuta and Kramer, 2017). In our study, age had no effect on the overall fixation duration of different types of emotions, but it is important to keep in mind that unlike in these studies, our sample did not include preschool children.

For happy faces, only Germans showed a bias for happy over angry or neutral faces. A possible explanation for this variability between groups is that positive emotions elicit more flexible reactions, since they broaden the recipient’ attention and facilitate social interactions (Fredrickson, 1998), but do not require an immediate and appropriate response like negative emotions. However, as only happy faces have been presented, it remains unclear if this finding can be generalized to other positive emotions (Sauter, 2010). Clear evidence for a positivity bias was found in studies using a greater variety of positive information (e.g., erotic pictures, food, babies; Pool et al., 2016).

Together, our findings suggest a culture-independent bias toward fearful, but not angry facial expressions, while there is more variability between cultural groups with regard to their responses to positive faces. This may suggest that the different, emotion-specific responses to negative facial expressions are adaptive and most likely shared across humans, since unlike positive emotions, fearful and angry faces require very specific, but different responses, to avoid potential harmful consequences.

Although this conclusion is tempting, it is important to consider some limitations of this study. First, based on the comparison of small samples from only two cultural groups, it is impossible to conclude that our finding of an attentional bias toward fearful, but not angry faces represents a universal pattern. Second, unlike Germans, the ≠Akhoe Hai||om did not look at facial expressions from their own ethnical group, which could have influenced their emotion recognition (Elfenbein and Ambady, 2002). Therefore, we cannot rule out the possibility that differences between cultural groups regarding happy faces are caused by the varying familiarity with the stimuli. This, however, seems not very likely given the similarities between groups with regard to their processing of negative emotions. Still, we have to consider the possibility that the ≠Akhoe Hai||om interpret facial emotions differently compared to Germans, as the universality hypothesis of basic emotion perception has been repeatedly questioned (Jack et al., 2012; Gendron et al., 2014), and since contextual information seems crucial for interpreting facial expressions (Crivelli et al., 2016). Third, only faces of adults, but not peers, were presented, which might have influenced the young participants’ fixations on anger. However, it is important to note that this is not a limitation unique to our study, as most research with children and adolescents is not using age-matched facial expressions. Fourth, since our sample only included adolescents and children, it remains unclear if our findings can be generalized to adults, as adolescents’ face processing capabilities may not fully resemble those of adults yet (Batty and Taylor, 2006), and since their bias toward negative stimuli develops throughout adolescence (Yurgelun-Todd and Killgore, 2006; Thomas et al., 2007). Despite these shortcomings, this study is a first step toward a better understanding of the cross-cultural similarities and differences in human emotion perception and provides evidence for the special relevance of fearful faces compared to facial expressions of other emotions.

Data Availability Statement

All datasets generated for this study are included in the article/Supplementary Material.

Ethics Statement

Ethical approval was not provided for this study on human participants because it did not involve any invasive techniques, ethically problematic procedures, or deception. Therefore, according to the regulations on freedom of research in the German Constitution [§5 (3)], it did not require approval by an Institutional Review Board. However, since our sample included participants from an indigenous group in Namibia, we obtained approval by the “Working Group of Indigenous Minorities in Southern Africa,” located in Windhoek, Namibia. In Germany, parents’ informed consent was obtained in the written form. In Namibia, however, this was problematic, since most parents were not able to read or write. Therefore, since this study was conducted at the school, we obtained permission by the principal of this school and obtained the participant’s additional consent either verbally (“yes”) or visually (“nodding”).

Author Contributions

CM and CP were responsible for designing this study and for collecting and analyzing the data. CM also contributed to writing the manuscript. IW and ST were involved in designing this study and contributed to drafting this manuscript. KL contributed to planning the study, collected parts of the data and was involved in writing the manuscript.

Funding

This work was funded by the Excellence Cluster EXC 302 Languages of Emotion, Project number 39932880, and by the Freie Universität Berlin within the Excellence Initiative of the German Research Foundation (Deutsche Forschungsgemeinschaft, DFG).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We thank all participants of this study. We are very grateful to WIMSA for their permission to work with the ≠Akhoe Hai||om and we thank Principal Efraim Kavetuna (Khomxa Khoeda Primary School), the directors of the Marion-Dönhoff-Gymnasium, Hamburg-Blankenese and Bertolt-Brecht-Secondary School, Berlin-Spandau, as well as all teachers and parents for their permission to conduct this research. Many thanks to Daniel Haun and Christian Rapold for supporting data collection in Namibia, and Disney Tjizao for translating the instructions. Special thanks go to Martin Schultze, Roger Mundry, and Jan Ries for their statistical support. We acknowledge support by the Open Access Publication Initiative of Freie Universität Berlin.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2020.00795/full#supplementary-material

Footnotes

  1. ^ https://faces.mpdl.mpg.de

References

Adams, R. B. Jr., and Kleck, R. E. (2005). Effects of direct and averted gaze on the perception of facially communicated emotion. Emotion 5, 3–11. doi: 10.1037/1528-3542.5.1.3

PubMed Abstract | CrossRef Full Text | Google Scholar

Baayen, R. H. (2008). Analyzing Linguistic Data. A Practical Introduction to Statistics Using R. Cambridge: Cambridge University Press. doi: 10.1017/CBO9780511801686

CrossRef Full Text | Google Scholar

Bar-Haim, Y., Lamy, D., Pergamin, L., Bakermans-Kranenburg, M. J., and Van Ijzendoorn, M. H. (2007). Threat-related attentional bias in anxious and nonanxious individuals: a meta-analytic study. Psychol. Bull. 133, 1–24. doi: 10.1037/0033-2909.133.1.1

PubMed Abstract | CrossRef Full Text | Google Scholar

Bates, D., Mächler, M., Bolker, B., and Walker, S. (2014). Fitting linear mixed-effects models using lme4. J. Stat. Soft. [arXiv]. doi: 10.18637/jss.v067.i01

CrossRef Full Text | Google Scholar

Batty, M., and Taylor, M. J. (2006). The development of emotional face processing during childhood. Dev. Sci. 9, 207–220. doi: 10.1111/j.1467-7687.2006.00480.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Baumeister, R. F., Bratslavsky, E., Finkenauer, C., and Vohs, K. D. (2001). Bad is stronger than good. Rev. Gen. Psychol. 5, 323–370. doi: 10.1037/1089-2680.5.4.323

CrossRef Full Text | Google Scholar

Burris, J. L., Barry-Anwar, R. A., and Rivera, S. M. (2017). An eye tracking investigation of attentional biases towards affect in young children. Dev. Psychol. 53, 1418–1427. doi: 10.1037/dev0000345

PubMed Abstract | CrossRef Full Text | Google Scholar

Crivelli, C., Russell, J. A., Jarillo, S., and Fernández-Dols, J.-M. (2016). The fear gasping face as a threat display in a Melanesian society. Proc. Natl. Acad. Sci. U.S.A. 113, 12403–12407. doi: 10.1073/pnas.1611622113

PubMed Abstract | CrossRef Full Text | Google Scholar

Ebner, N. C., Riediger, M., and Lindenberger, U. (2010). FACES—A database of facial expressions in young, middle-aged, and older women and men: development and validation. Behav. Res. Methods 42, 351–362. doi: 10.3758/BRM.42.1.351

PubMed Abstract | CrossRef Full Text | Google Scholar

Elam, K. K., Carlson, J. M., DiLalla, L. F., and Reinke, K. S. (2010). Emotional faces capture spatial attention in 5-year-old children. Evol. Psychol. 8, 754–767. doi: 10.1177/147470491000800415

CrossRef Full Text | Google Scholar

Elfenbein, H. A., and Ambady, N. (2002). On the universality and cultural specificity of emotion recognition: a meta-analysis. Psychol. Bull. 128, 203–235. doi: 10.1037/0033-2909.128.2.203

PubMed Abstract | CrossRef Full Text | Google Scholar

Emery, N. J. (2000). The eyes have it: the neuroethology, function and evolution of social gaze. Neurosci. Biobehav. Rev. 24, 581–604. doi: 10.1016/s0149-7634(00)00025-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Fredrickson, B. L. (1998). What good are positive emotions? Rev. Gen. Psychol. 2, 300–319. doi: 10.1037/1089-2680.2.3.300

CrossRef Full Text | Google Scholar

Frischen, A., Eastwood, J. D., and Smilek, D. (2008). Visual search for faces with emotional expressions. Psychol. Bull. 134, 662–676. doi: 10.1037/0033-2909.134.5.662

PubMed Abstract | CrossRef Full Text | Google Scholar

Fung, H. H., Isaacowitz, D. M., Lu, A. Y., Wadlinger, H. A., Goren, D., and Wilson, H. R. (2008). Age-related positivity enhancement is not universal: older Chinese look away from positive stimuli. Psychol. Aging 23, 440–446. doi: 10.1037/0882-7974.23.2.440

CrossRef Full Text | Google Scholar

Gendron, M., Roberson, D., van der Vyver, J. M., and Barrett, L. F. (2014). Perceptions of emotion from facial expressions are not culturally universal: evidence from a remote culture. Emotion 14, 251–262. doi: 10.1037/a0036052

CrossRef Full Text | Google Scholar

Grose-Fifer, J., Rodrigues, A., Hoover, S., and Zottoli, T. (2013). Attentional capture by emotional faces in adolescence. Adv. Cogn. Psychol. 9, 81–91. doi: 10.2478/v10053-008-0134-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Grossmann, I., Ellsworth, P. C., and Hong, Y. Y. (2012). Culture, attention, and emotion. J. Exp. Psychol. Gen. 141, 31–36. doi: 10.1037/a0023817

CrossRef Full Text | Google Scholar

Grossmann, T., and Jessen, S. (2017). When in infancy does the “fear bias” develop? J. Exp. Child Psychol. 153, 149–154. doi: 10.1016/j.jecp.2016.06.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Hansen, C. H., and Hansen, R. D. (1988). Finding the face in the crowd: an anger superiority effect. J. Pers. Soc. Psychol. 54, 917–924. doi: 10.1037/0022-3514.54.6.917

CrossRef Full Text | Google Scholar

Henrich, J., Heine, S. J., and Norenzayan, A. (2010). The weirdest people in the world. Behav. Brain Sci. 33, 61–83. doi: 10.1017/S0140525X0999152X

CrossRef Full Text | Google Scholar

Isaac, L., Vrijsen, J. N., Rinck, M., Speckens, A., and Becker, E. S. (2014). Shorter gaze duration for happy faces in current but not remitted depression: evidence from eye movements. Psychiatry Res. 218, 79–86. doi: 10.1016/j.psychres.2014.04.002

CrossRef Full Text | Google Scholar

Jack, R. E., Garrod, O. G. B., Yu, H., Caldara, R., and Schyns, P. G. (2012). Facial expressions of emotion are not culturally universal. Proc. Natl. Acad. Sci. U.S.A. 109, 7241–7244. doi: 10.1073/pnas.1200155109

CrossRef Full Text | Google Scholar

Jessen, S., and Grossmann, T. (2014). Unconscious discrimination of social cues from eye whites in infants. Proc. Natl. Acad. Sci. U.S.A. 111, 16208–16213. doi: 10.1073/pnas.1411333111

CrossRef Full Text | Google Scholar

Kobayashi, H., and Kohshima, S. (1997). Unique morphology of the human eye. Nature 387, 767–768. doi: 10.1006/jhev.2001.0468

CrossRef Full Text | Google Scholar

Lagattuta, K. H., and Kramer, H. J. (2017). Try to look on the bright side: children and adults can (sometimes) override their tendency to prioritize negative faces. J. Exp. Psychol. Gen. 146, 89–101. doi: 10.1037/xge0000247

CrossRef Full Text | Google Scholar

Leppänen, J. M., and Nelson, C. A. (2012). Early development of fear processing. Curr. Dir. Psychol. Sci. 21, 200–204. doi: 10.1177/0963721411435841

CrossRef Full Text | Google Scholar

Lisk, S., Vaswani, A., Linetzky, M., Bar-Haim, Y., and Lau, J. Y. (2019). Systematic review and meta-analysis: eye-tracking of attention to threat in child and adolescent anxiety. J. Am. Acad. Child Adolesc. Psychiatry 59, 88–99.e1. doi: 10.1016/j.jaac.2019.06.006

CrossRef Full Text | Google Scholar

Marsh, A. A., Ambady, N., and Kleck, R. E. (2005). The effects of fear and anger facial expressions on approach- and avoidance-related behaviors. Emotion 5, 119–124. doi: 10.1037/1528-3542.5.1.119

CrossRef Full Text | Google Scholar

Mogg, K., Garner, M., and Bradley, B. P. (2007). Anxiety and orienting of gaze to angry and fearful faces. Biol. Psychol. 76, 163–169. doi: 10.1016/j.biopsycho.2007.07.005

CrossRef Full Text | Google Scholar

Öhman, A. (2008). “Fear and anxiety,” in Handbook of Emotions, eds M. Lewis, J. M. Haviland-Jones, and L. Feldman Barrett (New York, NY: Guilford Press), 709–729.

Google Scholar

Öhman, A., Lundqvist, D., and Esteves, F. (2001). The face in the crowd revisited: a threat advantage with schematic stimuli. J. Pers. Soc. Psychol. 80, 381–396. doi: 10.1037//0022-3514.80.3.381

CrossRef Full Text | Google Scholar

Öhman, A., and Mineka, S. (2001). Fears, phobias, and preparedness: toward an evolved module of fear and fear learning. Psychol. Rev. 108, 483–522. doi: 10.1037/0033-295X.108.3.483

CrossRef Full Text | Google Scholar

Pool, E., Brosch, T., Delplanque, S., and Sander, D. (2016). Attentional bias for positive emotional stimuli: a meta-analytic investigation. Psychol. Bull. 142, 79–106. doi: 10.1037/bul0000026

CrossRef Full Text | Google Scholar

R Core Team (2013). R: A Language and Environment for Statistical Computing. Vienna: R Foundation for Statistical Computing.

Google Scholar

Rossignol, M., Campanella, S., Bissot, C., and Philippot, P. (2013). Fear of negative evaluation and attentional bias for facial expressions: an event-related study. Brain Cogn. 82, 344–352. doi: 10.1016/j.bandc.2013.05.008

CrossRef Full Text | Google Scholar

Rozin, P., and Royzman, E. B. (2001). Negativity bias, negativity dominance, and contagion. Pers. Soc. Psychol. Rev. 5, 296–320. doi: 10.1207/S15327957PSPR0504_2

CrossRef Full Text | Google Scholar

Sauter, D. (2010). More than happy: the need for disentangling positive emotions. Curr. Dir. Psychol. Sci. 19, 36–40. doi: 10.1177/0963721409359290

CrossRef Full Text | Google Scholar

Schmidt, K. L., and Cohn, J. F. (2001). Human facial expressions as adaptations: evolutionary questions in facial expression research. Am. J. Phys. Anthropol. 116, 3–24. doi: 10.1002/ajpa.20001

CrossRef Full Text | Google Scholar

Shechner, T., Jarcho, J. M., Britton, J. C., Leibenluft, E., Pine, D. S., and Nelson, E. E. (2013). Attention bias of anxious youth during extended exposure of emotional face pairs: an eye-tracking study. Depress. Anxiety 30, 14–21. doi: 10.1002/da.21986

CrossRef Full Text | Google Scholar

Thomas, L. A., De Bellis, M. D., Graham, R., and LaBar, K. S. (2007). Development of emotional facial recognition in late childhood and adolescence. Dev. Sci. 10, 547–558. doi: 10.1111/j.1467-7687.2007.00614.x

CrossRef Full Text | Google Scholar

Tottenham, N., Tanaka, J. W., Leon, A. C., McCarry, T., Nurse, M., Hare, T. A., et al. (2009). The NimStim set of facial expressions: judgments from untrained research participants. Psychiatry Res. 168, 242–249. doi: 10.1016/j.psychres.2008.05.006

CrossRef Full Text | Google Scholar

Vaish, A., Grossmann, T., and Woodward, A. (2008). Not all emotions are created equal: the negativity bias in social-emotional development. Psychol. Bull. 134, 383–403. doi: 10.1037/0033-2909.134.3.383

CrossRef Full Text | Google Scholar

van Rooijen, R., Ploeger, A., and Kret, M. E. (2017). The dot-probe task to measure emotional attention: a suitable measure in comparative studies? Psychon. Bull. Rev. 24, 1686–1717. doi: 10.3758/s13423-016-1224-1

CrossRef Full Text | Google Scholar

Whalen, P. J., Kagan, J., Cook, R. G., Davis, F. C., Kim, H., Polis, S., et al. (2004). Human amygdala responsivity to masked fearful eye whites. Science 306, 2061–2061. doi: 10.1126/science.1103617

CrossRef Full Text | Google Scholar

Whalen, P. J., Shin, L. M., McInerney, S. C., Fischer, H., Wright, C. I., and Rauch, S. L. (2001). A functional MRI study of human amygdala responses to facial expressions of fear versus anger. Emotion 1, 70–83. doi: 10.1037/1528-3542.1.1.70

CrossRef Full Text | Google Scholar

Widlok, T. (1999). Living on Mangetti: ‘Bushman’ Autonomy and Namibian Independence. Oxford: Oxford University Press.

Google Scholar

Williams, M., Moss, S., Bradshaw, J., and Mattingley, J. (2005). Look at me, I’m smiling: visual search for threatening and nonthreatening facial expressions. Vis. Cogn. 12, 29–50. doi: 10.1080/13506280444000193

CrossRef Full Text | Google Scholar

Yiend, J. (2010). The effects of emotion on attention: a review of attentional processing of emotional information. Cogn. Emot. 24, 3–47. doi: 10.1080/02699930903205698

CrossRef Full Text | Google Scholar

Yurgelun-Todd, D. A., and Killgore, W. D. (2006). Fear-related activity in the prefrontal cortex increases with age during adolescence: a preliminary fMRI study. Neurosci. Lett. 406, 194–199. doi: 10.1016/j.neulet.2006.07.046

CrossRef Full Text | Google Scholar

Keywords: attentional bias, fear bias, emotions, facial expressions, cross-cultural comparison, ≠Akhoe Hai||om, Germans, adolescents

Citation: Mühlenbeck C, Pritsch C, Wartenburger I, Telkemeyer S and Liebal K (2020) Attentional Bias to Facial Expressions of Different Emotions – A Cross-Cultural Comparison of ≠Akhoe Hai||om and German Children and Adolescents. Front. Psychol. 11:795. doi: 10.3389/fpsyg.2020.00795

Received: 19 December 2019; Accepted: 31 March 2020;
Published: 28 April 2020.

Edited by:

Katja Koelkebeck, LVR Hospital Essen, Germany

Reviewed by:

Mariska Esther Kret, Leiden University, Netherlands
Thomas Suslow, Leipzig University, Germany

Copyright © 2020 Mühlenbeck, Pritsch, Wartenburger, Telkemeyer and Liebal. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Katja Liebal, a2F0amEubGllYmFsQGZ1LWJlcmxpbi5kZQ==

These authors share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.