- 1School of Fundamental Science and Engineering, Waseda University, Tokyo, Japan
- 2Japan Society for the Promotion of Science, Tokyo, Japan
- 3Department of Management, The University of Tokyo, Tokyo, Japan
- 4Institute for Development and Ageing and Cancer, Tohoku University, Sendai, Japan
Animacy perception—discriminating between animate and inanimate visual stimuli—is the basis for engaging in social cognition and for our survival (e.g., avoiding potential danger). Previous studies indicate that factors in a target, such as the features or motion of a target, enhance animacy perception. However, factors in a perceiver, such as the visual attention of a perceiver to a target, have received little attention from researchers. Research on judgment, decision-making, and neuroeconomics indicates the active role of visual attention in constructing decisions. This study examined the role of visual attention in the perception of animacy by manipulating the exposure time of targets. Among Studies 1a to 1c conducted in this study, participants saw two face illustrations alternately; one of the faces was shown to be longer than the other. The participants chose the face that they considered more animated and rounder. Consequently, longer exposure time toward targets facilitated animacy perception and preference rather than the perception of roundness. Furthermore, preregistered Study 2 examined the underlying mechanisms. The results suggest that mere exposure, rather than orienting behavior, might play a vital role in the perception of animacy. Thus, in the reverse relationship between visual attention and animacy perception, animate objects capture attention—attention results in the perception of animacy.
1. Introduction
Animacy perception, which distinguishes animate from inanimate visual stimuli (Rutherford and Kuhlmeier, 2013), is a necessary component of social interaction. Evidence shows that such perceptions emerge even in infancy (Leslie, 1982; Gergely et al., 1995; Rochat et al., 1997) but are disrupted by developmental disorders (Rutherford et al., 2006) and amygdala damage (Heberlein and Adolphs, 2004).
Previous research on factors driving the perception of animacy mainly focused on the properties of target stimuli, such as human-like appearances (e.g., a face) and motion (e.g., interactive motion between two geometric shapes). For instance, people consider an object comparatively more animate when the object has unique human features such as eyes and mouth (Looser and Wheatley, 2010), intelligence (Bartneck et al., 2009), and facial expressions of happiness (Bowling and Banissy, 2017; Krumhuber et al., 2019; Saito et al., 2022). Furthermore, individuals consider moving objects more animate when the motion seems to have specific goals, such as chasing and helping (Heider and Simmel, 1944; Rochat et al., 1997; Castelli et al., 2000; Scholl and Tremoulet, 2000; Tremoulet and Feldman, 2000; Kuhlmeier et al., 2003).
Although most previous research has shown that factors in a target (e.g., human-like features of targets) play crucial roles in animacy perception, these factors do not necessarily facilitate it. According to the uncanny valley theory (Mori et al., 2012), inanimate objects (e.g., robots) resemblance to humans increases the perception of animacy. However, when the resemblance reaches a certain point, it provokes uncanny or strange feelings and hinders the perception. The uncanny valley theory suggests the importance of focusing not only on factors in a target but also on factors in a perceiver. However, minimal extant literature focuses on the factors (e.g., knowledge and mental state of participants) in animacy perception. For example, beliefs about the origin of moving objects (i.e., humans or robots; Cross et al., 2016) and the state of participants (i.e., loneliness) affect animacy perception (Powers et al., 2014). Thus, it is also essential to focus on factors affecting animacy perception in perceivers.
Notably, attention, which can be counted as a factor in perceivers, has a critical relationship with animacy perception and might be a causal effect in animacy perception. Previous studies show that animate objects capture attention (Pratt et al., 2010; Yang et al., 2012; Jackson and Calvillo, 2013; Calvillo and Jackson, 2014). For example, when individuals are tasked with finding a category exemplar and are unexpectedly exposed to either an animate or inanimate object, they are more likely to notice the animate object (Calvillo and Jackson, 2014). Thus, the authors concluded that these findings reflect that detecting animate objects is vital in ancestral hunter-gatherer environments and is consistent with the animate-monitoring hypothesis (New et al., 2007). As mentioned above, animate objects attract attention. However, is there also a reversal relationship between them? Specifically, does attracting attention lead to animacy perception? This prediction could be the case considering recent research on judgment and decision-making.
The growing body of research on judgment, decision-making, and neuroeconomics highlights the crucial role of visual attention in decision-making (Armel et al., 2008; Krajbich et al., 2010; Glöckner and Herbold, 2011; Krajbich and Rangel, 2011; Krajbich et al., 2012; Orquin and Loose, 2013; Cavanagh et al., 2014; Thomas et al., 2019). In particular, the attentional drift-diffusion model (aDDM), proposed by Krajbich and his peers, incorporates the role of visual attention in traditional decision-making models (i.e., the drift-diffusion model; Krajbich et al., 2010; Krajbich and Rangel, 2011; Krajbich et al., 2012; Krajbich, 2019). The aDDM is a decision-making model assuming that the evidence of an item for reaching a decision is amplified when the item receives more attention. Notably, assuming that visual attention modulates the accumulation of evidence to reach a threshold to decide, decision times and choices can accurately be predicted (Krajbich et al., 2010). Previous neural studies present supportive evidence that there was neural activity related to fixation-dependent value coding but did not examine the validity of aDDM (Lim et al., 2011; McGinty et al., 2016). Furthermore, numerous behavioral studies have shown that a longer gaze duration toward one option results in a higher choice probability for that option (Shimojo et al., 2003; Armel et al., 2008; Krajbich et al., 2010, 2012; Saito et al., 2017, 2020; Thomas et al., 2019; Motoki et al., 2021). Moreover, behavioral studies indicate the causal role of attention in decision-making by manipulating the gaze toward options and that the probabilities of choices have changed (Shimojo et al., 2003; Armel et al., 2008; Pärnamets et al., 2015). As mentioned above, visual attention plays a crucial role in decision-making. By employing this perspective, we sought to specify factors in a perceiver affecting animacy perception in the current study.
Gaze manipulation does not always bias decision-making (Shimojo et al., 2003). Though, it is reported that there is a consistent effect on simple perceptual choice (Tavares et al., 2017). According to Shimojo et al. (2003), gaze manipulation can influence subjective (e.g., preference) rather than objective judgments. In other words, gaze manipulation is likely to influence higher-level cognition (e.g., preference) rather than low-level perception (e.g., morphological perception). Given that preference for targets contributes to an uncanny valley feeling (Wang and Rochat, 2017), animacy perception may be influenced by gaze manipulation through a preference for targets. This study directly tests this hypothesis where gaze manipulation influences the animacy perception.
Exposure duration (i.e., mere exposure effect; Zajonc, 1968) and gaze shifting (i.e., gaze orienting) may be potential mechanisms influencing the role of visual attention in animacy perception. The account of exposure duration is based on the mere exposure effect (Zajonc, 1968). Specifically, the more people look at a stimulus, the more they like it. It has also been assumed that gaze orienting is a precursor to higher-level cognition (e.g., preferences; Shimojo et al., 2003; Simion and Shimojo, 2006; Simion and Shimojo, 2007). More prolonged exposure durations with orientation (i.e., gaze shifting) can induce a preference shift. In contrast, longer exposure durations without orientation do not result in a preference shift (Shimojo et al., 2003). We further elucidated the potential underlying mechanisms by manipulating exposure duration and gaze orientation.
This study examines whether stimuli exposure time influences the perception of animacy. In particular, we investigated whether the manipulation influenced high-level perceptions, animacy perception (Study 1a), and preference (Study 1b). Additionally, we examined the effect on roundness judgment (Study 1c), which we considered low-level perception. We expected high-level cognition (i.e., animacy and preference), rather than low-level perception, would be biased by gaze manipulation. Furthermore, Study 2 examined the underlying mechanisms by separating the factors of gaze manipulation into exposure duration and arbitrary eye movements. As we mentioned above, it is reported that arbitrary eye movement is necessary for biasing high-level cognition (Shimojo et al., 2003). However, contradicting findings reported that extended exposure duration, regardless of gaze orientation, biased decision-making (e.g., Nittono and Wada, 2009; Bird et al., 2012). Thus, we examined which factor of gaze manipulation, exposure duration, or arbitrary eye movements influence animacy perception in Study 2.
2. Study 1a to 1c
Study 1a examined the effects of gaze manipulation on animacy perception. The participants viewed two facial images with artificial features and then chose the image perceived as more animate. While viewing the images, participants’ eye movements were manipulated using the paradigm of a previous study (Shimojo et al., 2003). Study 1b was designed to replicate the effect of gaze manipulation on preference judgment (Shimojo et al., 2003) in the current experimental procedure. The procedure was similar, except that it made participants choose their preferred facial images. Study 1c was designed to confirm the specificity of gaze manipulation for both preference and animacy perception. The procedure was almost the same, except that it required participants to choose a rounder facial image.
2.1. Methods
2.1.1. Participants
To the best of our knowledge, because no prior study has examined the effect of exposure time on animacy perception, we did not formally calculate the sample size for Studies 1a to 1c. We recruited university students who participated in each study during the 1st wave of the recruitment period. Finally, 43 participants for Study 1a (11 women, 32 men; mean age, 20.78; SD of age, 1.38), 61 participants for Study 1b (20 women, 41 men; mean age, 21.13; SD of age, 2.75), and 29 participants for Study 1c (12 women, 17 men; mean age, 21.41; SD of age, 1.37) were selected. We considered those sample sizes (i.e., 29–61) almost sufficient to detect the effect given previous studies’ sample sizes ranged from 10 to 100 (Shimojo et al., 2003; Armel et al., 2008). The participants were all university students recruited via a university bulletin board and mailing list. After completing the study, participants received a small monetary compensation for their participation. This study was approved by the ethics committees of Tohoku University (Number: UMIN000025712) and Waseda University [Number 2019-357(1)] and conducted per the Declaration of Helsinki. For each study, the participants gave their free and informed consent.
2.1.2. Stimuli
In this study, we used 40 pairs of facial images (20 female and 20 male face pairs). To create these images, we selected 45 male and 45 female faces from the Chicago Face Database (Ma et al., 2015). All facial images displayed no emotional expression (i.e., neutral expression). The images depicted real human faces and might cause a ceiling effect on animacy perception that prevented the effects of gaze manipulation. Therefore, we modified these images to add artificial features using non-photorealistic rendering methods (Rosin and Lai, 2015). This method produces realistic cartoons from real images of the same identity (Figure 1). The images were resized to a uniform width of 600 pixels and height of 450 pixels.
Figure 1. Examples of modified face images used in the studies. Adapted with permission from Chicago Face Database, available at https://www.chicagofaces.org/ (Ma et al., 2015).
We further conducted an online pre-experiment to manipulate the attractiveness of the images using Qualtrics. We recruited 40 participants via Lancers1 and asked them to rate the attractiveness of the images on a 7-point Likert scale (1 = very unattractive to 7 = very attractive). Based on these ratings, we created 40 pairs of facial images. Stimulus codes for the exact stimuli employed are available in the online supplemental material.2 The average ratings of the faces in a pair were matched such that the difference in the average rating in each pair was <0.10 points. The average rating for all faces was 3.12 (SD = 0.49). The faces in a pair were also matched in terms of sex. There was an equal number of face pairs in each sex (20 male and 20 female face pairs).
2.1.3. Procedure
We used similar experimental procedures and conducted the experiments almost concurrently. Based on a previous study (Shimojo et al., 2003), we manipulated stimuli exposure time to participants while perceiving a pair of faces (Figure 2A).
Figure 2. Example of task flow in Studies 1a to 1c and Study 2. Note. In Studies 1b, 1c, and 2, the instruction for decision-making was changed depending on the conditions [which did you prefer?] for Study 1b; and which did you perceive as rounder?; for Studies 1c and 2. (A) the arbitrary eye movement condition in Studies 1a-1c, 2 and (B) the fixed eye movement condition in Study 2. Adapted with permission from Chicago Face Database, available at https://www.chicagofaces.org/ (Ma et al., 2015).
Participants completed the experiment individually on a computer (display resolution 14-inch, 1920 × 1080). The distance between the participant’s eyes and the display was approximately 60 cm. After showing the fixation cross for 500 ms, we presented each face six times to the participants. Faces alternated between the left and right halves of the screen. Therefore, participants had to shift their gaze toward the visible face on the screen. The presentation duration for each face in a pair was different, 900 ms for one face and 300 ms for another face. At one trial, one face was shown for 5400 ms (900 ms × 6 times) and another for 1800 ms (300 ms × 6 times). These durations were identical to those of the previous study (Shimojo et al., 2003). Faces that were shown longer than other faces were counterbalanced across the participants. After viewing a pair of faces, participants chose a face in which they perceived animacy more (Study 1a), preferred more (Study 1b), or perceived rounder (Study 1c) by pressing the corresponding keys. For instance, the “f” key for the left-sided face and the “j” key for the right-sided face. The reaction time was not constrained, and the order of face pair presentation was randomized across trials. The total number of trials was set to 40. Before the experiment, we explained the procedure to participants and confirmed their understanding of the instruction by asking them.
2.1.4. Statistical analysis
Through Studies 1a to 1c, we used mixed logistic models to predict the choice of the target presented on the left side (1: left-sided target, 0: right-sided target), with the left-sided target shown for a long or short duration (1: shown longer, 0: shown shorter) as a fixed effect, and participants and pairs of stimuli included as a random slope and a random intercept. All analyses were conducted using the lme4 package (Bates et al., 2014) in the R software (R Core Team, 2021). Regarding the analysis in Study 1c, we excluded 12 trials in which the stimuli were not presented for the intended duration owing to technical issues. In conclusion, we analyzed the data of 28 trials from each participant in Study 1c. The data analyzed in this study were made available at the Open Science Framework.3
2.2. Results
2.2.1. Study 1a (Animacy judgment)
The results of the analysis showed that participants tended to choose longer-shown faces as more animated faces in Study 1a (53.89, 95% CI [51.89–55.89]; b = 0.34, z = 3.19, p < 0.001). This result suggests that gaze bias influenced perceptions of animacy. The likelihoods of longer-shown stimuli chosen through Studies 1a to 1c are visualized in Figure 3. Table 1 further shows the details of the results.
Figure 3. The box plots of the likelihood of the longer-shown stimuli chosen by participants across Studies 1a to 1c. The dashed line represents the chance level (50%).
2.2.2. Study 1b (preference judgment)
The results of the analysis revealed that participants preferred the faces that were shown longer (57.03, 95% CI [54.65–59.39]; b = 0.53, z = 4.21, p < 0.001). This indicates that we successfully replicated the effect of eye movement on preference judgment (Shimojo et al., 2003; Krajbich et al., 2010; Saito et al., 2017).
2.2.3. Study 1c (roundness judgment)
The results of the analysis indicated that participants did not tend to choose longer-shown faces as rounder faces in Study 1c (49.88, 95% CI [46.38–53.37]; b = −0.02, z = −0.08, p = 0.98). This result suggests that gaze bias specifically influenced both preferences and animacy perception rather than morphological perception (i.e., roundness judgment).
2.3. Discussion
Through Studies 1a to 1c, we observed that gaze manipulation influences animacy and preference judgments, not roundness judgments. These findings suggest the specificity of the effect of gaze manipulation on animacy and preference perceptions and that these perceptions might be affected by gaze manipulation through the exact mechanism. However, regarding the mechanism, it is unclear what aspect of gaze manipulation we used affected the perceptions because we manipulated the presentation duration (i.e., mere exposure) and arbitrary eye movements (i.e., orienting behavior). Furthermore, we did not directly compare the effects of gaze manipulation on animacy and roundness judgment. Therefore, Study 2 was designed to address these questions.
3. Study 2
In Study 2, we sought to solve the issues mentioned above by directly comparing the effects of (1) present duration (i.e., mere exposure) and arbitrary eye movements (i.e., orienting behavior) and (2) animacy and roundness judgment. To specify the factors of gaze manipulation on animacy perception, we used a paradigm in which participants’ eye movements were fixed (Shimojo et al., 2003).
3.1. Method
3.1.1. Experimental design
This study included two independent variables: the type of judgment (two levels: animacy and roundness) and gaze manipulation (two levels: arbitrary eye movement and fixed eye movement). These variables were between-participant factors. The dependent variable was the choice of stimulus.
3.1.2. Participants and stimuli
We conducted a simulation-based power analysis using the SIMR package (Green and MacLeod, 2016) in R and the data from Study 1a to estimate the ideal sample size. This analysis determined the expected power to secure the fixed effect of gaze manipulation for various sample sizes. The results indicated the need for a sample size of 169 to achieve over 80% at an alpha level of 0.05. Considering that the average dropout rate of a typical web experiment is approximately 30% (Musch and Reips, 2000; Zhou and Fishbach, 2016), we determined the sample size to be 220 participants (55 participants for each condition). We recruited participants using Lancers. A total of 221 participants were recruited for the study. After excluding participants who failed the attention check, data from 205 participants (63 women, 136 men, 6 preferred not to disclose; mean age, 41.98; SD of age, 8.66) were analyzed. The participants received a small monetary compensation for their participation. This study was approved by AsPredicted.org.5 Further, we used the same stimuli as in Studies 1a to 1c.
3.1.3. Procedure
The procedure was almost identical to those of the previous studies, except as noted in the following text. We conducted the study online through Qualtrics6 because it was challenging to experiment in person due to the coronavirus disease 2019 pandemic at that time (October 18–25, 2021). At the beginning of each study, participants answered a question designed to check whether they read instructions as an attention check (Oppenheimer et al., 2009). In particular, participants had to ignore the standard response format and instead provide a confirmation that they had read the instruction in the question. Then, participants were randomly assigned to one of four conditions (two types of judgment: animacy, roundness × 2, gaze manipulation: arbitrary eye movement, and fixed eye movement). In Study 2, we established a fixed eye movement condition, where the stimuli face alternated at the center of the screen (Figure 2B). Thus, the participants did not have to shift their gaze toward the visible face on the screen. After viewing each pair of faces, participants chose the face they perceived as having more animacy (animacy condition) or rounder (roundness condition) by pressing the corresponding keys. For instance, the “f” key for the face presented at last and the “j” key for the face presented before the last. In conclusion, 104 participants were assigned to the animacy judgment condition (53 in arbitrary eye movement, 51 in fixed eye movement), and 101 participants were further assigned to the roundness judgment condition (52 participants in arbitrary eye movement, 49 participants in fixed eye movement).
3.2. Statistical analysis
Study 2 used the preregistered analysis, which was a linear mixed model predicting the choice of one target (arbitrary eye movement condition: 1 = left-sided target, 0 = right-sided target; fixed condition: 1 = the last-presented target, 0 = before the last-presented target), with the target was shown for a long or short duration (presentation duration: 1 = shown longer, 0 = shown shorter), gaze manipulation (1 = arbitrary eye movement, 0 = fixed eye movement), types of judgment (1 = animacy, 0 = morphological perception). Further, the interactions were included as fixed effects, and participants and pairs of stimuli were included as random effects. We used the performance package (Lüdecke et al., 2021) in the R software to investigate the variance inflation factors (VIFs) and overdispersion.
In addition to the preregistered analysis, we conducted similar analyses to the previous three studies as exploratory analyses for analytical consistency across studies. For each condition, we conducted an analysis that was a mixed logistic model predicting the choice of one target (arbitrary eye movement condition: 1 = left-sided target, 0 = right-sided target; fixed condition: 1 = the last-presented target, 0 = before the last presented target), with the target shown for a long or short duration (1 = shown longer, 0 = shown shorter), as a fixed effect. Further, participants and pairs of stimuli were included as a random slope and random intercept, respectively.
3.3. Results
Figure 4 shows the likelihood of choosing longer-shown faces. Regarding the preregistered analysis, we confirmed that multicollinearity was not a problem by inspecting the VIFs (VIFs < 3.91). Further, overdispersion was not a problem in the overdispersion test (χ2 = 7157.48, p = 1.00). The result from the preregistered analysis showed neither significant effects of presentation duration, gaze manipulation, and types of judgment nor those interactions (Table 2).
Figure 4. The box plots of the likelihood of the longer-shown stimuli chosen by participants in Study 2. The dashed line represents the chance level (50%).
Although we did not observe any significant results from the registered analysis, we conducted a mixed logistic model to predict the choice of one target for each condition, as in previous studies (Table 3).
In the arbitrary eye movement condition, we observed that participants tended to choose longer-shown faces when choosing more animate faces (54.86, 95% CI [52.71–56.99]; b = 0.42, z = 2.77, p = 0.01) than when choosing rounder faces (50.48, 95% CI [48.31–52.65]; b = 0.05, z = 1.14, p = 0.26). In the fixed eye movement condition, we also observed that participants tended to choose longer-shown faces when choosing more animate faces (53.43, 95% CI [51.24–55.61]; b = 0.29, z = 2.28, p = 0.02) but not when choosing rounder faces (50.71, 95% CI [48.48–52.95]; b = 0.23, z = 0.61, p = 0.54).
3.4. Discussion
In Study 2, the preregistered analysis showed neither significant effects nor interactions between the experimental conditions. Therefore, we failed to elucidate the factors of gaze manipulation (i.e., mere exposure and orienting behavior) that influence animacy perception. However, subsequent exploratory analyses were consistent with Studies 1a to 1c, showing that gaze manipulation in arbitrary and fixed eye movement conditions influenced only animacy perception rather than the perception of roundness. These results suggest that mere exposure may be critical in facilitating animacy perception.
4. General discussions
Factors in a perceiver have not received sufficient attention regarding the factors that drive animacy perception. We tested whether one of the primary factors, visual attention toward stimuli, affects animacy perception. Across Studies one and two, the participants felt that cartoon faces were more animated when manipulating their gaze to look at the faces longer. This effect of biased exposure duration was also observed in preference judgments (Study 1b) rather than in lower-level perception (i.e., roundness judgments, Studies 1c and 2). Furthermore, in the preregistered online study (Study 2), it was found that arbitrary eye movements were not necessarily needed to increase animacy perception. However, exposure duration played a crucial role in influencing it.
Our results provide evidence that gazing behavior influences the perception of animacy. In this study, manipulating the exposure duration in arbitrary and fixed eye movement conditions facilitated animacy perception. This finding is inconsistent with the claim that gaze orienting is necessary to bias higher-level cognition, such as preference judgment (Shimojo et al., 2003). Instead, this finding is consistent with studies that show that gaze orienting is not a necessary condition for forming higher-level cognition but instead demonstrates that a mere exposure effect underlies biased higher-level cognition by gaze manipulation (Glaholt and Reingold, 2009; Nittono and Wada, 2009; Glaholt and Reingold, 2011; Bird et al., 2012).
There are several potential explanations for why the mere exposure effect derives animacy perceptions. First, along with aDDM (Krajbich et al., 2010), attention would have facilitated evidence of animacy and preference. Second, mere exposure may have changed several psychological constructs, as mere exposure increases familiarity and saliency (Montoya et al., 2017). Familiarity seemed to be a crucial construct in this study, given the uncanny valley theory, where unfamiliarity or strangeness hinders the perception of animacy (Mori et al., 2012). Moreover, an empirical study indicated that people attribute fundamental capacities of the mind, which is a concept strongly related to animacy, to preferred targets (Kozak et al., 2006). Examining the relationship between the mere exposure effect and animacy will likely be a pivotal issue for future work.
We have observed that the choice probabilities of longer-shown stimuli in the animacy condition were greater than chance. Nonetheless, it is important to note that in Study 2’s preregistered analysis, we could not identify the gaze manipulation components that affected animacy perception. Given that the preregistered analysis did not reveal significant effects of judgment types and gaze manipulations on the choice probabilities, we cannot conclude that gaze manipulation uniquely affected animacy perception. Instead, we need to stress that the effect of gaze manipulation on animacy perception might be limited or relatively small. The degree and the uniqueness of the gaze manipulation effect on animacy perception should be further examined in future studies.
Future work would be needed to specify the relationship between visual attention and animacy perception in detail. Firstly, it is necessary to test whether the effect of gaze manipulation occurs for completely inanimate objects (e.g., simple geometrics). The facial stimuli in the present study seemed relatively animate; therefore, it is unclear whether the gaze manipulation effect can trigger animacy perception. Thus, testing whether exposure duration facilitates the perception of animacy, even when the targets are entirely inanimate, would be an interesting direction. It is crucial to inspect the underlying mechanisms biased by gaze manipulation directly influencing animacy perception. Notably, attentional bias results in changes in the target’s characteristics, such as saliency, liking (Mrkva and Van Boven, 2020), and familiarity (Montoya et al., 2017). Therefore, future studies need to examine the psychological mechanisms that mediate the relationship between gaze manipulation and animacy perception.
This study tested whether exposure time plays a role in the perception of animacy. We found evidence that biased exposure time of targets facilitated both animacy perception and preference toward targets rather than lower-level perception (i.e., morphological judgment). The underlying mechanisms biased by gaze manipulation directly influencing animacy perception are not clear. However, our findings suggest that biasing visual attention toward targets facilitates animacy perception, possibly because mere exposure increases familiarity or preference.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://osf.io/cr4yx/.
Ethics statement
The studies involving human participants were reviewed and approved by the ethics committees of Tohoku University. The patients/participants provided their written informed consent to participate in this study.
Author contributions
TS: conceptualization, methodology, investigation, software, formal analysis, writing—original draft preparation. KM: conceptualization, methodology, investigation, writing—reviewing and editing. RN: methodology, writing—reviewing and editing. MS: supervision, writing—review and editing. All authors contributed to the article and approved the submitted version.
Funding
This research was supported by a Grant-in-Aid for JSPS Fellows (21J01224; TS) from the Japan Society for the Promotion of Science and JSPS KAKENHI 19H01760 [Grant-in-Aid for Scientific Research (B)] and JSPS KAKENHI 19H05003 [Grant-in-Aid for Scientific Research on Innovative Areas (Research in a Proposed Research Area)] (RN).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
2. ^https://osf.io/cr4yx/?view_only=633225a44c9f455993688e2c96ea382c
References
Armel, K. C., Beaumel, A., and Rangel, A. (2008). Biasing simple choices by manipulating relative visual attention. Judgm. Decis. Mak. 3, 396–403.
Bartneck, C., Kanda, T., Mubin, O., and Al Mahmud, A. (2009). Does the design of a robot influence its animacy and perceived intelligence? Adv. Robot. 1, 195–204. doi: 10.1007/s12369-009-0013-7
Bates, D., Maechler, M., Bolker, B., and Walker, S. (2014). lme4: linear mixed-effects models using Eigen and S4. R Package Version, 1, 1–23.
Bird, G. D., Lauwereyns, J., and Crawford, M. T. (2012). The role of eye movements in decision making and the prospect of exposure effects. Vis. Res. 60, 16–21. doi: 10.1016/j.visres.2012.02.014
Bowling, N. C., and Banissy, M. J. (2017). Emotion expression modulates perception of animacy from faces. J. Exp. Soc. Psychol. 71, 83–95. doi: 10.1016/j.jesp.2017.02.004
Calvillo, D. P., and Jackson, R. E. (2014). Animacy, perceptual load, and inattentional blindness. Psychon. Bull. Rev. 21, 670–675. doi: 10.3758/s13423-013-0543-8
Castelli, F., Happé, F., Frith, U., and Frith, C. (2000). Movement and mind: a functional imaging study of perception and interpretation of complex intentional movement patterns. NeuroImage 12, 314–325. doi: 10.1006/nimg.2000.0612
Cavanagh, J. F., Wiecki, T. V., Kochar, A., and Frank, M. J. (2014). Eye tracking and pupillometry are indicators of dissociable latent decision processes. J. Exp. Psychol. Gen. 143, 1476–1488. doi: 10.1037/a0035813
Cross, E. S., Ramsey, R., Liepelt, R., Prinz, W., and Hamilton, A. F. D. C. (2016). The shaping of social perception by stimulus and knowledge cues to human animacy. Philos. Trans. R. Soc. B Biol. Sci 371:20150075. doi: 10.1098/rstb.2015.0075
Gergely, G., Nádasdy, Z., Csibra, G., and Bíró, S. (1995). Taking the intentional stance at 12 months of age. Cognition 56, 165–193. doi: 10.1016/0010-0277(95)00661-H
Glaholt, M. G., and Reingold, E. M. (2009). Stimulus exposure and gaze bias: a further test of the gaze cascade model. Atten. Percept. Psychophys. 71, 445–450. doi: 10.3758/APP.71.3.445
Glaholt, M. G., and Reingold, E. M. (2011). Eye movement monitoring as a process tracing methodology in decision making research. J. Neurosci. Psychol. Econ. 4, 125–146. doi: 10.1037/a0020692
Glöckner, A., and Herbold, A. K. (2011). An eye-tracking study on information processing in risky decisions: evidence for compensatory strategies based on automatic processes. J. Behav. Decis. Mak. 24, 71–98. doi: 10.1002/bdm.684
Green, P., and MacLeod, C. J. (2016). SIMR: an R package for power analysis of generalized linear mixed models by simulation. Methods Ecol. Evol. 7, 493–498. doi: 10.1111/2041-210X.12504
Heberlein, A. S., and Adolphs, R. (2004). Impaired spontaneous anthropomorphizing despite intact perception and social knowledge. Proc. Natl. Acad. Sci. U. S. A. 101, 7487–7491. doi: 10.1073/pnas.0308220101
Heider, F., and Simmel, M. (1944). An experimental study of apparent behavior. Am. J. Psychol. 57, 243–259. doi: 10.2307/1416950
Jackson, R. E., and Calvillo, D. P. (2013). Evolutionary relevance facilitates visual information processing. Evol. Psychol. 11, 1011–1026. doi: 10.1177/147470491301100506
Kozak, M. N., Marsh, A. A., and Wegner, D. M. (2006). What do I think you’re doing? Action identification and mind attribution. J. Pers. Soc. Psychol. 90, 543–555. doi: 10.1037/0022-3514.90.4.543
Krajbich, I. (2019). Accounting for attention in sequential sampling models of decision making. Curr. Opin. Psychol. 29, 6–11. doi: 10.1016/j.copsyc.2018.10.008
Krajbich, I., Armel, C., and Rangel, A. (2010). Visual fixations and the computation and comparison of value in simple choice. Nat. Neurosci. 13, 1292–1298. doi: 10.1038/nn.2635
Krajbich, I., Lu, D., Camerer, C., and Rangel, A. (2012). The attentional drift-diffusion model extends to simple purchasing decisions. Front. Psychol. 3:193. doi: 10.3389/fpsyg.2012.00193
Krajbich, I., and Rangel, A. (2011). Multialternative drift-diffusion model predicts the relationship between visual fixations and choice in value-based decisions. Proc. Natl. Acad. Sci. U. S. A. 108, 13852–13857. doi: 10.1073/pnas.1101328108
Krumhuber, E. G., Lai, Y. K., Rosin, P. L., and Hugenberg, K. (2019). When facial expressions do and do not signal minds: the role of face inversion, expression dynamism, and emotion type. Emotion 19, 746–750. doi: 10.1037/emo0000475
Kuhlmeier, V., Wynn, K., and Bloom, P. (2003). Attribution of dispositional states by 12-month-olds. Psychol. Sci. 14, 402–408. doi: 10.1111/1467-9280.01454
Leslie, A. M. (1982). The perception of causality in infants. Perception 11, 173–186. doi: 10.1068/p110173
Lim, S. L., O’Doherty, J. P., and Rangel, A. (2011). The decision value computations in the vmPFC and striatum use a relative value code that is guided by visual attention. J. Neurosci. 31, 13214–13223. doi: 10.1523/JNEUROSCI.1246-11.2011
Looser, C. E., and Wheatley, T. (2010). The tipping point of animacy: how, when, and where we perceive life in a face. Psychol. Sci. 21, 1854–1862. doi: 10.1177/0956797610388044
Lüdecke, D., Ben-Shachar, M., Patil, I., Waggoner, P., and Makowski, D. (2021). Performance: an R package for assessment, comparison and testing of statistical models. J. Open Source Softw. 6:3139. doi: 10.21105/joss.03139
Ma, D. S., Correll, J., and Wittenbrink, B. (2015). The Chicago face database: a free stimulus set of faces and norming data. Behav. Res. Methods 47, 1122–1135. doi: 10.3758/s13428-014-0532-5
McGinty, V. B., Rangel, A., and Newsome, W. T. (2016). Orbitofrontal cortex value signals depend on fixation location during free viewing. Neuron 90, 1299–1311. doi: 10.1016/j.neuron.2016.04.045
Montoya, R. M., Horton, R. S., Vevea, J. L., Citkowicz, M., and Lauber, E. A. (2017). A re-examination of the mere exposure effect: the influence of repeated exposure on recognition, familiarity, and liking. Psychol. Bull. 143, 459–498. doi: 10.1037/bul0000085
Mori, M., MacDorman, K., and Kageki, N. (2012). The uncanny valley [from the field]. IEEE Robot. Autom. Mag. 19, 98–100. doi: 10.1109/mra.2012.2192811
Motoki, K., Saito, T., and Onuma, T. (2021). Eye-tracking research on sensory and consumer science: a review, pitfalls, and future directions. Food Res. Int. 145:110389. doi: 10.1016/j.foodres.2021.110389
Mrkva, K., and Van Boven, L. (2020). Salience theory of mere exposure: relative exposure increases liking, extremity, and emotional intensity. J. Pers. Soc. Psychol. 118, 1118–1145. doi: 10.1037/pspa0000184
Musch, J., and Reips, U. D. (2000). “A brief history of web experimenting” in Psychological Experiments on the Internet. ed. M. H. Birnbaum (Cambridge, MA: Academic Press)
New, J., Cosmides, L., and Tooby, J. (2007). Category-specific attention for animals reflects ancestral priorities, not expertise. Proc. Natl. Acad. Sci. U. S. A. 104, 16598–16603. doi: 10.1073/pnas.0703913104
Nittono, H., and Wada, Y. (2009). Gaze shifts do not affect preference judgments of graphic patterns. Percept. Mot. Skills 109, 79–94. doi: 10.2466/pms.109.1.79-94
Oppenheimer, D. M., Meyvis, T., and Davidenko, N. (2009). Instructional manipulation checks: detecting satisficing to increase statistical power. J. Exp. Soc. Psychol. 45, 867–872.
Orquin, J. L., and Loose, S. M. (2013). Attention and choice: a review on eye movements in decision making. Acta Psychol. 144, 190–206. doi: 10.1016/j.actpsy.2013.06.003
Pärnamets, P., Johansson, P., Hall, L., Balkenius, C., Spivey, M. J., and Richardson, D. C. (2015). Biasing moral decisions by exploiting the dynamics of eye gaze. Proc. Natl. Acad. Sci. U. S. A. 112, 4170–4175. doi: 10.1073/pnas.1415250112
Powers, K. E., Worsham, A. L., Freeman, J. B., Wheatley, T., and Heatherton, T. F. (2014). Social connection modulates perceptions of animacy. Psychol. Sci. 25, 1943–1948. doi: 10.1177/0956797614547706
Pratt, J., Radulescu, P. V., Guo, R. M., and Abrams, R. A. (2010). It’s alive! Animate motion captures visual attention. Psychol. Sci. 21, 1724–1730. doi: 10.1177/0956797610387440
R Core Team. (2021). R: A language and environment for statistical computing. Available at: https://www.R-project.org/ (Accessed March 1, 2021).
Rochat, P., Morgan, R., and Carpenter, M. (1997). Young infants’ sensitivity to movement information specifying social causality. Cogn. Dev. 12, 537–561. doi: 10.1016/S0885-2014(97)90022-8
Rosin, P. L., and Lai, Y. K. (2015). “Non-photorealistic rendering of portraits” in Proceedings of the Workshop on Computational Aesthetics (Goslar: Eurographics Association), 159–170.
Rutherford, M. D., and Kuhlmeier, V. A. (2013). Social perception: Detection and interpretation of Animacy, agency, and intention MIT Press.
Rutherford, M. D., Pennington, B. F., and Rogers, S. J. (2006). The perception of animacy in young children with autism. J. Autism Dev. Disord. 36, 983–992. doi: 10.1007/s10803-006-0136-8
Saito, T., Almaraz, S., and Hugenberg, K. (2022). Happy = human: A feeling of belonging modulates the “expression-to-mind” effect. Soc. Cogn. 40, 213–227. doi: 10.1521/soco.2022.40.3.213
Saito, T., Nouchi, R., Kinjo, H., and Kawashima, R. (2017). Gaze bias in preference judgments by younger and older adults. Front. Aging Neurosci. 9:285. doi: 10.3389/fnagi.2017.00285
Saito, T., Sudo, R., and Takano, Y. (2020). The gaze bias effect in toddlers: preliminary evidence for the developmental study of visual decision-making. Dev. Sci. 23:e12969. doi: 10.1111/desc.12969
Scholl, B. J., and Tremoulet, P. D. (2000). Perceptual causality and animacy. Trends Cogn. Sci. 4, 299–309. doi: 10.1016/S1364-6613(00)01506-0
Shimojo, S., Simion, C., Shimojo, E., and Scheier, C. (2003). Gaze bias both reflects and influences preference. Nat. Neurosci. 6, 1317–1322. doi: 10.1038/nn1150
Simion, C., and Shimojo, S. (2006). Early interactions between orienting, visual sampling and decision making in facial preference. Vis. Res. 46, 3331–3335. doi: 10.1016/j.visres.2006.04.019
Simion, C., and Shimojo, S. (2007). Interrupting the cascade: orienting contributes to decision making even in the absence of visual stimulation. Percept. Psychophys. 69, 591–595. doi: 10.3758/BF03193916
Tavares, G., Perona, P., and Rangel, A. (2017). The attentional drift diffusion model of simple perceptual decision-making. Front. Neurosci. 11:468.
Thomas, A. W., Molter, F., Krajbich, I., Heekeren, H. R., and Mohr, P. N. C. (2019). Gaze bias differences capture individual choice behaviour. Nat. Hum. Behav. 3, 625–635. doi: 10.1038/s41562-019-0584-8
Tremoulet, P. D., and Feldman, J. (2000). Perception of animacy from the motion of a single object. Perception 29, 943–951. doi: 10.1068/p3101
Wang, S., and Rochat, P. (2017). Human perception of animacy in light of the uncanny valley phenomenon. Perception 46, 1386–1411. doi: 10.1177/0301006617722742
Yang, J., Wang, A., Yan, M., Zhu, Z., Chen, C., and Wang, Y. (2012). Distinct processing for pictures of animals and objects: evidence from eye movements. Emotion 12, 540–551. doi: 10.1037/a0026848
Zajonc, R. B. (1968). Attitudinal effects of mere exposure. J. Pers. Soc. Psychol. 9, 1–27. doi: 10.1037/h0025848
Keywords: animacy perception, visual attention, gaze manipulation, mere exposure, orienting behavior
Citation: Saito T, Motoki K, Nouchi R and Sugiura M (2023) Facilitating animacy perception by manipulating stimuli exposure time. Front. Psychol. 13:1017685. doi: 10.3389/fpsyg.2022.1017685
Edited by:
Michael J. Serra, Texas Tech University, United StatesCopyright © 2023 Saito, Motoki, Nouchi and Sugiura. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Toshiki Saito, ✉ t.saito18@kurenai.waseda.jp