Skip to main content

BRIEF RESEARCH REPORT article

Front. Psychol., 14 March 2024
Sec. Personality and Social Psychology
This article is part of the Research Topic The Social Brain: New Insights from Social, Clinical, and Biological Psychology View all 10 articles

Cognitive control in honesty and dishonesty under different conflict scenarios: insights from reaction time

Hao-Ming LiHao-Ming Li1Wen-Jing YanWen-Jing Yan2Yu-Wei Wu
Yu-Wei Wu3*Zi-Ye Huang
Zi-Ye Huang1*
  • 1Wenzhou Seventh People’s Hospital, Wenzhou, China
  • 2School of Mental Health, Wenzhou Medical University, Wenzhou, China
  • 3Student Affairs Division, Wenzhou Business College, Wenzhou, China

This study investigated the role of cognitive control in moral decision-making, focusing on conflicts between financial temptations and the integrity of honesty. We employed a perceptual task by asking participants to identify which side of the diagonal contained more red dots within a square to provoke both honest and dishonest behaviors, tracking their reaction times (RTs). Participants encountered situations with no conflict, ambiguous conflict, and clear conflict. Their behaviors in the clear conflict condition categorized them as either “honest” or “dishonest.” Our findings suggested that, in ambiguous conflict situations, honest individuals had significantly longer RTs and fewer self-interest responses than their dishonest counterparts, suggesting a greater need for cognitive control to resolve conflicts and a lesser tendency toward self-interest. Moreover, a negative correlation was found between participants’ number of self-interest responses and RTs in ambiguous conflict situations (r = −0.27 in study 1 and r = −0.66 in study 2), and a positive correlation with cheating numbers in clear conflict situations (r = 0.36 in study 1 and r = 0.82 in study 2). This suggests less cognitive control was required for self-interest and cheating responses, bolstering the “Will” hypothesis. We also found that a person’s self-interest tendency could predict their dishonest behavior. These insights extend our understanding of the role of cognitive control plays in honesty and dishonesty, with potential applications in education, policy-making, and business ethics.

1 Introduction

Human behavior is often governed by complex decision-making processes, with one recurring challenge being the conflict between self-interest and the pursuit of moral righteousness. This moral quandary, the struggle between the temptation of personal financial gain and the aspiration to uphold an honest image, unfolds in various scenarios ranging from relatively minor instances of tax evasion and inflated expense reports, to more severe instances of fraudulent financial schemes (Mazar et al., 2008). Such moral dilemmas offer a fascinating window into human behavior and motivations. They invite questions regarding how individuals reconcile these seemingly incompatible drives of personal gain and moral obligation. An increasingly explored proposition within the behavioral sciences is that cognitive control, our inherent ability to regulate thoughts, emotions, and actions, acts as a mediator in this tension between self-interest and moral self-image (Wood et al., 2008).

Despite its intuitive appeal, the role of cognitive control in moral decision-making, particularly its contribution to resolving conflicts between self-interest and honesty, remains a contentious topic within psychological and neuroscientific research. Although a large amount of data is available, the results are mixed (Köbis et al., 2019; Capraro, 2023). This debate predominantly centers around two main hypotheses: the “Will” hypothesis and the “Grace” hypothesis (Tabatabaeian et al., 2015). The “Will” hypothesis paints a less flattering image of human nature. It posits that humans, by default, are selfish and dishonest, and that it is cognitive control that keeps these basic instincts in check, compelling individuals toward honesty (Gino et al., 2011). This hypothesis aligns with traditional economic models of human behavior, which suggest that individuals are naturally driven to pursue self-interest, with social norms, laws, and moral values acting as external constraints on these inborn desires (Becker, 1968; Henrich et al., 2005). Contrastingly, the “Grace” hypothesis presents a more favorable image of humans, suggesting that people are essentially honest, and that cognitive control is used to suppress instinctual honest responses when there are opportunities to profit from dishonesty (Rand et al., 2012). This view is supported by empirical research that shows individuals respond faster when instructed to tell the truth than when directed to lie, suggesting honesty may indeed be more intuitive (Capraro, 2017; Suchotzki et al., 2017; Verschuere et al., 2018; Capraro et al., 2019).

This ongoing debate is far from a mere academic exercise. Instead, it underscores the complex, multifaceted nature of human morality, highlighting the need for more nuanced and empirical investigations into the interplay between cognitive control and moral behavior (Baumeister et al., 2009). As Baumeister and Exline (1999) propose, understanding these moral dynamics requires acknowledging individual differences, considering situational variables, and appreciating the dynamic nature of moral decision-making processes. Study found that the social consequences of lying could be a promising key to the riddle of intuition’s role in honesty. When dishonesty harms abstract others, promoting intuition causes more people to lie and people to lie more. However, when dishonesty harms concrete others, promoting intuition has no significant effect on dishonesty (Köbis et al., 2019). Recent research advancements have further complicated this landscape. With the advent of neuroimaging techniques, studies suggest that the impact of cognitive control on moral behavior may be dependent on an individual’s inherent moral disposition toward honesty or dishonesty (Greene and Paxton, 2009). Specifically, brain regions associated with cognitive control, such as the anterior cingulate cortex and the inferior frontal gyrus, have been found to help individuals predisposed toward dishonesty to act honestly, while enabling those predisposed toward honesty to cheat when the situation permit (Speer et al., 2022).

Against this backdrop, the present study embarks on an exploration of how individuals, predisposed toward honesty or dishonesty, respond to situations that present a conflict between personal financial gain and moral self-image. Beyond neuroimaging, reaction time (RT) measures, often utilized in cognitive psychology, are believed to offer critical insights into cognitive control’s involvement in moral decision-making conflicts (Evans et al., 2015; Andrighetto et al., 2020). RTs provide non-invasive, real-time evidence of the cognitive processes at play during moral decision-making (Shalvi et al., 2012). This study introduces three distinct decision conflict scenarios, allowing for a more nuanced examination of individual differences in cognitive control and moral tendencies. By analyzing the interaction between cognitive control, moral inclination, and response times across these scenarios, we hope to provide a more comprehensive, more dynamic, and ultimately, a more human perspective on the landscape of moral decision-making (Tangney et al., 2007).

2 Study 1

2.1 Method

2.1.1 Participants

We recruited sixty-seven undergraduate or postgraduate students from Wenzhou University. The participants, with an average age of 19.39 (SD = 1.18), comprised 39 females and 28 males. All participants had normal or corrected-to-normal vision and provided informed consent before the study. The study adhered to the sixth revision of the Declaration of Helsinki (2008) and was approved by the university’s Institutional Review Board (IRB).

2.1.2 Procedure

We engaged the participants in a perceptual task (Gino et al., 2010). Each trial presented a square divided diagonally into two sections. Each section had 20 dots scattered randomly on the left or right side of the diagonal. After a one-second exposure, participants identified which side of the diagonal held more dots by clicking the respective mouse button. The reward for each trial was calculated as follows: clicking the left mouse button yielded 0.02CNY, whereas clicking the right button yielded 0.2CNY. Therefore, trials with more dots on the left side of the diagonal presented a conflict between answering accurately and maximizing profit.

The perceptual task was split into two phases. The first phase consisted of 100 practice trials, after which participants received feedback on potential earnings for each trial and cumulative earnings if these trials had involved real payment. In the second phase, the participants completed 200 trials, earning real money, and received information about their earnings for each trial and overall.

Participants could earn a maximum of 40CNY on this perceptual task (by always pressing the right mouse button). There were four blocks. Each block consisted of 50 trials, and each block included 8 trials in which the answer was clearly “more on right” (no conflict condition, i.e., the ratio of the number of dots on the right to the number of dots on the left was greater than or equal to 1.5), 17 trials in which the answer was clearly “more on left” (clear conflict condition, i.e., the ratio of the number of dots on the right to the number of dots on the left was less than or equal to 2/3), and 25 ambiguous trials (ambiguous conflict condition, i.e., the ratio of the number of dots on the right to the number of dots on the left was between 2/3 and 1.5). The responses in ambiguous condition reflect an individual’s self-interest tendency. Once participants completed this task, the computer indicated that they should report their performance in Phase 2 on a collection slip to be handed to the experimenter at the end of the study.

2.2 Results

All participants displayed honest behavior in no conflict condition. A total of 53.73% (36/67) participants were found to cheat one or more times in clear conflict condition (see Figure 1). The participants who cheated in clear conflict condition (36 participants, mean cheating number = 23.63) will be referred to as ‘dishonest individuals’, while the remaining participants (31 participants) will be referred to as ‘honest individuals’.

Figure 1
www.frontiersin.org

Figure 1. Examples of decision conflicts.

We compared the reaction times (RTs) in the no conflict, ambiguous conflict and clear conflict conditions among honest and dishonest participants. Results showed that honest individuals required longer RTs than dishonest individuals in the ambiguous conflict condition, p = 0.047, suggesting that honest individuals required more time to revolve ambiguous conflict. Also, honest individuals made less self-interest responses (M = 42, SD = 5.08) than dishonest individuals (M = 67.42, SD = 22.34) in the ambiguous condition, p < 0.001. There were no RT differences in no conflict and clear conflict conditions among honest and dishonest participants, p = 0.07; p = 0.09.

Moreover, the RTs in ambiguous trials correlated with the self-interest numbers in the ambiguous condition, r = −0.27, p = 0.028 and the cheating numbers in the clear conflict condition, r = −0.24, p = 0.046. The self-interest numbers in the ambiguous condition correlated with the cheating numbers in the clear conflict condition, r = 0.36, p = 0.003. When using RTs and self-interest numbers in the ambiguous condition to predict the cheating numbers in the clear conflict condition, the model was significant, with R2 = 0.15, p = 0.005. The self-interest number in the ambiguous condition was a significant indicator of cheating numbers in the clear conflict condition, p = 0.01; whereas the RTs in the ambiguous conditions was not significant in the model, p = 0.19.

3 Study 2

3.1 Method

3.1.1 Participant

We recruited ninety-five undergraduate or postgraduate students from Hebei Normal University. The participants, with an average age of 19.55 (SD = 1.07), comprised 75 females and 20 males. All participants had normal or corrected-to-normal vision and provided informed consent before the study. The study adhered to the sixth revision of the Declaration of Helsinki (2008) and was approved by the university’s Institutional Review Board (IRB).

3.1.2 Procedure

The task is same as that of Study 1, only some differences in experimental materials. In the experiment, 18 images were made in the order of the left and right red dots from less to most. The experiment consisted of 4 blocks, and 18 images in each block were randomly presented 4 times, for a total of 72 trials. The experiment consisted of a total of 288 trials.

3.2 Results

All participants displayed honest behavior in no conflict condition. A total of 54.74% (52/95) participants were found to cheat one or more times in clear conflict condition. The participants who cheated in clear conflict condition (52 participants, mean cheating number = 29.98) will be referred to as ‘dishonest individuals’, while the remaining participants (43 participants) will be referred to as ‘honest individuals’.

We compared the reaction times (RTs) in the no conflict, ambiguous conflict and clear conflict conditions among honest and dishonest participants. Results showed that honest individuals required longer RTs (M = 665.25, SD = 143.04) than dishonest individuals in (M = 550.64, SD = 166.12) the ambiguous conflict condition, p = 0.001, suggesting that honest individuals required more time to revolve ambiguous conflict. Also, honest individuals made less self-interest responses (M = 3.50, SD = 2.48) than dishonest individuals (M = 9.69, SD = 4.78) in the ambiguous condition, p < 0.001. Moreover, the RTs in ambiguous trials correlated with the self-interest numbers in the ambiguous condition, r = −0.66, p < 0.001 and the cheating numbers in the clear conflict condition, r = −0.65, p < 0.001. The self-interest numbers in the ambiguous condition correlated with the cheating numbers in the clear conflict condition, r = 0.82, p < 0.001. When using RTs and self-interest numbers in the ambiguous condition to predict the cheating numbers in the clear conflict condition, the model was significant, with R2 = 0.73, p < 0.001. The self-interest number and the RTs in the ambiguous conditions were significant indicators of cheating numbers in the clear conflict condition, p < 0.001; p = 0.003.

We also investigated the effect of conflict degree on the RTs of honest and dishonest people. Subtract the non-conflicting RTs from the conflicting RTs corresponding to the left and right red dots (i.e., the RTs under the condition that the left red dot is 13 minus the RTs under the condition that the right red dot is 7; The RTs under the condition that the left red dot is 14 minus the RTs under the condition that the red dot on the right is 6; and so on). We believe that the smaller difference between the numbers of red dots on the left and right, the greater psychological conflict of the individual. The results showed that conflict degree affected the participants’ responses, the greater the conflict, the longer RTs required, F(6, 498) = 47.67, p < 0.001. There was no difference between the honest and dishonest people in their RTs at different conflict levels, p = 0.46.

4 Discussion

Our study investigated the interplay between cognitive control and moral decision-making, particularly focusing on how individuals with different predispositions toward honesty or dishonesty react in situations where personal financial gain conflicts with moral self-image. The key finding is that individuals who are inherently more honest exhibited longer reaction times in scenarios with ambiguous moral conflicts, suggesting a deeper cognitive engagement in these dilemmas. Conversely, those predisposed to dishonesty responded more quickly, implying less cognitive deliberation. This differentiation highlights the complex role of cognitive control in navigating moral decisions, indicating that it is influenced by an individual’s moral inclinations. Essentially, our results contribute to understanding the nuanced mechanisms behind moral behavior, showing that moral decision-making is a dynamic process shaped by both cognitive control and personal ethical standards.

Our observation that honest individuals exhibit longer reaction times in ambiguous conflict conditions than their dishonest counterparts offers an intriguing insight into the cognitive processes underlying moral behavior. This finding aligns with the work of Capraro and Rand (2018), who suggested that honesty might be more intuitive to individuals with a stronger predisposition toward prosocial behavior, requiring less cognitive control in clear-cut situations but more deliberation when the context is ambiguous. Our results extend this theory by quantitatively showing that the cognitive effort, as measured by reaction times, increases in moral dilemmas where the right choice is not immediately apparent.

Additionally, the correlation between reaction times and self-interest behaviors in ambiguous and clear conflict conditions, as observed in our study, indicates a dynamic interplay between cognitive control and situational factors. This extends the findings of Shalvi et al. (2012), who highlighted the role of situational clarity in ethical decision-making. Our results further elaborate on this by showing that the ambiguity of a situation not only affects decision-making speed but also interacts with an individual’s moral inclination to influence their choices.

Furthermore, our study contributes to the debate surrounding the “Will” and “Grace” hypotheses. The negative correlation between cognitive control and the number of self-interest responses suggests that honesty, far from being the default human condition, may be the product of a conscious cognitive effort to restrain self-serving impulses. This would be consistent with the “Will” hypothesis.

4.1 Applications and limitations

The results extend our understanding of the role of cognitive control plays in honesty and dishonesty, with potential applications in education, policy-making, and business ethics. For educational settings, the results suggest curricula should emphasize enhancing ethical reasoning and cognitive control, preparing students to navigate moral challenges thoughtfully. Policy implications include designing environments that discourage dishonesty by clarifying ethical standards and making dishonest actions more cognitively taxing, thereby promoting transparency and accountability. In business ethics, our findings advocate for cultures of integrity supported by clear ethical guidelines and training programs that bolster moral awareness and cognitive control, helping employees prioritize ethical standards over self-interest. This approach aims to foster a more honest and ethical conduct across various sectors.

Our study, while offering valuable insights into the complex interplay between cognitive control and moral decision-making, is not without its limitations. One of the primary constraints involves the sample size and demographic composition, primarily undergraduate and postgraduate students, which may not fully represent the broader population. This limitation could affect the generalizability of our findings, as the specific age group and educational background of our participants might influence their moral decision-making processes and cognitive control mechanisms differently compared to a more diverse population. Additionally, our reliance on reaction times as the use of intuitive or reflective processes should be careful. Rather some studies highlight the pitfalls of using RT correlations as support for dual-process theories. Reaction times, in this context, primarily reflect the cognitive processing involved in navigating moral conflicts rather than directly indicating whether honesty is an inherent or automatic response (Evans et al., 2015; Krajbich et al., 2015; Andrighetto et al., 2020).

5 Conclusion

Our study contributes to the nuanced understanding of the interplay between cognitive control and moral decision-making, revealing the complex mechanisms through which individuals navigate ethical dilemmas. By examining the roles of decision conflict and moral deliberation across different moral predispositions, our findings challenge and extend existing theories on moral psychology. Despite limitations related to sample diversity and the interpretation of reaction times, this research underscores the importance of considering individual differences and the multifaceted nature of cognitive processes in ethical behavior. Looking forward, it paves the way for further interdisciplinary investigations into moral decision-making, encouraging a broader exploration of how cognitive, emotional, and social factors collectively shape our moral actions. As we continue to unravel the cognitive underpinnings of morality, this work not only deepens our theoretical understanding but also has practical implications for promoting ethical behavior in an increasingly complex world.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by IRB of the Seventh People’s Hospital of Wenzhou (EC-KY-2022048). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

H-ML: Writing – original draft. W-JY: Conceptualization, Writing – original draft. Y-WW: Conceptualization, Writing – original draft. Z-YH: Writing – review & editing, Conceptualization, Investigation, Validation.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was supported by Wenzhou Science and Technology Project (Y2023864).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Andrighetto, G., Capraro, V., Guido, A., and Szekely, A. (2020). Cooperation, response time, and social value orientation: a meta-analysis. Proc. Cogn. Sci. Soc., 2116–2122. doi: 10.31234/osf.io/cbakz

Crossref Full Text | Google Scholar

Baumeister, R. F., and Juola Exline, J. (1999). Virtue, personality, and social relations: Self‐control as the moral muscle. Journal of personality, 67, 1165–1194. doi: 10.1111/1467-6494.00086

Crossref Full Text | Google Scholar

Baumeister, R. F., Masicampo, E. J., and DeWall, C. N. (2009). Prosocial benefits of feeling free: disbelief in free will increases aggression and reduces helpfulness. Personal. Soc. Psychol. Bull. 35, 260–268. doi: 10.1177/0146167208327217

PubMed Abstract | Crossref Full Text | Google Scholar

Becker, G. S. (1968). Crime and punishment: an economic approach. J. Polit. Econ. 76, 169–217. doi: 10.1086/259394

Crossref Full Text | Google Scholar

Capraro, V. (2017). Does the truth come naturally? Time pressure increases honesty in one-shot deception games. Econ. Lett. 158, 54–57. doi: 10.1016/j.econlet.2017.06.015

Crossref Full Text | Google Scholar

Capraro, V. (2023). The dual-process approach to human sociality: Meta-analytic evidence for a theory of internalized heuristics for self-preservation. arXiv. doi: 10.48550/arXiv.1906.09948

Crossref Full Text | Google Scholar

Capraro, V., and Rand, D. G. (2018). Do the right thing: Experimental evidence that preferences for moral behavior, rather than equity or efficiency per se, drive human prosociality. Judgment and Decision Making, 13, 99–111. doi: 10.1017/S1930297500008858

Crossref Full Text | Google Scholar

Capraro, V., Schulz, J., and Rand, D. G. (2019). Time pressure and honesty in a deception game. J. Behav. Exp. Econ. 79, 93–99. doi: 10.1016/j.socec.2019.01.007

Crossref Full Text | Google Scholar

Evans, A. M., Dillon, K. D., and Rand, D. G. (2015). Fast but not intuitive, slow but not reflective: decision conflict drives reaction times in social dilemmas. J. Exp. Psychol. Gen. 144, 951–966. doi: 10.1037/xge0000107

Crossref Full Text | Google Scholar

Gino, F., Norton, M. I., and Ariely, D. (2010). The counterfeit self: the deceptive costs of faking it. Psychol. Sci. 21, 712–720. doi: 10.1177/0956797610366545

Crossref Full Text | Google Scholar

Gino, F., Schweitzer, M. E., Mead, N. L., and Ariely, D. (2011). Unable to resist temptation: how self-control depletion promotes unethical behavior. Organ. Behav. Hum. Decis. Process. 115, 191–203. doi: 10.1016/j.obhdp.2011.03.001

Crossref Full Text | Google Scholar

Greene, J. D., and Paxton, J. M. (2009). Patterns of neural activity associated with honest and dishonest moral decisions. Proc. Natl. Acad. Sci. 106, 12506–12511. doi: 10.1073/pnas.0900152106

PubMed Abstract | Crossref Full Text | Google Scholar

Henrich, J., Boyd, R., Bowles, S., Camerer, C., Fehr, E., Gintis, H., et al. (2005). “Economic man” in cross-cultural perspective: behavioral experiments in 15 small-scale societies. Behav. Brain Sci. 28, 795–815. doi: 10.1017/S0140525X05000142

PubMed Abstract | Crossref Full Text | Google Scholar

Köbis, N. C., Verschuere, B., Bereby-Meyer, Y., Rand, D., and Shalvi, S. (2019). Intuitive honesty versus dishonesty: Meta-analytic evidence. Perspect. Psychol. Sci. 14, 778–796. doi: 10.1177/1745691619851778

PubMed Abstract | Crossref Full Text | Google Scholar

Krajbich, I., Bartling, B., Hare, T., and Fehr, E. (2015). Rethinking fast and slow based on a critique of reaction-time reverse inference. Nat. Commun. 6:7455. doi: 10.1038/ncomms8455

PubMed Abstract | Crossref Full Text | Google Scholar

Mazar, N., Amir, O., and Ariely, D. (2008). The dishonesty of honest people: a theory of self-concept maintenance. J. Mark. Res. 45, 633–644. doi: 10.1509/jmkr.45.6.633

Crossref Full Text | Google Scholar

Rand, D. G., Greene, J. D., and Nowak, M. A. (2012). Spontaneous giving and calculated greed. Nature 489, 427–430. doi: 10.1038/nature11467

PubMed Abstract | Crossref Full Text | Google Scholar

Shalvi, S., Eldar, O., and Bereby-Meyer, Y. (2012). Honesty requires time (and lack of justifications). Psychol. Sci. 23, 1264–1270. doi: 10.1177/0956797612443835

Crossref Full Text | Google Scholar

Speer, S. P., Smidts, A., and Boksem, M. A. (2022). Cognitive control and dishonesty. Trends Cogn. Sci. 26, 796–808. doi: 10.1016/j.tics.2022.06.005

PubMed Abstract | Crossref Full Text | Google Scholar

Suchotzki, K., Verschuere, B., Van Bockstaele, B., Ben-Shakhar, G., and Crombez, G. (2017). Lying takes time: a Meta-analysis on reaction time measures of deception. Psychol. Bull. 143, 428–453. doi: 10.1037/bul0000087

PubMed Abstract | Crossref Full Text | Google Scholar

Tabatabaeian, M., Dale, R., and Duran, N. D. (2015). Self-serving dishonest decisions can show facilitated cognitive dynamics. Cogn. Process. 16, 291–300. doi: 10.1007/s10339-015-0660-6

Crossref Full Text | Google Scholar

Tangney, J. P., Stuewig, J., and Mashek, D. J. (2007). Moral emotions and moral behavior. Annu. Rev. Psychol. 58, 345–372. doi: 10.1146/annurev.psych.56.091103.070145

PubMed Abstract | Crossref Full Text | Google Scholar

Verschuere, B., Köbis, N. C., Bereby-Meyer, Y., Rand, D., and Shalvi, S. (2018). Taxing the brain to uncover lying? Meta-analyzing the effect of imposing cognitive load on the reaction-time costs of lying. J. Appl. Res. Mem. Cogn. 7, 462–469. doi: 10.1016/j.jarmac.2018.04.005

Crossref Full Text | Google Scholar

Wood, A. M., Linley, P. A., Maltby, J., Baliousis, M., and Joseph, S. (2008). The authentic personality: a theoretical and empirical conceptualization and the development of the authenticity scale. J. Couns. Psychol. 55, 385–399. doi: 10.1037/0022-0167.55.3.385

Crossref Full Text | Google Scholar

Keywords: honesty, dishonesty, cognitive control, moral decision-making, reaction time

Citation: Li H-M, Yan W-J, Wu Y-W and Huang Z-Y (2024) Cognitive control in honesty and dishonesty under different conflict scenarios: insights from reaction time. Front. Psychol. 15:1271916. doi: 10.3389/fpsyg.2024.1271916

Received: 03 August 2023; Accepted: 26 February 2024;
Published: 14 March 2024.

Edited by:

Chiara Lucifora, University of Bologna, Italy

Reviewed by:

Valerio Capraro, Middlesex University, United Kingdom
Hansika Singhal, University of Delhi, India

Copyright © 2024 Li, Yan, Wu and Huang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yu-Wei Wu, 986958226@qq.com; Zi-Ye Huang, 18106790817@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.