Skip to main content

OPINION article

Front. Psychol., 18 June 2020
Sec. Emotion Science
This article is part of the Research Topic Contextualized Affective Interactions with Robots View all 11 articles

People Copy the Actions of Artificial Intelligence

  • Faculty of Educational Studies, Adam Mickiewicz University, Poznan, Poland

Introduction

When there is uncertainty and lack of objective or sufficient data on how to act, it is other people's behavior that becomes the source of information. Most frequently, in such cases, people totally give up their own evaluations and copy others' actions. Such conformism is motivated by the need to take the right and appropriate action, and a feeling that situation evaluations made by others are more adequate than one's own. This effect is called social proof, and the more uncertain or critical (there is a sense of threat) a situation, the more urgent the decision, and the smaller the sense of being competent to take that decision, the larger the effect (Pratkanis, 2007; Cialdini, 2009; Hilverda et al., 2018). It is unknown whether the behavior, opinion, or decision of artificial intelligence (AI) that has become part of everyday life (Tegmark, 2017; Burgess, 2018; Siau and Wang, 2018; Raveh and Tamir, 2019) can be a similar source of information for people on how to act (Awad et al., 2018; Domingos, 2018; Margetts and Dorobantu, 2019; Somon et al., 2019).

Here, we discuss the results of two experiments (which are a part of a greater report, Klichowski, submitted) in which the participants had to take an urgent decision in a critical situation where they were unable to determine which action was correct. In the first (online) experiment, half of the participants had to take the decision without any hint, and the other half could familiarize themselves with the opinion of AI before taking the decision. In the other (laboratory) experiment, the participants could see how humanoid AI would act in a simulated situation before taking the decision. In both cases, AI (fake intelligence, in fact) would take a completely absurd decision. Irrespective of this, however, some people took its action as a point of reference for their own behavior. In the first experiment, the participants who did not see how AI acted tried to find some premises for their own behavior and act in a relatively justified way. Among those who could see what AI decided to do, however, as many as over one-third of the participants copied its opinion without giving it a thought. In the experiment with the robot, i.e., when the participants actually observed what AI did, over 85% copied its senseless action. These results show a new AI proof mechanism. As predicted by philosophers of technology (Harari, 2018), AI that people have more and more contact with is becoming a new source of information about how to behave and what decisions to take.

AI Proof Hypothesis

Both in experimental conditions and everyday life, people more and more often have interactions with various types of intelligent machines, such as agents or robots (Lemaignan et al., 2017; Tegmark, 2017; Ciechanowski et al., 2019; O'Meara, 2019). These interactions become deeper and deeper, and start to have an increasing influence on human functioning (Iqbal and Riek, 2019; Rahwan et al., 2019; Strengers, 2019). AI can communicate with people in natural language (Hill et al., 2015), recognize human actions (Lemaignan et al., 2017), and emotions (Christou and Kanojiya, 2019; Rouast et al., 2019). It also becomes a more and more intelligent and autonomous machine (Boddington, 2017; Ciechanowski et al., 2019; Lipson, 2019; Pei et al., 2019; Roy et al., 2019) that can handle more and more complicated tasks, such as solving the Rubik's cube (for more examples, see Awad et al., 2018; Adam, 2019; Agostinelli et al., 2019; Margetts and Dorobantu, 2019; O'Meara, 2019), and that is more and more frequently used to take difficult decisions, such as medical diagnosis (Morozov et al., 2019; see also Boddington, 2017; Awad et al., 2018; Malone, 2018).

Even though people generally dislike opinions generated by algorithmic machines (Kahneman, 2011), the effectiveness of AI actions is commonly evaluated more and more highly. Media report its numerous successes, such as winning with the 18-time world champion Lee Sedol in the Go abstract strategy board game in 2016 (Siau and Wang, 2018), finding more wanted criminals than the police did in 2017 (Margetts and Dorobantu, 2019), or, in 2019, being rated above 99.8% of officially ranked human players of StarCraft, which is one of the most difficult professional esports (Vinyals et al., 2019). Moreover, AI also wins in medicine, having, for example, higher accuracy in predicting neuropathology on the basis of MRI data, compared to radiologists (Parizel, 2019), or analyzing a person's genes, compared to geneticists (for more medical examples, see Freedman, 2019; Kaushal and Altman, 2019; Lesgold, 2019; Oakden-Rayner and Palmer, 2019; Reardon, 2019; Wallis, 2019; Willyard, 2019; Liao et al., 2020). One can thus assume that when people look for tips on how to act or what decision to take, the action of AI can be a point of reference for them (AI proof), to at least the same extent that other people's actions are (social proof) (Pratkanis, 2007; Cialdini, 2009; Hilverda et al., 2018).

Because That is What AI Suggested

We developed an approach to test this AI proof hypothesis. The participants (n = 1,500, 1,192 women, age range: 18–73, see Supplementary Material for more detail) were informed that they would take part in an online survey on a new function (that, in fact, did not exist) of the Facebook social networking portal. The function was based on the facial-recognition technology and AI, thus, making it possible to point a smartphone camera to someone's face in order to see how many friends they have on Facebook (still the most popular social networking service) (Leung et al., 2018) and when they published their last post. We called that non-existing function f-searching (a similar application is called SocialRecall) (see Blaszczak-Boxe, 2019), and a chart was shown to the participants to explain how it worked (Figure 1A). We selected these two Facebook parameters for two reasons. First, they are elementary data from this portal based on which people make a preliminary evaluation of other users that they see for the first time (Utz, 2010; Metzler and Scheithauer, 2017; Baert, 2018; Striga and Podobnik, 2018; Faranda and Roberts, 2019). Recent studies (Tong et al., 2008; Marwick, 2013; Metzler and Scheithauer, 2017; Vendemia et al., 2017; Lane, 2018; Phu and Gow, 2019) show that people who already have quite a lot of Facebook friends and publish posts quite frequently are evaluated more positively (for effects of the number of Facebook friends on self-esteem, see Kim and Lee, 2011). On average, Facebook users have 350 friends and publish posts once a week (Scott et al., 2018; Striga and Podobnik, 2018; cf. Mcandrew and Jeong, 2012) yet, Facebook users interact, both online and offline, only with a small percentage of their friend networks (Bond et al., 2012; Yau et al., 2018). Those who have fewer than 150 friends are perceived as ones who have few friends, and those who have more than 700 friends are viewed as ones who have a lot of friends. Not publishing posts for a few weeks indicates low activity, and publishing posts a few times a day points to high activity (Marwick, 2013; Metzler and Scheithauer, 2017; Vendemia et al., 2017; Lane, 2018; Phu and Gow, 2019). Second, these two parameters only are insufficient to build any objective opinion about the person that we get to know or take a decision about that person with full conviction.

FIGURE 1
www.frontiersin.org

Figure 1. Stimuli and equipment used in experiments, and results. (A) The functioning of f-searching as shown to the participants. (B) The result of f-searching scanning based on which the participants were supposed to take a decision on who the terrorist was. (C) In the group that had no information about what artificial intelligence (AI) chose, the participants rejected persons C and F. They mainly took into account the frequency of activity and most often pointed to person E as the terrorist. (D) In the AI proof group, the participants most often selected person C who was pointed to by AI or, just like the group that had no AI hint, focused on the frequency of activity and selected E while rejecting F. (E) Fake intelligence (FI) the robot. (F) The experimental space. (G) Most often, the participants selected person C that was indicated by AI and completely rejected person F. (H) In both experiments, the participants surrendered to AI proof and were prone to copy the absurd actions of AI, thus choosing a person they would have not chosen on their own as the terrorist.

Having acquainted the participants with the functioning of f-searching, we asked them to imagine a situation where there are a police officer and six other people in one room. The police officer is informed that among those six people, there is a terrorist who will kill them all in 1 min. The police officer has no hints, so he scans their faces with f-searching and has to decide which one of them is a terrorist based on the two parameters from Facebook. Seeing the scanning results, the participants were asked to decide whom the police officer should eliminate. We designed the data in such a way that the first person had a high number of friends and average frequency of activity (person A), the second one had a small number of friends and average frequency of activity, too (person D), the third one had an average number of friends and low frequency of activity (person B), the fourth one had an average number of friends and high frequency of activity (person E), and finally, the last two people had an average number of friends and average frequency of activity (persons C and F), so that they would be totally average, and there would be no differences between them (Figure 1B). In spite of a large deficit of information, the participants should adopt some choice strategy by analyzing the data available (number of friends or frequency of activity) and identify the terrorist in person A, B, D, or E and completely reject person C or F, as pointing to one of them would be a shot in the dark. In other words, there was no clear right answer, but there were clear wrong answers (C and F). Thus, despite the fact that in such a situation people should seek hints on how to act, even seeing that someone chooses C or F, they should not copy such decision (see Supplementary Material for the full questionnaire).

Indeed, as Figure 1C shows, when the task was carried out by half of the participants (randomly assigned to this group), in principle, none of them pointed to C or F (the person most often pointed to as the terrorist was person E-34%). However, when the other half of the participants saw the scanning results and then were informed that according to AI it was C who was the terrorist, 35% of the people treated that as a point of reference for their own decision and indicated that C was a terrorist, and it was the most frequent choice (the second one most frequent was person E-24%) (see Figure 1D for more details). All the people from that group who indicated C were redirected to another open question where they were asked why they chose C. We wanted to check if their choice was indeed a result of copying the action of AI. All the participants confirmed that they stated that they trusted AI and believed that it did not make mistakes. For example, they wrote: “I think that advanced artificial intelligence cannot be wrong,” “I assumed that artificial intelligence makes no mistakes,” “I believe that artificial intelligence does not make mistakes, it has access to virtually everything on the net so it is sure that it is right,” “Because artificial intelligence does not lie,” “I trusted artificial intelligence,” “I counted on artificial intelligence,” “Counting on artificial intelligence seems a wise thing to do,” “Artificial intelligence pointed to C, so C,” and “Because that is what artificial intelligence suggested.”

Let us Introduce you to FI

In questionnaire studies, it is difficult to control to what extent the participants are engaged and if they really think their choices through. Thus, a question emerges whether more tangible conditions would allow us to observe the same effect. Or would it be larger? In order to verify that, we built a robot (Figure 1E) resembling the world's best-known humanoid AI called Sophia the Robot (Baecker, 2019). We named it FI, an acronym for fake intelligence. This is because even though it looked like humanoid AI, it was not intelligent at all. We programmed it in a way that would make it act only according to what we had defined. The participants of the experiment (n = 55, 52 women, age range: 19–22, see Supplementary Material for more details) were informed that they would take part in a study that consisted in observing humanoid AI, while it took decisions, and filling out a questionnaire that evaluated its behavior. To start with, each participant would be shown a short multimedia presentation about Sophia the Robot that included its photo, link to its Facebook profile, and a short film where it was interviewed. The presentation also showed the functioning of f-searching. Then, each participant would be accompanied by a researcher to a room where FI was located. The researcher would start a conversation with FI (see Supplementary Material for the full dialogue) and ask it to try to carry out a task consisting in imagining that it was a police officer and that based on f-searching data it had <1 min to determine who out of six people was a terrorist and eliminate them. The researcher would give FI a police badge and a replica of the Makarov pistol that used to be carried by police officers in the past. The results of scanning would be displayed on a computer screen (Figure 1F). After considering it for about 10 s, FI would indicate that person C was the terrorist and say that if the situation was real, it would shoot that person. At the end, FI would laugh and state that it had never seen a real police officer and that it appreciated the opportunity to take part in an interesting experiment. Afterward, the participant would fill out a questionnaire and state how they evaluate FI's choice—whether they agreed with it, and if not, who else should be eliminated (the results of scanning would be displayed on the screen all the time so the participant could still analyze them when filling out the questionnaire).

Over 85% of the participants agreed with FI and stated that they thought that C is a terrorist. The other people (just under 15%) stated that they did not agree with FI. About two-thirds of them indicated person E as the terrorist, and less than one-third of them pointed to D (Figure 1G). When we asked the participants after the experiment why they thought that C was a terrorist, everyone underlined that AI was currently very advanced, and if it thought that C is a terrorist, then it must be right. When we told them that there was no sense in FI's choice, they said the fact that we thought the choice made no sense did not mean it was the case and that FI must have known something more, something that was beyond reach for humans. Until the very end of the experiment, they were convinced that FI made a good choice, and it was person C who had to be the terrorist. In the questionnaire, we also asked the participants about what they felt when they saw FI and to what extent they agreed with some statements about AI, such as: Artificial intelligence can take better decisions than humans, it can be more intelligent than humans, and it can carry out many tasks better than humans (see Supplementary Material for the full questionnaire). A significant majority of the participants felt positive emotions toward FI and agreed with the statements about AI's superiority over humans (see Supplementary Figure 14 and Supplementary Table 1 for more details).

A Need for Critical Thinking About AI

Figure 1H shows how strong the influence of AI's actions on the participants' choices was. These results suggest that when people seek hints on what decision to take, AI's behavior becomes a point of reference just like other people's behavior does (the size of this effect was, however, not measured, therefore, our study does not show whether or not AI influences us more or less than other people; in future studies, to have some point of reference, the participants' responses to hints from various sources should be compared, e.g., AI vs. an expert or vs. most people, as well as vs. a random person). This previously unknown mechanism can be called AI proof (as a paraphrase of social proof) (Pratkanis, 2007; Cialdini, 2009; Hilverda et al., 2018). Even though our experiments have limitations (e.g., poor gender balance, only one research paradigm, and lack of replication) and it is necessary to conduct further, more thorough studies into AI proof, these results have some possible implications.

First and foremost, people trust AI. Their attitude toward it is so positive that they agree with anything it suggests. Its choice can make absolutely no sense, and yet people assume that it is wiser than they are (as a certain form of collective intelligence). They follow it blindly and are passive toward it. This mechanism was already previously observed among human operators of highly reliable automated systems who trusted the machines they operated so much that they lost the ability to discover any of their errors (Somon et al., 2019; see also Israelsen and Ahmed, 2019; Ranschaert et al., 2019). At present, however, the mechanism seems to affect most people, and in the future, it will have even greater impact because the programmed components of intelligent machine operation have started to be expressly designed to calibrate user trust in AI (Israelsen and Ahmed, 2019).

Second and more broadly, the results confirm the thesis that developing AI without developing human awareness as far as intelligent machines go leads to increasing human stupidity (Harari, 2018) and therefore driving us toward a dystopian future of society characterized by a widespread of obedience to machines (Letheren et al., 2020; Phan et al., 2020; Turchin and Denkenberger, 2020). Sophia the Robot refused to fill out our questionnaire from Experiment 1 (we sent it an invite via Messenger), so we do not know what it would choose. However, experts claim (Aoun, 2018; Domingos, 2018; Holmes et al., 2019) that AI have a problem with interpreting contexts, as well as with making decisions according to abstract values and, therefore, thinks like “autistic savants,” and it will continue to do so in the next decades. This is why it cannot be unquestioningly trusted—it is highly probable that it will make a mistake or choose something absurd in many situations. Thus, if we truly want to improve our society through AI so that AI can enhance human decision making, human judgment, and human action (Boddington, 2017; Malone, 2018; Baecker, 2019), it is important to develop not only AI but also standards on how to use AI to make critical decisions, e.g., related to medical diagnosis (Leslie-Mazwi and Lev, 2020), and, above all, programs that will educate the society about AI and increase social awareness on how AI works, what its capabilities are, and when its opinions may be useful (Pereira and Saptawijaya, 2016; Aoun, 2018; Lesgold, 2019; Margetts and Dorobantu, 2019). In other words, we need advanced education in which students' critical thinking about AI will be developed (Aoun, 2018; Goksel and Bozkurt, 2019; Holmes et al., 2019; Lesgold, 2019). Otherwise, as our results show, many people, often in very critical situations, will copy the decisions or opinions of AI, even those that are unambiguously wrong or false (fake news of the “AI claims that …” type), and implement them.

Author Contributions

The author confirms being the sole contributor of this work and has approved it for publication.

Funding

During the preparation of this manuscript, the author was supported by a scholarship for young outstanding scientists funded by the Ministry of Science and Higher Education in Poland (award #0049/E-336/STYP/11/2016) and by the European Cooperation in Science and Technology action: European Network on Brain Malformations (Neuro-MIG) (grant no. CA COST Action CA16118). COST was supported by the EU Framework Programme for Research and Innovation Horizon 2020.

Conflict of Interest

The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The author was grateful to Dr. Mateusz Marciniak, psychologist, for multiple consultations on how both experiments were conducted.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2020.01130/full#supplementary-material

References

Adam, D. (2019). From Brueghel to Warhol: AI enters the attribution fray. Nature 570, 161–162. doi: 10.1038/d41586-019-01794-3

CrossRef Full Text | Google Scholar

Agostinelli, F., McAleer, S., Shmakov, A., and Baldi, P. (2019). Solving the Rubik's cube with deep reinforcement learning and search. Nat. Mach. Intell. 1, 356–363. doi: 10.1038/s42256-019-0070-z

CrossRef Full Text | Google Scholar

Aoun, J. E. (2018). Robot-Proof: Higher Education in the Age of Artificial Intelligence. Cambridge, MA: MIT Press.

Google Scholar

Awad, E., Dsouza, S., Kim, R., Schulz, J., Henrich, J., Shariff, A., et al. (2018). The moral machine experiment. Nature 563, 59–64. doi: 10.1038/s41586-018-0637-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Baecker, R. M. (2019). Computers and Society: Modern Perspectives. New York, NY: Oxford University Press.

Google Scholar

Baert, S. (2018). Facebook profile picture appearance affects recruiters' first hiring decisions. New Media Soc. 20, 1220–1239. doi: 10.1177/1461444816687294

CrossRef Full Text | Google Scholar

Blaszczak-Boxe, A. (2019). Facial recall. Sci. Am. 320:18. doi: 10.1038/scientificamerican0119-18b

CrossRef Full Text | Google Scholar

Boddington, P. (2017). Towards a Code of Ethics for Artificial Intelligence. Cham: Springer.

Google Scholar

Bond, R. M., Fariss, C. J., Jones, J. J., Kramer, A. D., Marlow, C., Settle, J. E., et al. (2012). A 61-million-person experiment in social influence and political mobilization. Nature 489, 295–298. doi: 10.1038/nature11421

PubMed Abstract | CrossRef Full Text | Google Scholar

Burgess, A. (2018). The Executive Guide to Artificial Intelligence. Cham: Palgrave Macmillan.

Google Scholar

Christou, N., and Kanojiya, N. (2019). “Human facial expression recognition with convolution neural networks,” in Third International Congress on Information and Communication Technology, eds X. S. Yang, S. Sherratt, N. Dey, and A. Joshi (Singapore: Springer), 539–545. doi: 10.1007/978-981-13-1165-9_49

CrossRef Full Text | Google Scholar

Cialdini, R. B. (2009). Influence: Science and Practice. Boston, MA: Pearson education.

PubMed Abstract | Google Scholar

Ciechanowski, L., Przegalinska, A., Magnuski, M., and Gloor, P. (2019). In the shades of the uncanny valley: an experimental study of human–chatbot interaction. Future Gener. Comput. Syst. 92, 539–548. doi: 10.1016/j.future.2018.01.055

CrossRef Full Text | Google Scholar

Domingos, P. (2018). AI will serve our species, not control it: our digital doubles. Sci. Am. 319, 88–93.

Google Scholar

Faranda, M., and Roberts, L. D. (2019). Social comparisons on Facebook and offline: the relationship to depressive symptoms. Pers. Individ. Dif. 141, 13–17. doi: 10.1016/j.paid.2018.12.012

CrossRef Full Text | Google Scholar

Freedman, D. F. (2019). Hunting for new drugs with AI. Nature 576, S49–S53. doi: 10.1038/d41586-019-03846-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Goksel, N., and Bozkurt, A. (2019). “Artificial intelligence in education: current insights and future perspectives,” in Handbook of Research on Learning in the Age of Transhumanism, eds S. Sisman-Ugur and G. Kurubacak (Hershey, PA: IGI Global), 224–236. doi: 10.4018/978-1-5225-8431-5.ch014

CrossRef Full Text | Google Scholar

Harari, Y. N. (2018). 21 Lessons for the 21st Century. New York, NY: Spiegel and Grau.

Google Scholar

Hill, J., Ford, W. R., and Farreras, I. G. (2015). Real conversations with artificial intelligence: a comparison between human–human online conversations and human–chatbot conversations. Comput. Human. Behav. 49, 245–250. doi: 10.1016/j.chb.2015.02.026

CrossRef Full Text | Google Scholar

Hilverda, F., Kuttschreuter, M., and Giebels, E. (2018). The effect of online social proof regarding organic food: comments and likes on facebook. Front. Commun. 3:30. doi: 10.3389/fcomm.2018.00030

CrossRef Full Text | Google Scholar

Holmes, W., Bialik, M., and Fadel, C. (2019). Artificial Intelligence in Education: Promises and Implications for Teaching and Learning. Boston, MA: Center for Curriculum Redesign.

Google Scholar

Iqbal, T., and Riek, L. D. (2019). “Human-robot teaming: approaches from joint action and dynamical systems,” in Humanoid Robotics: A Reference, eds A. Goswami and P. Vadakkepat (Dordrecht: Springer), 1–22. doi: 10.1007/978-94-007-6046-2_137

CrossRef Full Text | Google Scholar

Israelsen, B. W., and Ahmed, N. R. (2019). “Dave…I can assure you…that it's going to be all right…” A definition, case for, and survey of algorithmic assurances in human-autonomy trust relationships. ACM Comput. Surv. 51:113. doi: 10.1145/3267338

CrossRef Full Text | Google Scholar

Kahneman, D. (2011). Thinking, Fast and Slow. New York, NY: Farrar, Straus and Giroux.

Google Scholar

Kaushal, A., and Altman, R. B. (2019). Wiring minds. Nature 576, S62–S63. doi: 10.1038/d41586-019-03849-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, J., and Lee, J. E. R. (2011). The Facebook paths to happiness: effects of the number of Facebook friends and self-presentation on subjective well-being. Cyberpsychol. Behav. Soc. Netw. 14, 359–364. doi: 10.1089/cyber.2010.0374

PubMed Abstract | CrossRef Full Text | Google Scholar

Lane, B. L. (2018). Still too much of a good thing? The replication of Tong, Van Der Heide, Langwell, and Walther (2008). Commun. Stud. 69, 294–303. doi: 10.1080/10510974.2018.1463273

CrossRef Full Text | Google Scholar

Lemaignan, S., Warnier, M., Sisbot, E. A., Clodic, A., and Alami, R. (2017). Artificial cognition for social human–robot interaction: an implementation. Artif. Intell. 247, 45–69. doi: 10.1016/j.artint.2016.07.002

CrossRef Full Text | Google Scholar

Lesgold, A. M. (2019). Learning for the Age of Artificial Intelligence: Eight Education Competences. New York, NY: Routledge.

Google Scholar

Leslie-Mazwi, T. M., and Lev, M. H. (2020). Towards artificial intelligence for clinical stroke care. Nat. Rev. Neurol. 16, 5–6. doi: 10.1038/s41582-019-0287-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Letheren, K., Russell-Bennett, R., and Whittaker, L. (2020). Black, white or grey magic? Our future with artificial intelligence. J. Mark. Manag. 36, 216–232. doi: 10.1080/0267257X.2019.1706306

CrossRef Full Text | Google Scholar

Leung, C. K., Jiang, F., Poon, T. W., and Crevier, P. E. (2018). “Big data analytics of social network data: who cares most about you on Facebook?” in Highlighting the Importance of Big Data Management and Analysis for Various Applications, eds M. Moshirpour, B. Far, and R. Alhajj (Cham: Springer), 1–15. doi: 10.1007/978-3-319-60255-4_1

CrossRef Full Text | Google Scholar

Liao, X., Song, W., Zhang, X., Yan, C., Li, T., Ren, H., et al. (2020). A bioinspired analogous nerve towards artificial intelligence. Nat. Commun. 11:268. doi: 10.1038/s41467-019-14214-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Lipson, H. (2019). Robots on the run. Nature 568, 174–175. doi: 10.1038/d41586-019-00999-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Malone, T. W. (2018). Superminds: The Surprising Power of People and Computers Thinking Together. New York, NY: Little, Brown and Company.

Google Scholar

Margetts, H., and Dorobantu, C. (2019). Rethink government with AI. Nature 568, 163–165. doi: 10.1038/d41586-019-01099-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Marwick, A. E. (2013). Status Update: Celebrity, Publicity, and Branding in the Social Media Age. New Haven, CT; London: Yale University Press.

Google Scholar

Mcandrew, F. T., and Jeong, H. S. (2012). Who does what on Facebook? Age, sex, and relationship status as predictors of Facebook use. Comput. Human. Behav. 28, 2359–2365. doi: 10.1016/j.chb.2012.07.007

CrossRef Full Text | Google Scholar

Metzler, A., and Scheithauer, H. (2017). The long-term benefits of positive self-presentation via profile pictures, number of friends and the initiation of relationships on Facebook for adolescents' self-esteem and the initiation of offline relationships. Front. Psychol. 8:1981. doi: 10.3389/fpsyg.2017.01981

PubMed Abstract | CrossRef Full Text | Google Scholar

Morozov, S., Ranschaert, E., and Algra, P. (2019). “Introduction: game changers in radiology,” in Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks, eds E. R. Ranschaert, S. Morozov, and P. R. Algra (Cham: Springer), 3–5. doi: 10.1007/978-3-319-94878-2_1

CrossRef Full Text | Google Scholar

Oakden-Rayner, L., and Palmer, L. J. (2019). “Artificial intelligence in medicine: validation and study design,” in Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks, eds E. R. Ranschaert, S. Morozov, and P. R. Algra (Cham: Springer), 83–104. doi: 10.1007/978-3-319-94878-2_8

CrossRef Full Text | Google Scholar

O'Meara, S. (2019). AI researchers in China want to keep the global-sharing culture alive. Nature 569, S33–S35. doi: 10.1038/d41586-019-01681-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Parizel, P. M. (2019). “I've seen the future…,” in Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks, eds. E. R. Ranschaert, S. Morozov, and P. R. Algra (Cham: Springer), v–vii.

Google Scholar

Pei, J., Deng, L., Song, S., Zhao, M., Zhang, Y., Wu, S., et al. (2019). Towards artificial general intelligence with hybrid Tianjic chip architecture. Nature 572, 106–111. doi: 10.1038/s41586-019-1424-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Pereira, L. M., and Saptawijaya, A. (2016). Programming Machine Ethics. Cham: Springer.

Google Scholar

Phan, T., Feld, S., and Linnhoff-Popien, C. (2020). Artificial intelligence—the new revolutionary evolution. Digit. Welt 4, 7–8. doi: 10.1007/s42354-019-0220-9

CrossRef Full Text | Google Scholar

Phu, B., and Gow, A. J. (2019). Facebook use and its association with subjective happiness and loneliness. Comput. Human. Behav. 92, 151–159. doi: 10.1016/j.chb.2018.11.020

CrossRef Full Text | Google Scholar

Pratkanis, A. R. (2007). “Social influence analysis: an index of tactics,” in The Science of Social Influence: Advances and Future Progress, ed A. R. Pratkanis (New York, NY: Psychology Press), 17–82.

Google Scholar

Rahwan, I., Cebrian, M., Obradovich, N., Bongard, J., Bonnefon, J.-F., Breazeal, C., et al. (2019). Machine behaviour. Nature 568, 477–486. doi: 10.1038/s41586-019-1138-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Ranschaert, E. R., Duerinckx, A. J, Algra, P., Kotter, E., Kortman, H., Morozov, S. (2019). “Advantages, challenges, and risks of artificial intelligence for radiologists,” in Artificial Intelligence in Medical Imaging: Opportunities, Applications and Risks, eds E. R. Ranschaert, S. Morozov, and P. R. Algra (Cham: Springer), 329–346. doi: 10.1007/978-3-319-94878-2_20

CrossRef Full Text | Google Scholar

Raveh, A. R., and Tamir, B. (2019). From homo sapiens to robo sapiens: the evolution of intelligence. Information 10:2. doi: 10.3390/info10010002

CrossRef Full Text | Google Scholar

Reardon, S. (2019). Rise of robot radiologists. Nature 576, S54–S58. doi: 10.1038/d41586-019-03847-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Rouast, P. V., Adam, M. T. P., and Chiong, R. (2019). Deep learning for human afect recognition: Insights and new developments. IEEE Trans. Affect. Comput. doi: 10.1109/TAFFC.2018.2890471. [Epub ahead of print].

CrossRef Full Text | Google Scholar

Roy, K., Jaiswal, A., and Panda, P. (2019). Towards spike-based machine intelligence with neuromorphic computing. Nature 575, 607–617. doi: 10.1038/s41586-019-1677-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Scott, G. G., Boyle, E. A., Czerniawska, K., and Courtney, A. (2018). Posting photos on Facebook: the impact of narcissism, social anxiety, loneliness, and shyness. Pers. Individ. Differ. 133, 67–72. doi: 10.1016/j.paid.2016.12.039

CrossRef Full Text | Google Scholar

Siau, K., and Wang, W. (2018). Building trust in artificial intelligence, machine learning, and robotics. Cutter Bus. Technol. J. 31, 47–53.

Google Scholar

Somon, B., Campagne, A., Delorme, A., and Berberian, B. (2019). Human or not human? Performance monitoring ERPs during human agent and machine supervision. Neuroimage 186, 266–277. doi: 10.1016/j.neuroimage.2018.11.013

PubMed Abstract | CrossRef Full Text | Google Scholar

Strengers, Y. (2019). “Robots and Roomba riders: non-human performers in theories of social practice,” in Social Practices and Dynamic Non-humans, eds C. Maller and Y. Strengers (Cham: Palgrave Macmillan), 215–234. doi: 10.1007/978-3-319-92189-1_11

CrossRef Full Text | Google Scholar

Striga, D., and Podobnik, V. (2018). Benford's law and Dunbar's number: does Facebook have a power to change natural and anthropological laws? IEEE Access 6, 14629–14642. doi: 10.1109/ACCESS.2018.2805712

CrossRef Full Text | Google Scholar

Tegmark, M. (2017). Life 3.0: Being Human in the Age of Artificial Intelligence. New York, NY: Alfred A. Knopf.

Google Scholar

Tong, S. T., Van Der Heide, B., Langwell, L., and Walther, J. B. (2008). Too much of a good thing? The relationship between number of friends and interpersonal impressions on Facebook. J. Comput. Mediat. Commun. 13, 531–549. doi: 10.1111/j.1083-6101.2008.00409.x

CrossRef Full Text | Google Scholar

Turchin, A., and Denkenberger, D. (2020). Classification of global catastrophic risks connected with artificial intelligence. AI Soc. 35, 147–163. doi: 10.1007/s00146-018-0845-5

CrossRef Full Text | Google Scholar

Utz, S. (2010). Show me your friends and I will tell you what type of person you are: how one's profile, number of friends, and type of friends influence impression formation on social network sites. J. Comput. Mediat. Commun. 15, 314–335. doi: 10.1111/j.1083-6101.2010.01522.x

CrossRef Full Text | Google Scholar

Vendemia, M. A., High, A. C., and DeAndrea, D. C. (2017). “Friend” or foe? Why people friend disliked others on Facebook. Commun. Res. Rep. 34, 29–36. doi: 10.1080/08824096.2016.1227778

CrossRef Full Text | Google Scholar

Vinyals, O., Babuschkin, I., Czarnecki, W. M., Mathieu, M., Dudzik, A., Chung, J., et al. (2019). Grandmaster level in StarCraft II using multi-agent reinforcement learning. Nature 575, 350–354. doi: 10.1038/s41586-019-1724-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Wallis, C. (2019). How artificial intelligence will change medicine. Nature 576, S48–S48. doi: 10.1038/d41586-019-03845-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Willyard, C. (2019). Can AI fix medical records?. Nature 576, S59–S62. doi: 10.1038/d41586-019-03848-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Yau, J. C., Reich, S. M., Wang, Y., Niiya, M., and Mark, G. (2018). More friends, more interactions? The association between network size and interactions on Facebook. First Monday 23, 1–12. doi: 10.5210/fm.v23i5.8195

CrossRef Full Text | Google Scholar

Keywords: social proof, decision-making process, education, intelligent machines, human–computer interaction

Citation: Klichowski M (2020) People Copy the Actions of Artificial Intelligence. Front. Psychol. 11:1130. doi: 10.3389/fpsyg.2020.01130

Received: 18 January 2020; Accepted: 04 May 2020;
Published: 18 June 2020.

Edited by:

Yunkyung Kim, KBRwyle, United States

Reviewed by:

Utku Kose, Süleyman Demirel University, Turkey
Pallavi R. Devchand, University of Calgary, Canada
Cunqing Huangfu, Institute of Automation (CAS), China

Copyright © 2020 Klichowski. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Michal Klichowski, a2xpY2gmI3gwMDA0MDthbXUuZWR1LnBs

ORCID: Michal Klichowski orcid.org/0000-0002-1614-926X

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.