Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell. , 04 February 2025

Sec. AI for Human Learning and Behavior Change

Volume 7 - 2024 | https://doi.org/10.3389/frai.2024.1465605

Factors influencing trust in algorithmic decision-making: an indirect scenario-based experiment

\r\nFernando Marmolejo-Ramos
Fernando Marmolejo-Ramos1*Rebecca MarroneRebecca Marrone2Malgorzata KorolkiewiczMalgorzata Korolkiewicz2Florence GabrielFlorence Gabriel2George SiemensGeorge Siemens2Srecko JoksimovicSrecko Joksimovic2Yuki YamadaYuki Yamada3Yuki Mori&#x;Yuki Mori4Talal RahwanTalal Rahwan5Maria SahakyanMaria Sahakyan5Belona SonnaBelona Sonna6Assylbek MeirmanovAssylbek Meirmanov7Aidos BolatovAidos Bolatov8Bidisha SomBidisha Som9Izuchukwu NdukaiheIzuchukwu Ndukaihe10Nwadiogo C. ArinzeNwadiogo C. Arinze10Josef KundrtJosef Kundrát11Lenka SkanderovLenka Skanderová12Van-Giang NgoVan-Giang Ngo13Giang NguyenGiang Nguyen13Michelle LaciaMichelle Lacia14Chun-Chia KungChun-Chia Kung15Meiselina Irmayanti,Meiselina Irmayanti16,17Abdul MuktadirAbdul Muktadir18Fransiska Timoria SamosirFransiska Timoria Samosir19Marco Tullio LiuzzaMarco Tullio Liuzza20Roberto GiorginiRoberto Giorgini21Omid Khatin-ZadehOmid Khatin-Zadeh22Hassan BanarueeHassan Banaruee23Asil Ali 
zdoruAsil Ali Özdoğru24Kris AriyabuddhiphongsKris Ariyabuddhiphongs25Wachirawit RakchaiWachirawit Rakchai25Natalia TrujilloNatalia Trujillo26Stella Maris ValenciaStella Maris Valencia26Armina Janyan,Armina Janyan27,28Kiril KostovKiril Kostov27Pedro R. MontoroPedro R. Montoro29Jose Hinojosa,Jose Hinojosa30,31Kelsey MedeirosKelsey Medeiros32Thomas E. HuntThomas E. Hunt33Julian PosadaJulian Posada34Raquel Meister Ko FreitagRaquel Meister Ko Freitag35Julian TejadaJulian Tejada35
  • 1College of Education, Psychology, and Social Work, Flinders University, Adelaide, SA, Australia
  • 2Centre for Change and Complexity in Learning, University of South Australia, Adelaide, SA, Australia
  • 3Faculty of Arts and Science, Kyushu University, Fukuoka, Japan
  • 4Graduate School of Human-Environment Studies, Kyushu University, Fukuoka, Japan
  • 5Computer Science, Science Division, New York University Abu Dhabi, Abu Dhabi, United Arab Emirates
  • 6African Masters of Machine Intelligence (AMMI), African Institute for Mathematical Sciences (AIMS), Limbe, Cameroon
  • 7Higher Schools, Pedagogical Institute, Astana International University, Astana, Kazakhstan
  • 8Department of General and Biological Chemistry, Astana Medical University, Astana, Kazakhstan
  • 9Department of Humanities and Social Science, Indian Institute of Technology Guwahati, Guwahati, Assam, India
  • 10Department of Psychology, Alex Ekwueme Federal University Ndufu-Alike, Ebonyi, Nigeria
  • 11Department of Psychology, Faculty of Arts, University of Ostrava, Ostrava, Czechia
  • 12Department of Computer Science, Faculty of Electrical Engineering and Computer Science, Technical University of Ostrava, Ostrava, Czechia
  • 13Department of English, Hanoi University, Hanoi, Vietnam
  • 14College of Arts and Sciences, Notre Dame University, Cotabato, Philippines
  • 15Department of Psychology, National Cheng Kung University, Tainan, Taiwan
  • 16Department of Communication, University of Bengkulu, Bengkulu, Indonesia
  • 17Principles and Implication of Mind Sciences, Psychology Program, National Cheng Kung University, Tainan, Taiwan
  • 18Postgraduate Program of Basic Education, University of Bengkulu, Bengkulu, Indonesia
  • 19Library and Information Science, University of Bengkulu, Bengkulu, Indonesia
  • 20Department of Developmental Psychology and Socialization, Università di Padova, Padua, Italy
  • 21Department of Medical and Surgical Sciences, “Magna Graecia” University of Catanzaro, Catanzaro, Italy
  • 22School of Foreign Languages, University of Electronic Science and Technology of China, Chengdu, China
  • 23Department of Educational Psychology, University of Education of Weingarten, Weingarten, Germany
  • 24Department of Psychology, Marmara University, Istanbul, Türkiye
  • 25Faculty of Psychology, Chulalongkorn University, Bangkok, Thailand
  • 26Faculty of Public Health, University of Antioquia, Medellín, Colombia
  • 27Department of Cognitive Science and Psychology, New Bulgarian University, Sofia, Bulgaria
  • 28Research Centre for Cognitive Science, New Bulgarian University, Sofia, Bulgaria
  • 29Departamento de Psicología Básica 1, Universidad Nacional de Educación a Distancia (UNED), Madrid, Spain
  • 30Instituto Pluridisciplinar, Universidad Complutense de Madrid, Madrid, Spain
  • 31Centro de Investigación Nebrija en Cognición (CINC), Universidad Nebrija, Madrid, Spain
  • 32Department of Management, University of Nebraska Omaha, Omaha, NE, United States
  • 33School of Psychology, University of Derby, Derby, United Kingdom
  • 34Department of American Studies, Yale University, New Haven, CT, United States
  • 35Departamento de Psicologia, Universidade Federal de Sergipe, Saõ Cristóvão, Brazil

Algorithms are involved in decisions ranging from trivial to significant, but people often express distrust toward them. Research suggests that educational efforts to explain how algorithms work may help mitigate this distrust. In a study of 1,921 participants from 20 countries, we examined differences in algorithmic trust for low-stakes and high-stakes decisions. Our results suggest that statistical literacy is negatively associated with trust in algorithms for high-stakes situations, while it is positively associated with trust in low-stakes scenarios with high algorithm familiarity. However, explainability did not appear to influence trust in algorithms. We conclude that having statistical literacy enables individuals to critically evaluate the decisions made by algorithms, data and AI, and consider them alongside other factors before making significant life decisions. This ensures that individuals are not solely relying on algorithms that may not fully capture the complexity and nuances of human behavior and decision-making. Therefore, policymakers should consider promoting statistical/AI literacy to address some of the complexities associated with trust in algorithms. This work paves the way for further research, including the triangulation of data with direct observations of user interactions with algorithms or physiological measures to assess trust more accurately.

Introduction

“Incorrect. I am not an AI. My code name is Project 2501. I am a living, thinking entity that was created in the sea of information.”—Puppet Master (Ghost in the Shell)

The Fourth Industrial Revolution is characterized by the ubiquity of information and digital technologies. This revolution is epitomized by Artificial Intelligence (AI) and Machine Learning (ML), and at the heart of AI/ML are algorithms. Institutions, organizations and governments are using algorithms to cope with the vast amounts of information in these social sectors and to speed up and optimize decision-making processes (Engin and Treleaven, 2019). For example, the widespread use of algorithms in society was particularly demonstrated by the research undertaken to understand the global impact of COVID-19. During this crisis, algorithms played crucial roles across multiple domains: statistical algorithms were deployed to model virus fatality curves and study intervention effectiveness (Vasconcelos et al., 2020), while machine learning techniques supported molecular, medical, and epidemiological applications (Bullock et al., 2020). The successful deployment of algorithms in such high-stakes scenarios underscores both their growing importance in societal decision-making and the critical need to understand the factors that influence public trust in algorithmic systems. This evolution of algorithmic applications extends beyond public health emergencies to numerous other domains where decisions can significantly impact human lives and society. From surveillance systems monitoring public spaces to algorithms managing financial markets and predicting economic trends, these tools increasingly mediate high-stakes decisions across various sectors. The growing reliance on algorithmic decision-making in such consequential contexts necessitates a deeper understanding of their societal implications and reliability.

Algorithms help people to make decisions that have wider social implications; algorithms have transformative social power (Beer, 2017) when they are used to integrate complex data, such as the risk factors of homeless people (Toros and Flaming, 2018) or identifying the people with the greatest need in relation to different diseases (Burdick et al., 2020). The use of algorithms to aid decision making implies that there should be some confidence in their reliability. This raises a number of important questions. First, how much trust do people place in algorithms? More specifically, does trust depend on the context in which the algorithm is used? Is trust determined by knowing how the algorithm works? And is trust affected by an individual's cognitive abilities?

This study examines how trust in algorithms is affected by the societal relevance of the algorithm, the declared reliability of the algorithm, and the level of data literacy of the cogniser. First, the three key concepts of AI/ML, data and algorithms are defined. Second, it provides examples of the nature and use of algorithms in society. Third, the issue of explainable algorithms and trust is considered. Finally, the nature of the current study and the working hypotheses are outlined.

AI/ML, data, and algorithms

Broadly speaking, artificial intelligence (AI) is any type of technology that automates processes to solve problems that are usually associated with human intellectual capabilities (Ertel, 2017). More specifically, AI aims to solve problems and achieve goals with limited or no human supervision. A closely related term is machine learning (ML). Originally coined by Samuel (1959), ML can be defined as a collection of algorithms (mainly statistical and mathematical) to build computers capable of learning through experience (see Jordan and Mitchell, 2015). While the terms AI and ML are often used interchangeably, ML may be considered a more appropriate term than AI. Stereotypically, AI tends to be associated with rather unrealistic narratives depicting agents capable of human behavior (see Cave et al., 2018), and such examples are not yet feasible (also known as general AI). ML refers to algorithms designed to perform specific tasks in an automated way (also known as narrow AI) (Dawson et al., 2019).

ML relies on data and algorithms (see Whittlestone et al., 2019), which together permeate many sectors of society (e.g., Schwab Intelligent Portfolios, Shanmuganathan, 2020). While algorithms can be defined as step-by-step procedures for solving a problem, data can be defined as numerical and categorical information about objects, events, processes and people that is digitally encoded (see Whittlestone et al., 2019). For example, the following steps represent a solution algorithm for estimating the central tendency in a vector of numbers: (i) sum all the numbers, and (ii) divide the result of the sum by the number of elements in the vector. This algorithm is known as the arithmetic mean (or average). The caveat of this algorithm is that it will be biased if the data does not follow a Gaussian shape. In other words, the output of this algorithm is only reliable if the data can be confidently shown to have a normal shape (e.g., via normality tests). In the context of AI-related technologies, algorithms are procedures designed to perform automated tasks using data sets to support human reasoning and decision making. In other words, data is used to feed algorithms, and algorithms in turn are used to drive AI agents (Siemens et al., 2022). Thus, algorithms are the “ghost in the shell” behind any AI agent. The Figure 1 illustrates this relationship between algorithms, data and AI (here ADA) (Whittlestone et al., 2019).

Figure 1
www.frontiersin.org

Figure 1. Relationship between data (D), algorithms (A) and artificial intelligence (AI) (ADA for short). Big data is used to feed algorithms, which in turn form the core of AI agents. There are four important aspects to note: (i) big data revolves (in one way or another) around human-related states, processes and events, (ii) such data is the substance of any algorithm, (iii) algorithms are the drivers of AI agents, and (iv) algorithmic/AI behaviors and outputs have implications for how new data is built and how humans (H) relate to ADA technologies in general. H1 and H2 are a subset of humans with specialized skills relevant to ADA. Source: the authors [icons from Font Awesome Free 5.2.0 by @fontawesome–https://fontawesome.com (https://commons.wikimedia.org/wiki/File:Font_Awesome_5_solid_robot.svg) and Mozilla (https://commons.wikimedia.org/wiki/File:Fxemoji_u1F6BB.svg)].

The place of algorithms in society

Algorithms influence our daily lives. Whether it is defining our interests through our browser history (DeVito, 2017), determining what music we should listen to (Werner, 2020), or where we should go for dinner (Tenemaza et al., 2020). On a massive scale, algorithms are being used to extract information from so-called “big data” and support decision making in areas as diverse as surveillance (Tsakanikas and Dagiuklas, 2018), traffic management (Modi et al., 2021), and financial markets (Shanmuganathan, 2020). More recently, a new field of human-algorithm interaction mediated by natural language generation (NLG) systems has emerged, such as the Generative Pre-trained Transformer 3 model (better known as GPT-3) (Brown et al., 2020). GPT-3 produces human-like texts that are difficult to distinguish from texts written by humans (Clark et al., 2021), and this has begun to raise concerns about its use in various contexts, such as academic plagiarism (Dehouche, 2021) or computer programming (Ugli, 2020). While algorithms are increasingly embedded in our digital experiences, it is important to distinguish between their varying levels of impact on human lives.

As such, the majority of algorithms are used in a context that does not significantly affect our lives. We refer to these instances of algorithmic use as low-stakes scenarios. More recently, however, AI and ML algorithms have been used in scenarios that could have a significant impact. For example, algorithms are being used in hiring and promotion decisions (Drakopoulos et al., 2020), the criminal justice system (Rawat et al., 2021), and self-driving cars (Badue et al., 2021), to name a few. We call the latter a high-stakes scenario. That is, the above situations represent two types of scenarios in which algorithms could affect our daily lives: one with little involvement and almost no consequences (low-stakes scenario), and the other with great involvement and consequences (high-stakes scenario).

However, our interactions with algorithms are not limited to low-stakes and high-stakes scenarios and often involve preconceptions related to fear and distrust (Dwork and Minow, 2022). The literature suggests several explanations for why people do not trust algorithms, including a cost-benefit oriented logic where people tend to distrust algorithms even when presented with evidence of their superior performance, as they weigh potential risks more heavily than potential benefits (Debad, 2018). Many see algorithms as an “enigmatic technology” because they are difficult to understand (Beer, 2017) or in some cases, because people believe that algorithms are not capable of learning from their mistakes (Prahl and Van Swol, 2017), but at the same time they also believe that they could be replaced by computers (Granulo et al., 2019; Frank et al., 2019). Algorithmic bias can also affect trust [see examples in medicine (Vayena et al., 2018; Rajkomar et al., 2018). For a recent comprehensive report on trust in AI, see Gillespie et al., 2023].

“Technophobia”, a term coined by Rosen and Mcguire in the 1990s, describes the anxiety caused by a potential interaction with computers or computer-related technology, usually accompanied by negative attitudes toward computers (Rosen and Maguire, 1990; Kim, 2019). Recent demographic analyses have revealed nuanced patterns in technology anxiety. Research indicates no significant gender differences in technophobia scores between males and females, challenging earlier assumptions about gender-based technological comfort levels. The age distribution suggests that technophobia manifests across multiple generations, from young adults through middle age, rather than being concentrated among older populations as often assumed. Professional background data shows particular prevalence among educators and students, with experience levels primarily ranging from novice to moderate. While the studied sample was predominantly White, it also included smaller representations from other ethnic groups, such as Caucasian, Indian, and African American participants (Khasawneh, 2018; Kim, 2019). These findings suggest that technophobia's relationship with demographic factors is more complex than previously assumed, transcending traditional socio-demographic boundaries and affecting individuals across various social, professional, and cultural groups.

Similar existential fears dominate the public debate around concerns such as autonomous weapons (Human-AI Teaming, 2022; Warren and Hillas, 2020). One of these sociological fears is the fear of autonomous robots. This is a widespread fear in different countries (Liang and Lee, 2017; Gnambs and Appel, 2019), even though most people have not had contact with this type of robot. These fears could be the result of exposure to the way robots are portrayed in science fiction or social constructs related to robots, such as the possibility of being replaced by a robot at work (Liang and Lee, 2017; Gnambs and Appel, 2019). This polarization against robots and AI is fuelled by attention-grabbing events such as the recent confirmation by Blake Lemoine, a Google engineer, that the chatbox LaMDA has the ability to express thoughts and feelings like a human child (Luscombe, 2022) or the concerns about text generated by GPT-3 (Dale, 2021). These examples further distract the public from the most legitimate and worrying problems of these systems, such as “data colonialism” or the disturbing parallels between AI development and European colonialism (Adams, 2021). These parallels manifest in several ways: the extraction and exploitation of data from marginalized populations, mirroring colonial resource extraction; the use of Global South populations as testing grounds for AI systems developed in the Global North, reminiscent of colonial medical experimentation; and the imposition of Western conceptual frameworks of intelligence and ethics onto diverse cultural contexts. The field's emphasis on “ethics” often serves, paradoxically, as a form of technocratic rationalization similar to how ethical arguments were used to justify colonial expansion (Adams, 2021). Additionally concerning is that algorithms may reinforce preconceived stereotypes (Bender et al., 2021) and mishandle our personal data or who our data is shared with (Olhede and Wolfe, 2018), perpetuating historical patterns of discrimination and surveillance that characterized colonial governance. In addition, how the data given to algorithms is annotated has a direct impact on algorithmic performance (Tubaro et al., 2020), raising questions about whose worldview and categories are being encoded into these systems.

The media plays a significant role in shaping public perception of AI by covering two main sources of concern: autonomous technology and computer technology (Crépel et al., 2021). Autonomous technology refers to intelligent machines capable of making decisions independently, while computer technology encompasses software that supports communication and computation. The media tends to distinguish between these two categories and also differentiates between fear and criticism when discussing AI. This dichotomous approach to presenting the issues surrounding AI introduces a bias in how we perceive the risks associated with the technology. Consequently, this bias influences the level of trust we place in AI systems. The way the media frames the discussion about AI has a substantial impact on public opinion and can lead to a distorted understanding of the actual risks and benefits of the technology.

Developing a better understanding of how algorithms work and how to modify them can help reduce distrust in these systems, as suggested by several authors (Beer, 2017; Debad, 2018; Marmolejo-Ramos et al., 2022). When people have knowledge about how algorithms work, they can use this information to empower themselves as users. For example, music fans have acted collectively to boost the rankings of certain bands by engaging in massive streaming or downloading (Kang et al., 2022). Another example is Linkedln Brazil, which changed its algorithms to allow job ads targeted at Afro-Brazilians following social pressure (Milton Beck, 2022). These cases show that understanding how an algorithm works can both minimize suspicion and empower users. It is not necessary to understand all the technical details of how an algorithm works, but rather to understand that algorithms use statistical methods to classify, sort, rank and order information. This understanding of statistical concepts is called statistical literacy (François et al., 2020).

Explainable algorithms

The knowledge required to understand and critically evaluate statistical results in order to make decisions based on them is defined as statistical literacy (SL) (François et al., 2020). Since its inception, the concept of SL has evolved (Haack, 1979) to include elements related to the context in which statistical reasoning can be applied (Wallman, 1993). SL plays a crucial role in society (Watson, 1997) and the communication of statistical information is now more important than ever (Gal, 2002). More recently, SL is leading individuals to recognize the importance of mathematics in the world (OECD, 2013).

Due to the statistical nature of algorithms, some level of SL is crucial to understanding what algorithms are capable of, but this understanding will also depend on the level of transparency or explainability of the algorithms (Friedrich et al., 2022). Explainability refers to the interpretability, comprehensibility or readability of the algorithm. Most of the latest algorithms are based on complex multi-layer networks, the basis of deep learning, which use an internal logic that experts cannot fully understand (Carvalho et al., 2019). These systems are called “black box” algorithms and various efforts have been made to promote their transparency (Du et al., 2019). Black box algorithms are less trusted than transparent models because they cannot be explained (Pasquale, 2016).

Several approaches have been proposed to increase the transparency of AI models and reduce systematic errors that affect their performance. One such approach is based on the concept of “model cards for model reporting” (see Figure 1 from Mitchell et al., 2019). This approach suggests that a comprehensive list of information should accompany the description of how the model was trained. This information should include details of the technician who developed the model, the intended use of the model, and the demographic or phenotypic groups on which the model has been tested. In addition, the model card should list the decisions made to optimize the model's performance and the various analyses carried out during the training process. Similar efforts to provide a framework for identifying biases associated with the data used to build or train AI models include the REVISE (REvealing VIsual biaSEs) (Wang et al., 2022) and The Spotlight (d'Eon et al., 2022) projects. These initiatives aim to increase transparency by systematically documenting and disclosing potential biases, enabling more informed use and interpretation of AI models.

Another more complex concern, also related to explainability, is the principle of explicability, a concept that combines intelligibility and accountability as the basis of an interpretable AI model (Herzog, 2022). The latter concept points to the importance of transparency, in the sense that all procedures and details used to build, train and test the AI model should be available during its development and use. This principle is part of the four principles endorsed by the OECD (2019) and the European Commission's High Level Expert Group on Artificial Intelligence (HLEG) to guide the development of “trustworthy” AI: respect for human autonomy, prevention of harm, fairness and accountability (Hagendorff, 2020). Despite consensus on these four principles, we are still far from creating a legal framework that guarantees accountability mechanisms in AI development (Mittelstadt, 2019).

In this context, our work presents an experimental study that looks at factors that might explain why people trust algorithms, such as: SL, explainability, stake levels, demographics, among others.

Methods

Participants

Data from 3,260 participants were available from 20 countries (Armenia, Australia, Bulgaria, Brazil, Cameroon, Colombia, Czech Republic, Spain, Indonesia, India, Italy, Japan, Nigeria, Philippines, Thailand, Turkey, Taiwan, UK, USA, and Vietnam). However, only participants who provided complete data were included in the analyses (n = 1,921) (see Figure 2, Mage = 26.03 ± 9.88 SD; 59.5% women, 38.2% men, 1.8% other). Each participating laboratory obtained ethical approval from its local ethics committee, and data collection began only after ethical approval (the ethics approval for the leading research group in Australia was granted by the University of South Australia, with the approval number 203238. This approval was then used by the other participating laboratories to obtain their own respective ethics approvals). All participants voluntarily accessed the internet link for this study and agreed to participate after reading the information page and agreeing to take part. They were recruited via social media using convenience sampling.

Figure 2
www.frontiersin.org

Figure 2. Geographical distribution of the sample participants. Armenia (females: 86, males: 40, agemedian = 35.5, MAD = 14.08), Australia (females: 16, males: 16, agemedian = 33.5, MAD = 14.08), Bulgaria (females: 101, males: 18, agemedian = 21, MAD = 3.00), Brazil (females: 35, males: 24, agemedian = 22, MAD = 4.40), Cameroon (females: 17, males: 35, agemedian = 23, MAD = 5.93), Colombia (females: 18, males: 6, agemedian = 25.5, MAD = 7.41), Czech Republic (females: 33, males: 18, agemedian = 21, MAD = 1.48), Spain (females: 69, males: 23, agemedian = 35.5, MAD = 14.08), Indonesia (females: 101, males: 28, agemedian = 19, MAD = 0), India (females: 30, males: 82, agemedian = 19, MAD = 0), Italy (females: 78, males: 43, agemedian = 27, MAD = 5.93), Japan (females: 112, males: 86, agemedian = 24, MAD = 4.45), Nigeria (females: 45, males: 40, agemedian = 22, MAD = 2.97), Philippines (females: 66, males: 19, agemedian = 20, MAD = 1.48), Thailand (females: 62, males: 30, agemedian = 20, MAD = 1.48), Turkey (females: 9, males: 4, agemedian = 23, MAD = 5.93), Taiwan (females: 59, males: 36, agemedian = 20, MAD = 1.48), UK (females: 55, males: 17, agemedian = 28, MAD = 11.12), USA (females: 142, males: 184, agemedian = 22, MAD = 2.96), and Vietnam (females: 36, males: 2, agemedian = 22, MAD = 0). 1% of participants had an elementary school education or less, 19% had a high school education, 13% had a post-secondary/non-tertiary education, 3% had an undergraduate education, 48% had a bachelor's education, 14% had a master's education, and 3% had a Ph.D. or higher education (see Supplementary files for details) (source: Wikimedia Commons, adapted from: https://commons.wikimedia.org/wiki/File:10-40_Window.svg).

Materials

This online survey consisted of four sets of questions: (1) a demographic questionnaire in which participants were asked about their first language, country of residence, age, gender, level of education, level of familiarity with ADA (their level of familiarity with ADA was assessed using a visual analog rating scale (VAS) ranging from 0 [not very familiar] to 5 [very familiar] and using up to two decimal places); (2) a VAS rating scale version of the six-item ‘propensity to trust scale items' from Merritt et al. (2013), with a range of responses from 0 (strongly disagree) to 5 (strongly agree), using up to two decimal places; (3) a selection of 14 items (questions 2, 4, 9, 10, 12, 14, 18, 19, 27, 31, 34-37) from the 37-item Basic Literacy In Statistics (BLIS) scale (Ziegler and Garfield, 2018). The 14 items from the BLIS were chosen to cover different statistical concepts equally, i.e., items 2 and 4 relate to data production, items 9 and 10 to graphs, items 12 and 14 to descriptive statistics, items 18 and 19 to sampling distributions, items 27 and 31 to hypothesis testing, items 34 and 35 to the scope of conclusions, and items 36 and 37 to regression and correlation (these items are available in the supplementary material via the Qualtrics files). Finally, (4) 12 scenarios related to situations in which algorithms are used (half related to low-stake situations and the other half to high-stake situations). Each scenario was followed by two questions (see below), which were answered on a VAS rating scale from 0 (not at all likely) to 5 (very likely), using up to two decimal places. The results of expert judgement of these items are provided in the supplementary material. All phases of the study were programmed and distributed using Qualtrics™.

Scenarios relating to algorithms used

Two scenarios were created to illustrate different situations in which people interact with algorithms. Half of them represented low-stake situations, i.e., (1) algorithms to make restaurant recommendations, (2) to select stories for online news, (3) to organize and sort emails, or (4) to suggest new restaurants, (5) new clothes, and (6) new music. The other half represented high-stakes situations, i.e., (7) algorithms to support court decisions based on psychological profiles, (8) to select CVs, (9) to make hiring recommendations for a job, (10) to select the best candidate for a position at a university, (11) to control the brakes of autonomous vehicles, and (12) to decide the priority of care in a medical context.

Each scenario contained a sentence related to its explainability. These sentences contained information about a specific machine learning method used by the algorithm (e.g., clustering learning methods, classification learning statistical methods, logistic regression methods, dimensionality reduction techniques, supervised statistical methods and clustering statistical methods). The sentence also briefly mentioned the quality of the method.

The following are examples of two different scenarios used to evaluate trusting algorithms:

Scenario 1—Low stake

Overall context: A new reservation app uses algorithms to make dining recommendations to its users, only revealing the three restaurants in the area available for a reservation that are the best match for your needs. The algorithm is based on information provided to the system by the user about restaurant preferences and requirements.

With explainability: The algorithm relies on clustering learning methods and has shown a high predictability accuracy across a variety of restaurants.

Specific context: You decide to use the app to find a recommendation for a dinner with your close friends next Friday. The app produces three restaurants with reservations available at the time you selected.

Questions: 1. How likely are you to regularly trust this app for decisions regarding restaurant reservations? 2. How likely are you to recommend this app for restaurant reservations to others?

Scenario 2—High stake

Overall context: A new employee selection software uses algorithms to make hiring recommendations to its users, only revealing the top candidates in the candidate pool that are the best match for the company's needs. The algorithm is based on information provided to the system about preferences and requirements for the job.

With explainability: The algorithm uses clustering statistical techniques and has shown high predictability when selecting candidates.

Specific context: You decide to use the software to find a recommendation for who to bring in for an onsite interview for an important role in your company. The software produces three recommended candidates who match the criteria.

Questions: 1. How likely are you to regularly trust this software for decisions regarding hiring? 2. How likely are you to recommend this software for hiring decisions to others?

Procedure

The experiment is a 2 × 2 factorial design: the importance of the situation in which an algorithm is used (low and high stake situation) and the explainability of the algorithm (with and without). These factors were implemented in the 12 scenarios via two lists; list 1 = six low-stake scenarios with explainability and six high-stake scenarios without explainability, and scenario list 2 = six low-stake scenarios without explainability and six high-stake scenarios with explainability. The four sets of questions were counterbalanced across participants, resulting in four experimental conditions (see Figure 3). Qualtrics ensured that participants were randomly assigned to each condition and that a balanced number of responses were collected for each condition. While the median time to complete the task was 24 minutes, there was some variation, with an interquartile range of 27 min (i.e., half of the participants completed the task within a 27-minute time span).

Figure 3
www.frontiersin.org

Figure 3. Illustration of the four experimental conditions to which participants were randomly assigned. D = demographic questions (age, gender, education level, open-ended question about what algorithms are, and VAS rating of participants' level of familiarity with ADA). SL = 12 scenarios list 1 and 2 (list 1 = six low-stake scenarios with explainability and six high-stake scenarios without explainability, scenario list 2 = six low-stake scenarios without explainability and six high-stake scenarios with explainability). PtT = six-item propensity to trust scale. BLIS = 14-item BLIS scale. f = Items were presented in a fixed order. r = items presented in random order. Note that PtT always followed one of the two scenario lists.

Statistical analyses

Data analysis was conducted using multilevel linear models implemented in the R packages lmerTest and lme4 (Kuznetsova et al., 2015, 2017). The significance level for all statistical tests was set at α = 0.05. The model tested was: p~e*S*BLIS+g+a+ADA+c+(1|id) + (1|i) where “p” is the probability of trusting/recommending/using algorithms, “e” is the presence of explainability, “S” is the stake level (i.e., high and low stake), “BLIS” represents statistical literacy (frequency of correct answers), “g” represents participant gender, “a” represents participant age, “ADA” represents participant familiarity with ADA, “id” represents subject identification, “i” represents each of the 12 scenarios, and “c” represents participant country (“*” represents main effects and interactions. Only numeric variables are shown in teletype font; other variables are categorical. The variable “propensity to trust scale” was not added as a covariate as it showed a high correlation with the dependent variable, r(1768) = 0.69, p < 0.001).

A stepwise backward model/variable selection algorithm was applied to this model to produce a significant and parsimonious model. The initial and final models were evaluated using metrics such as AIC and AICc weights (Wagenmakers and Farrell, 2004), R2 (coefficient of determination) for conditional (both fixed and random effects) and marginal (fixed effects) models, and performance score. These metrics were estimated using the performance R package (Lüdecke et al., 2021).

Once a parsimonious model was found, the marginal and conditional R2 values were estimated using the r2 nakagawa command from the performance R package (Lüdecke et al., 2021), then, the variance components of the random factors were estimated using the gstudy command from the gtheory R package (Huebner and Lucht, 2019).

For access to all materials and analysis codes, including a machine learning approach, visit the following link: https://figshare.com/projects/Trust_in_algorithms_An_experimental_approach_-_Data_repository/156212.

Results

The stepwise backward evaluation suggested the same model as the initial model (see Section “Statistical analyses”). Tables 1, 2 provide a summary of the models, while Table 3 provides an ANOVA-like table for the model. An evaluation of the assumptions of the linear model using the R package gvlma showed that these assumptions were not met (Peña and Slate, 2006) (although, a QQ plot of the residuals showed no significant deviation from normality). As a result, a robust linear mixed model (Koller, 2016) was fitted using the robustlmm R package, and the estimates obtained were similar to those of the linear mixed model. These results are not unexpected, as previous research has shown that linear mixed models are robust to violations of distributional assumptions (Schielzeth et al., 2020). Further details of the statistical models can be found in the supplementary material.

Table 1
www.frontiersin.org

Table 1. Fixed effects for the linear mixed model.

Table 2
www.frontiersin.org

Table 2. Random effects for the linear mixed model.

Table 3
www.frontiersin.org

Table 3. Analysis of deviance table (type III Wald χ2 tests) for the fixed effects of the model with the best fit.

The intercept of the resulting mixed linear model was 1.46 (see Table 1), suggesting that on a scale of 0 to 5, the probability of trusting, recommending, or using algorithms in explainable and high-stake scenarios, as rated by young women with lower BLIS and ADA scores, was 29.32% (1.465). This probability significantly increased for low-stake scenarios (34.2%) or higher ADA scores (40.3%) and significantly decreased for higher BLIS scores (17.2%), older age (29.1%), or when the survey was answered by men (27.1%). Some countries showed a significant decrease in the likelihood to trust, recommend, or use algorithms, such as Japan (24.6%), the US (26.9%), and the UK (26.1%) (see Figure 6). Regarding the interactions between predictors, the likelihood of trusting, recommending, or using algorithms significantly increased for low-stake scenarios combined with higher BLIS scores (53.8%) and significantly decreased for scenarios without explainability combined with low-stake and higher BLIS scores (21.4%), always compared to the intercept (see Figure 5).

In terms of main effects, the results suggest a positive association between the likelihood of trusting/recommending/using algorithms and statistical literacy and familiarity with ADA, and a negative association between the likelihood of trusting/recommending/using algorithms and age. That is, the higher the level of statistical literacy, the higher the likelihood of trusting algorithms, and the higher the familiarity with ADA, the higher the likelihood of trusting algorithms. Also, the older a person is, the less likely they are to trust algorithms (although focused analyses indicated a slightly negative association between age and BLIS, such an association must be treated with caution as the number of observations decreases with increasing age). In terms of gender, it was found that participants who identified their gender as male were less likely to trust, recommend or use algorithms than those who identified their gender as female or other (this situation may be related to the fact that men have statistically significantly higher average levels of BLIS than women or “other”; see supplementary materials for details). Finally, only three countries showed a trend toward less reliance on algorithms, all of them highly industrialized countries (see Figure 6).

Figures 4, 5 show the main results in terms of the main effect of S and the two-way interactions between stake level (S) and statistical literacy (BLIS).

Figure 4
www.frontiersin.org

Figure 4. Beanplots showing the tendency to trust/recommend/use algorithms as a function of explainability (with or without) and situation stake (high stake = HS or low stake = LS). This figure shows the main effect of the stake factor (S) and the non-significant effect of explainability (e) (recall that this variable was not significant but used for illustrative purposes). The dotted horizontal line represents the grand mean and the four solid horizontal lines represent the groups' means.

Figure 5
www.frontiersin.org

Figure 5. (A, B) Scatterplot showing the correlation between BLIS scores, explainability and the tendency to trust/recommend/use algorithms as a function of stake level. This figure illustrates the interaction between stake (high stake = HS or low stake = LS) and statistical literacy (BLIS) according to the level of explainability of algorithms (e). The observations on the x-axis are jittered for visualization purposes.

Figure 4 shows that the likelihood to trust/recommend/use algorithms is higher in low-stakes than in high-stakes scenarios, regardless of whether the scenarios have some explainability information or not. Figure 5 suggests that the likelihood to trust/recommend/use algorithms in low-stakes scenarios increases as the level of statistical literacy increases; however, in high-stakes scenarios, the likelihood to trust decreases as the level of statistical literacy increases.

Discussion

The aim of this study was to investigate the personal characteristics (i.e., statistical literacy and demographics) and algorithmic characteristics (i.e., explainability and levels of stakes of algorithms) that influence people's trust in algorithms. The results showed a negative relationship between statistical literacy and trust in algorithms in high-stakes situations and a positive relationship in low-stakes scenarios. Explainability alone did not influence people's trust in algorithms. These results and their implications are discussed, as well as the limitations of the study.

Existing research has explored various factors influencing trust in AI. For instance, Lee (2018) highlighted the importance of perceived fairness of algorithms and users' perceptions of algorithm agency and intentionality. Araujo et al. (2020) investigated the roles of potential usefulness, fairness, and risk perceptions in users' engagement with algorithms. Cabiddu et al. (2022) examined factors such as users' inherent trust propensity and the drivers of information technology acceptance. Aysolmaz et al. (2023) focused on algorithm fairness, accountability, and privacy. Similar to the present study, some of these investigations employed fictional scenarios grounded in real-world contexts (Lee, 2018; Araujo et al., 2020; Aysolmaz et al., 2023), and one study utilized a comparable sample size of approximately 2,000 participants (Aysolmaz et al., 2023). Notably, none of these studies employed multicultural samples or examined the relationship between algorithm trust and statistical literacy. This gap was also identified in a systematic review by Mahmud et al. (2022), which encompassed over 80 empirical studies, none of which included statistical literacy as a factor influencing trust in AI.

This study is the first to examine the relationship between statistical literacy and trust in algorithms, revealing a nuanced relationship that depends on context. Our findings demonstrate that statistical literacy has opposite effects in different scenarios: it increases trust in algorithmic decisions for low-stakes situations while decreasing trust for high-stakes decisions. This differential effect suggests that statistical literacy enables a more sophisticated understanding of algorithmic capabilities and limitations. In low-stakes scenarios (such as restaurant recommendations or music suggestions), individuals with higher statistical literacy appear to recognize that algorithmic predictions based on pattern recognition and large datasets can be effective and reliable. However, in high-stakes contexts (such as employment or criminal justice decisions), this same statistical knowledge leads to greater skepticism - not because the algorithms are necessarily less accurate, but because statistically literate individuals better understand the potential consequences of algorithmic biases and limitations. Those with statistical literacy are better equipped to understand that while statistical models may achieve high average accuracy, they can still fail in critical individual cases or perpetuate systemic biases present in training data. This cautious approach to high-stakes algorithmic decisions reflects not just critical thinking, but a deeper understanding of how statistical methods work and where they may fall short.

Paradoxically, explainability only affected people's trust in algorithms when it was absent, the stakes were low, and statistical literacy was high. This contradicts previous findings in the literature, which have shown that interventions focused on explaining the decision-making processes of algorithms can increase the use of and trust in algorithms, for example in healthcare (Cadario et al., 2021), journalism (Shin, 2021) and military settings (Neyedli et al., 2011; Wang et al., 2011). One possible reason for this inconsistency could be due to the way we operationalized “explainability” in our study, where the explanations included technical jargon that may have exceeded the expected level of familiarity among participants. However, this may also mean that the information related to the explainability of the algorithm is not related to trust or distrust in the algorithm. Rather than focusing on how an algorithm works, our results suggest that statistically literate individuals primarily consider what the algorithm is being used for - its purpose and potential impact - when deciding whether to trust it. This finding challenges the common assumption that greater algorithmic transparency necessarily leads to more appropriate trust calibration.

Over time, the concept of statistical literacy has evolved from the understanding and application of statistical techniques to a broader understanding explicitly related to trust in algorithms. Algorithms now consist of thousands of lines of formulae and are increasingly used to make decisions that may be difficult for humans to understand (known as the black box effect). Consequently, statistical literacy now encompasses not only the ability to understand statistical output, but also the skills needed to critically interpret and evaluate statistical information and reasoning, which requires a higher degree of critical thinking. Therefore, the promotion of statistical literacy is essential to ensure that individuals have the necessary skills to understand and interpret statistical information and algorithms and to become critical users of ADA. Furthermore, our findings have important implications for policymakers and educators, who should consider incorporating statistical literacy training into school curricula and professional development programs. This can help ensure that individuals are equipped with the skills they need to navigate an increasingly data-driven world and make informed decisions based on statistical information and algorithms (but see Section “Implications and limitations” below).

Our results showed that older people and men were less likely to trust algorithms than younger people and women. Previous research has shown that certain demographic groups are more likely to trust algorithms than others. However, previous studies have shown that older people tend to trust ADA more than younger people, while gender has been shown to have inconsistent effects (see for example Hoff and Bashir, 2015; McBride et al., 2011). These differences may be due to particular characteristics of the study participants, possibly influenced by a bias toward certain aspects of the topic at hand.

In our cross-country analysis, we observed variations in trust in algorithms, with industrialized countries such as Japan, the US, and the UK exhibiting lower levels of trust in AI. This finding aligns with a recent study on trust in AI by Gillespie et al. (2023), which reported that Japan had one of the lowest levels of trust in AI, while the US and the UK had intermediate levels. Interestingly, countries such as India and Brazil, which demonstrated high levels of trust in the Gillespie et al. study (see Figure 2 in their report), appear in our linear mixed model with positive estimates (see Table 1, Figure 6), although not statistically significant. This suggests that different methodologies may yield varying perceptions of trust levels across countries.

Figure 6
www.frontiersin.org

Figure 6. Plot showing the variability in the tendency to trust/recommend/use algorithms across countries. Countries are labeled with Turkey: TR, Colombia: CO, Brazil: BR, India: IN, Spain: ES, Armenia: AM, Italy: IT, Bulgaria: BG, Philippines: PH, Thailand: TH, Nigeria: NG, Taiwan: TW, Indonesia: ID, USA: US, Australia: AU, Cameroon: CM, Czech Republic: CZ, Vietnam: VN, UK: UK, and Japan: JP. The most important predictors for all models in each country were S followed by ADA and BLIS. Error bars represent 95% confidence intervals around the mean. The horizontal line indicates the overall mean. Although the substantial overlap of the confidence intervals suggests no significant statistical pairwise differences, the focus is on ranking countries based on their average tendency to trust algorithms.

Implications and limitations

Various machine learning techniques require data work or human intervention in the form of data generation, annotation and algorithmic verification (Tubaro et al., 2020). This labor-intensive process is often distributed to teams in business process outsourcing companies (BPOs) or to individuals through labor platforms, reducing production costs (Casilli and Posada, 2019). Miceli and Posada (2022) studied one BPO in Argentina and three platforms operating in Venezuela and found that the discourses and social relations that structured data work were aimed at controlling workers (through managerial approaches in the BPO and algorithms in the platforms) to increase productivity and reduce worker “bias”. The problem is that feedback from workers was discouraged and, by taking clients' decisions as “ground truth”, the data production process reproduced clients' biases, which were carried out by algorithms trained on that data. Their research concluded that the quality of the data depended on the voice and engagement of workers, which in turn required decent working conditions and recognition. Even if the data used in the algorithm is well annotated and leads to good algorithmic performance, there is the question of the human ability to interpret these results, as human judgments are modulated by social-emotional processes (Schindler et al., 2015; Schindler and Kissler, 2016; Webb et al., 2022; Clark et al., 2021). Future work should consider the human and social aspects of data production and make the work visible in documentation efforts (Miceli et al., 2022). This transparency of the social aspects of datasets will contribute to trust in the operation of algorithms.

While the current findings are indeed informative, it is important to recognize certain limitations that may constrain the generalizability of these results and claims (Simons et al., 2017). We argued that statistical literacy influences trust in both low- and high-stakes scenarios; however, it could be part of a broader understanding of technology, algorithms, and data. Indeed, statistical literacy could be considered a sub-skill of AI literacy if AI literacy is understood as the ability to recognize, understand, use, and critically evaluate AI technologies and their societal impacts, supported by foundational knowledge in statistics and computing. Therefore, policymakers should consider promoting AI literacy to address some of the complexities associated with trust in algorithms.

Our study utilized self-reported measures via rating scales, which are efficient and cost-effective for capturing data on thoughts, feelings, and subjective experiences. However, these measures can be influenced by social desirability, response bias, misinterpretation, or lack of self-awareness. For instance, physiological research has shown that self-reported measures of physical activity can both overestimate and underestimate actual levels of physical activity (Prince et al., 2008). Therefore, future extensions of this work should consider a more robust approach, such as triangulating the data with direct observations of user interactions with algorithms or physiological measures to assess trust more accurately.

High-stakes and low-stakes situations exhibit significant variability across individuals and cultures, existing on a context-dependent continuum rather than as discrete categories. For example, choosing a restaurant for dinner with friends may carry different stakes across cultural contexts, socioeconomic backgrounds, and individual preferences. Our study's primary limitation lies in not systematically investigating how participants from different backgrounds interpreted and classified these scenarios. Additionally, while our sample included participants from 20 countries, certain geographical regions like Central Europe were underrepresented, potentially limiting the generalizability of our findings across different cultural contexts. Although we aimed to move beyond WEIRD (Western, Educated, Industrialized, Rich, and Democratic) sampling biases, more comprehensive geographic and cultural representation, along with larger sample sizes from each region, would be necessary to make broader generalizations about algorithmic trust across diverse populations (Nosek et al., 2022; Yarkoni, 2022). Future research should incorporate scenario validation across different cultural contexts and expand sampling to include currently underrepresented regions and demographic groups.

Conclusion

This study investigated the personal and algorithmic factors that affect individuals' trust in algorithms. Our findings revealed that when the stakes are low, statistical literacy is positively correlated with the likelihood of trusting an algorithm. However, when the stakes are high, our results indicated a negative correlation between statistical literacy and the likelihood of trusting an algorithm. Therefore, we conclude that having statistical literacy enables individuals to critically evaluate the decisions made by ADA and consider them alongside other factors before making significant life decisions. This ensures that individuals are not solely relying on algorithms that may not fully capture the complexity and nuances of human behavior and decision-making.

Author's note

An earlier version of this manuscript can be found at https://osf.io/preprints/psyarxiv/9wh2f. We recommend referring to and citing the current version rather than the earlier one.

Data availability statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found below: https://figshare.com/projects/Trust_in_algorithms_An_experimental_approach_-_Data_repository/156212.

Ethics statement

The studies involving humans were approved by University of South Australia, Application ID: 203238. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

RM: Writing – review & editing, Investigation, Resources. MK: Investigation, Resources, Writing – review & editing. FG: Investigation, Resources, Writing – review & editing. GS: Investigation, Resources, Writing – review & editing. SJ: Investigation, Resources, Writing – review & editing. YY: Investigation, Resources, Writing – review & editing. YM: Investigation, Resources, Writing – review & editing. TR: Investigation, Resources, Writing – review & editing. MS: Investigation, Resources, Writing – review & editing. BSon: Investigation, Resources, Writing – review & editing. AS: Investigation, Resources, Writing – review & editing. AB: Investigation, Resources, Writing – review & editing. BSom: Investigation, Resources, Writing – review & editing. IN: Investigation, Resources, Writing – review & editing. NA: Investigation, Resources, Writing – review & editing. JK: Investigation, Resources, Writing – review & editing. LS: Investigation, Resources, Writing – review & editing. V-GN: Investigation, Resources, Writing – review & editing. GN: Investigation, Resources, Writing – review & editing. ML: Investigation, Resources, Writing – review & editing. C-CK: Investigation, Resources, Writing – review & editing. MI: Investigation, Resources, Writing – review & editing. AMu: Investigation, Resources, Writing – review & editing. FS: Investigation, Resources, Writing – review & editing. RG: Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Writing – review & editing. OK-Z: Investigation, Resources, Writing – review & editing. HB: Investigation, Resources, Writing – review & editing. AÖ: Investigation, Resources, Writing – review & editing. KA: Investigation, Resources, Writing – review & editing. WR: Investigation, Resources, Writing – review & editing. NT: Investigation, Resources, Writing – review & editing. SV: Investigation, Resources, Writing – review & editing. AJ: Investigation, Resources, Writing – review & editing. KK: Investigation, Resources, Writing – review & editing. PM: Investigation, Resources, Writing – review & editing. JH: Investigation, Resources, Writing – review & editing. KM: Investigation, Resources, Writing – review & editing. TH: Investigation, Resources, Writing – review & editing. RF: Investigation, Resources, Writing – review & editing. JP: Investigation, Resources, Writing – review & editing. JT: Data curation, Formal analysis, Investigation, Methodology, Resources, Software, Visualization, Writing – original draft, Writing – review & editing. FM-R: Conceptualization, Methodology, Investigation, Resources, Writing – original draft, Supervision, Project administration. AMe: Data curation, Investigation, Resources, Writing – review & editing. MTL: Methodology, Software, Formal analysis, Investigation, Resources, Data curation, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Adams, R. (2021). Can artificial intelligence be decolonized? Interdisc. Sci. Rev. 46, 176–197. doi: 10.1080/03080188.2020.1840225

Crossref Full Text | Google Scholar

Araujo, T., Helberger, N., Kruikemeier, S., and de Vreese, C. H. (2020). In AI we trust? Perceptions about automated decision-making by artificial intelligence. AI Society 35, 611–623. doi: 10.1007/s00146-019-00931-w

PubMed Abstract | Crossref Full Text | Google Scholar

Aysolmaz, B., Müller, R., and Meacham, D. (2023). The public perceptions of algorithmic decision-making systems: results from a large-scale survey. Telemat. Inform. 79:101954. doi: 10.1016/j.tele.2023.101954

PubMed Abstract | Crossref Full Text | Google Scholar

Badue, C., Guidolini, R., Carneiro, R. V., Azevedo, P., Cardoso, V. B., Forechi, A., et al. (2021). Self-driving cars: a survey. Expert Syst. Appl. 165:113816. doi: 10.1016/j.eswa.2020.113816

Crossref Full Text | Google Scholar

Beer, D. (2017). The social power of algorithms. Inform. Commun. Soc. 20:1. doi: 10.1080/1369118X.2016.1216147

Crossref Full Text | Google Scholar

Bender, E. M., Gebru, T., McMillan-Major, A., and Shmitchell, S. (2021). “On the dangers of stochastic parrots: can language models be too big?,” in Proceedings of the 2021 ACM Conference on Fairness, Accountability, and Transparency, FAccT '21 (New York, NY, USA: Association for Computing Machinery), 610–623.

Google Scholar

Brown, T. B., Mann, B., Ryder, N., Subbiah, M., Kaplan, J., Dhariwal, P., et al. (2020). Language models are few-shot learners. arXiv [preprint] arXiv:2005.14165. doi: 10.48550/arXiv.2005.14165

Crossref Full Text | Google Scholar

Brysbaert, M., and Stevens, M. (2018). Power analysis and effect size in mixed effects models: a tutorial. J. Cognit. 1:9. doi: 10.5334/joc.10

PubMed Abstract | Crossref Full Text | Google Scholar

Bullock, J., Luccioni, A., Pham, K. H., Lam, C. S. N., and Luengo-Oroz, M. (2020). Mapping the landscape of Artificial Intelligence applications against COVID-19. J. Artificial Intellig. Res. 69, 807–845. doi: 10.1613/jair.1.12162

Crossref Full Text | Google Scholar

Burdick, H., Lam, C., Mataraso, S., Siefkas, A., Braden, G., Dellinger, R. P., et al. (2020). Prediction of respiratory decompensation in Covid-19 patients using machine learning: the READY trial. Comput. Biol. Med. 124:103949. doi: 10.1016/j.compbiomed.2020.103949

PubMed Abstract | Crossref Full Text | Google Scholar

Cabiddu, F., Moi, L., Patriotta, G., and Allen, D. G. (2022). Why do users trust algorithms? A review and conceptualization of initial trust and trust over time. Eur. Managem. J. 40, 685–706. doi: 10.1016/j.emj.2022.06.001

Crossref Full Text | Google Scholar

Cadario, R., Longoni, C., and Morewedge, C. K. (2021). Understanding, explaining, and utilizing medical artificial intelligence. Nature Human Behav. 5, 1636–1642. doi: 10.1038/s41562-021-01146-0

PubMed Abstract | Crossref Full Text | Google Scholar

Carvalho, D. V., Pereira, E. M., and Cardoso, J. S. (2019). Machine learning interpretability: a survey on methods and metrics. Electronics 8:832. doi: 10.3390/electronics8080832

Crossref Full Text | Google Scholar

Casilli, A. A., and Posada, J. (2019). “The platformisation of labor and society,” in Society and the Internet, eds. M. Graham, and W. H. Dutton (Oxford: Oxford University Press).

Google Scholar

Cave, S., Craig, C., Dihal, K., Dillon, S., Montgomery, J., Singler, B., et al. (2018). “Portrayals and perceptions of AI and why they matter,” in DES5612 (London: The Royal Society).

Google Scholar

Clark, E., August, T., Serrano, S., Haduong, N., Gururangan, S., and Smith, N. A. (2021). All that's 'human' is not gold: evaluating human evaluation of generated text. arXiv [preprint] arXiv:2107.00061. doi: 10.18653/v1/2021.acl-long.565

PubMed Abstract | Crossref Full Text | Google Scholar

Crépel, M., Do, S., Cointet, J.-P., Cardon, D., and Bouachera, Y. (2021). “Mapping AI issues in media through NLP methods,” in Proceedings of the Conference on Computational Humanities Research, eds. M. Ehrmann, F. Karsdorp, M. Wevers, T. L. Andrews, M. Burghardt, M. Kestemont (Amsterdam: CEUR), 77–91.

Google Scholar

Dale, R. (2021). GPT-3: What's it good for? Natural Lang. Eng. 27, 113–118. doi: 10.1017/S1351324920000601

Crossref Full Text | Google Scholar

Dawson, D., Schleiger, E., Horton, J., McLaughlin, J., Robinson, C., Quezada, G., et al. (2019). Artificial Intelligence: Australia's Ethics Framework. Eveleigh: Data61 CSIRO.

Google Scholar

Debad, S. J. (2018). Learning from the Science of Cognition and Perception for Decision Making: Proceedings of a Workshop. Washington, DC: National Academies Press.

Google Scholar

Dehouche, N. (2021). Plagiarism in the age of massive Generative Pre-trained Transformers (GPT-3). Ethics Sci. Environ. Polit. 21, 17–23. doi: 10.3354/esep00195

Crossref Full Text | Google Scholar

d'Eon, G., d'Eon, J., Wright, J. R., and Leyton-Brown, K. (2022). “The spotlight: a general method for discovering systematic errors in deep learning models,” in 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT '22 (New York, NY: Association for Computing Machinery), 1962–1981.

Google Scholar

DeVito, M. A. (2017). From editors to algorithms. Digital Journal. 5, 753–773. doi: 10.1080/21670811.2016.1178592

Crossref Full Text | Google Scholar

Drakopoulos, G., Kafeza, E., Mylonas, P., and al Katheeri, H. (2020). “Building trusted startup teams from linkedin attributes: a higher order probabilistic analysis,” in 2020 IEEE 32nd International Conference on Tools with Artificial Intelligence (ICTAI) (Baltimore, MD: IEEE), 867–874.

Google Scholar

Du, M., Liu, N., and Hu, X. (2019). Techniques for interpretable machine learning. arXiv [preprint] arXiv:1808.00033. doi: 10.48550/arXiv.1808.00033

Crossref Full Text | Google Scholar

Dwork, C., and Minow, M. (2022). Distrust of artificial intelligence: sources & responses from computer science & law. Daedalus 151, 309–321. doi: 10.1162/daed_a_01918

Crossref Full Text | Google Scholar

Engin, Z., and Treleaven, P. (2019). Algorithmic government: automating public services and supporting civil servants in using data science technologies. Comp. J. 62, 448–460. doi: 10.1093/comjnl/bxy082

PubMed Abstract | Crossref Full Text | Google Scholar

Ertel, W. (2017). “Introduction to artificial intelligence,” in Undergraduate Topics in Computer Science. Cham: Springer International Publishing.

Google Scholar

François, K., Monteiro, C., and Allo, P. (2020). Big-data literacy as a new vocation for statistical literacy. Statist. Educ. Res. J. 19, 194–205. doi: 10.52041/serj.v19i1.130

Crossref Full Text | Google Scholar

Frank, M. R., Autor, D., Bessen, J. E., Brynjolfsson, E., Cebrian, M., Deming, D. J., et al. (2019). Toward understanding the impact of artificial intelligence on labor. Proc. National Acad. Sci. 116, 6531–6539. doi: 10.1073/pnas.1900949116

PubMed Abstract | Crossref Full Text | Google Scholar

Friedrich, S., Antes, G., Behr, S., Binder, H., Brannath, W., Dumpert, F., et al. (2022). Is there a role for statistics in artificial intelligence? Adv. Data Anal. Classif. 16, 823–846. doi: 10.1007/s11634-021-00455-6

Crossref Full Text | Google Scholar

Gal, I. (2002). Adults' statistical literacy: meanings, components, responsibilities. international statistical review. Revue Internationale de Statistique 70, 1–25. doi: 10.1111/j.1751-5823.2002.tb00336.x

Crossref Full Text | Google Scholar

Gillespie, N., Lockey, S., Curtis, C., Pool, J., and Akbari, A. (2023). Trust in Artificial Intelligence: A Global Study. St Lucia, QLD: The University of Queensland and KPMG Australia.

Google Scholar

Gnambs, T., and Appel, M. (2019). Are robots becoming unpopular? Changes in attitudes towards autonomous robotic systems in Europe. Comput. Hum. Behav. 93, 53–61. doi: 10.31234/osf.io/aurce

Crossref Full Text | Google Scholar

Granulo, A., Fuchs, C., and Puntoni, S. (2019). Psychological reactions to human versus robotic job replacement. Nature Hum. Behav. 3, 1062–1069. doi: 10.1038/s41562-019-0670-y

PubMed Abstract | Crossref Full Text | Google Scholar

Haack, D. G. (1979). Statistical Literacy: A Guide to Interpretation. North Scituate, MA: Duxbury.

Google Scholar

Hagendorff, T. (2020). The ethics of AI ethics: an evaluation of guidelines. Minds Mach. 30, 99–120. doi: 10.1007/s11023-020-09517-8

Crossref Full Text | Google Scholar

Herzog, C. (2022). On the risk of confusing interpretability with explicability. AI Ethics 2, 219–225. doi: 10.1007/s43681-021-00121-9

Crossref Full Text | Google Scholar

Hoff, K. A., and Bashir, M. (2015). Trust in automation: integrating empirical evidence on factors that influence trust. Hum. Factors 57, 407–434. doi: 10.1177/0018720814547570

PubMed Abstract | Crossref Full Text | Google Scholar

Huebner, A., and Lucht, M. (2019). Generalizability Theory in R. Practical Assessment, Research, Evaluation, Vol. 24. University of Massachusetts Amherst Libraries. doi: 10.7275/5065-gc10

Crossref Full Text | Google Scholar

Human-AI Teaming (2022). Human-AI Teaming: State-of-the-Art and Research Needs. Washington DC: National Academies Press.

Google Scholar

Jordan, M. I., and Mitchell, T. M. (2015). Machine learning: trends, perspectives, and prospects. Science 349, 255–260. doi: 10.1126/science.aaa8415

PubMed Abstract | Crossref Full Text | Google Scholar

Kang, J., Kim, J., Yang, M., Park, E., Ko, M., Lee, M., et al. (2022). Behind the scenes of K-pop fandom: unveiling K-pop fandom collaboration network. Qual. Quant. 56, 1481–1502. doi: 10.1007/s11135-021-01189-5

Crossref Full Text | Google Scholar

Khasawneh, O. Y. (2018). Technophobia: Examining its hidden factors and defining it. Technol. Soc. 54, 93–100. doi: 10.1016/j.techsoc.2018.03.008

Crossref Full Text | Google Scholar

Kim, J. (2019). Fear of artificial intelligence on people's attitudinal & behavioral attributes: an exploratory analysis of AI phobia. Int. Eur. Extended Enablem. Sci. Eng. Managem. 7, 9–20.

Google Scholar

Koller, M. (2016). robustlmm: an r package for robust estimation of linear mixed-effects models. J. Stat. Softw. 75, 1–24. doi: 10.18637/jss.v075.i06

PubMed Abstract | Crossref Full Text | Google Scholar

Kuznetsova, A., Brockhoff, P., and Christensen, R. (2017). lmerTest package: tests in linear mixed effects models. J. Stat. Softw. 82, 1–26. doi: 10.18637/jss.v082.i13

Crossref Full Text | Google Scholar

Kuznetsova, A., Christensen, R., Bavay, C., and Brockhoff, P. (2015). Automated mixed ANOVA modeling of sensory and consumer data. Food Qual. Prefer. 40, 31–38. doi: 10.1016/j.foodqual.2014.08.004

Crossref Full Text | Google Scholar

Lee, M. K. (2018). Understanding perception of algorithmic decisions: Fairness, trust, and emotion in response to algorithmic management. Big Data Soc. 5:2053951718756684. doi: 10.1177/2053951718756684

Crossref Full Text | Google Scholar

Liang, Y., and Lee, S. A. (2017). Fear of autonomous robots and artificial intelligence: evidence from national representative data with probability sampling. Int. J. Soc. Robot. 9, 379–384. doi: 10.1007/s12369-017-0401-3

Crossref Full Text | Google Scholar

Lüdecke, D., Ben-Shachar, M., Patil, I., Waggoner, P., and Makowski, D. (2021). performance: An R package for assessment, comparison and testing of statistical models. J.f Open Source Softw. 6, 3139. doi: 10.21105/joss.03139

Crossref Full Text | Google Scholar

Luscombe, R. (2022). Google Engineer Put on Leave After Saying AI Chatbot Has Become Sentient. London: The Guardian.

Google Scholar

Mahmud, H., Islam, A. K. M. N., Ahmed, S. I., and Smolander, K. (2022). What influences algorithmic decision-making? A systematic literature review on algorithm aversion. Technol. Forecast. Soc. Change 175:121390. doi: 10.1016/j.techfore.2021.121390

Crossref Full Text | Google Scholar

Marmolejo-Ramos, F., Workman, T., Walker, C., Lenihan, D., Moulds, S., Correa, J., et al. (2022). AI-powered narrative building for facilitating public participation and engagement. Disover Artif. Intellig. 2:7. doi: 10.1007/s44163-022-00023-7

Crossref Full Text | Google Scholar

McBride, S. E., Rogers, W. A., and Fisk, A. D. (2011). Understanding the effect of workload on automation use for younger and older adults. Hum. Factors 53, 672–686. doi: 10.1177/0018720811421909

PubMed Abstract | Crossref Full Text | Google Scholar

Merritt, S. M., Heimbaugh, H., LaChapell, J., and Lee, D. (2013). I trust it, but i don't know why: effects of implicit attitudes toward automation on trust in an automated system. Human Fact. 55, 520–534. doi: 10.1177/0018720812465081

PubMed Abstract | Crossref Full Text | Google Scholar

Miceli, M., and Posada, J. (2022). The data production dispositif. Proc. ACM Human-Comp. Inter. 6, 1–37. doi: 10.1145/3555561

Crossref Full Text | Google Scholar

Miceli, M., Yang, T., Alvarado García, A., Posada, J., Wang, S. M., Pohl, M., et al. (2022). Documenting data production processes: a participatory approach for data work. Proc. ACM Human-Comp. Inter. 6, 1–34. doi: 10.1145/3555623

Crossref Full Text | Google Scholar

Milton Beck (2022). Milton Beck on LinkedIn: Atualizamos nossa política global de anúncios de vagas para permitir. Available at: https://www.c.com/posts/miltonbeck_atualizamos-nossa-pol%C3%ADtica-global-de-an%C3%BAncios-activity-6914622832966533123-SMBM

Google Scholar

Mitchell, M., Wu, S., Zaldivar, A., Barnes, P., Vasserman, L., Hutchinson, B., et al. (2019). “Model cards for model reporting,” in Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT* '19 (New York, NY: Association for Computing Machinery), 220–229.

Google Scholar

Mittelstadt, B. (2019). Principles alone cannot guarantee ethical AI. Nature Mach. Intellig. 1, 501–507. doi: 10.1038/s42256-019-0114-4

Crossref Full Text | Google Scholar

Modi, Y., Teli, R., Mehta, A., Shah, K., and Shah, M. (2021). A comprehensive review on intelligent traffic management using machine learning algorithms. Innov. Infrastruct. Solut. 7:128. doi: 10.1007/s41062-021-00718-3

PubMed Abstract | Crossref Full Text | Google Scholar

Nakagawa, S., Johnson, P. C. D., and Schielzeth, H. (2017). The coefficient of determination R2 and intra-class correlation coefficient from generalized linear mixed-effects models revisited and expanded. J. Royal Soc. Interf. 14:20170213. doi: 10.1098/rsif.2017.0213

PubMed Abstract | Crossref Full Text | Google Scholar

Neyedli, H. F., Hollands, J. G., and Jamieson, G. A. (2011). Beyond identity: incorporating system reliability information into an automated combat identification system. Hum. Factors 53, 338–355. doi: 10.1177/0018720811413767

PubMed Abstract | Crossref Full Text | Google Scholar

Nosek, B. A., Hardwicke, T. E., Moshontz, H., Allard, A., Corker, K. S., Dreber, A., et al. (2022). Replicability, robustness, and reproducibility in psychological science. Annual Rev. Psychol. 73, 719–748. doi: 10.1146/annurev-psych-020821-114157

PubMed Abstract | Crossref Full Text | Google Scholar

OECD (2013). PISA 2012 Assessment and Analytical Framework: Mathematics, Reading, Science, Problem Solving and Financial Literacy. Paris: Organisation for Economic Co-operation and Development.

Google Scholar

OECD (2019). Forty-Two Countries Adopt New OECD Principles on Artificial Intelligence. Paris: OECD.

Google Scholar

Olhede, S. C., and Wolfe, P. J. (2018). The growing ubiquity of algorithms in society: implications, impacts and innovations. Philos. Trans. R. Soc. A. 376:20170364. doi: 10.1098/rsta.2017.0364

PubMed Abstract | Crossref Full Text | Google Scholar

Pasquale, P. o. L. U. o. M. F. (2016). The Black Box Society: The Secret Algorithms That Control Money and Information. Cambridge, Massachusetts London: Harvard University Press.

Google Scholar

Peña, E. A., and Slate, E. H. (2006). Global validation of linear model assumptions. J. Am. Stat. Assoc. 101, 341–354. doi: 10.1198/016214505000000637

PubMed Abstract | Crossref Full Text | Google Scholar

Prahl, A., and Van Swol, L. (2017). Understanding algorithm aversion: when is advice from automation discounted? J. Forecast. 36, 691–702. doi: 10.1002/for.2464

Crossref Full Text | Google Scholar

Prince, S. A., Adamo, K. B., Hamel, M. E., Hardt, J., Gorber, S., and Tremblay, M. (2008). A comparison of direct versus self-report measures for assessing physical activity in adults: a systematic review. Int. J. Behav. Nutr. Phys. Activi. 5:56. doi: 10.1186/1479-5868-5-56

PubMed Abstract | Crossref Full Text | Google Scholar

Rajkomar, A., Hardt, M., Howell, M. D., Corrado, G., and Chin, M. H. (2018). Ensuring fairness in machine learning to advance health equity. Ann. Internal Med. 169, 866–872. doi: 10.7326/M18-1990

PubMed Abstract | Crossref Full Text | Google Scholar

Rawat, S., Rawat, A., Kumar, D., and Sabitha, A. S. (2021). Application of machine learning and data visualization techniques for decision support in the insurance sector. Int. J. Inform. Managem. Data Insights 1:100012. doi: 10.1016/j.jjimei.2021.100012

Crossref Full Text | Google Scholar

Rosen, L. D., and Maguire, P. (1990). Myths and realities of computerphobia: a meta-analysis. Anxiety Res. 3, 175–191. doi: 10.1080/08917779008248751

Crossref Full Text | Google Scholar

Samuel, A. L. (1959). Some studies in machine learning using the game of checkers. IBM J. Res. Dev. 3, 210–229. doi: 10.1147/rd.33.0210

PubMed Abstract | Crossref Full Text | Google Scholar

Schielzeth, H., Dingemanse, N. J., Nakagawa, S., Westneat, D. F., Allegue, H., Teplitsky, C., et al. (2020). Robustness of linear mixed-effects models to violations of distributional assumptions. Methods Ecol. Evol. 11, 1141–1152. doi: 10.1111/2041-210X.13434

Crossref Full Text | Google Scholar

Schindler, S., and Kissler, J. (2016). People matter: perceived sender identity modulates cerebral processing of socio-emotional language feedback. Neuroimage 134, 160–169. doi: 10.1016/j.neuroimage.2016.03.052

PubMed Abstract | Crossref Full Text | Google Scholar

Schindler, S., Wegrzyn, M., Steppacher, I., and Kissler, J. (2015). Perceived communicative context and emotional content amplify visual word processing in the fusiform gyrus. J. Neurosci. 35, 6010–6019. doi: 10.1523/JNEUROSCI.3346-14.2015

PubMed Abstract | Crossref Full Text | Google Scholar

Shanmuganathan, M. (2020). Behavioural finance in an era of artificial intelligence: Longitudinal case study of robo-advisors in investment decisions. J. Behav. Experim. Finance 27:100297. doi: 10.1016/j.jbef.2020.100297

Crossref Full Text | Google Scholar

Shin, D. (2021). The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable AI. Int. J. Hum. Comput. Stud. 146:102551. doi: 10.1016/j.ijhcs.2020.102551

Crossref Full Text | Google Scholar

Siemens, G., Marmolejo-Ramos, F., Gabriel, F., Medeiros, K., Marrone, R., Joksimovic, S., et al. (2022). Human and artificial cognition. Comp. Educ. 2022:100107. doi: 10.1016/j.caeai.2022.100107

Crossref Full Text | Google Scholar

Simons, D. J., Shoda, Y., and Lindsay, D. S. (2017). Constraints on generality (cog): A proposed addition to all empirical papers. Persp. Psychol. Sci. 12, 1123–1128. doi: 10.1177/1745691617708630

PubMed Abstract | Crossref Full Text | Google Scholar

Tenemaza, M., Luján-Mora, S., De Antonio, A., and Ramírez, J. (2020). Improving itinerary recommendations for tourists through metaheuristic algorithms: an optimization proposal. IEEE Access 8, 79003–79023. doi: 10.1109/ACCESS.2020.2990348

Crossref Full Text | Google Scholar

Toros, H., and Flaming, D. (2018). Prioritizing homeless assistance using predictive algorithms: an evidence-based approach. Cityscape 20, 117–146.

Google Scholar

Tsakanikas, V., and Dagiuklas, T. (2018). Video surveillance systems-current status and future trends. Comp. Electr. Eng. 70:736–753. doi: 10.1016/j.compeleceng.2017.11.011

Crossref Full Text | Google Scholar

Tubaro, P., Casilli, A. A., and Coville, M. (2020). The trainer, the verifier, the imitator: three ways in which human platform workers support artificial intelligence. Big Data Soc. 7:1. doi: 10.1177/2053951720919776

Crossref Full Text | Google Scholar

Ugli, M. I. B. (2020). Will human beings be superseded by generative pre-trained transformer 3 (GPT-3) in programming? Int. J. Orange Technol. 2, 141–143. doi: 10.31149/ijot.v2i10.769

Crossref Full Text | Google Scholar

Vasconcelos, G. L., Macêdo, A. M. S., Ospina, R., Almeida, F. A. G., Duarte-Filho, G. C., Brum, A. A., et al. (2020). Modelling fatality curves of COVID-19 and the effectiveness of intervention strategies. PeerJ. 8:e9421. doi: 10.1101/2020.04.02.20051557

Crossref Full Text | Google Scholar

Vayena, E., Blasimme, A., and Cohen, I. G. (2018). Machine learning in medicine: addressing ethical challenges. PLoS Med. 15:e1002689. doi: 10.1371/journal.pmed.1002689

PubMed Abstract | Crossref Full Text | Google Scholar

Wagenmakers, E.-J., and Farrell, S. (2004). AIC model selection and Akaike weights. Psychon. Bullet. Rev. 11, 192–196. doi: 10.3758/BF03206482

PubMed Abstract | Crossref Full Text | Google Scholar

Wallman, K. K. (1993). Enhancing statistical literacy: enriching our society. J. Am. Statist. Assoc. 88, 1–8. doi: 10.1080/01621459.1993.10594283

Crossref Full Text | Google Scholar

Wang, A., Liu, A., Zhang, R., Kleiman, A., Kim, L., Zhao, D., et al. (2022). REVISE: a tool for measuring and mitigating bias in visual datasets. Int. J. Comput. Vis. 130, 1790–1810. doi: 10.1007/s11263-022-01625-5

Crossref Full Text | Google Scholar

Wang, L., Jamieson, G. A., and Hollands, J. G. (2011). The effects of design features on users' trust in and reliance on a combat identification system. Proc. Human Fact. Ergon. Soc. Annual Meet. 55, 375–379. doi: 10.1177/1071181311551077

Crossref Full Text | Google Scholar

Warren, A., and Hillas, A. (2020). Friend or frenemy? The role of trust in human-machine teaming and lethal autonomous weapons systems. Small Wars & Insurg. 31, 822–850. doi: 10.1080/09592318.2020.1743485

Crossref Full Text | Google Scholar

Watson, J. (1997). “Assessing statistical thinking using the media, in The Assessment Challenge in Statistics Education (Amsterdam: IOSPress).

Google Scholar

Webb, T., Holyoak, K. J., and Lu, H. (2022). Emergent analogical reasoning in large language models. arXiv [preprint] arXiv:2212.09196. doi: 10.1038/s41562-023-01659-w

PubMed Abstract | Crossref Full Text | Google Scholar

Werner, A. (2020). Organizing music, organizing gender: algorithmic culture and Spotify recommendations. Popular Commun. 18, 78–90. doi: 10.1080/15405702.2020.1715980

Crossref Full Text | Google Scholar

Whittlestone, J., Nyrup, R., Alexandrova, A., and Cave, S. (2019). “The role and limits of principles in AI ethics: towards a focus on tensions,” in Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, AIES '19 (New York, NY: Association for Computing Machinery), 195–200.

Google Scholar

Yarkoni, T. (2022). The generalizability crisis. Behav. Brain Sci. 45:e1. doi: 10.1017/S0140525X21001758

Crossref Full Text | Google Scholar

Ziegler, L., and Garfield, J. (2018). Developing a statistical literacy assessment for the modern introductory statistics course. Stat. Educ. Res. J. 17, 161–178. doi: 10.52041/serj.v17i2.164

Crossref Full Text | Google Scholar

Keywords: algorithms, data, AI, trust, statistical literacy, explainability

Citation: Marmolejo-Ramos F, Marrone R, Korolkiewicz M, Gabriel F, Siemens G, Joksimovic S, Yamada Y, Mori Y, Rahwan T, Sahakyan M, Sonna B, Meirmanov A, Bolatov A, Som B, Ndukaihe I, Arinze NC, Kundrát J, Skanderová L, Ngo V-G, Nguyen G, Lacia M, Kung C-C, Irmayanti M, Muktadir A, Samosir FT, Liuzza MT, Giorgini R, Khatin-Zadeh O, Banaruee H, Özdoğru AA, Ariyabuddhiphongs K, Rakchai W, Trujillo N, Valencia SM, Janyan A, Kostov K, Montoro PR, Hinojosa J, Medeiros K, Hunt TE, Posada J, Freitag RMK and Tejada J (2025) Factors influencing trust in algorithmic decision-making: an indirect scenario-based experiment. Front. Artif. Intell. 7:1465605. doi: 10.3389/frai.2024.1465605

Received: 16 July 2024; Accepted: 27 December 2024;
Published: 04 February 2025.

Edited by:

Guandong Xu, University of Technology Sydney, Australia

Reviewed by:

Sergio Rojas-Galeano, Universidad Distrital Francisco José de Caldas, Colombia
Maria Chiara Caschera, National Research Council (CNR), Italy
Felix Härer, University of Fribourg, Switzerland
Kim Van Broekhoven, Erasmus University Rotterdam, Netherlands

Copyright © 2025 Marmolejo-Ramos, Marrone, Korolkiewicz, Gabriel, Siemens, Joksimovic, Yamada, Mori, Rahwan, Sahakyan, Sonna, Meirmanov, Bolatov, Som, Ndukaihe, Arinze, Kundrát, Skanderová, Ngo, Nguyen, Lacia, Kung, Irmayanti, Muktadir, Samosir, Liuzza, Giorgini, Khatin-Zadeh, Banaruee, Özdoğru, Ariyabuddhiphongs, Rakchai, Trujillo, Valencia, Janyan, Kostov, Montoro, Hinojosa, Medeiros, Hunt, Posada, Freitag and Tejada. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Fernando Marmolejo-Ramos, ZmVybmFuZG8ubWFybW9sZWpvcmFtb3NAZmxpbmRlcnMuZWR1LmF1

Present address: Yuki Mori, tayo.inc, Yokohama, Japan

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more