- 1Healthcare Programme, Group Research and Development, DNV, Høvik, Norway
- 2School of Digital Technologies, Tallinn University, Tallinn, Estonia
Editorial on the Research Topic
Trustworthy AI for healthcare
Artificial Intelligence (AI) integration in healthcare has been met with enthusiasm but also caution (1). The expectation of AI to revolutionize healthcare is high, with its potential to enhance access, improve quality, and streamline efficiency. The landscape of AI in healthcare is rapidly evolving, with significant advancements in diagnostics, decision support systems, patient monitoring, robotics, personalized medicine, drug discovery, clinical trials, monitoring, and organizational workflow management (2). However, the adoption of AI in healthcare has not kept pace with its development. This is often attributed to the lack of various aspects of trustworthiness in AI systems (3–5).
This research topic explores the concept of “trustworthy AI for healthcare,” which stands at the intersection of technology, ethics, and clinical practice. Trustworthy AI for healthcare refers to the development and deployment of AI systems in healthcare that are reliable, safe, and transparent, and that respect ethical principles and values. It delves into the current state of AI in healthcare, the challenges impeding its adoption, and the paramount importance of trust. It underscores the need for transparency in AI algorithms, the ability to interpret and explain AI decisions, and the collaborative efforts required to achieve these goals (6, 7).
The article “Dicing with data: the risks, benefits, tensions and tech of health data in the iToBoS project” reviews the iToBoS project, which created an AI tool for early melanoma detection. Key challenges identified in the project include (1) a small clinical trial cohort, raising anonymization concerns; (2) difficulty in obtaining informed consent due to the necessity to explain the involved complex technology; (3) an open data commitment, necessitating extra privacy measures for diverse data types; and (4) communicating algorithmic results to stakeholders. The authors reflect on the tensions that these issues cause, considering the broader health sector challenges.
The authors of the article “Developing machine learning systems worthy of trust for infection science: A requirement for future implementation into clinical practice” discuss the critical role of infection science, particularly during the SARS-CoV-2 pandemic, and emphasize the potential of AI in improving patient outcomes, optimizing clinical workflows, and enhancing public health management. Despite promising research, the lack of trustworthy AI systems hinders the transition of AI models from research to clinical practice. The paper advocates for developing systems that meet user, stakeholder, and regulatory requirements and highlights the need for a systematic approach to trustworthiness in AI applications.
In the article “The unmet promise of trustworthy AI in healthcare: why we fail at clinical translation,” the authors state that the clinical application of AI often fails, primarily due to the lack of a precise definition of “trust” and “trustworthiness.” This deficiency leads to unintentional misuse and the possibility of intentional “thics washing” by industry stakeholders. The paper contends that these barriers hinder the realization of trustworthy medical AI’s potential and advocates for reassessing the meaning of trust in healthcare AI to close the gap between theoretical guidelines and practical application.
The authors of the article “A trustworthy AI reality-check: the lack of transparency of artificial intelligence products in healthcare” state that trustworthiness hinges on transparent algorithm development and testing to pinpoint biases and communicate harm risks. Publicly available information on the risks of medical AI products is often insufficient. This study assessed 14 CE-certified AI radiology products in the EU, examining their transparency based on a custom survey aligned with AI trust guidelines. The transparency scores varied widely; and significant gaps in training data documentation, ethical considerations, and deployment limitations were found. The authors call for establishing transparency requirements to uphold AI’s trustworthiness in healthcare.
Trust among various stakeholders—societies, organizations, and businesses—is crucial for ensuring smooth operations, creating value, and minimizing disruptions or accidents. However, existing regulations often do not fully cater to technological advancements, leading to gaps that may introduce challenges. In response, newly introduced regulations, such as the European Union’s Artificial Intelligence Act (AI Act) and the U.S. Executive Order on Artificial Intelligence, aim to set requirements for trustworthy and responsible AI. The development of trustworthy AI is imperative to foster a healthcare environment where AI can be safely, scalably, and sustainably adopted.
Standards play a crucial role in bridging the gap between the high-level principles outlined in regulations that cover all sectors and the concrete technical specifications needed for AI-enabled systems to achieve compliance in specific sectors. To build trust in these systems, regulations and legislation may require third-party audits to provide assurance that entities comply with the established standards. As regulations and standards evolve, the role of ongoing research also becomes increasingly critical. Research on Trustworthy AI is essential not only for shaping these emerging regulations but also for providing the vital insights and empirical data needed to inform and refine both regulations and the corresponding standards.
Collaboration between academia and industry is crucial in this endeavor. Each brings to the table a wealth of knowledge and experience that is both distinct and complementary. Industry provides practical insights from deploying and implementing AI systems, while academia contributes through rigorous research and theoretical frameworks. This synergy can accelerate the adoption of AI systems that are not only innovative but also reliable and understandable.
In conclusion, integrating AI into healthcare is complex and fraught with challenges. Yet, the pursuit of trustworthy AI is a journey worth undertaking. It promises a future where healthcare is not only powered by intelligence but also grounded in trust—a future where AI acts as a partner in healthcare delivery, augmenting clinicians’ capabilities and enhancing patient care. This editorial calls for a multidisciplinary approach to realize this vision, urging stakeholders across industry and academia to unite in the quest for an AI-enabled healthcare system that is as trustworthy as it is transformative.
Author contributions
OA: Writing – original draft, Writing – review & editing. AB: Writing – original draft, Writing – review & editing. SS: Writing – review & editing. SA: Writing – review & editing.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. European Parliamentary Research Service. Artificial Intelligence in Healthcare. Applications, Risks, and Ethical and Societal Impacts. Tech. rep. Brussels: European Parliament (2022). doi: 10.2861/568473
2. Alowais SA, Alghamdi SS, Alsuhebany N, Alqahtani T, Alshaya AI, Almohareb SN et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med Educ. (2023) 23:1–15. doi: 10.1186/S12909-023-04698-Z
3. Albahri AS, Duhaim AM, Fadhel MA, Alnoor A, Baqer NS, Alzubaidi L et al. A systematic review of trustworthy and explainable artificial intelligence in healthcare: assessment of quality, bias risk, and data fusion. Inf Fusion. (2023) 96:156–91. doi: 10.1016/J.INFFUS.2023.03.008
4. Rajpurkar P, Chen E, Banerjee O, Topol EJ. AI in health and medicine. Nat Med. (2022) 28:31–8. doi: 10.1038/s41591-021-01614-0
5. Van De Sande D, Van Genderen ME, Smit JM, Huiskens J, Visser JJ, Veen RE et al. Developing, implementing and governing artificial intelligence in medicine: a step-by-step approach to prevent an artificial intelligence winter. BMJ Health Care Inf. (2022) 29:e100495. doi: 10.1136/bmjhci-2021-100495
Keywords: systems, patient monitoring, robotics, personalized medicine, drug discovery, clinical trials, monitoring
Citation: Agafonov O, Babic A, Sousa S and Alagaratnam S (2024) Editorial: Trustworthy AI for healthcare. Front. Digit. Health 6:1427233. doi: 10.3389/fdgth.2024.1427233
Received: 3 May 2024; Accepted: 6 May 2024;
Published: 3 June 2024.
Edited and Reviewed by: Uwe Aickelin, The University of Melbourne, Australia
© 2024 Agafonov, Babic, Sousa and Alagaratnam. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Oleg Agafonov, oleg.agafonov@dnv.com