Skip to main content

OPINION article

Front. Med., 19 December 2024
Sec. Regulatory Science
This article is part of the Research Topic Ethical and Legal Implications of Artificial Intelligence in Public Health: Balancing Innovation and Privacy View all articles

Bridging the gap in AI integration: enhancing clinician education and establishing pharmaceutical-level regulation for ethical healthcare

\r\nAlessandro Perrella
Alessandro Perrella1*Francesca F. Bernardi,Francesca F. Bernardi2,3Massimo BisognoMassimo Bisogno4Ugo Trama,Ugo Trama2,3
  • 1I UOC Emergin Infectious Disease and High Countagiousness, AORN Ospedali dei Colli—P.O. D. Cotugno, Naples, Italy
  • 2Coordination of the Regional Health System, General Directorate for Health Protection, Naples, Italy
  • 3Department of Experimental Medicine, University of Campania Luigi Vanvitelli, Naples, Italy
  • 4Regional Special Office for Digital Transformation, Campania Region, Naples, Italy

1 Introduction

Recently the role of AI in healthcare has been deeply studied and discussed in scientific literature. Promising applications of artificial intelligence and machine learning (AI/ML) are revolutionizing both clinical and administrative domains, with significant advancements demonstrated in drug discovery, precise analysis and interpretation of radiological images, early and accurate sepsis detection, efficient hospital resource management, automated documentation of clinical encounters and decision support system (DSS). These use cases underscore the immense potential of AI/ML to enhance efficiency, accuracy, and outcomes across the healthcare spectrum (1). However, a very recent article raises essential considerations about the adoption and regulation of AI in clinical settings (2). Therefore, the integration of Artificial Intelligence (AI) in healthcare presents numerous opportunities and challenges. Conversely, there is a significant gap between clinician education regarding AI and the regulatory measures necessary for ethical deployment. To address this gap effectively, a structured, organized approach must be followed, encompassing clearly defined steps for both the education of clinicians and the establishment of rigorous regulatory frameworks. This paper argues that bridging this gap requires a dual approach: enhancing clinicians' understanding of AI technologies and treating AI systems as rigorously as pharmaceuticals through strict regulatory processes. By doing so, we can foster ethical and effective AI integration into clinical practice, ensuring patient safety and better healthcare outcomes. Here, we outline four key considerations that should guide the planning and regulation of AI integration in healthcare.

2 Expanding AI education among clinicians

Looking at the current clinical practice in healthcare according to the increase diffusion of AI is arguable the knowledge of physician about this new technology. In fact despite the increasing prevalence of AI in healthcare, many clinicians remain inadequately educated about what AI entails, its limitations and its implications too (1). Given that clinicians play a central role in patient care, a comprehensive AI education program should be a priority. Education initiatives should focus not only on how AI systems operate but also on how their regulatory framework should move on as for pharmaceuticals. This lack of understanding represents a significant barrier to the responsible adoption of AI technologies in clinical practice. To effectively bridge this gap, clinician education must cover both technical and ethical aspects of AI and should be part of medical degree course as well. A deeper understanding of the processes involved in validating AI tools can empower clinicians to participate meaningfully in discussions about the safety and efficacy of these systems. By building a foundational knowledge of AI, clinicians will be better prepared to evaluate and use AI tools within an ethical context and advocate for appropriate use and safety measures (3). This approach could be compared to that currently used to improve clinical or healthcare activity against antimicrobial resistance (AMR) or as for antimicrobial stewardship (4).

3 Understanding legal and ethical implications

Legal and ethical education must be extended beyond decision-makers and regulators to the clinicians who interact directly with AI systems. For clinicians to use AI responsibly, they must understand the legal implications, such as data privacy, accountability, and the ethical risks of biases1,2 (5). The recent European Union Artificial Intelligence Act lays out a regulatory framework for AI; however, this information is not often communicated to those on the front lines of healthcare (3). Educating clinicians about such regulatory initiatives is crucial to align their practical use of AI with ethical guidelines.1 As they navigate the clinical environment, an understanding of legal parameters will help them mitigate risks and ensure that AI integration prioritizes patient safety and respects privacy rights.

4 AI should be managed like a drug

There is a growing consensus that AI systems used in healthcare should be regulated in a manner similar to pharmaceuticals. Just as pharmaceuticals undergo a series of clinical trials to evaluate safety, efficacy, and ethical considerations, AI technologies should also follow rigorous validation processes before widespread implementation.

Parallels to pharmaceutical validation:

Phases of testing: AI could adopt phases similar to drug development—preclinical (testing in controlled environments), Phase I (safety in small clinical settings), Phase II (efficacy trials), and Phase III (large-scale clinical testing). This structure would ensure that AI is tested for both safety and effectiveness in diverse, real-world environments.

Classification: AI tools could be classified as drugs according to a system like ATC. For instance according to the type of AI tools type [drug discovery, precise analysis and interpretation of analysis, decision support system (DSS)] we could give them a code and a related regulatory activity.

Risk assessments: Like drugs, AI must undergo thorough risk assessments, which include evaluating potential biases, unintended outcomes, and ethical implications.

Regulatory oversight: A dedicated regulatory agency—potentially within the European Medicines Agency (EMA)—should combine expertise from both medical and engineering domains. This mixed approach would ensure a balanced evaluation of both the medical efficacy and technical performance of AI tools.

Finally, as in pharmaceuticals, AI systems should come with comprehensive documentation, including a “Summary of Product Characteristics” and “Package Leaflet” that outlines their intended use, limitations, and instructions for safe implementation. This approach will standardize AI information, enabling healthcare providers to make informed decisions based on clear guidance (6).

However, the two frameworks Pharmaceuticals and AI have some substantial differences (Table 1). In fact, while for innovative drug the time required to be approved tor clinical use is usually 9.1 years (7) AI technologies have shorter development cycles compared to traditional drug due to iterative model improvements and less dependency on long-term biological testing. This rapid pace requires an expedited but still rigorous framework for validation and deployment. Unlike drugs, AI tools benefit from iterative deployment, allowing for updates and enhancements after initial deployment based on real-world performance and user feedback, which means ongoing evaluation is critical. AI tools also offer significant advantages in scalability and adaptability. They can rapidly scale across diverse clinical environments and adjust to new datasets as they become available, setting them apart from more static pharmacological solutions. Therefore, AI should be managed and regulated as a Drug but according to a Phase of Testing, Risk Assessment and Regulatory Oversight being based to on specific tools for AI (8) (Figure 1).

Table 1
www.frontiersin.org

Table 1. Key differences between drug and AI evaluation.

Figure 1
www.frontiersin.org

Figure 1. Comparison of drug and AI frameworks. Figure shows difference in frameworks of Drugs and AI development with several differences that highlight the need to find for AI quick as secure system for validation and deployment. The drug development timetable process is based on Brown et al. (7).

Figure shows difference in frameworks of Drugs and AI development with several differences that highlight the need to find for AI quick as secure system for validation and deployment.

5 Economic factors should not compromise ethical standards

While the commercialization of AI in healthcare is inevitable, ethical standards must not be compromised for economic gain. Proper governance, including stringent oversight by a mixed medical-engineering regulatory body, will ensure that AI systems adhere to healthcare's core ethical principles. Regulatory frameworks must ensure that economic motivations should drive innovation but must always be secondary to patient safety, quality of care, and ethical obligations. This balance will foster trust in AI technologies among both clinicians and patients (9).

6 Conclusion

Artificial intelligence (AI) offers immense potential to transform healthcare, from improving patient outcomes to enhancing clinical workflows and driving medical innovation. However, its successful integration into healthcare requires overcoming several key challenges, including gaps in clinician education, the need for ethical governance, and the absence of tailored regulatory frameworks. This paper argues for regulating AI with the same rigor as pharmaceuticals, incorporating validation phases, risk assessments, and detailed documentation, while adapting these processes to AI's fast and iterative development cycles.

A critical first step is educating clinicians about AI. Many healthcare professionals lack a clear understanding of how AI systems work, their limitations, and the ethical and legal issues they raise. Comprehensive training programs are needed to build this knowledge. Such education should focus not only on technical aspects but also on teaching clinicians how to assess AI tools for safety, effectiveness, and ethical implications. This would empower clinicians to confidently use AI in their practice, ensuring that its benefits are fully realized while safeguarding patient trust.

Equally important is the development of regulatory frameworks that match AI's unique characteristics. Unlike pharmaceuticals, which follow a linear path from development to deployment, AI evolves continuously. Regulations must therefore balance rapid innovation with robust oversight, ensuring AI systems are safe, effective, and free from bias. This approach requires collaboration between healthcare professionals, technology developers, and regulators to create guidelines that are both practical and ethical.

Economic pressures should not overshadow the ethical responsibilities involved in AI integration. While commercialization drives innovation, patient safety and quality of care must always come first. By fostering partnerships between industry and healthcare that prioritize ethical principles, we can build trust in AI technologies and their use in clinical practice.

To address these challenges, the following steps should guide AI integration:

1. Establish clear and practical processes for validating and monitoring AI systems, drawing inspiration from pharmaceutical regulation but tailoring these to AI's specific needs.

2. Develop accessible training programs for clinicians, focusing on building their confidence and competence in using AI tools.

3. Support pilot projects and real-world case studies to demonstrate how AI can be safely and effectively used in different healthcare settings.

By taking this structured and adaptive approach, we can bridge the current gaps in AI integration. This will ensure that AI evolves responsibly, supporting healthcare providers and benefiting patients while maintaining the highest ethical standards. Ultimately, this balanced strategy will enable AI to fulfill its promise of transforming healthcare in a way that is safe, effective, and equitable (10, 11).

Author contributions

AP: Conceptualization, Formal analysis, Supervision, Writing – original draft, Writing – review & editing. FB: Methodology, Writing – review & editing. MB: Conceptualization, Writing – review & editing. UT: Resources, Writing – review & editing.

Funding

The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

1. Rajpurkar P, Chen E, Banerjee O, Topol EJ. AI in health and medicine. Nat Med. (2022) 28:31–8. doi: 10.1038/s41591-021-01614-0

PubMed Abstract | Crossref Full Text | Google Scholar

2. Busch F, Kather JN, Johner C, Moser M, Truhn D, Adams LC, et al. Navigating the European Union Artificial Intelligence Act for Healthcare. npj Digit Med. (2024) 7:210. doi: 10.1038/s41746-024-01213-6

PubMed Abstract | Crossref Full Text | Google Scholar

3. Gilbert S. The EU passes the AI Act and its implications for digital medicine are unclear. NPJ Digit Med. (2024) 7:135. doi: 10.1038/s41746-024-01116-6

PubMed Abstract | Crossref Full Text | Google Scholar

4. Chetty S, Swe-Han KS, Mahabeer Y, Pillay A, Essack SY. Interprofessional education in antimicrobial stewardship, a collaborative effort. JAC Antimicrob Resist. (2024) 6:dlae054. doi: 10.1093/jacamr/dlae054

PubMed Abstract | Crossref Full Text | Google Scholar

5. Mennella C, Maniscalco U, De Pietro G, Esposito M. Ethical and regulatory challenges of AI technologies in healthcare: a narrative review. Heliyon. (2024) 10:e26297. doi: 10.1016/j.heliyon.2024.e26297

PubMed Abstract | Crossref Full Text | Google Scholar

6. Palaniappan K, Lin EYT, Vogel S. Global regulatory frameworks for the use of artificial intelligence (AI) in the healthcare services sector. Healthcare. (2024) 12:562. doi: 10.3390/healthcare12050562

PubMed Abstract | Crossref Full Text | Google Scholar

7. Brown DG, Wobst HJ, Kapoor A, Kenna LA, Southall N. Clinical development times for innovative drugs. Nat Rev Drug Discov. (2022) 21:793–4. doi: 10.1038/d41573-021-00190-9

PubMed Abstract | Crossref Full Text | Google Scholar

8. Ben Charif A, Zomahoun HTV, Gogovor A, Abdoulaye Samri M, Massougbodji J, Wolfenden L, et al. Tools for assessing the scalability of innovations in health: a systematic review. Health Res Policy Sys. (2022) 20:34. doi: 10.1186/s12961-022-00830-5

PubMed Abstract | Crossref Full Text | Google Scholar

9. Dankwa-Mullan I. Health equity and ethical considerations in using artificial intelligence in public health and medicine. Prev Chronic Dis. (2024) 21:240245. doi: 10.5888/pcd21.240245

PubMed Abstract | Crossref Full Text | Google Scholar

10. Cheatham B, Javanmardian K, Samandari H. Confronting the risks of artificial intelligence. McKinsey (2019). Available at: https://perma.cc/T2CX-HYZF

Google Scholar

11. Brundage M, Avin S, Clark J, Toner H, Eckersley P, Garfinkel B, et al. The Malicious Use of Artificial Intelligence. Future of Humanity Institute (2018). Available at: https://arxiv.org/ftp/arxiv/papers/1802/1802.07228.pdf (accessed October 1, 2024).

Google Scholar

Keywords: regulatory drug development, European Medicines Agency (EMA), artificial intelligence, antimicrobial stewardship (AMS), ethic and development, ethic, medical AI governance, pharmaceutical validation

Citation: Perrella A, Bernardi FF, Bisogno M and Trama U (2024) Bridging the gap in AI integration: enhancing clinician education and establishing pharmaceutical-level regulation for ethical healthcare. Front. Med. 11:1514741. doi: 10.3389/fmed.2024.1514741

Received: 21 October 2024; Accepted: 09 December 2024;
Published: 19 December 2024.

Edited by:

Giovanni Maio, University of Freiburg, Germany

Reviewed by:

Adamantios Koumpis, University Hospital of Cologne, Germany
Luk Arbuckle, IQVIA Applied AI Science, Canada

Copyright © 2024 Perrella, Bernardi, Bisogno and Trama. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Alessandro Perrella, YWxlc3NhbmRyby5wZXJyZWxsYSYjeDAwMDQwO29zcGVkYWxpZGVpY29sbGkuaXQ=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.