Skip to main content

OPINION article

Front. Artif. Intell., 05 April 2023
Sec. Medicine and Public Health

Impact of ChatGPT on medical chatbots as a disruptive technology

\nJames C. L. Chow,
James C. L. Chow1,2*Leslie SandersLeslie Sanders3Kay LiKay Li4
  • 1Radiation Medicine Program, Princess Margaret Cancer Centre, University Health Network, Toronto, ON, Canada
  • 2Department of Radiation Oncology, University of Toronto, Toronto, ON, Canada
  • 3Department of Humanities, York University, Toronto, ON, Canada
  • 4Department of English, University of Toronto, Toronto, ON, Canada

1. Introduction

Sweeping changes in artificial intelligence (AI) have been brought about in recent years, resulting in remarkable progress taking a number of forms, such as AI chatbots. ChatGPT (Chat Generative Pre-trained Transformer) is a language model for dialogue. This chatbot, developed by Open AI, was released in prototype form on November 30, 2022 (ChatGPT, 2023). Since then, ChatGPT has attracted numerous users from various fields, because it can provide detailed answers and humanlike responses to almost any question. ChatGPT is reputed to be serving various medical functions, ranging from uses in medical writing and documentation to medical education. Recently, ChatGPT has been reported to be capable of passing the gold-standard US medical exam, suggesting that is has potentially significant applications in the field of medicine (Kung et al., 2023).

However, it is questionable whether ChatGPT can consistently provide reliable health information for patients or healthcare providers interacting with it. A reliable medical chatbot could constitute a seamless interface to information for both patients and healthcare providers. As a patient-oriented tool, it would allow users to obtain disease-related information or book medical appointments (Bates, 2019; Khadija et al., 2021). Simultaneously, it could serve healthcare professionals by providing essential information relating to their work, such as medical protocols for treatment of common and rare diseases or hospital policies (Bates, 2019; Gupta et al., 2021; Wan, 2021).

Unlike a specific medical chatbot, ChatGPT has not been trained on a finely-tuned dataset created by medical professionals (Sallam, 2023). This raises concerns, as patients may initially turn to ChatGPT for assistance. While this tool has the potential to educate and expedite care, there is also a risk that it may provide inaccurate diagnoses or recommendations (Cascella et al., 2023). Furthermore, the chatbot's machine learning and data search algorithms are still in the prototype phase, and the development of related ethical policies and regulations is ongoing (Liebrenz et al., 2023).

As researchers studying the design and creation of medical chatbots, we expect that ChatGPT will be able to evolve into a reliable and practical medical chatbot. Here, we would like to explore some obstacles to the achievement of this goal and potential solutions to them, by considering ChatGPT as a disruptive technology.

2. ChatGPT as a disruptive technology

Disruptive technologies often begin as niche solutions or products with limited initial market appeal. Over time, they gain acceptance and transform the industry or market they are a part of (Kostoff et al., 2004). A prime example is the digital camera, which eliminated the need for film and traditional film processing. Before digital cameras, film cameras dominated the photography market. However, digital cameras disrupted this market by offering a more convenient and cost-effective alternative. ChatGPT is also a disruptive technology with the potential to fundamentally change how we interact with technology and perhaps to revolutionize the way medical professionals engage with patients. If ChatGPT can function as a professionally trained medical chatbot, it may be able to operate more quickly than existing medical chatbots, draw on a larger database, reduce medical errors, and assist doctors in improving their performance.1 A concern with any disruptive technology is that, while it may be innovative, timely, and impactful, it also carries risks and it may not immediately integrate well with complex, professional applications, such as those that are relevant in medicine.

Currently, several obstacles hinder ChatGPT from functioning fully as a medical chatbot. For instance, its database may not be entirely up to date; the current knowledge cutoff is September 2021. Additionally, medical information sourced from the internet might not be entirely accurate, posing a risk of providing misinformed answers.2 Numerous ethical concerns exist, including patient safety, privacy, data content, and cybersecurity (Xu et al., 2021; Parviainen and Rantala, 2022). Caution is necessary for clinical applications, and medical professionals are working to verify and fine-tune the chatbot. User feedback influences the chatbot's training, but users may not understand the interaction model, making adoption more difficult. Shifting the culture of medical service from human-to-human to machine-to-human interactions will take time. Finally, rapid AI advancements will continuously modify the ethical framework (Parviainen and Rantala, 2022). This process is expected to be lengthy and time-consuming for various stakeholders, such as medical service providers, AI developers, and users.

We must admit that the chatbot is extremely popular, as it is very user-friendly, and patients as users may focus more on the benefits of convenience and efficiency, rather than the reliability and accuracy of the tool (Chin et al., 2023). ChatGPT also presents an air of authority and so sounds rather trustworthy. This is particularly noteworthy during the period of the recent pandemic, during which medical resources have been limited, and virtual chats have become quite the norm. At its present stage of development, caution probably needs to be exercised by both medical service providers and patients in making serious and ethical use of the medication information provided by ChatGPT, given its nature as a disruptive technology. Medical service providers also need to acquire a detailed understanding from AI developers of the data and conversational flow algorithm underlying the AI chatbot.

Nevertheless, although ChatGPT is currently still imperfect as a humanlike medical chatbot, we believe that it is bound to change healthcare systems in the near future. Below, we explore some obstacles to that goal and discuss potential solutions to each obstacle.

3. Some obstacles on the path to ChatGPT becoming a medical chatbot

3.1. Current medical chatbots vs. ChatGPT

ChatGPT may not be as accurate as medical chatbots developed by dedicated medical professionals. Medical chatbots that have been built over the years often use AI features like natural language processing (Hirschberg and Manning, 2015) and focus on specific tasks, such as answering users' inquiries on various medical topics (Kovacek and Chow, 2021; Rebelo et al., 2022). Chatbot creators can build the underlying software, link it to a maintained database, and easily modify the conversational flow and data. In contrast, ChatGPT was not trained by a specialized team of medical professionals (Chow, 2021; Siddique and Chow, 2021). Its strength lies in its natural language processing (NLP) capabilities, based on the Generative Pre-trained Transformer (GPT-3.5) form of AI. This allows ChatGPT to extract information from unstructured data sources, such as electronic health records (EHRs), identify patterns and recurrences, such as particular symptoms, and generate diagnostic reports.3 This can potentially reduce the workload of frontline health workers during routine medical checks and could help to alleviate shortages of healthcare workers. Recently, ChatGPT has been upgraded with new features and improvements. The new language model (GPT-4), which is linked to Bing Chat, can process up to 25,000 words and boasts advanced capabilities in creativity and visual input, as well as longer contextual memory, compared to its predecessor, GPT-3.5. According to internal testing, ChatGPT now reportedly provides 40% more factual responses, making it a safer and more accurate tool for tasks such as music composition, screenwriting, and technical writing.4

Traditional medical chatbots use AI and natural language processing to predict user intent and provide appropriate responses (Chow et al., 2023). They are continuously improved through user feedback and performance data. These processes are controlled by chatbot creators using a well-maintained, human-designed database. However, ChatGPT, as a disruptive technology, draws information from the internet, making the accuracy and currency of the medical information it supplies questionable and sometimes uncontrollable. Although this approach saves time and effort in database preparation, ChatGPT requires careful training from medical professionals, as it may be trained by any user, which can lead to inaccurate information. Therefore, it is crucial to test and evaluate ChatGPT's performance, as its responses may be unpredictable and dependent on the data used for training. Development of a robust quality assurance system and a systematic approach to monitoring of database updates and maintenance can help to ensure the accuracy and precision of the information provided by ChatGPT. In addition, the chatbot creator can partner with a team of medical experts to review and validate the dataset used to train ChatGPT, so that a custom training dataset can be built to improve the accuracy and relevance of ChatGPT's medical knowledge. Continuous monitoring and improvement are also necessary to monitor the performance of ChatGPT as a medical chatbot, to ensure that it can remain current and relevant through regular updating of the training data with new and accurate information.

3.2. Human doctors vs. AI doctors

The current medical system relies on certified professionals to provide reliable services to patients. These professionals need to maintain their certifications, ensuring quality care. However, AI-based chatbots such as ChatGPT do not undergo any similar verification process, raising ethical concerns. AI chatbots could provide a quick solution to the high demand for medical care during situations like pandemics. The fact that ChatGPT has passed the Medical Boards examination may increase public acceptance and trust in AI systems in the healthcare domain. As people become more familiar with AI technologies, they might be more open to incorporating AI-based tools into their healthcare routines. This increased acceptance may lead to further integration of AI in the medical field, enhancing the efficiency and effectiveness of healthcare services. However, it is important to remember that passing the Medical Boards examination does not necessarily make ChatGPT a complete substitute for human medical professionals. Practical experience, empathy, and interpersonal skills are essential components of healthcare that AI systems do not easily replicate. Additionally, ChatGPT's performance on the examination may not fully represent its ability to handle complex and nuanced medical situations in real-world settings.

To address the challenges of using ChatGPT in medicine, medical professional organizations should consider establishing suitable frameworks to monitor and assess the quality of ChatGPT for applications in healthcare. This will involve the provision of clear guidelines for users on how to use ChatGPT correctly and guidance for service providers on safely implementing ChatGPT as a medical chatbot. A major consideration should involve setting parameters for the safe usage of ChatGPT. For example, its functions could be limited to particular areas where ChatGPT has demonstrated accuracy, such as diagnosis, education, and healthcare. Through implementation of these measures, ChatGPT could become an invaluable asset to the medical profession.

3.3. Ethical concerns

Since the pandemic period, medical chatbots have been rapidly developed as conversational agents for use by patients, thus accelerating the development and deployment by medical ethicists of corresponding ethical frameworks for this disruptive technology. To date, many legal and ethical challenges have already emerged regarding medical chatbots that need to be addressed and dealt with (Liebrenz et al., 2023). These include the data content of the chatbot, cybersecurity, data use, privacy and integration, patient safety, and trust and transparency between all participants. The construction of such ethical frameworks will take time because it is dependent on patients' feedback and robust updating of the chatbot itself. It also involves a great deal of negotiation among various stakeholders, for example, concerning patient data and their ownership. The present progress in the deployment of such ethical frameworks cannot keep pace with the rapid advancement of ChatGPT as a medical chatbot. This will exert an increasing amount of pressure on medical professionals when they want to implement this type of disruptive technology in the medical system within such a short period of time.

Continued systematic research into the ethical implications of ChatGPT for its users is necessary, and international collaborations should be pursued to establish a global standard for ethics surrounding the use of ChatGPT as a medical chatbot. In order to address issues related to data content, cybersecurity, privacy, and integration, various stakeholders (including medical professionals, patient and hospital representatives, and computer security experts) should convene to establish policies regarding patient data ownership and security. Adequate protection of patient data must be implemented as a standard regulation to ensure patient privacy when using the chatbot on the Internet of Things. In order to promote patient safety, trust, and transparency, ethical guidelines and protocols should be developed and put in place to govern the appropriate use of AI-generated medical advice. Users should receive education on the limitations of and potential risks associated with using ChatGPT as a medical chatbot. We also recommend that users verify any critical information with healthcare professionals before making any decisions related to their health. Medical ethics must serve as guiding principles in shaping the ethical framework for ChatGPT as a provider of medical information and medical “practitioner”.

4. Summary

While ChatGPT has the potential, as a disruptive technology, to improve access to healthcare services, there are also concerns relating to its use as a medical chatbot. One concern is the accuracy and reliability of the medical information provided by ChatGPT, as it is not a licensed medical professional and may not have access to up-to-date medical knowledge. Additionally, there are concerns about the transparency of the chatbot model and the ethics of making use of user information, as well as the potential for biases in the data used to train ChatGPT's algorithms. As such, it is important to carefully consider the potential risks and benefits of using ChatGPT as a medical chatbot, and to ensure that appropriate safeguards are put in place to address these concerns. As we believe that ChatGPT will be further developed into a humanlike medical chatbot in the future, we urge relevant stakeholders to continue studying and improving the chatbot. Moreover, we urge them to engage in serious examination of the obstacles to achieving this goal as soon as possible, in order that related standards for quality assurance systems and regulations can be established, so as to keep pace with the challenges posed by this disruptive technology.

Author contributions

JC and KL wrote the article. JC, KL, and LS revised the article. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by a Canadian Institutes of Health Research Planning and Dissemination Grant—Institute Community Support (CIHR PCS−168296).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Footnotes

References

Bates, M. (2019). Health care chatbots are here to help. IEEE Pulse. 10, 12–14. doi: 10.1109/MPULS.2019.2911816

PubMed Abstract | CrossRef Full Text | Google Scholar

Cascella, M., Montomoli, J., Bellini, V., and Bignami, E. (2023). Evaluating the feasibility of ChatGPT in healthcare: an analysis of multiple clinical and research scenarios. J Med Syst. 47, 1–5. doi: 10.1007/s10916-023-01925-4

PubMed Abstract | CrossRef Full Text | Google Scholar

ChatGPT (2023). ChatGPT: Optimizing Language Models for Dialogue. Available online at: https://openai.com/blog/chatgpt/ (accessed March 28, 2023).

Chin, H., Lima, G., Shin, M., Zhunis, A., Cha, C., Choi, J., et al. (2023). User-chatbot conversations during the COVID-19 pandemic: study based on topic modeling and sentiment analysis. J Med Internet Res. 25, e40922. doi: 10.2196/40922

PubMed Abstract | CrossRef Full Text | Google Scholar

Chow, J. C. (2021). “Artificial intelligence in radiotherapy and patient care,” in Artificial Intelligence in Medicine 2021 Jul 11. Cham: Springer International Publishing. p. 1–13. doi: 10.1007./978-3-030-58080-3_143-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Chow, J. C., Sanders, L., and Li, K. (2023). Design of an educational chatbot using artificial intelligence in radiotherapy. AI. 4, 319–332. doi: 10.3390/ai4010015

CrossRef Full Text | Google Scholar

Gupta, J., Singh, V., and Kumar, I. (2021). “Florence-a health care chatbot,” in 2021 7th International Conference on Advanced Computing and Communication Systems (ICACCS) 2021 Mar (Coimbatore: IEEE), 19. p. 504–508. doi: 10.1109/ICACCS52021, 9442006.

CrossRef Full Text | Google Scholar

Hirschberg, J., and Manning, C. D. (2015). Advances in natural language processing. Science. 349, 261–266. doi: 10.1126/science.aaa8685

PubMed Abstract | CrossRef Full Text | Google Scholar

Khadija, A., Zahra, F. F., and Naceur, A. (2021). AI-powered health chatbots: toward a general architecture. Procedia. Comp. Sci. 191, 355–360. doi: 10.1016/j.procs.07.048

CrossRef Full Text | Google Scholar

Kostoff, R. N., Boylan, R., and Simons, G. R. (2004). Disruptive technology roadmaps. Technol Forecast Soc Change. 71, 141–159. doi: 10.1016/S0040-1625(03)00048-9

CrossRef Full Text | Google Scholar

Kovacek, D., and Chow, J. C. (2021). An AI-assisted chatbot for radiation safety education in radiotherapy. IOP SciNotes. 2, 034002. doi: 10.1088/2633-1357/ac1f88

CrossRef Full Text | Google Scholar

Kung, T. H., Cheatham, M., Medenilla, A., Sillos, C., De Leon, L., Elepaño, C., et al. (2023). Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. PLOS Digital Health. 2, e0000198. doi: 10.1371/journal.pdig.0000198

PubMed Abstract | CrossRef Full Text | Google Scholar

Liebrenz, M., Schleifer, R., Buadze, A., Bhugra, D., and Smith, A. (2023). Generating scholarly content with ChatGPT: ethical challenges for medical publishing. Lancet Digital Health. 5, e105–e106. doi: 10.1016/S2589-7500(23)00019-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Parviainen, J., and Rantala, J. (2022). Chatbot breakthrough in the 2020s? An ethical reflection on the trend of automated consultations in health care. Med. Health Care Philos. 25, 61–71. doi: 10.1007/s11019-021-10049-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Rebelo, N., Sanders, L., Li, K., and Chow, J. C. (2022). Learning the treatment process in radiotherapy using an artificial intelligence–assisted chatbot: development study. JMIR Formative Res. 6, e39443. doi: 10.2196/39443

PubMed Abstract | CrossRef Full Text | Google Scholar

Sallam, M. (2023). The Utility of ChatGPT as an example of large language models in healthcare education, research and practice: systematic review on the future perspectives and potential limitations. medRxiv. (2023) 2023:02. doi: 10.1101/02, 19.23286155

CrossRef Full Text | Google Scholar

Siddique, S., and Chow, J. C. (2021). Machine learning in healthcare communication. Encyclopedia. 1, 220–239. doi: 10.3390/encyclopedia1010021

CrossRef Full Text | Google Scholar

Wan, E. (2021). I'm like a wise little person: notes on the metal performance of Woebot the mental health Chatbot. Theatre J. 73, E−21. doi: 10.1353./tj.2021.0068

CrossRef Full Text | Google Scholar

Xu, L., Sanders, L., Li, K., and Chow, J. C. (2021). Chatbot for health care and oncology applications using artificial intelligence and machine learning: systematic review. JMIR Cancer. 7, e27850. doi: 10.2196/27850

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: ChatGPT, medical chatbot, artificial intelligence, disruptive technology, healthcare, nature language processing, knowledge base, big data

Citation: Chow JCL, Sanders L and Li K (2023) Impact of ChatGPT on medical chatbots as a disruptive technology. Front. Artif. Intell. 6:1166014. doi: 10.3389/frai.2023.1166014

Received: 14 February 2023; Accepted: 23 March 2023;
Published: 05 April 2023.

Edited by:

Steve Jiang, University of Texas Southwestern Medical Center, United States

Reviewed by:

Michael Dohopolski, University of Texas Southwestern Medical Center, United States
Wei Liu, Mayo Clinic Arizona, United States

Copyright © 2023 Chow, Sanders and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: James C. L. Chow, james.chow@rmp.uhn.ca

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.