Skip to main content

ORIGINAL RESEARCH article

Front. Public Health, 31 May 2024
Sec. Digital Public Health

Using Google web search to analyze and evaluate the application of ChatGPT in femoroacetabular impingement syndrome

Yifan ChenYifan Chen1Shengqun ZhangShengqun Zhang1Ning TangNing Tang2Daniel M. GeorgeDaniel M. George3Tianlong Huang
Tianlong Huang1*JinPing Tang
JinPing Tang4*
  • 1Orthopaedic Department, The Second Xiangya Hospital of Central South University, Changsha, China
  • 2Orthopaedic Department, The Third Xiangya Hospital of Central South University, Changsha, China
  • 3Royal Adelaide Hospital, Adelaide, SA, Australia
  • 4Department of Orthopaedics, The Third People's Hospital of Chenzhou, Chenzhou, Hunan, China

Background: Chat Generative Pre-trained Transformer (ChatGPT) is a new machine learning tool that allows patients to access health information online, specifically compared to Google, the most commonly used search engine in the United States. Patients can use ChatGPT to better understand medical issues. This study compared the two search engines based on: (i) frequently asked questions (FAQs) about Femoroacetabular Impingement Syndrome (FAI), (ii) the corresponding answers to these FAQs, and (iii) the most FAQs yielding a numerical response.

Purpose: To assess the suitability of ChatGPT as an online health information resource for patients by replicating their internet searches.

Study design: Cross-sectional study.

Methods: The same keywords were used to search the 10 most common questions about FAI on both Google and ChatGPT. The responses from both search engines were recorded and analyzed.

Results: Of the 20 questions, 8 (40%) were similar. Among the 10 questions searched on Google, 7 were provided by a medical practice. For numerical questions, there was a notable difference in answers between Google and ChatGPT for 3 out of the top 5 most common questions (60%). Expert evaluation indicated that 67.5% of experts were satisfied or highly satisfied with the accuracy of ChatGPT’s descriptions of both conservative and surgical treatment options for FAI. Additionally, 62.5% of experts were satisfied or highly satisfied with the safety of the information provided. Regarding the etiology of FAI, including cam and pincer impingements, 52.5% of experts expressed satisfaction or high satisfaction with ChatGPT’s explanations. Overall, 62.5% of experts affirmed that ChatGPT could serve effectively as a reliable medical resource for initial information retrieval.

Conclusion: This study confirms that ChatGPT, despite being a new tool, shows significant potential as a supplementary resource for health information on FAI. Expert evaluations commend its capacity to provide accurate and comprehensive responses, valued by medical professionals for relevance and safety. Nonetheless, continuous improvements in its medical content’s depth and precision are recommended for ongoing reliability. While ChatGPT offers a promising alternative to traditional search engines, meticulous validation is imperative before it can be fully embraced as a trusted medical resource.

Background

Computer technology is rapidly advancing, paving the way for the use of artificial intelligence in medicine. Specifically, machine learning (ML), a subset of artificial intelligence, allows computer algorithms to learn from experience using large datasets without explicit programming. The potential benefits of machine learning include automating repetitive tasks, enhancing treatment decisions, predicting adverse outcomes, and streamlining perioperative healthcare management (1). In the field of orthopedics, ML has been used for multiple purposes. These include detecting mechanical loosening in hip prostheses (2), diagnosing meniscus tears (3), grading osteoarthritis (4, 5), diagnosing bone tumors, and identifying fractures (612). As machine learning methods continue to advance and larger datasets become more accessible, there is a growing interest in the applications of machine learning (13). OpenAI (San Francisco, California) released ChatGPT on their official website on November 30th, 2022. They described it as a sibling model to Instruct GPT that is specifically trained to understand instructions in prompts and provide detailed responses. This advanced language model has sparked a surge of interest in machine learning due to its ability to engage in complex conversations and provide information on a wide range of topics (14). In an evolving healthcare model that prioritizes consumer needs, patients now have unprecedented access to information. This includes the use of ChatGPT to gain a better understanding of medical issues. A recent study revealed that 89% of American citizens search their health symptoms on Google before scheduling a doctor’s appointment (15). The quality of online health information remains problematic, and consumers’ evaluation of such information is often subjective and misleading (1619).

Femoroacetabular impingement (FAI) is a well-known clinical syndrome deriving from morphologic alterations of the acetabulum or femur and is a common cause of hip pain and limited hip range of motion. With the advancement of imaging technology and improved understanding of FAI, diagnosis and evaluation of this condition have become more accurate. The use of advanced imaging techniques like Magnetic Resonance Imaging (MRI) and hip arthroscopy has provided researchers with a deeper understanding of the pathogenesis, prevention, and treatment methods for FAI. Collaboration among multiple disciplines has also contributed to progress in FAI research, as professionals from various fields work together to comprehensively study FAI from different perspectives. As a result, FAI has emerged as a current research are of interest (20).

ChatGPT has already been explored in the field of orthopedics and Total Joint Arthroplasty (21). However, it has not yet been explored in the field of Femoroacetabular Impingement (FAI). Since FAI patients are often younger and more accustomed to using emerging scientific technologies like ChatGPT, we have decided to explore its appropriate and inappropriate use for collecting online health information from patients in the FAI domain. In order to assess ChatGPT in this regard, we aimed to compare it to a Google web search (Menlo Park, California), which implements well known deep learning algorithms, RankBrain, and Bidirectional Encoder Representations from Transformers (BERT), a natural language processing system that identifies search patterns quite accurately (22, 23). Google is the most widely used search engine in the United States and is also the preferred search engine for patients seeking information on orthopedic diagnosis and treatment (24). These algorithms offer a comprehensive list of frequently asked questions (FAQs) and provide links to websites that address each relevant question. Research has shown that up to 65% of patients considering elective orthopedic surgery use the internet: however, specific information they seek remains largely undefined. Additionally, the quality and readability of the available information are often poor quality (25, 26).

This study aims to evaluate the effectiveness of ChatGPT as an online health information resource for patients, specifically in the context of femoroacetabular impingement syndrome (FAI). ChatGPT, introduced in 2022 as a novel machine learning tool, is designed to provide fluent conversational responses. This study compares the capabilities of ChatGPT with those of Google Web Search, the most prevalent search engine in the United States, in three key aspects:

i. Identifying and replicating the most common questions regarding FAI, with attention to the types of questions and topics covered,

ii. Analyzing the responses provided to these frequently asked questions, and.

iii. Evaluating the accuracy and relevance of numerical responses in the FAQs.The underlying hypothesis of this research is that both Google Web Search and ChatGPT will demonstrate comparable proficiency in addressing these aspects, irrespective of the specificity of the question type or topic.

Methods

On February 6th, 2024, we conducted a series of searches on google.com using the latest version of Google Chrome browser (Version 119.0.6045.160, Menlo Park, California). The primary aim of these searches was to minimize the influence of personalized search algorithms (27). To achieve this, we implemented a ‘clean browser installation’—a process which entailed removing browsing history, cookies, website data, cached images, and files, as well as filtering out sponsored websites and advertisements. This setup was essential to ensure that the search results were not skewed by previous browsing behavior or targeted advertising. The specific search term employed was ‘femoroacetabular impingement syndrome’. Through this approach, we aimed to capture a more objective snapshot of the information available to a typical user querying this condition. This search term is entered separately to extract the top 10 frequently asked questions along with their respective sources from relevant websites. These questions are obtained from the “People also ask” section in Google web search by clicking on individual questions to reveal additional related queries. Inclusion criteria encompass any questions pertaining to “hip impingement syndrome, femoroacetabular impingement syndrome, hip impingement, hip collision, Femoro-Acetabular Impingement, Impingement, Femoracetabular.” Exclusion criteria involve eliminating duplicate questions and those unrelated to femoroacetabular impingement syndrome (for example: “What happens after gluteal muscle contracture surgery?”).

ChatGPT is a chatbot developed by OpenAI and released in November 2022. It is a freely accessible and replicable language model designed for various natural language processing tasks, including text generation and language translation. GPT-4, the successor to GPT-3.5, significantly expands upon its predecessor’s capabilities, featuring an even more sophisticated architecture with approximately 175 billion parameters. While the exact number of parameters for GPT-4 has not been publicly disclosed, it is known for its enhanced understanding and generation of natural language, improved contextual awareness, and the ability to generate more accurate and nuanced responses. GPT-4 represents a substantial leap forward in AI language models, pushing the boundaries of machine learning and natural language processing technologies (28). ChatGPT uses a Transformer model network that incorporates the concept of “attention” to generate responses based on word associations. This enables the deep learning model to assign varying levels of importance to different parts of a sentence. As a result, it produces realistic and human-like responses, which can be particularly useful in static environments or situations where the question/response remains unchanged over time. For instance, consider the non-fluctuating problem of asking how to say hello in Spanish. To demonstrate ChatGPT’s capabilities, we inputted the following statement: “Perform a google search with the search term ‘Femoroacetabular Impingement Syndrome’ and record the 10 most frequently asked questions related to this topic.” We then recorded the top 10 questions and their corresponding answers for each output.

To compare the ability of ChatGPT and Google in providing information to patients, we conducted a study using the search term “Femoroacetabular Impingement Syndrome” for both platforms. We identified the top 10 most common questions and the top 5 most frequently asked numerical questions (Tables 13) from Google web search results. These questions were then inputted into ChatGPT to record question-answer pairs. This analysis helps us understand the commonly asked questions by patients and allows us to assess how well ChatGPT and Google provide relevant information. It is worth noting that although Google is widely used by patients to obtain online health information, the quality of information obtained through Google searches varies. Among the information provided by Google, sources like PubMed may offer more reliable answers (15, 23). PubMed can be a valuable resource for future comparisons. This study aims to replicate patients’ internet searches and evaluate the suitability of ChatGPT. Compared to Google Web Search, ChatGPT offers fluent conversational responses to queries, making it an important online health information resource for patients.

Table 1
www.frontiersin.org

Table 1. Top 10 most popular questions for FAI according to Google web search.

Table 2
www.frontiersin.org

Table 2. Top 10 most popular questions for FAI according to ChatGPT.

Table 3
www.frontiersin.org

Table 3. Top 5 questions and answer comparison of search engines for numerical questions about femoroacetabular impingement syndrome.

The questions retrieved in our research are classified into problem themes: fact, policy, and value. This classification is based on previous literature that utilized the modified Rothwell system (29, 30). These issues are divided into 9 themes related to femoroacetabular impingement syndrome: specific activities, rehabilitation timeline, concepts, technical details, costs, treatment guidelines and options, potential consequences and complications, curability and quality of life, as well as patient preferences and decision-making. The website is categorized into the following groups: commercial category, academic category, medical practices (or private clinics), individual surgeons specializing in a single field, government institutions, and social media (Table 4) (31, 32). The Cohen’s kappa coefficient is utilized to assess the inter-rater reliability of website classification. We used IBM SPSS Statistics (Version R24.0.0.0, Armonk, New York) to conduct statistical analysis on website categorization and problem classification. In the case of website classification, the kappa value for inter-rater reliability is 1 (p < 0.01). For problem classification, the kappa value for inter-rater reliability is 0.702 (p < 0.01).

Table 4
www.frontiersin.org

Table 4. Question topics and website categorization by Rothwell’s classification.

To rigorously assess the accuracy and applicability of information sourced from ChatGPT, we enlisted 40 orthopedic specialists with extensive experience in hip arthroscopy. Their evaluations provide critical insights into the practical utility of ChatGPT responses in clinical settings.

The expert panel was diverse, consisting of 32.5% chief physicians, 57.5% associate chief physicians, and 10% attending physicians. Their annual surgical volume varied, with 30% performing fewer than 10 hip arthroscopies per year and 12.5% performing over 100, indicating a broad range of clinical activity.

Experts were asked to assess the quality of ChatGPT’s answers to questions related to femoroacetabular impingement syndrome (FAI) based on accuracy, completeness, relevance, and understandability. The evaluation used a five-point scale ranging from very satisfied to very dissatisfied.

Results

Most popular questions for Google web search and ChatGPT

When searching for “Femoroacetabular Impingement Syndrome” on both Google web search and ChatGPT search, 8 out of 20 questions (40%) were found to be similar. The most common theme category was Fact, with 5 out of 10 questions in both Google web search and ChatGPT falling into this category. Based on themes, the more common subcategories were Specific Activities (3 out of 10 questions) for Google web search, and Timeline of Recovery (3 out of 10 questions) as well as Technical Details (3 out of 10 questions) for ChatGPT (Tables 1, 2).

Answers to the most popular questions

In Google web search, Medical Practice is the most common source of answers for the top 10 questions, accounting for 7 out of 10. There are a total of 15 answers, including those for the top 5 numerical questions. The specific questions asked are as follows: (1) “How many people have femoroacetabular impingement?,” (2) “How long does femoroacetabular impingement last?,” (3) “How long is FAI recovery?,” (4) “Can hip impingement be cured?,” (5) “Is cycling bad for hip impingement?,” (6) “Can you live with hip impingement?,” (7) “Does walking make hip impingement worse?” (Tables 36). It should be noted that Google and ChatGPT provide noticeably different suggestions.

Most popular questions for discrete questions

Medical Practice provided answers for 7 out of the top 10 questions searched on Google. Among the top 5 most frequently asked numerical questions, there were 3 (60%) that had different answers between Google search and ChatGPT (Table 3). The question with the largest discrepancy in data was “How many people have femoroacetabular impingement?.” For the top 10 most common questions about femoroacetabular impingement, there were different responses for 4 out of 10 questions (40%) (Tables 5, 6). Only two replies obtained from Google search for these 10 questions (20%) used sources from academic websites (Tables 5, 6).

Table 5
www.frontiersin.org

Table 5. Answers to most popular questions for FAI.

Table 6
www.frontiersin.org

Table 6. Answers to most popular questions for FAI.

Expert evaluation results

Expert evaluation insights on ChatGPT’s handling of femoroacetabular impingement syndrome (FAI)

Treatment methods

A substantial 67.5% of experts reported satisfaction or high satisfaction with the accuracy of ChatGPT’s descriptions of both conservative and surgical treatment options for FAI. Furthermore, 62.5% of experts indicated their satisfaction or high satisfaction with the safety of the information, suggesting a significant trust in ChatGPT’s responses to safely guide patients. This result confirms the strong approval of the content’s relevance and practicality, with a similar percentage also acknowledging that the information was comprehensively suitable for patient use.

Etiology of FAI: In terms of explaining the etiology of FAI, specifically cam and pincer impingements, 52.5% of experts expressed satisfaction or very satisfaction with ChatGPT’s explanations. This satisfaction rate emphasizes ChatGPT’s ability to handle complex medical topics and translate intricate medical knowledge into comprehensible information.

Overall potential

When assessing the potential of ChatGPT to act as a reliable medical resource for initial information retrieval, 62.5% of experts confirmed that ChatGPT could effectively fulfill this role. This endorsement underscores ChatGPT’s adaptability in medical settings and solidifies its position as a valuable educational tool in healthcare.

Reliability analysis

The internal consistency of the responses, as measured by Cronbach’s alpha, was notably high (α = 0.908). This high level indicates a reliable assessment across different evaluators, further validating the dependability of the feedback received.

Discussion

Publi cations on machine learning in the field of orthopedics have increased tenfold in the past 20 years (33). The increasing availability of information and modern data in today’s value-based era of medicine may be the reason for this (34). ChatGPT is a machine learning tool that has gained attention for its accessibility, user-friendliness, engaging interface, and limitless possibilities. Femoroacetabular impingement (FAI), also known as hip impingement syndrome or hip joint impingement syndrome, is a condition affecting the hip joint. It involves symptoms caused by friction, collision, or compression between the femoral head and acetabulum. This syndrome usually results from morphological changes in the acetabulum or femur and commonly leads to hip pain and limited range of motion (20). Femoroacetabular impingement syndrome is a common condition that significantly affects patients’ quality of life, the overall incidence of FAI diagnosis was 54.4 per 100,000 person-years (35). It has been extensively studied for many years due to its prevalence, particularly among young individuals who make up a significant portion of ChatGPT users. Young people are generally tech-savvy and comfortable with using artificial intelligence (AI) assistants and chatbots like ChatGPT for various purposes such as entertainment, learning, information-seeking, and problem-solving. Considering the overlap between the user base of ChatGPT and the population affected by FAI, it is important to note that people’s use of ChatGPT to search for information about FAI can contribute to both the development of ChatGPT and influence their healthcare decision-making process accordingly.

ChatGPT differs from traditional search engines like Google in its learning mechanisms. Firstly, ChatGPT utilizes deep learning techniques, specifically the Transformer model, trained on large-scale corporate data. In contrast, traditional search engines primarily rely on information retrieval techniques such as keyword matching and ranking algorithms. Secondly, ChatGPT is a conversational language model that allows real-time interaction and dialogue. It can understand user questions, responses, and instructions while generating coherent and personalized replies. Traditional search engines require users to input keywords and return web page links and search results related to those keywords. Furthermore, ChatGPT can comprehend and remember previous conversation history while considering context when answering questions. This enables it to provide more coherent and personalized responses and always suggests following the guidance of professionals at the end of each answer, demonstrating a strong sense of rigor and risk avoidance ability. Traditional search engines mainly rely on phrase matching without taking into account a user’s prior queries or conversations. Lastly, ChatGPT can answer complex questions through inference and text generation while making educated guesses based on contextual cues. It possesses a certain level of creativity and flexibility in generating new textual content. Conversely, traditional search engines heavily depend on pre-established indexes and webpage content without actively reasoning or generating new text themselves.

The purpose of this study is to assess the suitability of patients’ internet searches as a means of accessing online health information, specifically comparing ChatGPT to Google web search. Our findings reveal notable distinctions between Google web search and ChatGPT in terms of the most frequently asked questions and corresponding answers about FAI. Furthermore, we observed that while Google web search relies on Medical Practice as a primary source for answer references, ChatGPT does not directly access the internet or specific websites like Google does. Consequently, we were unable to obtain specific website sources for ChatGPT.

There are a few limitations to consider in this study. Firstly, the number of questions included is limited, which may affect the generalizability of the findings. However, despite this limitation, the study provides valuable insights into the potential of this emerging machine learning (ML) tool due to its novelty and dedicated research efforts. It’s important to note that ChatGPT is a deep learning tool trained on text data from 2021 and does not have internet access like Google Web Search. While ChatGPT utilizes various web text sources during training such as web pages, books, papers, and publicly available content, it cannot retrieve specific information from websites like Google Web Search can. The Google Web Search used in this study was conducted On February 6th, 2024. Differences in databases between both searches could explain any discrepancies found in common questions and answers obtained during our research. Additionally, while ChatGPT remains unchanged throughout the study period, Google’s algorithm may have undergone changes that cannot be replicated. It’s worth noting that medical care standards and treatment methods evolve over time. This evolution might lead to inconsistencies with ChatGPT’s ability to generate accurate human-like responses since accuracy relies on a training set that reflects current issues and addresses non-stationary environments or non-volatile problems (where questions/responses remain consistent over time). An example of a non-volatile problem would be “What is the normal human body temperature?” In this analysis, we chose to compare ChatGPT with Google Web Search because it is widely used as a search engine in the United States (23), but if you use other search engines like Bing (Redmond, Washington), DuckDuckGo (Paoli, Pennsylvania), or Yahoo (Sunnyvale, California), the results should be similar. A study was conducted to compare the performance of natural language search engines and keyword-based search engines in accurately retrieving answers to natural learning queries. The author discovered that there was no significant difference in the accuracy of retrieving answers among four search engines: Google, Ask.com (Oakland, California), Hakia (New York City) and Bing (Redmond, Washington) (36). However, to the author’s knowledge, Google web search is the only search engine that provides additional FAQs when specific questions are clicked on. This feature sets it apart from Bing, DuckDuckGo, and Yahoo. By utilizing this functionality, the author can gather the top 10 most common questions without restricting future analysis to a limited number of inquiries. It should be noted that we have not considered any selection biases related to demographic data of Google search engine or ChatGPT users. It is widely recognized that factors such as age, gender, education level, and income can influence internet usage patterns and may also impact the utilization of ChatGPT (37, 38). Furthermore, the results obtained from ChatGPT may vary depending on whether more specific or less specific questions are inputted. However, it is important to note that this study’s main objective is to assess ChatGPT’s ability to replicate Google web search queries. To the best of the author’s knowledge, this analysis represents the first attempt to incorporate ChatGPT into a research design about FAI for evaluating its suitability in providing common questions and answers sought by patients compared to a search engine like Google. It focuses on its role in addressing patient queries rather than its application in scientific writing, commentary, and editing (39).

When searching for “Femoroacetabular Impingement Syndrome” on both Google web search and ChatGPT search, 8 out of 20 questions (40%) were found to be similar. However, it is important to note that although the questions are different, most of the queries made using each search tool are related to “facts.” Out of the 10 common questions found on Google, only 3 pertain to curability and quality of life, which ChatGPT attempts to replicate but does not appear in Google’s top 10 questions. This distinction is significant because curability and quality of life greatly influence patients’ treatment decision-making process. Further investigation may be necessary to understand why these different questions arise from these search tools. It should also be noted that due to its learning model, ChatGPT may be more sensitive to unique word choices used in queries compared to Google web search (40). ChatGPT can be effectively used to enhance patients’ online searches, providing them with valuable information that may not be readily accessible. However, it is important to note that relying solely on ChatGPT for answers may lead to open-ended and ambiguous responses. Therefore, it is recommended that patients consult their doctors for more complex and evolving questions.

Google web search and ChatGPT provide different answers to the most popular questions. ChatGPT provides longer responses to the most popular questions and often suggest consulting a doctor for further evaluation (Tables 5, 6 show shortened ChatGPT responses due to space constraints). This is how ChatGPT offers balanced and impartial information to patients. However, excessively long replies can affect readability.

Discrete questions and answers were similarly disparate between the search tools. For instance, when asking “How many people have femoroacetabular impingement?,” Google web search yields a result of 0.0544%, while ChatGPT provides a range of 10 to 15%. According to previous research, Google’s answer is more reliable, the overall incidence of FAI diagnosis was 54.4 per 100,000 person-years, and it consistently increased between 2000 and 2016 (35). These are just two out of seven different responses that offer varying answers. Such variations can impact patient decisions, considering factors like symptoms at different stages post-injury, diverse outcomes of conservative and surgical treatments, and the effects of various activities on rehabilitation. It’s worth noting that the response from Google web search is sourced from Pubmed, whereas ChatGPT relies on its pre-trained dataset. However, it is important to acknowledge that information obtained from internet sources for evaluation may be unfiltered, unreviewed or inaccurate.

After analyzing the most frequently asked questions about FAI, it was observed that the majority of inquiries revolve around conservative or non-surgical treatments. Interestingly, the cost category in Rothwell’s classification (Table 4) did not feature prominently in these popular questions. This suggests that public concerns regarding FAI primarily center on avoiding surgical intervention and understanding the potential consequences of untreated FAI. Furthermore, the relationship between FAI and physical activity is one of the most concerning topics for the public (Tables 1, 2). Understanding the relationship between physical activity and recovery in FAI is crucial for patients seeking health information online. Our research has shown that individuals who engage in high-impact sports such as soccer, dance, gymnastics, basketball, and ice hockey have a higher incidence of FAI compared to those who are not physically active. (41) Furthermore, sedentary individuals are also at a higher risk group. Femoroacetabular impingement syndrome can lead to structural abnormalities or functional impairments in the hip joint, which may affect comfort and mobility during physical activities. Engaging in appropriate exercises can be beneficial in managing and rehabilitating FAI, highlighting the importance of accessing accurate and tailored health information. This underscores the relevance of our study in evaluating ChatGPT as a potential online health information resource, offering insights into how patients with FAI might use it to seek advice on effective physical activities and recovery strategies. By understanding the information needs of FAI patients, especially those related to exercise and recovery, our study aims to assess the accuracy and utility of ChatGPT in providing relevant and reliable health information, aligning closely with the broader purpose of our research. Notably, when it comes to determining the incidence rate of FAI among populations (Table 3), Google search results and ChatGPT responses provided conflicting information. Based on previous research, the overall incidence of FAI diagnosis was 54.4 per 100,000 person-years, and it consistently increased between 2000 and 2016 (8).

Several studies have shown that patients who discuss their online health survey results during consultations experience improved doctor-patient relationships (17, 42, 43). According to a randomized controlled study, patient education websites have been found to enhance patients’ comprehension of their surgery and satisfaction with the consent process when compared to standard discussions with treating surgeons (44). In conclusion, these findings suggest that patients can benefit from using search tools and actively communicating accurate and useful information with their doctors. However, the full extent of this combination’s effectiveness is still not fully understood.

Since its inception, ChatGPT has rapidly become a popular resource for health-related information, attracting over 100 million users within just 2 months of its launch. Although Google continues to be the primary source for medical queries, used by up to 89% of American citizens (45), ChatGPT provides a more user-friendly interface. It is devoid of the frequent advertisements and commercial content that often clutter traditional search engines. Consequently, ChatGPT has emerged as a compelling alternative for patients seeking diverse health information, especially in contexts where immediate, comprehensible guidance is essential.

The expert evaluation of ChatGPT’s responses to questions concerning Femoroacetabular Impingement Syndrome (FAI) substantiates its potential as a healthcare resource. An impressive 67.5% of orthopedic specialists reported either satisfaction or high satisfaction with the accuracy of ChatGPT’s descriptions of both conservative and surgical treatment options. Furthermore, a similar proportion of experts confirmed the safety and comprehensiveness of the information provided, affirming that ChatGPT can reliably deliver valuable medical guidance to patients. While the expert feedback on ChatGPT has been predominantly positive, emphasizing its capability to effectively handle complex medical topics, it also highlights the necessity for continuous improvements. Experts have advocated for further enhancements in the depth and precision of the medical content, aiming to ensure that ChatGPT consistently provides high-quality information that adheres to professional medical standards.

In conclusion, although ChatGPT shows significant promise as a medical resource, its use should still be approached with caution. It is crucial to ensure the integrity of the information and minimize the risk of disseminating misleading content. Thus, ongoing validation and development are essential before ChatGPT can be fully integrated into medical practice as a trusted resource. The enthusiastic yet cautious reception by medical professionals suggests a future where AI tools like ChatGPT could transform patient education and preliminary medical consultation, provided that their development is consistently guided by rigorous expert feedback and stringent medical accuracy standards.

What is known about the subject

Femoroacetabular impingement (FAI) is a well-known clinical syndrome deriving from morphologic alterations of the acetabulum or femur and is a common cause of hip pain and limited hip range of motion. Existing literature has extensively investigated FAI’s etiology, diagnostic methods, and treatment strategies, highlighting the need for accurate and timely interventions to mitigate its impact on patients’ quality of life. Nevertheless, the field of FAI-related research currently exhibits a paucity of in-depth exploration into the realm of artificial intelligence (AI) and machine learning. Specifically, comprehensive inquiries investigating the assimilation of AI technologies within this context are still in their incipient stages. This prevailing gap calls for an increased focus on unveiling the potential of AI-driven methodologies within the FAI landscape.

Computer technology is rapidly advancing, paving the way for the use of artificial intelligence in medicine. Specifically, machine learning (ML), a subset of artificial intelligence, allows computer algorithms to learn from experience using large datasets without explicit programming. In the field of orthopedics, ML has been used for multiple purposes. ChatGPT is a model that is specifically trained to understand instructions in prompts and provide detailed responses.

As such, our study addresses this gap by assessing the application of ChatGPT and Google in delivering accurate and relevant medical information about FAI. By comparing the responses provided by these platforms, our research contributes to an enhanced comprehension of the potential benefits and limitations of AI-based resources in the orthopedic domain.

What this study adds to existing knowledge

ChatGPT has already been explored in the field of orthopedics and Total Joint Arthroplasty. However, it has not yet been explored in the field of Femoroacetabular Impingement (FAI). Given the typically younger age demographic of FAI patients, who readily embrace emerging scientific technologies like ChatGPT, coupled with the inclination of these youthful individuals to proactively seek out disease-related information online both prior to and following medical consultations, our study is designed to delve into the nuanced exploration of ChatGPT’s suitable and inappropriate applications within the realm of FAI. This inquiry aims to illuminate the judicious utilization of ChatGPT as a conduit for furnishing online health information to FAI patients, thereby contributing to a comprehensive comprehension of its utility and constraints within this specific medical context. Therefore, this study employs a Cross-Sectional Study design to investigate the application of ChatGPT and Google in the context of Femoroacetabular Impingement (FAI). This approach addresses the existing gap regarding the utilization of ChatGPT in the realm of FAI, enabling an assessment of its viability as an online health information resource for patients. By comparing the responses provided by these platforms, our research contributes to an enhanced comprehension of the potential benefits and limitations of AI-based resources in the orthopedic domain. Furthermore, our research serves as a reference for healthcare professionals in effectively engaging with patients who possess a wealth of online medical knowledge during doctor-patient communication.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

YC: Writing – original draft. SZ: Writing – original draft. NT: Methodology, Software, Supervision, Writing – original draft. DG: Data curation, Formal analysis, Resources, Visualization, Writing – original draft. TH: Writing – review & editing. JT: Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Hunan Provincial Natural Science Foundation of China for the year 2021 (Grant No. 2021JJ30932).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpubh.2024.1412063/full#supplementary-material

References

1. Groot, OQ, Ogink, PT, Lans, A, Twining, PK, Kapoor, ND, DiGiovanni, W, et al. Machine learning prediction models in orthopedic surgery: a systematic review in transparent reporting. J Orthop Res. (2022) 40:475–83. doi: 10.1002/jor.25036

PubMed Abstract | Crossref Full Text | Google Scholar

2. Karnuta, JM, Haeberle, HS, Luu, BC, Roth, AL, Molloy, RM, Nystrom, LM, et al. Artificial intelligence to identify arthroplasty implants from radiographs of the hip. J Arthroplast. (2021) 36:S290–S294.e1. doi: 10.1016/j.arth.2020.11.015

PubMed Abstract | Crossref Full Text | Google Scholar

3. Roblot, V, Giret, Y, Bou Antoun, M, Morillot, C, Chassin, X, Cotten, A, et al. Artificial intelligence to diagnose meniscus tears on MRI. Diagn Interv Imaging. (2019) 100:243–9. doi: 10.1016/j.diii.2019.02.007

Crossref Full Text | Google Scholar

4. Chen, P, Gao, L, Shi, X, Allen, K, and Yang, L. Fully automatic knee osteoarthritis severity grading using deep neural networks with a novel ordinal loss. Comput Med Imaging Graph. (2019) 75:84–92. doi: 10.1016/j.compmedimag.2019.06.002

PubMed Abstract | Crossref Full Text | Google Scholar

5. Antony, J., McGuinness, K., O'Connor, N. E., and Moran, K., "Quantifying radiographic knee osteoarthritis severity using deep convolutional neural networks," 2016 23rd international conference on pattern recognition (ICPR), Cancun, Mexico, (2016), pp. 1195–1200.

Google Scholar

6. Bien, N, Rajpurkar, P, Ball, RL, Irvin, J, Park, A, Jones, E, et al. Deep-learning-assisted diagnosis for knee magnetic resonance imaging: development and retrospective validation of MRNet. PLoS Med. (2018) 15:e1002699. doi: 10.1371/journal.pmed.1002699

PubMed Abstract | Crossref Full Text | Google Scholar

7. Chung, SW, Han, SS, Lee, JW, Oh, KS, Kim, NR, Yoon, JP, et al. Automated detection and classification of the proximal humerus fracture by using deep learning algorithm. Acta Orthop. (2018) 89:468–73. doi: 10.1080/17453674.2018.1453714

PubMed Abstract | Crossref Full Text | Google Scholar

8. Haeberle, HS, Helm, JM, Navarro, SM, Karnuta, JM, Schaffer, JL, Callaghan, JJ, et al. Artificial intelligence and machine learning in lower extremity arthroplasty: a review. J Arthroplast. (2019) 34:2201–3. doi: 10.1016/j.arth.2019.05.055

PubMed Abstract | Crossref Full Text | Google Scholar

9. Kim, DH, and MacKinnon, T. Artificial intelligence in fracture detection: transfer learning from deep convolutional neural networks. Clin Radiol. (2018) 73:439–45. doi: 10.1016/j.crad.2017.11.015

Crossref Full Text | Google Scholar

10. Lalehzarian, SP, Gowd, AK, and Liu, JN. Machine learning in orthopaedic surgery. World J Orthop. (2021) 12:685–99. doi: 10.5312/wjo.v12.i9.685

PubMed Abstract | Crossref Full Text | Google Scholar

11. Mont, MA, Krebs, VE, Backstein, DJ, Browne, JA, Mason, JB, Taunton, MJ, et al. Artificial intelligence: influencing our lives in joint arthroplasty. J Arthroplast. (2019) 34:2199–200. doi: 10.1016/j.arth.2019.08.017

PubMed Abstract | Crossref Full Text | Google Scholar

12. Yi, PH, Wei, J, Kim, TK, Sair, HI, Hui, FK, Hager, GD, et al. Automated detection & classification of knee arthroplasty using deep learning. Knee. (2020) 27:535–42. doi: 10.1016/j.knee.2019.11.020

Crossref Full Text | Google Scholar

13. Wang, VM, Cheung, CA, Kozar, AJ, and Huang, B. Machine learning applications in Orthopaedic imaging. J Am Acad Orthop Surg. (2020) 28:e415–7. doi: 10.5435/JAAOS-D-19-00688

PubMed Abstract | Crossref Full Text | Google Scholar

14. Susnjak, T.. ChatGPT: The end of online exam integrity? 19, (2022). Available at: http://arxiv.org/abs/2212.09292 (Accessed July 15, 2023).

Google Scholar

15. Stokel-Walker, C. ChatGPT listed as author on research papers: many scientists disapprove. Nature. (2023) 613:620–1. doi: 10.1038/d41586-023-00107-z

PubMed Abstract | Crossref Full Text | Google Scholar

16. Cline, RJ, and Haynes, KM. Consumer health information seeking on the internet: the state of the art. Health Educ Res. (2001) 16:671–92. doi: 10.1093/her/16.6.671

PubMed Abstract | Crossref Full Text | Google Scholar

17. Daraz, L, Morrow, AS, Ponce, OJ, Beuschel, B, Farah, MH, Katabi, A, et al. Can patients trust online health information? A Meta-narrative systematic review addressing the quality of health information on the internet. J Gen Intern Med. (2019) 34:1884–91. doi: 10.1007/s11606-019-05109-0

PubMed Abstract | Crossref Full Text | Google Scholar

18. Sun, Y, Zhang, Y, Gwizdka, J, and Trace, CB. Consumer evaluation of the quality of online health information: systematic literature review of relevant criteria and indicators. J Med Internet Res. (2019) 21:e12522. doi: 10.2196/12522

Crossref Full Text | Google Scholar

19. Zhang, Y, Sun, Y, and Xie, B. Quality of health information for consumers on the web: a systematic review of indicators, criteria, tools, and evaluation results. J Assoc Inf Sci Technol. (2015) 66:2071–84. doi: 10.1002/asi.23311

Crossref Full Text | Google Scholar

20. Tang, N, Zhang, W, Su, Y, Han, Z, Deng, L, Li, Y, et al. Femoroacetabular impingement and labral tear: from the Most highly cited articles to research interests. Orthop Surg. (2021) 13:1922–33. doi: 10.1111/os.13037

PubMed Abstract | Crossref Full Text | Google Scholar

21. Dubin, JA, Bains, SS, Chen, Z, Hameed, D, Nace, J, Mont, MA, et al. Using a Google web search analysis to assess the utility of ChatGPT in Total joint arthroplasty. J Arthroplast. (2023) 38:1195–202. doi: 10.1016/j.arth.2023.04.007

Crossref Full Text | Google Scholar

22. Devlin, J, Chang, MW, Lee, K, and Toutanova, K. BERT: pre-training of deep bidirectional transformers for language understanding. In: Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers); Association for Computational Linguistics (2019):4171–4186.

Google Scholar

23. Nayak, P. Understanding searches better than ever before. The keyword (2019). p. 295. Available at: https://blog.google/products/search/search-language-understandingbert/ (Accessed February 5, 2023).

Google Scholar

24. Statista Research Department Number of search engine users in the United States from 2014 to 2020 (in millions) (2016). https://www.statista.com/statistics/253795/number-of-search-engine-users-in-the-united-states/ (Accessed February 5, 2023).

Google Scholar

25. Cassidy, JT, and Baker, JF. Orthopaedic patient information on the world wide web: an essential review. J Bone Joint Surg Am. (2016) 98:325–38. doi: 10.2106/JBJS.N.01189

PubMed Abstract | Crossref Full Text | Google Scholar

26. Fraval, A, Ming Chong, Y, Holcdorf, D, Plunkett, V, and Tran, P. Internet use by orthopaedic outpatients - current trends and practices. Australas Med J. (2012) 5:633–8. doi: 10.4066/AMJ.2012.1530

PubMed Abstract | Crossref Full Text | Google Scholar

27. Tripathi, S, Singh, C, Kumar, A, Pandey, C, and Jain, N. Bidirectional transformer based multi-task learning for natural language understanding In: E Métais, F Meziane, S Vadera, V Sugumaran, and M Saraee, editors. Natural language processing and information systems: Lecture Notes in Computer Science. Cham Switzerland: Springer International Publishing (2019). 54–65.

Google Scholar

28. Welsh, M. The end of programming. Commun ACM. (2022) 66:34–5. doi: 10.1145/3570220

Crossref Full Text | Google Scholar

29. Kanthawala, S, Vermeesch, A, Given, B, and Huh, J. Answers to health questions: internet search results versus online health community responses. J Med Internet Res. (2016) 18:e95. doi: 10.2196/jmir.5369

PubMed Abstract | Crossref Full Text | Google Scholar

30. Rothwell, JD. In mixed company: Communicating in small groups. Wadsworth: Cengage Learning (2013).

Google Scholar

31. López-Jornet, P, and Camacho-Alonso, F. The quality of internet sites providing information relating to oral cancer. Oral Oncol. (2009) 45:e95–8. doi: 10.1016/j.oraloncology.2009.03.017

PubMed Abstract | Crossref Full Text | Google Scholar

32. Starman, JS, Gettys, FK, Capo, JA, Fleischli, JE, Norton, HJ, and Karunakar, MA. Quality and content of internet-based information for ten common orthopaedic sports medicine diagnoses. J Bone Joint Surg Am. (2010) 92:1612–8. doi: 10.2106/JBJS.I.00821

PubMed Abstract | Crossref Full Text | Google Scholar

33. Cabitza, F, Locoro, A, and Banfi, G. Machine learning in orthopedics: a literature review. Front Bioeng Biotechnol. (2018) 6:75. doi: 10.3389/fbioe.2018.00075

PubMed Abstract | Crossref Full Text | Google Scholar

34. Helm, JM, Swiergosz, AM, Haeberle, HS, Karnuta, JM, Schaffer, JL, Krebs, VE, et al. Machine learning and artificial intelligence: definitions, applications, and future directions. Curr Rev Musculoskelet Med. (2020) 13:69–76. doi: 10.1007/s12178-020-09600-8

PubMed Abstract | Crossref Full Text | Google Scholar

35. Hale, RF, Melugin, HP, Zhou, J, LaPrade, MD, Bernard, C, Leland, D, et al. Incidence of Femoroacetabular impingement and surgical management trends over time. Am J Sports Med. (2021) 49:35–41. doi: 10.1177/0363546520970914

Crossref Full Text | Google Scholar

36. Hariri, N. Do natural language search engines really understand what users want? A comparative study on three natural language search engines and google. Online Inf Rev. (2013) 37:287–303. doi: 10.1108/OIR-12-2011-0210

Crossref Full Text | Google Scholar

37. Bidmon, S, and Terlutter, R. Gender differences in searching for health information on the internet and the virtual patient-physician relationship in Germany: exploratory results on how men and women differ and why. J Med Internet Res. (2015) 17:e156. doi: 10.2196/jmir.4127

PubMed Abstract | Crossref Full Text | Google Scholar

38. Dehghani, M, Johnson, KM, Garten, J, Boghrati, R, Hoover, J, Balasubramanian, V, et al. TACIT: an open-source text analysis, crawling, and interpretation tool. Behav Res Methods. (2017) 49:538–47. doi: 10.3758/s13428-016-0722-4

PubMed Abstract | Crossref Full Text | Google Scholar

39. Gilat, R, and Cole, BJ. How will artificial intelligence affect scientific writing, reviewing and editing? The future is Here …. Arthroscopy. (2023) 39:1119–20. doi: 10.1016/j.arthro.2023.01.014

PubMed Abstract | Crossref Full Text | Google Scholar

40. Guo, B., Zhang, X., Wang, Z., Jiang, M., Nie, J., Ding, Y., et al. How close is ChatGPT to human experts? Comparison Corpus, evaluation, and detection, arXiv preprint [Epub ahead of preprint] (2023). doi: 10.48550/arXiv.2301.07597

Crossref Full Text | Google Scholar

41. Siebenrock, KA, Ferner, F, Noble, PC, Santore, RF, Werlen, S, and Mamisch, TC. The cam-type deformity of the proximal femur arises in childhood in response to vigorous sporting activity. Clin Orthop Relat Res. (2011) 469:3229–40. doi: 10.1007/s11999-011-1945-4

PubMed Abstract | Crossref Full Text | Google Scholar

42. Ellis, DJ, Mallozzi, SS, Mathews, JE, Moss, IL, Ouellet, JA, Jarzem, P, et al. The relationship between preoperative expectations and the short-term postoperative satisfaction and functional outcome in lumbar spine surgery: a systematic review. Global Spine J. (2015) 5:436–51. doi: 10.1055/s-0035-1551650

PubMed Abstract | Crossref Full Text | Google Scholar

43. Neuprez, A, Delcour, JP, Fatemi, F, Gillet, P, Crielaard, JM, Bruyère, O, et al. Patients' expectations impact their satisfaction following Total hip or knee arthroplasty. PLoS One. (2016) 11:e0167911. doi: 10.1371/journal.pone.0167911

PubMed Abstract | Crossref Full Text | Google Scholar

44. Fraval, A, Chandrananth, J, Chong, YM, Coventry, LS, and Tran, P. Internet based patient education improves informed consent for elective orthopaedic surgery: a randomized controlled trial. BMC Musculoskelet Disord. (2015) 16:14. doi: 10.1186/s12891-015-0466-9

Crossref Full Text | Google Scholar

45. Polesie, S, and Larkö, O. Use of large language models: editorial comments. Acta Derm Venereol. (2023) 103:adv00874. doi: 10.2340/actadv.v103.9593

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: ChatGPT, machine learning, femoroacetabular impingement syndrome (FAI), large language model, Google

Citation: Chen Y, Zhang S, Tang N, George DM, Huang T and Tang J (2024) Using Google web search to analyze and evaluate the application of ChatGPT in femoroacetabular impingement syndrome. Front. Public Health. 12:1412063. doi: 10.3389/fpubh.2024.1412063

Received: 05 April 2024; Accepted: 23 May 2024;
Published: 31 May 2024.

Edited by:

Mitchell K. Ng, Maimonides Medical Center, United States

Reviewed by:

Qi Zuo, Wuhan University, China
Eric David Ornos, University of the Philippines Manila, Philippines

Copyright © 2024 Chen, Zhang, Tang, George, Huang and Tang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tianlong Huang, tianlong.huang@csu.edu.cn; JinPing Tang, 737555088@qq.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.