Skip to main content

SYSTEMATIC REVIEW article

Front. Oncol., 15 August 2022
Sec. Cancer Imaging and Image-directed Interventions
This article is part of the Research Topic Reviews in Cancer Imaging and Image-directed Interventions View all 17 articles

Application of artificial intelligence in the diagnosis of subepithelial lesions using endoscopic ultrasonography: a systematic review and meta-analysis

Xin-Yuan LiuXin-Yuan LiuWen SongWen SongTao MaoTao MaoQi ZhangQi ZhangCuiping ZhangCuiping ZhangXiao-Yu Li*Xiao-Yu Li*
  • Department of Gastroenterology, The Affiliated Hospital of Qingdao University, Qingdao, China

Endoscopic ultrasonography (EUS) is the most common method for diagnosing gastrointestinal subepithelial lesions (SELs); however, it usually requires histopathological confirmation using invasive methods. Artificial intelligence (AI) algorithms have made significant progress in medical imaging diagnosis. The purpose of our research was to explore the application of AI in the diagnosis of SELs using EUS and to evaluate the diagnostic performance of AI-assisted EUS. Three databases, PubMed, EMBASE, and the Cochrane Library, were comprehensively searched for relevant literature. RevMan 5.4.1 and Stata 17.0, were used to calculate and analyze the combined sensitivity, specificity, positive likelihood ratio (PLR), negative likelihood ratio (NLR), diagnostic odds ratio (DOR), and summary receiver-operating characteristic curve (SROC). Eight studies were selected from 380 potentially relevant studies for the meta-analysis of AI-aided EUS diagnosis of SELs. The combined sensitivity, specificity, and DOR of AI-aided EUS were 0.92 (95% CI, 0.85-0.96), 0.80 (95% CI, 0.70-0.87), and 46.27 (95% CI, 19.36-110.59), respectively). The area under the curve (AUC) was 0.92 (95% CI, 0.90-0.94). The AI model in differentiating GIST from leiomyoma had a pooled AUC of 0.95, sensitivity of 0.93, specificity of 0.88, PLR of 8.04, and NLR of 0.08. The combined sensitivity, specificity, and AUC of the AI-aided EUS diagnosis in the convolutional neural network (CNN) model were 0.93, 0.81, and 0.94, respectively. AI-aided EUS diagnosis using conventional brightness mode (B-mode) EUS images had a combined sensitivity of 0.92, specificity of 0.79, and AUC of 0.92. AI-aided EUS diagnosis based on patients had a combined sensitivity, specificity, and AUC of 0.95, 0.83, and 0.96, respectively. Additionally, AI-aided EUS was superior to EUS by experts in terms of sensitivity (0.93 vs. 0.71), specificity (0.81 vs. 0.69), and AUC (0.94 vs. 0.75). In conclusion, AI-assisted EUS is a promising and reliable method for distinguishing SELs, with excellent diagnostic performance. More multicenter cohort and prospective studies are expected to be conducted to further develop AI-assisted real-time diagnostic systems and validate the superiority of AI systems.

Systematic Review Registration: PROSPERO (https://www.crd.york.ac.uk/PROSPERO/), identifier CRD42022303990.

Introduction

Gastrointestinal subepithelial lesions (SELs) are tumors that originate from the muscularis mucosa, submucosa, or muscularis propria (1). According to statistics, one SEL is found in every 300 endoscopy examinations (2). SELs, including gastrointestinal stromal tumors (GIST), leiomyomas, schwannomas, neuroendocrine tumors (NET), lipomas, and ectopic pancreas, are asymptomatic and difficult to distinguish due to their similar morphology in size, shape, surface color, contour, and margin (1). GISTs are the most prevalent SELs, with a prevalence of 14–20 cases per million, and have the potential to evolve into malignancies (3, 4). Approximately 60% of patients with GISTs can be cured by surgery (5). Therefore, it is crucial to differentiate GISTs from other benign tumors.

With the development of endoscopic ultrasonography (EUS), fine-needle aspiration biopsy (FANB), immunohistochemical staining methods, and various new imaging technologies, such as contrast-enhanced harmonic EUS (CH-EUS), the approaches for diagnosing and treating SELs have improved (6). EUS as a useful tool has recently become the conventional inspection method for the discovery and diagnosis of SELs. However, the diagnostic accuracy of EUS is limited and closely related to the professional level and experience of the endoscopists (7). EUS-FNAB can be used to obtain tissue specimens for immunohistochemical staining and is the gold standard for diagnosing SELs. Nevertheless, the diagnostic yield of EUS-FNAB for SELs is not ideal, ranging from 60% to 85% (810). FNAB is an invasive and risky operation, and the limited sampling sites are subjectively determined by endoscopists, which may lead to missed diagnoses. Therefore, alternative methods are needed for the accurate diagnosis of SELs to avoid surgical resection of benign lesions as GISTs with malignant potential.

Recently, artificial intelligence (AI) has been extensively used in medical imaging technology, owing to its superior performance. Machine learning (ML) involves the fields of computer science and statistics, generating algorithms to analyze various types of data, and building appropriate descriptive and predictive models (11). Artificial neural networks (ANN), as mathematical models of information processing, are supervised ML models inspired by the structure of brain synaptic connections (11). A convolutional neural network (CNN) is a deep learning algorithm that shows strong performance in image recognition, classification, and processing (12). AI-aided EUS diagnostic tools have been widely applied to differentiate various types of pancreatic diseases, such as pancreatic tumors, chronic pancreatitis, and autoimmune pancreatitis (1315). In recent years, several studies have explored the value of CNN in distinguishing SELs based on EUS images, mainly in differentiating GIST from benign lesions. In this systematic review and meta-analysis, we aimed to assess the effectiveness and accuracy of AI in diagnosing SELs using EUS images and focused on the performance of computer-aided diagnosis models in differentiating GIST from other benign lesions by comparing AI and EUS experts.

Methods

Search strategy

This study followed the preferred reporting items for systematic reviews and meta-analyses (PRISMA) guidelines (16). The PubMed, Embase, and Cochrane Library databases were systematically and comprehensively searched for studies on the AI-aided diagnostic accuracy of SELs under EUS with or without EUS experts as controls published until February 2022. Search terms in the title, abstract, and keywords are as follows: (“artificial intelligence” OR “AI” OR “machine learning” OR “deep learning” OR “convolutional neural network” OR “computer-assisted” OR “computer-aided” OR “neural network” OR “digital image analysis” OR “digital image processing”) AND (“endoscopic ultrasound” OR “endoscopic ultrasonography” OR “EUS”). To avoid omissions, the SELs were not included in the retrieval strategy. The retrieved articles were screened independently by two investigators (Xin-Y L and WS). Disagreements were discussed and resolved by a third researcher (TM). This protocol was registered with PROSPERO (CRD42022303990).

Inclusion and exclusion criteria

The inclusion criteria for studies were as follows (1): prospective or retrospective study design; (2) study subjects were adult participants (≥18 years old); (3) all SELs patients were diagnosed based on histopathological diagnosis after surgical or endoscopic resection or EUS-FNAB; (4) AI algorithm was applied to the diagnosis of patients with SELs using EUS images; (5) study results demonstrated the diagnostic performance of computer-aided diagnosis (CAD) algorithms, including area under the curve (AUC), sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), diagnostic odds ratio (DOR), or accuracy, enabling the calculation of true positive (TP), false positive (FP), true negative (TN), and false negative (FN); and (6) the manuscript was written in English. Conference proceedings, case reports, narrative and systematic reviews, meta-analyses, and studies with incomplete data (TP, FP, TN, and FN could not be calculated) were excluded. Studies with failed randomization and significant differences in baseline data between groups were also excluded from this systematic review.

Data extraction and quality assessment

The number of histologically confirmed SELs that were true-positive (GIST considered to be GIST by AI or experts), true-negative (non-GIST considered to be non-GIST by AI or experts), false-positive (non-GIST considered to be GIST), or false-negative (GIST considered to be non-GIST) were extracted. Additionally, the first author’s name; year of publication; country where the study was conducted; study type; number of samples in the training, validation, and test sets; imaging modality; AI model; and video were also retrieved.

The Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool was used to evaluate the quality and potential bias of all included studies in four aspects: patient selection, index test, reference standard, and flow and timing quality (17). Regarding the problem of pre-specified thresholds, we referred to the study by Thaninee et al. and modified the problem as to whether the performance of the AI-assisted diagnostic system was validated in another cohort (18). Two reviewers (Xin-Y L and WS) independently assessed the eight studies, and conflicts were discussed and resolved with a third reviewer (TM).

Statistical analysis

RevMan 5.4.1 (The Cochrane Collaboration, 2020, London, United Kingdom) and Stata 17.0 (StataCorp, College Station, TX, USA) were used for diagnostic meta-analysis. Published data were extracted, including the reported TP, FP, FN, TN, sensitivity, and specificity of the test datasets. The pooled sensitivity, specificity, positive likelihood ratio (PLR), negative likelihood ratio (NLR), diagnostic score, and DOR with 95% confidence intervals (CIs) were calculated and analyzed using the bivariate mixed-effects model. A summary receiver-operating characteristic curve (SROC) was also constructed, and the AUC was calculated to assess diagnostic accuracy. A funnel plot and its symmetric distribution were used to evaluate the risk of publication bias. Subgroup and meta-regression analyses were performed to explore the sources of heterogeneity. Heterogeneity among the studies was determined using I2 and Cochran’s Q tests. P < 0.1 generally suggests significant heterogeneity, and I2 >50% indicates substantial heterogeneity.

Result

Literature search and bias assessment

The literature retrieval process and screening results are shown in Figure 1. Initially, 380 potentially relevant studies were retrieved from the three databases, and 98 duplicates were removed. Subsequently, 268 studies were excluded after reviewing the titles and abstracts, as they were irrelevant articles and were not suitable for the research topic or type. After screening the full text of 14 eligible studies, two studies that did not meet the eligibility criteria and four studies related to GIST malignant potential were excluded. Finally, eight studies were selected for the meta-analysis of AI-aided diagnosis of SELs according to the PRISMA flowchart (1926).

FIGURE 1
www.frontiersin.org

Figure 1 Flowchart of literature search.

The characteristics of all included studies are summarized in Table 1. A total of eight studies with 339 patients with GIST and 194 patients without GIST were included in the meta-analysis, seven of which were within the last three years. They were all retrospective studies, and one of them used both retrospective and prospective test sets in the stage of testing AI models (22). Three studies were conducted in Japan, two in South Korea, and three in China, Turkey, and the United States. Only one study developed an AI model based on contrast-enhanced harmonic EUS (CH-EUS) images, whereas the others used the conventional brightness mode (B-mode) of EUS. Considering computer-aided models, except for one study that used the ANN model, the remaining studies applied the CNN model. Only one study did not use EUS experts as controls (19). The training, validation, and testing datasets of the included studies are presented in Supplementary Table 1. All the studies trained and developed AI models using a large number of EUS images. One of the studies used videos from each patient divided into 0.1s intervals, yielding images to train the AI model (24).

TABLE 1
www.frontiersin.org

Table 1 Characteristics of included studies.

The quality and risk of bias of the included studies determined using the QUADAS-2 tool are presented in Figure 2. One meta-analysis of AI-aided diagnosis of GIST identified a high-risk bias in patient selection (19).

FIGURE 2
www.frontiersin.org

Figure 2 Quality assessment of included studies using QUADAS-2.

The slope coefficient of the Deeks’ funnel plot was symmetrical (p = 0.14) (Figure 3), indicating that publication bias was insignificant.

FIGURE 3
www.frontiersin.org

Figure 3 Deeks’ funnel plot of publication bias.

Diagnostic performance of AI-assisted EUS in GIST

We incorporated data from all retrospective diagnostic test sets and performed a meta-analysis of the eight included studies. The pooled sensitivity of AI-aided EUS diagnosis of GIST was 0.92 (95% CI, 0.85-0.96) (Figure 4A) and specificity was 0.80 (95% CI, 0.70-0.87) (Figure 4B). The pooled PLR and NLR were 4.61 (95% CI, 3.00-7.08) (Figure 4C) and 0.10 (95% CI, 0.05-0.19) (Figure 4D), respectively. The diagnostic score and DOR were 3.83 (95% CI, 2.96-4.71) and 46.27 (95% CI, 19.36-110.59), respectively (Supplementary Figure 1). Figure 5A shows the SROC curve of AI-aided EUS, with an AUC of 0.92 (95% CI, 0.90-0.94).

FIGURE 4
www.frontiersin.org

Figure 4 Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS diagnosis of GIST.

FIGURE 5
www.frontiersin.org

Figure 5 SROC curves of AI-assisted EUS diagnosis of GIST. (A) The SROC curve of eight studies. (B) SROC curve of nine datasets including prospective diagnostic test set.

Subsequently, we expanded the sample size by including the data from a prospective diagnostic test set. The combined results of AI-assisted EUS diagnosis of GIST were shown as follows: AUC of 0.92 (95% CI, 0.89-0.94) (Figure 5B), sensitivity 0.92 (95% CI, 0.85-0.95), specificity 0.78 (95% CI, 0.69-0.86), PLR 4.23 (95% CI, 2.88-6.22), and NLR 0.11 (95% CI, 0.06-0.19) (Supplementary Figure 2). The diagnostic score and DOR were 3.67 (95% CI, 2.90-4.45) and 39.40 (95% CI, 18.20-85.30), respectively (Supplementary Figure 3).

To investigate the clinical application of AI in the diagnosis of GIST, we generated a Fagan diagram (Figure 6). Assuming a 20% prevalence of GIST, the diagram shows a posterior probability of 54% for GIST if the test is positive, and approximately 2% for a negative test.

FIGURE 6
www.frontiersin.org

Figure 6 Fagan normogram for the prediction of GISTs in EUS images.

Subgroup analysis of AI-assisted EUS

The specific types of SELs in the included studies are shown in Supplementary Table 2. One study involved five SELs, including GIST, leiomyomas, schwannomas, NET, and ectopic pancreas (24). Four studies developed AI only for the differential diagnosis of GIST and leiomyoma (22, 23, 25, 26), and a subgroup analysis of these four studies was conducted to explore the discriminating ability of the two diseases. The AI model had a pooled AUC of 0.95 (95% CI, 0.93-0.97), sensitivity of 0.93 (95% CI, 0.87-0.97), specificity of 0.88 (95% CI, 0.71-0.96), PLR of 8.04 (95% CI, 2.92-22.18), and NLR of 0.08 (95% CI, 0.04-0.15) (Supplementary Figures 4, 5).

We performed a subgroup analysis after excluding Nguyen’s study, as the AI model adopted was ANN. The combined sensitivity and specificity of AI-assisted EUS diagnosis of GIST on the CNN model were 0.93 (95% CI, 0.87-0.97) and 0.81 (95% CI, 0.68-0.89) (Supplementary Figures 6A, B), respectively. The pooled PLR was 4.85 (95% CI, 2.81-8.36) and NLR was 0.08 (95% CI, 0.04-0.17) (Supplementary Figures 6C, D). Figure 7A shows the SROC curve of the AI-assisted EUS, with an AUC of 0.94 (95% CI, 0.92-0.96). The I2 was 50.57% for PLR, 74.05% for sensitivity, 71.16% for specificity, and 73.61 for NLR, indicating that significant heterogeneity existed in the pooled sensitivity, specificity, and NLR.

FIGURE 7
www.frontiersin.org

Figure 7 SROC curves of AI-assisted EUS and EUS experts in diagnosis of GIST. (A) SROC curve of seven studies on CNN AI-models. (B) SROC curve of seven studies on imaging modality. (C) SROC curve of seven studies based on patients. (D)The SROC curve of the EUS experts.

We also conducted a subgroup analysis of seven studies on imaging modalities without CH-EUS. The AI model had a pooled AUC of 0.92 (95% CI, 0.89-0.94) (Figure 7B), sensitivity of 0.92 (95% CI, 0.84-0.97), specificity of 0.79 (95% CI, 0.68-0.87), PLR of 4.39 (95% CI, 2.85-6.78), and NLR of 0.10 (95% CI, 0.04-0.21) (Supplementary Figure 7). However, the heterogeneity within the subgroups was still significantly high.

A subgroup analysis of AI-assisted EUS diagnosis of GIST was performed on the study subjects, namely six studies based on patients and not regions of interest. As shown in Supplementary Figure 8, the combined sensitivity, specificity, PLR, and NLR were 0.95 (95% CI, 0.89-0.97), 0.83 (95% CI, 0.67-0.92), 5.43 (95% CI, 2.75-10.71), 0.07 (95% CI, 0.03-0.13), respectively. The SROC curve, with an AUC of 0.96 (95% CI, 0.94-0.97), is displayed in Figure 7C. I2 was 38.69% for PLR, 51.31% for NLR, 58.46% for sensitivity, and 71.06% for specificity, indicating a low degree of heterogeneity in PLR, whereas there was moderate heterogeneity in NLR, sensitivity, and specificity.

To further explore the source of heterogeneity, we performed meta-regression analysis. The number of samples was a major source of heterogeneity in univariate meta-regression analysis (p <0.001, Figure 8). Study quality (p = 0.03) and study subjects (p = 0.01) were major sources of heterogeneity in the joint meta-regression model (Table 2).

TABLE 2
www.frontiersin.org

Table 2 P-value of parameters in the joint model.

FIGURE 8
www.frontiersin.org

Figure 8 Univariate meta-regression for the reason of heterogeneity in sensitivity and specificity.

Comparison between AI and EUS experts

Seven studies simultaneously tested the accuracy of EUS experts in the diagnosis of GIST. All EUS experts performed more than 500 EUS examinations or had at least 5-year experience in evaluating gastrointestinal SELs. The SROC curve of the EUS experts, with an AUC of 0.75 (95% CI, 0.71-0.78), is displayed in Figure 7D. The pooled sensitivity of EUS experts in diagnosing GIST was 0.71 (95% CI, 0.63-0.78) (Figure 9A) and specificity was 0.69 (95% CI, 0.61-0.76) (Figure 9B). The combined PLR and NLR are 2.28 (95% CI, 1.85-2.82) (Figure 9C) and 0.42 (95% CI, 0.33-0.54) (Figure 9D), respectively. There was little heterogeneity in the specificity (p = 0.37), PLR (p = 0.69), and NLR (p = 0.12).

FIGURE 9
www.frontiersin.org

Figure 9 Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of EUS experts in diagnosis of GIST.

For diagnosis of GIST under EUS, AI was superior to EUS experts in terms of sensitivity [0.93 (95% CI, 0.87-0.97) vs. 0.71 (95% CI, 0.63-0.78)], specificity [0.81 (95% CI, 0.68-0.89) vs. 0.69 (95% CI, 0.61-0.76)], and PLR [4.85 (95% CI, 2.81-8.36) vs. 2.28 (95% CI, 1.85-2.82)], and NLR [0.08 (95% CI, 0.04-0.17) vs. 0.42 (95% CI, 0.33-0.54)]. Figure 10 shows the comparison of SROC curves between AI-assisted EUS models and EUS experts with AUC of 0.94 (95% CI, 0.92-0.96) vs. 0.75 (95% CI, 0.71-0.78), suggesting that AI-assisted EUS models have better diagnostic performance.

FIGURE 10
www.frontiersin.org

Figure 10 Comparison of SROC curves between AI-assisted EUS models and EUS experts.

Discussion

With the application of artificial intelligence in medical imaging technology, an increasing number of diseases have advanced their diagnosis and treatment methods. In this systematic review and meta-analysis, we explored the application of computer-aided diagnosis systems in gastrointestinal SELs and found that artificial intelligence algorithm models have excellent diagnostic performance with a sensitivity of 0.92 (95% CI, 0.85-0.96) and specificity of 0.80 (95% CI, 0.70-0.87). EUS is currently the most accurate and prevalent imaging modality for evaluating gastrointestinal SELs because of its ability to penetrate tissue layers and, thus, most likely identify the origin of the lesion (1). A previous study has shown that CH-EUS has better diagnostic performance than B-mode EUS in distinguishing leiomyomas from GIST and discriminating the risk stratification of GIST (27). In addition to improving the equipment performance and imaging technology of EUS, the application of artificial intelligence undoubtedly compensates for the limitations of EUS. With the help of the AI system, it is expected to shorten the diagnostic time, improve diagnostic efficiency, and reduce the misdiagnosis rate of GIST and other benign lesions, thus avoiding unnecessary EUS tests, invasive biopsies, and surgical operations.

In our initial literature search, we found that Kim and Lee used digital image analysis of objective information provided by EUS images to diagnose gastric stromal tumors (28, 29). We excluded these two studies because they were limited to analyzing the features of EUS images and did not develop corresponding AI models. We also found four studies that explored the application of AI in the malignancy stratification of GISTs, and the overall accuracy of the AI models in predicting the malignant potential of GISTs was 66.0%-83.4% (3033). During the literature search, we found that several studies have explored the application of AI in SELs, especially GISTs. Therefore, we systematically and comprehensively summarized the application of AI-assisted EUS for the diagnosis of SELs. Although there are many types of SELs, most studies classified SELs into two categories: GIST and non-GIST, to explore the accuracy of AI-assisted EUS. In four studies, the non-GISTs only referred to leiomyoma, and we performed a subgroup analysis (22, 23, 25, 26). Nguyen et al. developed an ANN with excellent performance for differentiating lipomas (AUC=0.92), carcinoids (AUC=0.86), and GISTs (AUC=0.89) (19). Despite the SELs involved in the Minoda’s research, including GIST, leiomyoma, schwannoma, and aberrant pancreas, the results section was still divided into GIST and non-GIST for exploration (20). Kim et al. utilized CNN-CAD to first classify SELs into GIST and non-GIST tumors, and then sub-classified the non-GIST tumors into leiomyomas and schwannomas. Accuracy of the CNN-CAD system in differentiating leiomyomas from schwannomas was 85.0% (95% CI: 81.6-87.7%) (21). In the Hirai’s study, accuracy of the AI system for five-category classification was 86.1%, including GIST, leiomyoma, schwannoma, NET, and ectopic pancreas (24).

Nguyen trained, constructed, and internally validated an ANN through unsupervised and supervised learning based on the features extracted through texture analysis (19). In the traditional sense, ANN is a type of machine learning (ML). As a computer application, ML can recognize patterns in training data and generate mathematical models to develop an AI system to realize the recognition and prediction function, similar to the learning behavior of humans (13). Other studies trained CNN models using deep-learning algorithms. Deep learning-based analysis does not need to measure characteristic values, as they can be automatically and accurately identified, thereby demonstrating greater diagnostic ability (34). This is consistent with our findings that the combined AUC of CNN model after excluding the ANN model was improved from 0.92 (95% CI, 0.90-0.94) to 0.94 (95% CI, 0.92-0.96).

Heterogeneity is a prominent issue in this meta-analysis. Although we performed subgroup analyses based on the AI models, imaging modalities, and study subjects, the heterogeneity was not completely eliminated. Possible reasons for this are as follows: First, we have to consider the diversity of clinical samples, as most of the included studies were from different countries, and the manufacturers and models of EUS were inconsistent. In addition, the sample size was not sufficiently large. Second, methodological diversity should be considered. The specific algorithms, tools used, and parameter settings were not uniform, despite the fact that seven studies applied the CNN deep-learning model. The EUS expert group had little heterogeneity, probably because all EUS experts were selected on the basis of having performed more than 500 EUS examinations or having at least 5 years of experience in assessing gastrointestinal SELs. Additionally, different trial designs also contributed to the heterogeneity. Only two studies applied training, validation, and test sets (22, 24). Others merely had two datasets: one set to develop the AI model and the other to validate it. Considering the existence of heterogeneity, we avoided directly adopting a fixed-effects model.

In this review and meta-analysis, the diagnostic performance of AI models was superior to EUS experts, with the accuracy of 0.94 (95% CI, 0.92-0.96) vs. 0.75 (95% CI, 0.71-0.78). Additionally, two studies also investigated the diagnostic accuracy of AI-assisted EUS according to the size of SELs, ≥ 20 mm and <20 mm. Minoda et al. found that the accuracy, sensitivity, and specificity of SELs ≥ 20 mm between AI-assisted EUS and EUS experts were 90.0% vs. 53.3%, 91.7% vs. 50.0%, and 83.3% vs. 83.3%, respectively. The diagnostic performance for SELs ≥ 20 mm of AI-assisted EUS was significantly better than that of EUS experts, with an AUC of 0.965 vs. 0.684 (p = 0.007) (20). Tanaka et al. discovered that the diagnostic performance of AI and experts was completely consistent for cases with lesions <20 mm, but the specificity and accuracy of AI in diagnosing GISTs ≥ 20 mm were superior to those of experts (87.5% vs. 75.0% and 88.9% vs. 86.1%, respectively) (26). Therefore, we need to further develop and improve artificial intelligence algorithms to improve their performance in the diagnosis of small lesions.

This is the first systematic review and meta-analysis of AI-assisted EUS for SEL diagnosis. We summarized recent advances in AI in the diagnosis and differential diagnosis of SELs and evaluated the overall diagnostic performance of AI. Our meta-analysis also has some limitations. Although no publication bias existed, the number of eligible studies was limited (n=8) and most of the included studies were retrospective. Future studies are expected to expand the sample size, supplement videos, add external validation datasets, and conduct prospective real-time clinical studies to further confirm the credibility of AI diagnostic performance. In addition, the issue of heterogeneity among studies is also discussed above.

In conclusion, AI-assisted EUS is a promising and reliable method for differentiating SELs with high accuracy, and may become an important tool to assist endoscopists in diagnosing SELs in the near future.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Author contributions

X-YLiu, WS, and TM was responsible for literature search and screening, data extraction, analysis and interpretation, and manuscript writing. QZ and CZ was responsible for the statistical analysis and revising the manuscript. X-YLi was responsible for revising the manuscript, financial support, and final approval of the manuscript. All authors have read and approved the final version of the manuscript.

Funding

The study was supported by the National Natural Science Foundation (No. 81802777), the “Clinical medicine + X” scientific research project of Affiliated Hospital of Qingdao University, and Qingdao Chinese Medicine Technology Project (2021-zyym26).

Acknowledgments

We would like to thank Editage (www.editage.com) for English language editing. And we thank all the authors for helping with the writing and publication of this article.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.915481/full#supplementary-material

Supplementary figure 1 | The diagnostic score and DOR of AI-assisted EUS diagnosis of GIST.

Supplementary figure 2 | Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS diagnosis of GIST including prospective test set.

Supplementary figure 3 | The diagnostic score and DOR of AI-assisted EUS diagnosis of GIST including prospective diagnostic test set.

Supplementary figure 4 | SROC curves of AI-assisted EUS of four studies for the diagnosis of GIST and leiomyoma.

Supplementary figure 5 | Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS for the differential diagnosis of GIST and leiomyoma.

Supplementary figure 6 | Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS of seven studies on CNN AI-models.

Supplementary figure 7 | Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS of seven studies on imaging modality.

Supplementary figure 8 | Sensitivity (A), specificity (B), positive likelihood ratio (C), negative likelihood ratio (D) of AI-assisted EUS of seven studies based on patients.

Supplementary Table 1 | The training, validation, and test datasets of included studies.

Supplementary Table 2 | Types of SELs of included studies.

References

1. Faulx AL, Kothari S, Acosta RD, Agrawal D, Bruining DH, Chandrasekhara V, et al. The role of endoscopy in subepithelial lesions of the GI tract. Gastrointest Endosc (2017) 85:1117–32. doi: 10.1016/j.gie.2017.02.022

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Papanikolaou IS, Triantafyllou K, Kourikou A, Rösch T. Endoscopic ultrasonography for gastric submucosal lesions. World J Gastrointest Endosc (2011) 3:86–94. doi: 10.4253/wjge.v3.i5.86

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Rubin BP, Heinrich MC, Corless CL. Gastrointestinal stromal tumour. Lancet (2007) 369:1731–41. doi: 10.1016/s0140-6736(07)60780-6

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Park CH, Kim EH, Jung DH, Chung H, Park JC, Shin SK, et al. Impact of periodic endoscopy on incidentally diagnosed gastric gastrointestinal stromal tumors: findings in surgically resected and confirmed lesions. Ann Surg Oncol (2015) 22:2933–9. doi: 10.1245/s10434-015-4517-0

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Joensuu H, Hohenberger P, Corless CL. Gastrointestinal stromal tumour. Lancet (2013) 382:973–83. doi: 10.1016/s0140-6736(13)60106-3

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Alkhatib AA, Faigel DO. Endoscopic ultrasonography-guided diagnosis of subepithelial tumors. Gastrointest Endosc Clin N Am (2012) 22:187–205. doi: 10.1016/j.giec.2012.04.006

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Lim TW, Choi CW, Kang DH, Kim HW, Park SB, Kim SJ. Endoscopic ultrasound without tissue acquisition has poor accuracy for diagnosing gastric subepithelial tumors. Med (Baltimore) (2016) 95:e5246. doi: 10.1097/md.0000000000005246

CrossRef Full Text | Google Scholar

8. Kim GH, Cho YK, Kim EY, Kim HK, Cho JW, Lee TH, et al. Comparison of 22-gauge aspiration needle with 22-gauge biopsy needle in endoscopic ultrasonography-guided subepithelial tumor sampling. Scand J Gastroenterol (2014) 49:347–54. doi: 10.3109/00365521.2013.867361

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Han JP, Lee TH, Hong SJ, Kim HK, Noh HM, Lee YN, et al. EUS-guided FNA and FNB after on-site cytological evaluation in gastric subepithelial tumors. J Dig Dis (2016) 17:582–7. doi: 10.1111/1751-2980.12381

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Zhang XC, Li QL, Yu YF, Yao LQ, Xu MD, Zhang YQ, et al. Diagnostic efficacy of endoscopic ultrasound-guided needle sampling for upper gastrointestinal subepithelial lesions: a meta-analysis. Surg Endosc (2016) 30:2431–41. doi: 10.1007/s00464-015-4494-1

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Le Berre C, Sandborn WJ, Aridhi S, Devignes MD, Fournier L, Smaïl-Tabbone M, et al. Application of artificial intelligence to gastroenterology and hepatology. Gastroenterology (2020) 158:76–94.e2. doi: 10.1053/j.gastro.2019.08.058

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Lee H, Kwon H. Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans Image Process (2017) 26:4843–55. doi: 10.1109/tip.2017.2725580

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Kuwahara T, Hara K, Mizuno N, Haba S, Okuno N, Koda H, et al. Current status of artificial intelligence analysis for endoscopic ultrasonography. Dig Endosc (2021) 33:298–305. doi: 10.1111/den.13880

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Tonozuka R, Itoi T, Nagata N, Kojima H, Sofuni A, Tsuchiya T, et al. Deep learning analysis for the detection of pancreatic cancer on endosonographic images: a pilot study. J Hepatobiliary Pancreat Sci (2021) 28:95–104. doi: 10.1002/jhbp.825

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Marya NB, Powers PD, Chari ST, Gleeson FC, Leggett CL, Abu Dayyeh BK, et al. Utilisation of artificial intelligence for the development of an EUS-convolutional neural network model trained to enhance the diagnosis of autoimmune pancreatitis. Gut (2021) 70:1335–44. doi: 10.1136/gutjnl-2020-322821

PubMed Abstract | CrossRef Full Text | Google Scholar

16. McInnes MDF, Moher D, Thombs BD, McGrath TA, Bossuyt PM, Clifford T, et al. Preferred reporting items for a systematic review and meta-analysis of diagnostic test accuracy studies: The PRISMA-DTA statement. JAMA (2018) 319:388–96. doi: 10.1001/jama.2017.19163

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ, Reitsma JB, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med (2011) 155:529–36. doi: 10.7326/0003-4819-155-8-201110180-00009

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Prasoppokakorn T, Tiyarattanachai T, Chaiteerakij R, Decharatanachart P, Mekaroonkamol P, Ridtitid W, et al. Application of artificial intelligence for diagnosis of pancreatic ductal adenocarcinoma by EUS: A systematic review and meta-analysis. Endosc Ultrasound (2022) 11:17–26. doi: 10.4103/eus-d-20-00219

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Nguyen VX, Nguyen CC, Li B, Das A. Digital image analysis is a useful adjunct to endoscopic ultrasonographic diagnosis of subepithelial lesions of the gastrointestinal tract. J Ultrasound Med (2010) 29:1345–51. doi: 10.7863/jum.2010.29.9.1345

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Minoda Y, Ihara E, Komori K, Ogino H, Otsuka Y, Chinen T, et al. Efficacy of endoscopic ultrasound with artificial intelligence for the diagnosis of gastrointestinal stromal tumors. J Gastroenterol (2020) 55:1119–26. doi: 10.1007/s00535-020-01725-4

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Kim YH, Kim GH, Kim KB, Lee MW, Lee BE, Baek DH, et al. Application of a convolutional neural network in the diagnosis of gastric mesenchymal tumors on endoscopic ultrasonography images. J Clin Med (2020) 9:3162. doi: 10.3390/jcm9103162

CrossRef Full Text | Google Scholar

22. Yang X, Wang H, Dong Q, Xu Y, Liu H, Ma X, et al. An artificial intelligence system for distinguishing between gastrointestinal stromal tumors and leiomyomas using endoscopic ultrasonography. Endoscopy (2022) 54:251–61. doi: 10.1055/a-1476-8931

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Oh CK, Kim T, Cho YK, Cheung DY, Lee BI, Cho YS, et al. Convolutional neural network-based object detection model to identify gastrointestinal stromal tumors in endoscopic ultrasound images. J Gastroenterol Hepatol (2021) 36:3387–94. doi: 10.1111/jgh.15653

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Hirai K, Kuwahara T, Furukawa K, Kakushima N, Furune S, Yamamoto H, et al. Artificial intelligence-based diagnosis of upper gastrointestinal subepithelial lesions on endoscopic ultrasonography images. Gastric Cancer (2022) 25:382–91. doi: 10.1007/s10120-021-01261-x

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Seven G, Silahtaroglu G, Seven OO, Senturk H. Differentiating gastrointestinal stromal tumors from leiomyomas using a neural network trained on endoscopic ultrasonography images. Dig Dis (2021) 40:427–35. doi: 10.1159/000520032

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Tanaka H, Kamata K, Ishihara R, Handa H, Otsuka Y, Yoshida A, et al. Value of artificial intelligence with novel tumor tracking technology in the diagnosis of gastric submucosal tumors by contrast-enhanced harmonic endoscopic ultrasonography. J Gastroenterol Hepatol (2022) 37:841–46. doi: 10.1111/jgh.15780

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Lefort C, Gupta V, Lisotti A, Palazzo L, Fusaroli P, Pujol B, et al. Diagnosis of gastric submucosal tumors and estimation of malignant risk of GIST by endoscopic ultrasound. comparison between b mode and contrast-harmonic mode. Dig Liver Dis (2021) 53:1486–91. doi: 10.1016/j.dld.2021.06.013

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Kim GH, Kim KB, Lee SH, Jeon HK, Park DY, Jeon TY, et al. Digital image analysis of endoscopic ultrasonography is helpful in diagnosing gastric mesenchymal tumors. BMC Gastroenterol (2014) 14:7. doi: 10.1186/1471-230x-14-7

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Lee MW, Kim GH, Kim KB, Kim YH, Park DY, Choi CI, et al. Digital image analysis-based scoring system for endoscopic ultrasonography is useful in predicting gastrointestinal stromal tumors. Gastric Cancer (2019) 22:980–7. doi: 10.1007/s10120-019-00928-w

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Seven G, Silahtaroglu G, Kochan K, Ince AT, Arici DS, Senturk H. Use of artificial intelligence in the prediction of malignant potential of gastric gastrointestinal stromal tumors. Dig Dis Sci (2022) 67:273–81. doi: 10.1007/s10620-021-06830-9

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Liu C, Qiao M, Jiang F, Guo Y, Jin Z, Wang Y. TN-USMA net: Triple normalization-based gastrointestinal stromal tumors classification on multicenter EUS images with ultrasound-specific pretraining and meta attention. Med Phys (2021) 48:7199–214. doi: 10.1002/mp.15172

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Li X, Jiang F, Guo Y, Jin Z, Wang Y. Computer-aided diagnosis of gastrointestinal stromal tumors: a radiomics method on endoscopic ultrasound image. Int J Comput Assist Radiol Surg (2019) 14:1635–45. doi: 10.1007/s11548-019-01993-3

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Seven G, Silahtaroglu G, Kochan K, Kiremitci S, Senturk H. Artificial intelligence in the prediction of malignant potential in gastrointestinal stromal tumors of stomach. United Eur Gastroenterol J (2020) 8:806. doi: 10.1177/2050640620927345

CrossRef Full Text | Google Scholar

34. Rawat W, Wang Z. Deep convolutional neural networks for image classification: a comprehensive review. Neural Comput (2017) 29:2352–449. doi: 10.1162/NECO_a_00990

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: artificial intelligence, computer-assisted diagnosis, endoscopic ultrasonography, subepithelial lesions, gastrointestinal stromal tumors

Citation: Liu X-Y, Song W, Mao T, Zhang Q, Zhang C and Li X-Y (2022) Application of artificial intelligence in the diagnosis of subepithelial lesions using endoscopic ultrasonography: a systematic review and meta-analysis. Front. Oncol. 12:915481. doi: 10.3389/fonc.2022.915481

Received: 08 April 2022; Accepted: 18 July 2022;
Published: 15 August 2022.

Edited by:

Jakub Nalepa, Silesian University of Technology, Poland

Reviewed by:

Gwang Ha Kim, Pusan National University Hospital, South Korea
Hakan Senturk, Bezmiâlem Vakıf Üniversitesi, Turkey

Copyright © 2022 Liu, Song, Mao, Zhang, Zhang and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xiao-Yu Li, lixiaoyu05@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.