- 1Department of Gastroenterology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- 2Shanghai Key Laboratory of Pancreatic Diseases, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- 3Department of Gastroenterology, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
- 4Department of Gastroenterology, Shanghai Tong Ren Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- 5School of Mechanical Engineering, Nanjing University of Science and Technology, Nanjing, China
- 6Department of Pathology, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
Background: Endoscopically visible gastric neoplastic lesions (GNLs), including early gastric cancer and intraepithelial neoplasia, should be accurately diagnosed and promptly treated. However, a high rate of missed diagnosis of GNLs contributes to the potential risk of the progression of gastric cancer. The aim of this study was to develop a deep learning-based computer-aided diagnosis (CAD) system for the diagnosis and segmentation of GNLs under magnifying endoscopy with narrow-band imaging (ME-NBI) in patients with suspected superficial lesions.
Methods: ME-NBI images of patients with GNLs in two centers were retrospectively analysed. Two convolutional neural network (CNN) modules were developed and trained on these images. CNN1 was trained to diagnose GNLs, and CNN2 was trained for segmentation. An additional internal test set and an external test set from another center were used to evaluate the diagnosis and segmentation performance.
Results: CNN1 showed a diagnostic performance with an accuracy, sensitivity, specificity, positive predictive value (PPV) and negative predictive value (NPV) of 90.8%, 92.5%, 89.0%, 89.4% and 92.2%, respectively, and an area under the curve (AUC) of 0.928 in the internal test set. With CNN1 assistance, all endoscopists had a higher accuracy than for an independent diagnosis. The average intersection over union (IOU) between CNN2 and the ground truth was 0.5837, with a precision, recall and the Dice coefficient of 0.776, 0.983 and 0.867, respectively.
Conclusions: This CAD system can be used as an auxiliary tool to diagnose and segment GNLs, assisting endoscopists in more accurately diagnosing GNLs and delineating their extent to improve the positive rate of lesion biopsy and ensure the integrity of endoscopic resection.
1. Introduction
Gastric cancer is one of the most prevalent malignant tumours with high morbidity and mortality worldwide (1). Early diagnosis and appropriate treatment are key measures in reducing the mortality rate of gastric cancer; however, most patients are diagnosed at a late stage. Intestinal-type gastric cancer develops through a “Correa cascade”: chronic atrophic gastritis, intestinal metaplasia, and gastric intraepithelial neoplasia (GIN) (2). GIN is the final stage of gastric carcinogenesis and is defined as intraepithelial neoplasia with no clear evidence of depth invasion (3). The newest Western guideline recommends that endoscopically visible low-grade intraepithelial neoplasia (LGIN) lesions should undergo endoscopic resection, similar to high-grade intraepithelial neoplasia (HGIN) and early gastric cancer (EGC), because of a high rate of histological upstaging after resection (4). Therefore, accurate diagnosis and prompt treatment of gastric neoplastic lesions (GNLs) including EGC and GIN, is necessary in clinical practice.
The application of magnifying endoscopy with narrow-band imaging (ME-NBI) dramatically improves the detection rate of EGC and its precancerous lesions compared to white light endoscopy (WLE) (5). According to Yao’s research, EGC can be diagnosed under ME-NBI if the lesion has a demarcation line (DL) and an irregular microsurface (MS) pattern and/or an irregular microvessel (MV) pattern, which is called the “vascular plus surface classification system (VSCS)” (6). However, in clinical practice, it is not easy for endoscopists to judge whether the MS or MV of some atypical superficial gastric lesions is regular, which depends on the experience and knowledge reserve of endoscopists with EGC. At present, there is no accepted standard for the endoscopic diagnosis of LGIN. Previous studies have shown that MS and MV are often regular in LGIN lesions, while DL sometimes exists because of the differences in VS morphology between lesion mucosa and background mucosa (7). In our previous study, an auxiliary index with a high accuracy of identifying LGIN under ME-NBI was proposed – endoscopic acanthosis nigricans appearance (EANA) (8). For LGIN lesions of type 0-IIa, studies have shown that the morphological characteristics of dense-type crypt opening or regular white opaque substance (WOS) under ME-NBI can be helpful in distinguishing them from carcinomas (9, 10). Some endoscopists also use the morphologic evolution of gastric pits based on the Sakaki classification system to identify precancerous lesions and EGC (11). However, the diagnosis of EGC and LGIN under endoscopy requires extensive experience and clinical practice, which is lacking for many endoscopists.
In recent years, with the development of deep learning, artificial intelligence (AI) has been increasingly used in medical image processing and has achieved excellent performance in recognising lesions under upper or lower gastrointestinal endoscopy in different modes, such as WLE, ME-NBI or blue laser imaging (12). Many computer-aided diagnostic (CAD) systems have been established for identifying lesions such as esophageal squamous cell carcinoma, EGC and intestinal metaplasia (13–17). However, most of the previous CAD systems were developed for the diagnosis of EGC, and there is no diagnostic system for GNLs (including LGIN and EGC). In addition, most CAD systems are used for targeting lesions, and few can accurately mark the boundary of the lesions at the same time. Due to the clinical requirements of follow-up endoscopic or surgical resection of GNLs, the extent of lesions should be accurately determined under image-enhanced endoscopy in many cases. Therefore, we developed a CAD system based on deep learning to diagnose and segment GNLs in patients with suspected superficial lesions.
2. Materials and methods
2.1. Study design and image capture
This retrospective study was performed at three endoscopy centers, located in Shanghai General Hospital – South (center 1), Shanghai Jiao Tong University Affiliated Sixth People’s Hospital (center 2) and Shanghai General Hospital – North (center 3). It should be noted that the center 1 and the center 3 are independently operated in different locations, with fixed staff and endoscopy equipment purchased in different years. This study was conducted in accordance with the Declaration of Helsinki and was approved by the Ethics Committee of the Shanghai General Hospital (2020KY236). ME-NBI images from October 2014 to June 2021 were retrospectively collected. Every patient underwent conventional upper gastrointestinal WLE, and then the suspected superficial lesions were carefully examined under ME-NBI. The ME-NBI images were captured by standard endoscopes (GIF-H260Z or GIF-H290Z, Olympus Co., Tokyo, Japan) with an EVIS LUCERA ELITE endoscopic system (CV-290, Olympus Co.) and an endoscopic cold light source (CLV-290SL, Olympus Co.). A black soft hood was attached to the tip of the endoscope to obtain stable images at maximum magnification.
2.2. Setting up the datasets
Suspected superficial lesions were confirmed by pathological data obtained from the targeted biopsy samples, endoscopic submucosal dissection (ESD) samples or surgical samples. All pathological data from the three centers were assessed by senior gastrointestinal pathologists according to the Vienna classification (18). Category 3 was defined as LGIN lesions, and Categories 4 and 5 were defined as EGC lesions. LGIN and EGC lesions were defined as GNLs.
To distinguish neoplastic lesions from their background mucosa, ME-NBI images of chronic gastritis and intestinal metaplasia were obtained as background training under the same imaging conditions. Images of esophageal lesions, duodenal lesions, submucosal tumours, those with a lack of pathological data, and images of poor quality (including bleeding, mucous adherence, presence of foreign bodies and out of focus) were excluded from the analysis.
In the ME-NBI images, the following two criteria should be met for labelling superficial gastric lesions as LGIN: (1) there are DLs that can be identified under ME-NBI; (2) the presence of EANA or types IV–VI pit pattern of the Sakaki classification or regular white opaque substance (WOS) or dense-type crypt opening (8–11). The following criteria should be met for labelling superficial gastric lesions as EGC: (1) an irregular MV pattern with DL and (or) (2) an irregular MS pattern with DL (6).
All images were JPEG files with various sizes such as 1020×764 pixels, 716×512 pixels and 764×572 pixels. Baseline data such as the date of diagnosis of GNLs, lesion locations, macroscopic types and maximum diameter, were collected. Then, the redundant parts of the images, such as the patient information and acquisition time, were cropped before labelling. Two senior endoscopists separately labelled the images as “GNLs” or “non-GNLs” according to the pathological data and marked the location of GNLs using rectangular box annotation by LabelImg (version 1.8.6, https://github.com/heartexlabs/labelImg), then saved as XML files. If they were in agreement, the image was considered to be acceptable. If they had any disagreement, then another senior endoscopist made the judgement. An expert endoscopist used polygon annotation to label the extent of the GNLs by Labelme (version 4.2.10, https://github.com/wkentaro/labelme), saved as a JSON file. Before training, the images were resized to 416×416 pixels, and the number of ME-NBI images in the training set was expanded by rotating and flipping. Figure 1 shows a pipeline diagram of this study.
Figure 1 Pipeline diagram of this study. ME-NBI images were collected as JPEG files. After pathological validation, the images were resized and expanded, and labelled by endoscopists for training the different deep learning models. CNN1 and CNN2 are two parallel models. CNN1 output the confidence of GNLs and CNN2 output the segmentation and localisation of GNLs. Finally, the performance of CNN1 and CNN2 were evaluated. ME-MBI, magnifying endoscopy with narrow band imaging; CNN, convolutional neural network; GNLs, gastric neoplastic lesions.
All of the images were divided into the following datasets:
1. A total of 3757 images from 392 patients with GNLs and 2420 images from 568 patients with non-GNLs from center 1 and center 2 were used to train and test the performance of CNN1 and CNN2. Patients were randomly assigned to the training group and the internal test group, according to the 85:15 ratio of patients with GNLs. The training set consisted of 3331 ME-NBI images of GNLs from 333 patients and 2220 images from 492 patients with non-GNLs from center 1 and center 2, which were used to train CNN1 and CNN2.
2. The internal test set consisted of 200 images from 59 patients with GNLs and 200 images from 76 patients with non-GNLs from center 1 and center 2, which were not involved in the training of the modules and were used to test the performance of CNN1, and another 226 high-definition GNLs images with sufficient magnification from the same 59 patients with GNLs were selected to test the segmentation performance of CNN2.
3. The external test set consisted of 800 images from 106 patients with GNLs and 800 images from 130 patients with non-GNLs. These images were from center 3 and were used to test the generalisation ability of CNN1.
4. The comparison test set consisted of 100 randomly picked GNLs images and 100 non-GNLs images from both the internal and external test sets, which were used to compare the diagnostic performance of CNN1 with that of the endoscopists.
A flowchart of setting the datasets is shown in Figure 2.
Figure 2 Flowchart of setting the datasets. WLE, white light endoscopy; ME-MBI, magnifying endoscopy with narrow band imaging; CNN, convolutional neural network; GNLs, gastric neoplastic lesions.
2.3. Development of deep learning algorithm
The CAD system consists of CNN1, which is responsible for diagnosis, and CNN2, which is responsible for the lesion segmentation and localisation. CNN1 is based on YOLO V3 (19). EfficientNet B2 was used as the backbone to replace the original DarkNet53 to obtain better feature extraction results (Figure 3A). The EfficientNet B2 model was pretrained on the Pascal VOC 2007 dataset and used as the initial weights. CNN2 is based on UNet, and a pretrained VGG-16 on ImageNet was used to replace the backbone of UNet (Figure 3B) (20, 21). Feature pyramid networks (FPN) (22) and an attention module (23) were additionally added to improve the feature extraction performance.
Figure 3 Structure of CNN1 and CNN2. (A) CNN1 is based on the YOLO v3 network. Based on the results of the pre-experiment, we chose EfficientNet B2 to replace Darknet53 as the backbone for feature extraction. An FPN structure was used for the fusion of feature layers to combine different information about the features. Three effective feature layers of different sizes obtained by EfficientNet B2 and FPN were regressed and classified by YOLO Head. (B) CNN2 is based on UNet. To obtain better feature extraction, we used pretrained VGG-16 instead of UNet’s backbone (left). After a series of upsampling and concatenation operations, the 5 effective feature layers extracted by VGG-16 were dimensionally superimposed and convolved twice (right). Finally, the final segmentation result was output after a softmax function. CNN, convolutional neural network; FPN, feature pyramid networks.
To improve the accuracy and the generalisation and to avoid overfitting, the shallow-level convolutional layers of the pretrained DarkNet 53 and VGG-16 were frozen, and the weights of the high-level convolutional layers were updated during the training. The initial image size was 416×416 pixels, 8 images were put into training at a time, and the learning rate was automatically adjusted for the rate of decline. The model was evaluated every 10 iterations to calculate the average accuracy of the validation set and save the weights. The total number of training iterations was 500 in CNN1 and 100 in CNN2. Training details for CNN1 and CNN2 are provided in the Supplementary Material.
2.4. Evaluation of the deep learning-based CAD system
In the CNN1 model, the Youden index was used to calculate the threshold for binary classification. If the output confidence value of a neoplastic image is higher than the specified threshold, the image is considered a true positive (TP), and if it is less than the specified threshold, it is considered a false negative (FN). A non-neoplastic image is considered a false positive (FP) if its confidence value is higher than the specified threshold or a true negative (TN) if it is less than the specified threshold.
The diagnostic performance of CNN1 was evaluated by accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), receiver operating characteristic (ROC) curve and area under the curve (AUC).
In the CNN2 model, intersection over union (IOU), precision, recall and the Dice coefficient were used to evaluate the segmentation performance. IOU means the ratio of the intersection and union of the AI segmentation region to the lesion labelled by an expert endoscopist (ground truth). If the IOU between the CNN2 segmentation and the ground truth is higher than a specified threshold, it is considered a TP; otherwise, it is an FP. If IOU = 0, it is considered an FN.
Precision, recall, and the Dice coefficients (24) were calculated as follows:
To test for differences in the diagnostic outcomes between the CAD system and the endoscopists, endoscopists of different experience levels were selected to judge the images in the test set. Three endoscopists were considered senior endoscopists with more than five years of endoscopy experience. Moreover, they have received education and training in identifying EGC under ME-NBI and had participated in EGC screening. Three endoscopists were regarded as junior endoscopists with less than five years of experience in endoscopy and no specific EGC training.
2.5. Statistical analysis
All statistical analyses were conducted using SPSS 22.0 (IBM, Armonk, NY, USA). The 95% confidence intervals (95% CI) were evaluated using the Wilson method. Continuous variables were compared by one-way ANOVA, and categorical variables were compared by the chi-square test or Fisher’s exact test. A chi-square test was conducted to compare the accuracy, sensitivity, specificity, PPV and NPV between CNN1 and the endoscopists. A p value less than 0.05 was considered as statistically significant.
3. Results
3.1. Clinicopathological characteristics of the patients and lesions
This study included 498 patients diagnosed with GNLs, including LGIN (n=296) and EGC (n=202), who were separated into a training set (n=333), internal test set (n=59) and external test set (n=106). The characteristics of the patients and lesions from the three sets are shown in Table 1. Patients in all three datasets showed no significant differences in terms of mean age, gender composition, lesion location or macroscopic type, except for a higher proportion of patients with EGC in the external test set. For EGC lesions, there were no significant differences in the mean diameter, depth of infiltration and degree of differentiation among patients from the three sets.
3.2. Diagnostic performance of CNN1
The diagnostic performance of CNN1 was tested in the internal and external test sets. The maximum Youden index is achieved when the classification threshold is set to 0.4920. In the internal test set, CNN1 correctly diagnosed 363 of 400 ME-NBI images with an AUC of 0.928 (95% CI, 0.899-0.956) and an accuracy of 90.8% (95%CI, 87.9%-93.6%). In the external test set, CNN1 diagnosed 1372 of 1600 ME-NBI images with an AUC of 0.918 (95% CI, 0.903-0.932) and an accuracy of 85.8% (95%CI, 84.0%-87.5%). The diagnostic performance of CNN1 is shown in Table 2, and the ROC curves of the internal and external test sets are shown in Figures 4A, B.
Figure 4 The ROC curve of CNN1 in the test sets. (A) ROC curve of CNN1 in the internal test set. (B) ROC curve of CNN1 in the external test set. ROC, receiver operating characteristic; CNN, convolutional neural network.
3.3. Comparison of the diagnostic performance between CNN1 and the endoscopists
To compare the diagnostic performance between CNN1 and the endoscopists, six endoscopists (three senior endoscopists and three junior endoscopists) who had not participated in the previous work on this study were selected to analyse the 100 images in the comparison test set. The diagnostic performance of GNLs by CNN1 and the endoscopists is shown in Table 3. The data showed that CNN1 outperformed the senior and junior endoscopists in average diagnostic performance.
Next, the performance of the endoscopists in diagnosing GNLs with CNN1 assistance was further evaluated. The endoscopists rediagnosed the same images with CNN1 assistance at least two weeks after their previous independent diagnosis. The results showed an average increase in diagnostic accuracy of 11.8%, with an average improvement of 10.3% for senior endoscopists and an average of 13.3% for junior endoscopists (Figure 5A). Because of differences in their diagnostic style, the improvement in sensitivity after CNN1 assistance varied considerably among the endoscopists, with an average improvement in sensitivity of 18.7%, including 9.3% for senior endoscopists and 28.0% for junior endoscopists (Figure 5B). The diagnostic performance of each endoscopist for GNLs is shown in Table 4.
Figure 5 Changes in the diagnostic performance for GNLs of endoscopists before and after CNN1 assistance. (A) Changes in accuracy for GNLs of endoscopists before and after CNN1 assistance. (B) Changes in sensitivity for GNLs of endoscopists before and after CNN1 assistance. (C) Changes in the accuracy for EGC of endoscopists before and after CNN1 assistance. (D) Changes in the accuracy for LGIN of endoscopists before and after CNN1 assistance. Red dots represent senior endoscopists, and blue dots represent junior endoscopists. AI, artificial intelligence; GNLs, gastric neoplastic lesions; CNN, convolutional neural network; EGC, early gastric cancer; LGIN, low-grade intraepithelial neoplasia.
Specifically, the average accuracy of diagnosing EGC increased 3.5% among the senior endoscopists, which may be because the participating senior endoscopists were already very good at diagnosing EGC under ME-NBI, and it increased 9.7% among the junior endoscopists (Figure 5C, Table 5). The use of CNN1 significantly improved the specificity of EGC diagnosis by senior endoscopists (68.1% to 80.6%) and the sensitivity of EGC diagnosis by junior endoscopists (75.0% to 91.7%) (Table 5). For LGIN lesions, the average accuracy was significantly increased for both senior and junior endoscopists (both at 16.7%) (Figure 5D, Table 6). This suggests that with CNN1 assistance, the endoscopist’s ability to diagnose LGIN is greatly improved and will reduce the rate of missed LGIN diagnoses.
3.4. Segmentation performance of CNN2
The results show that CNN2 has a better segmentation effect on GNLs. The average IOU between CNN2 and the expert label was 0.584. To assist in the endoscopic resection of GNLs, the IOU threshold was set at 0.5. The precision, recall and Dice coefficient were 0.776, 0.983 and 0.867, respectively. The performance of CNN2 and other segmentation modules was tested to compare the segmentation effect. CNN2 exhibits better segmentation than PSPNET and DEEPLAB in average IOU, precision, recall and Dice coefficient (Figure 6). The segmentation performance of different GNLs is shown in Figure 7.
Figure 6 Performance of CNN2 and other deep models on the test set. (A) The average IOU compared between UNet, PSPNet and DeepLab. (B) The precision, recall and the Dice coefficient of UNet, PSPNet and DeepLab at the IOU threshold of 0.5. CNN, convolutional neural network. IOU, Intersection over union.
Figure 7 Examples of segmentation performance of CNN2 in the test sets. The margins of the neoplastic lesions labelled by an expert endoscopist are shown as blue lines, and the margins of the lesions predicted by CNN2 are shown as red lines. CNN, convolutional neural network.
4. Discussion
In this study, a CAD system based on deep learning was developed for the diagnosis and segmentation of GNLs under ME-NBI. The CAD system achieves good diagnostic performance, with an accuracy, sensitivity, specificity, NPV and PPV of the system for diagnosing GNLs in the internal test set of 90.8%, 92.5%, 89.0%, 89.4% and 92.2%, respectively, and it was subjected to external testing. When the IOU threshold was set at 0.5, CNN2 showed a precision, recall and the Dice coefficient of 0.776, 0.983 and 0.867, respectively.
ESD has been widely used in the treatment of EGC and HGIN because of its low invasiveness and low risk of lymph node metastasis; however, whether LGIN lesions should be treated with endoscopic resection remains controversial (25–28). It has been reported that 15-26.9% of LGIN lesions develop into HGIN or gastric cancer (29, 30). In addition, for LGIN lesions, the overall discrepancy rate of pathological results between biopsy specimens and ESD specimens is high, and approximately 24% of lesions show elevated histological upstaging after ESD (31, 32). Depressed morphology, largest diameter of the lesion, surface unevenness and surface erythema had a higher odds ratio with potential histological upstaging (28, 32–34). Therefore, the latest Western guidelines recommend endoscopic resection of both endoscopically visible neoplastic lesions, including LGIN, to reduce the risk of malignant transformation (4).
The application of ME-NBI has significantly improved the diagnosis rate of superficial GNLs (35, 36). However, there are still some limitations that affect its clinical application. There is currently a well-established system for diagnosing EGC under ME-NBI, but endoscopists need to be well educated in ME-NBI. In addition, atypical LGIN is easily missed during an endoscopic examination because the diversity of lesions often makes it difficult to distinguish LGIN lesions from other non-neoplastic mucosa. Therefore, a deep-learning based CNN1 was developed to diagnose GNLs under ME-NBI. CNN1 showed a great performance in both the internal test set and the external test set.
In this study, the accuracy of senior endoscopists in diagnosing EGC was close to that of CNN1 in our test (Table 5). However, for all of the endoscopists tested, including senior and junior endoscopists, their diagnostic performance for diagnosing LGIN lesions was poor with a high rate of missed diagnosis (Table 6). With the assistance of CNN1, senior endoscopists had improved accuracy and sensitivity for diagnosing LGIN without reducing their diagnostic performance for EGC, while junior endoscopists had improved the diagnostic ability for both EGC and LGIN lesions. All endoscopists who participated in the test reported more confidence in their diagnosis of GNLs with the assistance of CNN1, especially for LGIN lesions. Moreover, all three junior endoscopists believed that the auxiliary value of CNN1 was primarily reflected in which lesions needed targeted biopsy for patients with multiple suspected gastric lesions.
In addition to correct diagnosis, accurate delineation of the lesion margins is necessary for accurate lesion biopsy and endoscopic resection (37). However, this requires sufficient clinical practice and experience and is difficult for junior endoscopists. To address this problem, a semantic segmentation-based CNN2 was developed to depict of the margins of GNLs. A preliminary endoscopic diagnosis of suspicious lesions was performed based on the characteristics of LGIN and EGC under ME-NBI as described in previous studies (6, 8–11) and endoscopic experts labelled the extent of the lesions based on the pathology data. Combining the feature extraction network VGG-16 with the improved UNet structure, CNN2 can effectively extract the vascular and texture details of the shallow feature layers and the colour details of the deep feature layers from the lesion images, which can perform better in lesion segmentation. Figure 8 shows the features extracted by CNN2. When the IOU threshold was set to 0.5, CNN2 showed a recall score of 0.983, which means that it can be used as a complement to the CNN1 diagnostic results to maximise the identification of the potential GNLs. In addition, the precision of 0.776 assists endoscopists in determining the margins of GNLs during endoscopic biopsy or ESD treatment, thus improving the accuracy of the biopsy or the complete resection rate of the lesions and reducing the risk of positive horizontal resection margins in pathological evaluation after ESD.
Figure 8 Characteristic layers extracted from endoscopist-labelled ME-NBI images by CNN2. (A) The margins of neoplastic lesions labelled by an expert endoscopist are shown as blue lines, and the margins predicted by CNN2 are shown as red lines. (B) Shallow characteristic layers of vascular and texture details extracted by CNN2. (C) Deep characteristic layers of colour details extracted by CNN2. A lighter colour indicates a high possibility of an abnormal area. ME-NBI, magnifying endoscopy with narrow-band imaging; CNN, convolutional neural network.
The advantages of our study are as follows: first, this study is a multicenter study, and images from three centers were used for training and testing. Second, the CAD system was trained and tested based on GNLs for the first time. Third, this CAD system combines two CNN modules to enable the diagnosis and segmentation of neoplastic lesions and it showed great performance.
However, there are still some defects in this CAD system. First, due to technical limitations, this model is still an “offline” system, which only allows the input of ME-NBI images for diagnosis and segmentation. This problem can be addressed by improving the data transmission and analysis speed, or by using other techniques to obtain the results of the AI analysis during endoscope examination to guide the diagnosis and treatment in real time. Second, this model cannot output the interpretability of the model yet. Futher studies will improve the performance of this model by adding a heatmap module. Third, low-quality images, such as bleeding and mucus coverage, were excluded during the training. However, these images may also have some value in assisting endoscopists with the diagnosis. In future studies, these images will be incorporated to achieve a generalisation of the model.
In summary, the newly developed CAD system can be used as a high-sensitivity auxiliary tool to diagnose and determine the margins of GNLs under ME-NBI in patients with suspected superficial lesions. This system can assist endoscopists in identifying neoplastic lesions quickly and accurately by marking the margins of the lesions to assist them in making correct decisions and achieving precise treatment. This will ultimately reduce the risk of progression to gastric cancer and avoid the serious consequences of gastric cancer.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
Conceptualization: HZ, XJW, and LGL; Methodology: LHL, JC, ZD, WN, JX, TM, and XWW; Formal analysis and investigation: LHL, XB, KQ, CY, JW, and WN; Writing - original draft preparation: LHL, JC, and ZD; Writing - review and editing: HZ, XJW, and LGL; Funding acquisition: HZ and XJW; Resources: HZ, XJW, LGL, and XB; Supervision: HZ, XB, and JW. All authors contributed to the article and approved the submitted version.
Funding
This research was funded by Shanghai Science and Technology Program (grant number 18411952900 and 19411951500) and Clinical Research Innovation Plan of Shanghai General Hospital (grant number CTCCR-2021B01).
Acknowledgments
The authors thank Professor Binjie Qin of School of Biomedical Engineering, Shanghai Jiao Tong University for his guidance in the development of the deep learning algorithm.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.1075578/full#supplementary-material
References
1. Smyth EC, Nilsson M, Grabsch HI, van Grieken NC, Lordick F. Gastric cancer. Lancet (2020) 396(10251):635–48. doi: 10.1016/S0140-6736(20)31288-5
2. Correa P. Human gastric carcinogenesis: a multistep and multifactorial process–first merican cancer society award lecture on cancer epidemiology and prevention. Cancer Res (1992) 52(24):6735–40.
3. Schlemper RJ, Kato Y, Stolte M. Review of histological classifications of gastrointestinal epithelial neoplasia: Differences in diagnosis of early carcinomas between Japanese and Western pathologists. J Gastroenterol (2001) 36(7):445–56. doi: 10.1007/s005350170067
4. Pimentel-Nunes P, Libânio D, Marcos-Pinto R, Areia M, Leja M, Esposito G, et al. Management of epithelial precancerous conditions and lesions in the stomach (MAPS II): European society of gastrointestinal endoscopy (ESGE), European helicobacter and microbiota study group (EHMSG), European society of pathology (ESP), and sociedade portuguesa de endoscopia digestiva (SPED) guideline update 2019. Endoscopy (2019) 51(4):365–88. doi: 10.1055/a-0859-1883
5. Kato M, Kaise M, Yonezawa J, Toyoizumi H, Yoshimura N, Yoshidaet Y, et al. Magnifying endoscopy with narrow-band imaging achieves superior accuracy in the differential diagnosis of superficial gastric lesions identified with white-light endoscopy: a prospective study. Gastrointest Endosc(2010) 72(3):523–29. doi: 10.1016/j.gie.2010.04.041
6. Yao K, Anagnostopoulos GK, Ragunath K. Magnifying endoscopy for diagnosing and delineating early gastric cancer. Endoscopy (2009) 41(5):462–7. doi: 10.1055/s-0029-1214594
7. Yao K. Analysis and interpretation of magnifying endoscopy with narrow band imaging (M-NBI) findings in gastric epithelial tumors (Early gastric cancer and adenoma) stratified for Paris classification of macroscopic appearance. In: Zoom gastroscopy. Tokyo, Japan: Springer (2014). p. 123–6.
8. Cheng J, Xu X, Zhuang Q, Luo S, Gong X, Wu X, et al. Endoscopic acanthosis nigricans appearance: A novel specific marker for diagnosis of low-grade intraepithelial neoplasia. J Gastroenterol Hepatol (2020) 35(8):1372–80. doi: 10.1111/jgh.15000
9. Kanesaka T, Sekikawa A, Tsumura T, Maruo T, Osaki Y, Wakasa T, et al. Dense-type crypt opening seen on magnifying endoscopy with narrow-band imaging is a feature of gastric adenoma. Dig Endosc. (2014) 26(1):57–62. doi: 10.1111/den.12076
10. Yao K, Iwashita A, Tanabe H, Nishimata N, Nagahama T, Maki S, et al. White opaque substance within superficial elevated gastric neoplasia as visualized by magnification endoscopy with narrow-band imaging: A new optical sign for differentiating between adenoma and carcinoma. Gastrointest Endosc (2008) 68(3):574–80. doi: 10.1016/j.gie.2008.04.011
11. Sakaki N, Iida Y, Okazaki Y, Kawamura S, Takemoto T. Magnifying endoscopic observation of the gastric mucosa, particularly in patients with atrophic gastritis. Endoscopy (1978) 10:269–74. doi: 10.1055/s-0028-1098307
12. Arribas J, Antonelli G, Frazzoni L, Fuccio L, Ebigbo A, van der Sommenet F, et al. Standalone performance of artificial intelligence for upper GI neoplasia: A meta-analysis. Gut (2021) 70:1458–68. doi: 10.1136/gutjnl-2020-321922
13. Cai SL, Li B, Tan WM, Niu XJ, Yu HH, Yao LQ, et al. Using a deep learning system in endoscopy for screening of early esophageal squamous cell carcinoma (with video). Gastrointest Endosc. (2019) 90(5):745–53.e2. doi: 10.1016/j.gie.2019.06.044
14. Ueyama H, Kato Y, Akazawa Y, Yatagai N, Komori H, Takeda T, et al. Application of artificial intelligence using a convolutional neural network for diagnosis of early gastric cancer based on magnifying endoscopy with narrow-band imaging. J Gastroenterol Hepatol (2021) 36(2):482–9. doi: 10.1111/jgh.15190
15. Ikenoyama Y, Hirasawa T, Ishioka M, Namikawa K, Yoshimizu S, Horiuchi Y, et al. Detecting early gastric cancer: Comparison between the diagnostic ability of convolutional neural networks and endoscopists. Dig Endosc. (2021) 33(1):141–50. doi: 10.1111/den.13688
16. Xu M, Zhou W, Wu L, Zhang J, Wang J, Mu G, et al. Artificial intelligence in the diagnosis of gastric precancerous conditions by image-enhanced endoscopy: A multicenter, diagnostic study (with video). Gastrointest Endosc. (2021) 94(3):540–8.e4. doi: 10.1016/j.gie.2021.03.013
17. Zhang L, Zhang Y, Wang L, Wang J, Liu Y. Diagnosis of gastric lesions through a deep convolutional neural network. Dig Endosc. (2021) 33(5):788–96. doi: 10.1111/den.13844
18. Dixon MF. Gastrointestinal epithelial neoplasia: Vienna revisited. Gut (2002) 51(1):130–1. doi: 10.1136/gut.51.1.130
19. Redmonm J, Farhadi A. Yolov3: An incremental improvement. arXiv Preprint (2018) arXiv:1804.02767. doi: 10.48550/arXiv.1804.02767
20. Ronneberger O, Fischer P, Brox T. (2015). U-Net: convolutional networks for biomedical image segmentation, in: International Conference on Medical Image Computing and Computer-Assisted Intervention. Cham: Springer (2015). p. 234–41. doi: 10.1007/978-3-319-24574-4_28
21. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv Preprint (2014). arXiv:1409.1556. doi:10.48550/arXiv.1409.1556
22. Lin TY, Dollar P, Girshick R, He K, Hariharan B, Belongie S. Feature pyramid networks for object detection. arXiv Preprint (2016) arXiv:1612.03144. doi: 10.1109/CVPR.2017.106
23. Bahdanau D, Cho K, Bengio Y. Neural machine translation by jointly learning to align and translate. arXiv Preprint (2014). arXiv:1409.0473. doi: 10.48550/arXiv.1409.0473
24. Dice LR. Measurements of amount of ecologic association between species. Ecol Ecol Soc America; (1945) 26:297–302. doi: 10.2307/1932409
25. Japanese Gastric Cancer Association. Japanese Gastric cancer treatment guidelines 2018 (5th edition). Gastric Cancer (2021) 24(1):1–21. doi: 10.1007/s10120-020-01042-y
26. Hirasawa T, Gotoda T, Miyata S, Kato Y, Shimoda T, Taniguchi H, et al. Incidence of lymph node metastasis and the feasibility of endoscopic resection for undifferentiated-type early gastric cancer. Gastric Cancer (2009) 12(3):148–52. doi: 10.1007/s10120-009-0515-x
27. Jeon JW, Kim SJ, Jang JY, Kim SM, Lim CH, Park JM, et al. Clinical outcomes of endoscopic resection for low-grade dysplasia and high-grade dysplasia on gastric pretreatment biopsy: Korea ESD study group. Gut Liver (2021) 15(2):225–31. doi: 10.5009/gnl19275
28. Sung JK. Diagnosis and management of gastric dysplasia. Korean J Intern Med (2016) 31(2):201–9. doi: 10.3904/kjim.2016.021
29. Farinati F, Rugge M, Valiante F, Baffa R, Di Mario F, Naccarato R. Gastric epithelial dysplasia. Gut (1991) 32(4):457. doi: 10.1136/gut.32.4.457-a
30. Park SY, Jeon SW, Jung MK, Cho CM, Tak WY, Kweon YO, et al. Long-term follow-up study of gastric intraepithelial neoplasias: Progression from low-grade dysplasia to invasive carcinoma. Eur J Gastroenterol Hepatol (2008) 20(10):966–70. doi: 10.1097/MEG.0b013e3283013d58
31. Zhao G, Xue M, Hu Y, Lai S, Chen S, Wang L. How commonly is the diagnosis of gastric low grade dysplasia upgraded following endoscopic resection? A meta-analysis. PloS One (2015) 10(7):e0132699. doi: 10.1371/journal.pone.0132699
32. Lim H, Jung HY, Park YS, Na HK, Ahn JY, Choi JY, et al. Discrepancy between endoscopic forceps biopsy and endoscopic resection in gastric epithelial neoplasia. Surg Endosc (2014) 28(4):1256–62. doi: 10.1007/s00464-013-3316-6
33. Cho YS, Chung IK, Jung Y, Han SJ, Yang JK, Lee TH, et al. Risk stratification of patients with gastric lesions indefinite for dysplasia. Korean J Intern Med (2021) 36(5):1074–82. doi: 10.3904/kjim.2018.285
34. Yang L, Jin P, Wang X, Zhang T, He YQ, Zhao XJ, et al. Risk factors associated with histological upgrade of gastric low-grade dysplasia on pretreatment biopsy. J Dig Dis (2018) 19(10):596–604. doi: 10.1111/1751-2980.12669
35. Muto M, Yao K, Kaise M, Kato M, Uedo N, Yagi K, et al. Magnifying endoscopy simple diagnostic algorithm for early gastric cancer (MESDA-G). Dig Endosc (2016) 28(4):379–93. doi: 10.1111/den.12638
36. Hwang JW, Bae YS, Kang MS, Kim JH, Jee SR, Lee SH, et al. Predicting pre- and post-resectional histologic discrepancies in gastric low-grade dysplasia: A comparison of white-light and magnifying endoscopy. J Gastroenterol Hepatol (2016) 31(2):394–402. doi: 10.1111/jgh.13195
37. Makazu M, Hirasawa K, Sato C, Ikeda R, Fukuchi T, Ishii Y, et al. Histological verification of the usefulness of magnifying endoscopy with narrow-band imaging for horizontal margin diagnosis of differentiated-type early gastric cancers. Gastric Cancer (2018) 21(2):258–66. doi: 10.1007/s10120-017-0734-5
Keywords: deep learning, suspected superficial lesions, magnifying endoscopy with narrow band imaging (ME-NBI), gastric neoplastic lesions, convolutional neural network (CNN)
Citation: Liu L, Dong Z, Cheng J, Bu X, Qiu K, Yang C, Wang J, Niu W, Wu X, Xu J, Mao T, Lu L, Wan X and Zhou H (2023) Diagnosis and segmentation effect of the ME-NBI-based deep learning model on gastric neoplasms in patients with suspected superficial lesions - a multicenter study. Front. Oncol. 12:1075578. doi: 10.3389/fonc.2022.1075578
Received: 20 October 2022; Accepted: 29 December 2022;
Published: 16 January 2023.
Edited by:
Yuan Zhang, McMaster University, CanadaReviewed by:
Yabin Xia, First Affiliated Hospital of Wannan Medical College, ChinaChao Shi, Ningxia Hui Autonomous Region People’s Hospital, China
Zhen Zhang, Maastro Clinic, Netherlands
Yubei Huang, Tianjin Medical University Cancer Institute and Hospital, China
Copyright © 2023 Liu, Dong, Cheng, Bu, Qiu, Yang, Wang, Niu, Wu, Xu, Mao, Lu, Wan and Zhou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Hui Zhou, mdzhouhui@163.com; Xinjian Wan, mdwanxinjian@126.com
†These authors have contributed equally to this work and share first authorship