Skip to main content

ORIGINAL RESEARCH article

Front. Med., 13 October 2022
Sec. Gastroenterology
This article is part of the Research Topic Capsule Endoscopy Procedure for the Diagnosis of Gastrointestinal Diseases View all 7 articles

Artificial intelligence to improve polyp detection and screening time in colon capsule endoscopy

  • 1Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Barcelona, Spain
  • 2Digestive System Research Unit, University Hospital Vall d'Hebron, Barcelona, Spain
  • 3Department of Medicine, Universitat Autònoma de Barcelona, Barcelona, Spain
  • 4Department of Colorectal Surgery, Raigmore Hospital, NHS Highland, Inverness, United Kingdom
  • 5CorporateHealth International ApS, Odense, Denmark

Colon Capsule Endoscopy (CCE) is a minimally invasive procedure which is increasingly being used as an alternative to conventional colonoscopy. Videos recorded by the capsule cameras are long and require one or more experts' time to review and identify polyps or other potential intestinal problems that can lead to major health issues. We developed and tested a multi-platform web application, AI-Tool, which embeds a Convolution Neural Network (CNN) to help CCE reviewers. With the help of artificial intelligence, AI-Tool is able to detect images with high probability of containing a polyp and prioritize them during the reviewing process. With the collaboration of 3 experts that reviewed 18 videos, we compared the classical linear review method using RAPID Reader Software v9.0 and the new software we present. Applying the new strategy, reviewing time was reduced by a factor of 6 and polyp detection sensitivity was increased from 81.08 to 87.80%.

Introduction

Colorectal cancer is the third most common type of cancer worldwide and ranks second on the list of most aggressive and deadly cancers (1). According to the Global Cancer Observatory, out of an estimated total of 1.9 million cases in 2020, this disease has caused the death of more than 935,000 people worldwide (1). This type of cancer has also been linked to unhealthy lifestyle habits such as smoking, alcohol consumption, unhealthy diets or sedentary lifestyle (2, 3). A clear relationship has been drawn between colon cancer and obesity (4, 5) along with genetic predisposition (6).

One of the initial signs of the development of colon cancer is the appearance of polyps in the colon that grow in an uncontrolled manner (7). The detection of polyps when they are still small is crucial to prevent their transformation into cancer. Screening programs are aimed to detect early-stage cancer, improving the patient's chances of survival (8, 9). With the disruption of colonoscopy screening programs due to the COVID-19 pandemic, an increase in incidence of 0.2–0.9% and deaths of 0.6–1.6% is predicted over the next 30 years (2020–2050) (10). To be able to mitigate these effects, well-resourced screening programs are urgently needed. Unfortunately, not all healthcare systems can afford a large increase in demand for colonoscopy which is usually a primary or secondary screening test. Therefore, other equally effective methods should be considered (10, 11). Among those alternatives, CCE has proven to be one of the most safe and effective tools, with an accuracy very similar to that of traditional colonoscopies (12). It is less invasive for the patient (13), it does not cause discomfort during the procedure [only minimal discomfort before the procedure due to required bowel preparation (13)] and no anesthesia is needed. However, the reading of CCE videos is time-consuming and requires qualified medical personnel (1416). Further studies have compared CCE sensitivity and reading time using different RAPID Reader viewing modes such as QuickView (1719), as well reading performance of expert and non-expert physicians (20). For CCE to be considered as an alternative procedure in colorectal cancer detection, the use of AI could streamline the process without compromising accuracy (2123).

AI has been extensively applied to medical imaging problems (24). In the colon capsule endoscopy field, multiple methods have been presented to automatically detect ulcers (25), polyps (26, 27), Crohn's disease (28), bowel cleanliness (29) or blood and mucosal lesions (30). These methods have shown promising results, but they are generally validated with limited data or biased datasets which does not guarantee a good generalization in clinical practice (22, 31). To our knowledge, few attempts have been presented with similar evaluation to the one we propose in this paper. In Aoki et al. (32) a CNN-based method is presented to detect erosions and ulcerations. In Beg et al. (33), a method to detect colonic lesions is introduced comparing two reviewing procedures. There is no doubt that AI has potential benefits to both doctors and patients, but its application to the clinical practice is challenging. AI is not yet at a point where it can completely replace the intervention by human experts (22). Although U.S. Food and Drug Administration (FDA) has approved some assistance algorithms (34), no guidelines establishing the role of AI currently exist. In order to achieve this, these systems would need to gain confidence of medical experts and ethical and regulatory issues would need to be solved (15, 35). This is why it is important to develop systems that cooperate with the experts, facilitating their decision making.

In this paper, we present a novel CNN-based system, AI-Tool, to assist physicians with the detection of colonic polyps. Given a video, it outputs a probability score per image frame to contain a polyp and a heatmap explaining the reasoning of the prediction. This heatmap, allows the expert to focus on the area of the image where the CNN suggests the polyp to be located. An experiment performed with 18 videos revised by 3 expert readers, has shown that the proposed system reduces screening time significantly while it also increases the sensitivity of polyp detection.

Methods

Study population

Eighteen videos of patients with at least one colon polyp obtained using the PillCam COLON 2 capsule (Medtronic) were randomly selected for this experiment following a Simple Randomization strategy. The data used in this study are retrospective CCE videos from patients that were conducted on behalf of the NHS Highland Raigmore Hospital in Inverness. All patients from this study came from referrals for symptoms or were on surveillance lists within the Highlands and Islands area of Scotland and had a positive Fecal Immunochemical Test (FIT). Referrals and final diagnoses were made based on local considerations, outside the influence of the teams conducting CCE procedures. Bowel preparation in accordance with a standardized, PEG-based, split-dose cleansing protocol was performed in all patients. All videos were obtained using PillCam COLON 2 which has two heads (front and rear). They were anonymized to protect patient information. Patients' mean age was 58.1 ± 18.7 years (range, 18–92 years) and mean colon transit time was 4 h 10 min (range, 0.17–14.2 h).

Before the experiments began, the two videos obtained from each of the heads of the capsule were meticulously reviewed by four independent CCE readers, experts from now on, in order to create the ground truth for the experiments (gold standard). Each detected polyp was assigned a unique identifier, the timestamp of the first and the last image where the polyp was visible, and from which head it was reported. The independent analysis of the experts was then shared with all experts, reaching a consensus in case of discrepancies. As experts, we have arbitrarily considered CCE readers with at least 3 months of experience in CCE. All of them have formal training in reviewing CCE videos and they analyze about 5–20 videos a week. On a daily basis, they follow the standard review protocol to ensure that each video is reviewed in a consistent, repeatable and well-documented manner. The results of all of them were validated by a medical doctor with 2 years of experience who created a final report about the results. In no case were any concerns reported back by that clinician either from the review nor from any possibly follow-up procedure about the quality of the report. Both, the final report and the gold standard used as ground-truth for the experiment in this study are the responsibility of the medical doctor that approved the results.

During this process, a total of 52 unique polyps were found. The video with the most polyps had 7, while there were 5 videos with only one polyp. The polyps' size was estimated with RAPID. A total of 23 polyps were identified as large (≥6 mm) and 29 polyps as small (<6 mm). There were 5 polyps larger than 10 mm and only one polyp smaller than 3 mm. The characterization of morphologies was done in accordance with the requirements from the referring clinicians, matching standard Paris classifications where possible. In no case polyps were selected or discarded based on size or morphology, all polyps reported by the physicians were included in the study.

Experimental design

Three experienced CCE readers reviewed the videos selected for this experiment. Each reader reviewed half of the videos using the standard RAPID Reader Software v9.0 (Medtronic) and the other half using the AI-Tool. Results obtained by the experts using each of the tools are reported in terms of number of polyps detected (sensitivity) as well as time needed to complete the reviews (screening time).

The experiment conducted in this study was restricted to the images of the colon. Identification of the entrance and exit of the colon was previously provided to the readers. When they used RAPID software, they were asked to perform the standard screening procedure without any screening time limitation. No information other than a single video identifier was provided to the reviewers during the analysis of the videos. For the AI-Tool, the review time was limited to 30 min regardless of the video's length.

For both tools, readers were required to review the videos without pauses or external stimuli that could lead to distractions. During the review, the readers labeled the images that they identified as a polyp using the tools provided within each of the applications. In the case of RAPID, experts were asked to tag all the unique polyps they found. When using the AI-Tool, experts were asked to make a decision, polyp, clear or other, for each sequence that was presented to them.

CCE readers

The expert readers are endoscopy nurses with at least 2 years of experience with CCE. They have a formal CCE training and conduct between 5 and 20 video analyses per week. On a daily basis, they follow a standard operating procedure to ensure that each video is analyzed in a consistent, repeatable way and documented according to common standards. None of these three readers were part of the gold standard creation process.

Screening tool: RAPID reader v9.0

RAPID Reader v9.0 is a Medtronic proprietary software, used to review and interpret wireless capsule endoscopy videos. This software is widely used by colon capsule endoscopy professionals designed to review the video in a comfortable way. The video is shown in temporal order and the app allows to label images, to add comments to images and to measure any part of the image with an integrated ruler function. The user can change the replay speed of the video at any time and the application shows the approximate area of the abdomen where the capsule is located. It also warns the user if the capsule is passing through an area at high speed, so the user can know if special attention is required. RAPID has several modes to review videos: single view with the frontal camera, twin head mode showing both cameras simultaneously (Figure 1) and collage mode which shows a batch of frames selected by traditional computer vision techniques.

FIGURE 1
www.frontiersin.org

Figure 1. RAPID Reader Software v9.0a: screen with images from both camera heads (green/yellow) and marked thumbnails. ahttps://www.medtronic.com/covidien/en-gb/products/capsule-endoscopy/pillcam-software.html.

Screening tool: AI-Tool

The AI-Tool is a software designed to assist clinicians in the detection of polyps, by complementing any proprietary video reviewing software, such as RAPID. It embeds a Convolutional Neural Network (CNN) into a web tool that presents images with potential polyps to the user in a sequence of declining certainty. Therefore, images that are very likely to contain a polyp will appear first. At the time we started the study, the CNN from Laiz et al. (26) was the state of the art for polyp detection so it was chosen among other candidates (3638) as the core of the AI-Tool. The hyperparameters of the model were fixed after a 5-Fold cross validation process using 120 CCE videos (2,080 polyp images and 246 k negative images). Different hyperparameters were tested (the same for the five models) and those that gave the best results in the validation sets were selected. A single model was then trained using the 120 CCE and embedded into the AI-Tool. The experimental validation of the network has shown a sensitivity over 90% at a specificity of 95% when evaluated in a fully automatic setting (when no expert is involved) using full videos. All 120 videos used in the training of the CNN were excluded for this experiment.

The AI-Tool computes two outputs using a CNN: a probability score per frame to contain a polyp and a heatmap to visualize the reasoning behind the score using CAM (39), an algorithm that uses the values of the latest CNN layers to display the image areas most relevant for classification. In this particular case, this method presents the most relevant image zones that allow the CNN to classify an image as polyp.

Each potential polyp image is displayed along with eight context frames, the four preceding and the four following it (Figure 2). For each frame, a colored square is shown to indicate the probability of it being a polyp using a colorblind friendly palette that can be customized when the application starts. Each image can be enlarged by clicking on it, then further information is presented such as the probability or the timestamp of the sequence. The image sequence can be also displayed as a video using the left and right keyboard arrows. Heatmaps can always be activated showing the most likely area to contain a polyp. A further benefit of the heatmap is that it helps readers to understand the reasoning behind the polyp probability and as a consequence, increase their trust in the system.

FIGURE 2
www.frontiersin.org

Figure 2. Candidate polyp sequence displayed in the AI-Tool. Each colored bar shows the probabilities for a polyp in one head of the capsule. The proposed image is presented in the center frame and 4 context images are placed by each side.

Results

Polyp detection

The overall sensitivity of polyp detection using RAPID as the screening procedure was 81.08% while using the AI-Tool the sensitivity increased to 87.80%. Table 1 shows the percentage of polyps found using both tools distinguishing between three categories: polyp size (in millimeters), visibility (in number of frames) and morphology. The sensitivity using RAPID turned out to be 76.92% for polyps smaller than 6 mm and 85.71% for larger polyps. Both numbers increased when the AI-Tool was used (85.42 and 91.18% for small and large polyps, respectively).

TABLE 1
www.frontiersin.org

Table 1. Detection of polyps distinguishing by size, visibility and morphology.

The biggest difference between both tools was observed in the visibility of the polyp. For polyps appearing in a few frames (low visibility), the table shows a significant improvement when using the AI-Tool. While RAPID achieved an accuracy of 58.33%, the AI-Tool reached 80.00%. This represents an increase of 21.67% in this category. Smaller improvements using the AI-Tool were also reported for polyps appearing in a larger number of frames.

These results show that small polyps and polyps that appear for only a few frames are more likely to be detected using the AI-Tool than using the RAPID application.

CCE screening time

One of the aims of this study was to compare the time needed for the detection of polyps using RAPID and our AI-Tool. The average time required for the experiments performed with RAPID was 47.11 min (11.6 min for each hour of CCE video reviewed) with a maximum of 126 min. Let us recall that the time for analysis using the AI-Tool was fixed at 30 min for all the experiments.

Figure 3 shows the average sensitivity curve as a function of time for both applications. AI-Tool took only 8.00 min to reach the same accuracy as RAPID (point A). Since the mean of RAPID experiments was 47.11 min we can state that the AI-Tool reduces the time needed to reach the same accuracy as RAPID by a factor of 47.11/8.00 ≈ 6. We can also see that RAPID experiments reached a 55.41% of sensitivity (point B) by minute 30 when the AI-Tool experiments finished.

FIGURE 3
www.frontiersin.org

Figure 3. Mean sensitivity curve of the experiments using both applications. In green the Super-Expert curve (gold standard) that represents the maximum value that the blue line could reach. This curve has been calculated simulating an expert who never makes mistakes when identifying a polyp while using the AI-Tool.

The shape of the curves is also an aspect worth considering. While the RAPID curve has an almost linear behavior, the AI-Tool curve shows an initial steep slope and, after minute 16, it is almost flat. In the first 10 min of the analysis 84.14% of the polyps are detected. In the next 10 min, this number rises to 86.59%, which represents an increase of 2.45% in this period. Finally, only 1.21% of the polyps are detected in the last 10 min of the experiment. This indicates that our application is proposing the relevant images in the first minutes of visualization. It is also worth mentioning that without limiting the time to 30 min as we did, the detection of polyps may slightly increase because those with a very low score would have been presented to the reviewers. However, this was not the objective of our study, since we aimed to see if the video review could be done better and in a shorter period of time.

Qualitative results

Heatmaps are a key element of our tool. They allow the medical staff to trust the system and make an informed decision on each image sequence. Figure 4 shows the heatmaps activation for three different categories. First, in the image on the left, we see high-scoring polyps. We observe how the heatmaps generated by the CAM algorithm are well defined and show the area where the polyp is located. In the center, we see polyps with a very low score, which the network erroneously classifies as negative. In this case, the heatmaps are not activated. Finally, in the image on the right we can see images that do not contain any polyp but to which the system assigns a high score.

FIGURE 4
www.frontiersin.org

Figure 4. Left: Images of correctly identified polyps with their respective heatmaps (True Positives). Center: Images of polyps with a very low score (False Negatives). Right: Images that do not contain a polyp but still have a very high score (False Positives).

It can also be seen that the false positive patterns correspond to textures and morphologies compatible with polyps. In almost all of them a rounded area is displayed which, without the context of the other images, can be difficult to classify as a polyp image or not. In addition, this is a valuable information for the reviewers of the video, as the heatmap shows the area on which it is important to focus their attention.

We now focus on showing the images of polyps that have not been detected with either of the two applications (specificity). Figure 5 shows those polyps not detected with the AI-Tool because of the imposed time restriction (30 min). The score given to these image does not exceed 15%, therefore, the experts never reviewed them. In fact, these four images are the polyps with the lowest score of this study. The polyps of these images are difficult to find since they are partially occluded or do not present a regular morphology.

FIGURE 5
www.frontiersin.org

Figure 5. Polyp frames to which the app has attributed a small score and, therefore, none of the experts have been able to review in the first 30 min. Polyps are circled in white.

Figure 6 shows images of polyps reviewed and discarded by all the experts even though the AI-Tool assigned them a remarkably high probability. These missed polyps are the result of human error or discrepancies between the video reviewer and the experts who generated the ground truth.

FIGURE 6
www.frontiersin.org

Figure 6. Images that all experts have reviewed and found not to be polyps. Polyps are circled in white.

Finally, Figure 7 shows some examples of polyp images missed in RAPID experiments but correctly detected using the AI-Tool. Due to the size of the polyp and the fast movement of the capsule that took few images, these polyps are especially difficult to find using RAPID. In contrast, they are easy to find for AI-Tool users as they are presented with the clearest image.

FIGURE 7
www.frontiersin.org

Figure 7. Example of polyps missed in RAPID experiments. Polyp frames are tagged with a red square. Polyps are circled in white.

Conclusions

In this paper, thanks to the AI enhancement, we mitigate one of the main drawbacks of CCE, the required time for the analysis while increasing the detection rate of the experts. redWe consider that these improvements further improve the value proposition of CCE as a clinically viable alternative to traditional imaging methods of the gastrointestinal tract.

The proposed method, AI-Tool, uses the output of a CNN architecture. It scores frames based on the probability of containing a polyp and then it reorders the video images to present the most relevant ones first. The validation was performed by three clinical experts that analyzed 18 videos, comparing the standard method with and without the proposed AI-enhanced application. With the assistance of the AI-Tool, the time required to review the videos was reduced by a factor of 6 and the sensitivity increased from 81.08 to 87.80%. In the case of small polyps (<6 mm), the improvement in sensitivity obtained by the AI-Tool was 8.50%, and, for polyps with low visibility (seen in <3 frames), the improvement in detection was 21.67%.

Data availability statement

The data used during the present study are not publicly available as they are property of National Services Scotland (NHS Highland) but are available through co-author AW (YW5ndXMud2F0c29uJiN4MDAwNDA7bmhzLnNjb3Q=) upon reasonable request.

Ethics statement

The studies involving human participants were reviewed and approved by University of Barcelona's Bioethics Commission, Institutional Review Board IRB00003099. The patients/participants provided their written informed consent to participate in this study.

Author contributions

Data collection was performed by HW. The first draft of the manuscript was written by PG. All authors commented on previous versions of the manuscript, study conception, design, read, and approved the final manuscript.

Funding

This work has been also supported by MINECO Grant RTI2018-095232-B-C21, SGR 1742, Instituto de Salud Carlos III as well as the Innovate UK project 104633.

Acknowledgments

The authors would like to thank the team from CorporateHealth International ApS for their feedback.

Conflict of interest

Author HW is co-founder of CorporateHealth International, a company that may be affected by the research reported in the enclosed manuscript.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin. (2021) 71:209–49. doi: 10.3322/caac.21660

PubMed Abstract | CrossRef Full Text | Google Scholar

2. O'Sullivan DE, Metcalfe A, Hillier TW, King WD, Lee S, Pader J, et al. Combinations of modifiable lifestyle behaviours in relation to colorectal cancer risk in Alberta's Tomorrow Project. Sci Rep. (2020) 10:1–11. doi: 10.1038/s41598-020-76294-w

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Ahmed M. Colon cancer: a clinician's perspective in 2019. Gastroenterol Res. (2020) 13:1. doi: 10.14740/gr1239

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Ye P, Xi Y, Huang Z, Xu P. Linking obesity with colorectal cancer: epidemiology and mechanistic insights. Cancers. (2020) 12:1408. doi: 10.3390/cancers12061408

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Lauby-Secretan B, Scoccianti C, Loomis D, Grosse Y, Bianchini F, Straif K. Body fatness and cancer–viewpoint of the IARC working group. N Engl J Med. (2016) 375:794–8. doi: 10.1056/NEJMsr1606602

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Valle L. Genetic predisposition to colorectal cancer: where we stand and future perspectives. World J Gastroenterol. (2014) 20:9828–49. doi: 10.3748/wjg.v20.i29.9828

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Bond J. Colon polyps and cancer. Endoscopy. (2003) 35:27–35. doi: 10.1055/s-2003-36410

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Levin TR, Corley DA, Jensen CD, Schottinger JE, Quinn VP, Zauber AG, et al. Effects of organized colorectal cancer screening on cancer incidence and mortality in a large community-based population. Gastroenterology. (2018) 155:1383–91. doi: 10.1053/j.gastro.2018.07.017

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Loveday C, Sud A, Jones ME, Broggio J, Scott S, Gronthound F, et al. Prioritisation by FIT to mitigate the impact of delays in the 2-week wait colorectal cancer referral pathway during the COVID-19 pandemic: a UK modelling study. Gut. (2021) 70:1053–60. doi: 10.1136/gutjnl-2020-321650

PubMed Abstract | CrossRef Full Text | Google Scholar

10. de Jonge L, Worthington J, van Wifferen F, Iragorri N, Peterse EF, Lew JB, et al. Impact of the COVID-19 pandemic on faecal immunochemical test-based colorectal cancer screening programmes in Australia, Canada, and the Netherlands: a comparative modelling study. Lancet Gastroenterol Hepatol. (2021) 6:304–14. doi: 10.1016/S2468-1253(21)00003-0

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Laghi L, Cameletti M, Ferrari C, Ricciardiello L. Impairment of colorectal cancer screening during the COVID-19 pandemic. Lancet Gastroenterol Hepatol. (2021) 6:425–6. doi: 10.1016/S2468-1253(21)00098-4

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Vuik FE, Nieuwenburg SA, Moen S, Spada C, Senore C, Hassan C, et al. Colon capsule endoscopy in colorectal cancer screening: a systematic review. Endoscopy. (2021) 53:815–24. doi: 10.1055/a-1308-1297

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Ismail MS, Murphy G, Semenov S, McNamara D. Comparing colon capsule endoscopy to colonoscopy; a symptomatic patient's perspective. BMC Gastroenterol. (2022) 22:1–8. doi: 10.1186/s12876-021-02081-0

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Maieron A, Hubner D, Blaha B, Deutsch C, Schickmair T, Ziachehabi A, et al. Multicenter retrospective evaluation of capsule endoscopy in clinical routine. Endoscopy. (2004) 36:864–8. doi: 10.1055/s-2004-825852

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Koulaouzidis A, Dabos K, Philipper M, Toth E, Keuchel M. How should we do colon capsule endoscopy reading: a practical guide. Therapeutic Adv Gastrointestinal Endoscopy. (2021) 14:26317745211001983. doi: 10.1177/26317745211001983

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Rondonotti E, Pennazio M, Toth E, Koulaouzidis A. How to read small bowel capsule endoscopy: a practical guide for everyday use. Endoscopy Int Open. (2020) 8:E1220–4. doi: 10.1055/a-1210-4830

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Farnbacher MJ, Krause HH, Hagel AF, Raithel M, Neurath MF, Schneider T. QuickView video preview software of colon capsule endoscopy: reliability in presenting colorectal polyps as compared to normal mode reading. Scand J Gastroenterol. (2014) 49:339–46. doi: 10.3109/00365521.2013.865784

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Günther U, Daum S, Zeitz M, Bojarski C. Capsule endoscopy: comparison of two different reading modes. Int J Colorectal Dis. (2012) 27:521–5. doi: 10.1007/s00384-011-1347-9

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Shiotani A, Honda K, Kawakami M, Kimura Y, Yamanaka Y, Fujita M, et al. Analysis of small-bowel capsule endoscopy reading by using Quickview mode: training assistants for reading may produce a high diagnostic yield and save time for physicians. J Clin Gastroenterol. (2012) 46:e92–5. doi: 10.1097/MCG.0b013e31824fff94

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Hausmann J, Linke JP, Albert JG, Masseli J, Tal A, Kubesch A, et al. Time-saving polyp detection in colon capsule endoscopy: evaluation of a novel software algorithm. Int J Colorectal Dis. (2019) 34:1857–63. doi: 10.1007/s00384-019-03393-0

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Byrne MF, Donnellan F. Artificial intelligence and capsule endoscopy: is the truly “smart” capsule nearly here? Gastrointest Endosc. (2019) 89:195–7. doi: 10.1016/j.gie.2018.08.017

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Yang YJ. The future of capsule endoscopy: the role of artificial intelligence and other technical advancements. Clin Endosc. (2020) 53:387. doi: 10.5946/ce.2020.133

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Chetcuti Zammit S, Sidhu R. Capsule endoscopy-Recent developments and future directions. Expert Rev Gastroenterol Hepatol. (2021) 15:127–37. doi: 10.1080/17474124.2021.1840351

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. (2019) 25:44–56. doi: 10.1038/s41591-018-0300-7

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Rahim T, Usman MA, Shin SY. A survey on contemporary computer-aided tumor, polyp, and ulcer detection methods in wireless capsule endoscopy imaging. Comput Med Imaging Graphics. (2020) 85:101767. doi: 10.1016/j.compmedimag.2020.101767

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Laiz P, Vitrià J, Wenzek H, Malagelada C, Azpiroz F, Seguí S. WCE polyp detection with triplet based embeddings. Comput Med Imaging Graphics. (2020) 86:101794. doi: 10.1016/j.compmedimag.2020.101794

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Jin EH, Lee D, Bae JH, Kang HY, Kwak MS, Seo JY, et al. Improved accuracy in optical diagnosis of colorectal polyps using convolutional neural networks with visual explanations. Gastroenterology. (2020) 158:2169–79.e8. doi: 10.1053/j.gastro.2020.02.036

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Goran L, Negreanu AM, Stemate A, Negreanu L. Capsule endoscopy: current status and role in Crohn's disease. World J Gastrointest Endosc. (2018) 10:184. doi: 10.4253/wjge.v10.i9.184

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Noorda R, Nevárez A, Colomer A, Beltrán VP, Naranjo V. Automatic evaluation of degree of cleanliness in capsule endoscopy based on a novel CNN architecture. Sci Rep. (2020) 10:1–13. doi: 10.1038/s41598-020-74668-8

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Mascarenhas M, Ribeiro T, Afonso J, Ferreira JP, Cardoso H, Andrade P, et al. Deep learning and colon capsule endoscopy: automatic detection of blood and colonic mucosal lesions using a convolutional neural network. Endoscopy Int Open. (2022) 10:E171–7. doi: 10.1055/a-1675-1941

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Saurin JC, Beneche N, Chambon C, Pioche M. Challenges and future of wireless capsule endoscopy. Clin Endosc. (2016) 49:26. doi: 10.5946/ce.2016.49.1.26

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Aoki T, Yamada A, Aoyama K, Saito H, Fujisawa G, Odawara N, et al. Clinical usefulness of a deep learning-based system as the first screening on small-bowel capsule endoscopy reading. Digest Endoscopy. (2020) 32:585–91. doi: 10.1111/den.13517

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Beg S, Wronska E, Araujo I, Suárez BG, Ivanova E, Fedorov E, et al. Use of rapid reading software to reduce capsule endoscopy reading times while maintaining accuracy. Gastrointest Endosc. (2020) 91:1322–7. doi: 10.1016/j.gie.2020.01.026

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Muehlematter UJ, Daniore P, Vokinger KN. Approval of artificial intelligence and machine learning-based medical devices in the USA and Europe (2015-20): a comparative analysis. Lancet Digital Health. (2021) 3:e195-e203. doi: 10.1016/S2589-7500(20)30292-2

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Ahmad OF, Stoyanov D, Lovat LB. Barriers and pitfalls for artificial intelligence in gastroenterology: ethical and regulatory issues. Tech Innovat Gastrointest Endoscopy. (2020) 22:80–4. doi: 10.1016/j.tgie.2019.150636

CrossRef Full Text | Google Scholar

36. Yuan Y, Meng MQH. Deep learning for polyp recognition in wireless capsule endoscopy images. Med Phys. (2017) 44:1379–89. doi: 10.1002/mp.12147

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Yuan Y, Qin W, Ibragimov B, Zhang G, Han B, Meng MQH, et al. Densely connected neural network with unbalanced discriminant and category sensitive constraints for polyp recognition. IEEE Trans Automat Sci Eng. (2019) 17:574–83. doi: 10.1109/TASE.2019.2936645

CrossRef Full Text | Google Scholar

38. Guo X, Yuan Y. Triple ANet: Adaptive abnormal-aware attention network for WCE image classification. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Shenzhen: Springer (2019). pp. 293–301.

Google Scholar

39. Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning deep features for discriminative localization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV: IEEE (2016). p. 2921–9.

Google Scholar

Keywords: colon capsule endoscopy, artificial intelligence, screening time, polyp detection, colorectal cancer prevention

Citation: Gilabert P, Vitrià J, Laiz P, Malagelada C, Watson A, Wenzek H and Segui S (2022) Artificial intelligence to improve polyp detection and screening time in colon capsule endoscopy. Front. Med. 9:1000726. doi: 10.3389/fmed.2022.1000726

Received: 22 July 2022; Accepted: 23 September 2022;
Published: 13 October 2022.

Edited by:

Zongxin Ling, Zhejiang University, China

Reviewed by:

Konstantinos Triantafyllou, National and Kapodistrian University of Athens, Greece
Colm Antoine O. Morain, Trinity College Dublin, Ireland

Copyright © 2022 Gilabert, Vitrià, Laiz, Malagelada, Watson, Wenzek and Segui. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Pere Gilabert, cGVyZS5naWxhYmVydCYjeDAwMDQwO3ViLmVkdQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.