- Department of Computer Science, King Abdulaziz University, Rabigh, Saudi Arabia
Although the detection procedure has been shown to be highly effective, there are several obstacles to overcome in the usage of AI-assisted cancer cell detection in clinical settings. These issues stem mostly from the failure to identify the underlying processes. Because AI-assisted diagnosis does not offer a clear decision-making process, doctors are dubious about it. In this instance, the advent of Explainable Artificial Intelligence (XAI), which offers explanations for prediction models, solves the AI black box issue. The SHapley Additive exPlanations (SHAP) approach, which results in the interpretation of model predictions, is the main emphasis of this work. The intermediate layer in this study was a hybrid model made up of three Convolutional Neural Networks (CNNs) (InceptionV3, InceptionResNetV2, and VGG16) that combined their predictions. The KvasirV2 dataset, which comprises pathological symptoms associated to cancer, was used to train the model. Our combined model yielded an accuracy of 93.17% and an F1 score of 97%. After training the combined model, we use SHAP to analyze images from these three groups to provide an explanation of the decision that affects the model prediction.
1 Introduction
The digestive system consists of the organs that make up the digestive system. Cellular mutations in at least one of these genes can lead to cancer and ultimately lead to the development of colon cancer. More importantly, colon cancer has a huge impact on the world, accounting for approximately 26.3% of all cancers (4.8 million cases) and 35.4% of blood cancer cases (3.4 million deaths) (1). As shown in Figure 1, the digestive system has a line that is approximately 25 feet long, starting from the mouth and ending at the anus. Many studies, including [Hospital] and (2), identified the most common types of cancer, including stomach, colon, colon, liver, and Cancer.
Latest studies have reported that a substantial proportion (over 50%) of gastrointestinal cancers can be attributed to risk factors that can be altered by adopting a healthier lifestyle alcohol intake, cigarette smoking, infection, unhealthy diet, and obesity (3, 4). Moreover, it has been observed that males have a higher susceptibility to gastrointestinal cancers compared to females, with the risk increasing with age, as indicated by (5). Unfortunately, due to late-stage diagnoses being predominant, the prognosis for such cancers is typically unfavorable (6), thus resulting in site-specific death rates that align with the incidence trends. However, if gastrointestinal cancers are detected in their early stages, the survival rate becomes higher in the five-year timeline (7). Nonetheless, a study conducted by (8) put forward that cognitive and technological issues contribute to significant diagnostic errors, despite the effectiveness of traditional screening procedures.
The Global-Cancer-Observatory (9) predicts a substantial increase in the global mortality and incidence rates of gastrointestinal (GI) cancers (10) by the year 2040. The mortality rate is projected to rise by 73%, reaching approximately 5.6 million cases, while the incidence rate is expected to increase by 58%, with an estimated 7.5 million new cases. These alarming statistics highlight the urgent need for the development of dependable systems to support medical facilities in obtaining accurate GI cancer diagnoses. Addressing this priority through innovative research endeavors becomes crucial to effectively combat the rising burden of GI cancers on a global scale.
Recent research has highlighted the potential of Artificial Intelligence (AI) in reducing misdiagnosis rates associated with conventional screening techniques, thereby enhancing overall diagnostic accuracy (11–23). The main reason for this accomplishment is the application of machines as well as deep-learning techniques. However, a significant hurdle faced by AI-supported systems is their perceived nature as computational” black boxes.” The lack of transparency in the decision-making processes of these AI models has resulted in hesitancy among healthcare institutions when it comes to adopting them for diagnostic purposes, despite their effectiveness (24–35). It is therefore important for AI researchers to integrate digestible explanations throughout the development of AI-aided medical applications, thus assuring healthcare practitioners while also clearing any doubts they might have.
In this context, XAI has emerged as a promising field that aims to address the computational difficulties posed by AI systems, warranting the provision of explanations for model predictions (36). By employing XAI techniques, AI researchers can enhance the interpretability and transparency of AI-driven diagnostic systems, thereby fostering trust and facilitating their integration into clinical practice. To address the challenges in AI driven diagnostic systems, this research work focuses on the investigation of SHAP. SHAP is an explanation approach for model predictions that was introduced by (37). In our study, we have utilized an ensemble model that we developed and trained on the pathology results obtained from the publicly accessible Kvasir dataset. By employing SHAP, we aim to provide interpretable explanations for the predictions made by our ensemble model, thereby enhancing the transparency and understandability of the AI-assisted diagnostic system.
To pinpoint the critical elements influencing the decision-making process, this study presents a unique approach for the categorization of gastrointestinal lesions. InceptionV3, Inception-ResNetV2, and VGG16 (Visual Geometry Group) architectures are used in the article to apply transfer learning. CNN Models are optimized for the goal of identifying gastrointestinal lesions like esophagitis, poylps and ulcerative-colitis by the application of enhancements and fine-tuning procedures. These improvements improve the precision and robustness of the models as compared with the latest techniques like (38–42). Creation of an Ensemble Model: By combining the predictions of every CNN model, the research project suggests and creates an ensemble model. By combining the advantages of several models, the ensemble model seeks to enhance classification performance by making use of the variety and complementary traits of its component models. The ensemble model is developed to classify gastrointestinal lesions in the dataset, and its performance is assessed. The study also thoroughly examines the characteristics that impact the classification procedure, illuminating the critical elements influencing precise lesion classification. Explainability (43, 44) characteristics made it possible to visualize the variables that contributed to each prediction in a comprehensible way, highlighting significant differences in performance that would not have been apparent otherwise.
The rest of the paper is organized as follows. Section II briefly gives a literature review; the framework of our novel technique is shown in Section III. Section IV presents experimental data, comments, and comparisons with existing techniques. Finally, the article is wrapped up in Section V.
2 Related work
Numerous research investigations have been carried out to develop automated models for detecting gastrointestinal cancer. According to (45), the detection of esophageal cancer using deep-learning (i.e., CNN) and machine-learning is becoming progressively prevalent. Preliminary screening of esophageal cancer has been made possible through the development of computer-assisted application by (46). Eventually, the researchers achieved the classification of esophageal images through the implementation of random forests as an ensemble classifier for esophageal image classification. Nonetheless, deep-learning models are being investigated.
In a study conducted in 2019 (47), developed a VGG16, InceptionV3, and ResNet50 model based on the transfer learning approach to classify endoscopic images into three classes: normal, benign-ulcer, and cancer using a custom dataset of 787 images including 367 samples of cancer, 200 samples of normal cases, and 220 samples of ulcers collected from Hospital. The images were first resized to 224×224 before they were preprocessed using Adaptive-Histogram-Equalization (AHE) to eliminate variations in the image brightness and contrast, thereby improving the local contrast and enhancing edge definition within each image region. Three binary classification tasks namely: normal vs. cancer, normal vs. ulcer, and cancer vs. ulcer were performed in this study and the accuracy, standard deviation, and Area-Under-Curve (AUC) values across the different CNN models. ResNet50 demonstrated the highest performance for all three-performance metrics. The model achieved an accuracy of above 92% for the classification tasks including the Normal images. However, for the cancer vs. ulcer task, a lower accuracy of 77.1% was noted. The authors conclude that this decrease is probably attributed due to the smaller visual differences between cancer and ulcer instances.
ResNet50 also achieved the lowest standard deviation, which indicates greater stability among the other models. In terms of AUC, ResNet50 reported an AUC of 0.97, 0.95, and 0.85, respectively for the normal vs. ulcer, normal vs. cancer, and cancer vs. ulcer tasks. The authors concluded that this proposed deep learning approach can be a valuable tool to complement traditional screening practices by medical practitioners thus reducing the risk of missing positive cases due to repetitive endoscopic frames or diminishing concentration.
The authors of (48) developed a deep CNN based on the UNet++ and Resnet50 architectures to classify between cases of gastritis (AG) and non-atrophic gastritis (non-AG) using white light endoscopy images. A total of 6,122 images (4,022 AG cases and 2,100 non-AG) were collected from 456 patients and were randomly partitioned into training (89%) and test sets (11%). For the binary classification task, the model achieved an accuracy of 83.70%, sensitivity of 83.77%, 13 and specificity of 83.75% while for the region segmentation task, an IOU score of 0.648 for the AG regions and 0.777 for the incisura region. The results suggest that the developed model based on the UNet++ and Resnet50 architectures can effectively distinguish between AG and non-AG cases, and it can also be used to delineate specific regions of interest within the endoscopic images.
Based on a research carried out by (49), images of non-cancerous lesions and Early-Gastric-Cancers (EGC) were used to evaluate CNN diagnostic potential. A dataset, comprising of 386 non-cancerous lesions images and 1702 ECG images, was used for the training of the CNN model. The analysis results showed a sensitivity level of 91.18% showing the model’s adeptness to rightly identify EGC cases and a specificity level of 90.64% indicating its ability to properly identify non-cancerous lesions. Substantially, reaching an accuracy level of 90.91% of the CNN model to diagnose both types of cases. Upon comparison, no remarkable differences were found between the specificity and accuracy levels of the AI-aided system and endoscopy specialists. However, the specificity and accuracy levels of the non-experts were below those of both the endoscopists and AI-aided system. According to the study findings, the CNN model exhibited exceptional EGC and non-cancerous lesions diagnostic performance. Consequently, this research demonstrates the potential of AI-aided systems in assisting medical practitioners.
In the study presented by (50), an automated detection approach utilizing Convolutional Neural Networks (CNN) was proposed to assist in the identification of Early-Gastric-Cancers (EGC) in endoscopic images. The method employed transfers learning on two distinct classes of image datasets: cancerous and normal. These datasets provided detailed information regarding the texture characteristics of the lesions and were obtained from a relatively limited data set. The CNN based network was trained using transfer learning techniques to leverage the knowledge acquired from pre-trained models. By utilizing this approach, the network achieved a notable accuracy of 87.6%. Subsequently, an external dataset was used for the evaluation of the model’s performance and an accuracy of 82.8% was attained.
The median filtering (MF) approach is used in the MSSADL-GITDC technique that is being presented to smooth images. The class attention layer (CAL) modifies the enhanced capsule network (CapsNet) model in the MSSADL-GITDC approach, which is offered for feature extraction. Deep Belief Network with Extreme Learning Machine (DBN-ELM) was utilized for GIT categorization. The accuracy of the suggested approaches was 98.03% (51). A unique approach for the automated identification and localization of gastrointestinal (GI) abnormalities in endoscopic video frame sequences is presented in this work (52). The photos used for training have poor annotations. The localization and anomaly detection performance obtained were both greater than 80% in terms of the area under the receiver operating characteristic (AUC).
These results suggest that the proposed automated detection method based on CNN, trained on the cancerous and normal image datasets, effectively aids in the identification of EGC in endoscopic images. The achieved accuracy of 87.6% on the training dataset demonstrates the model’s ability to discern between cancerous and normal instances. Furthermore, the comparable accuracy of 82.8% on the external dataset indicates the model’s generalizability and potential for practical application in clinical settings.
The idea of interpretable real-time deep-neural-network (SHAP) was first proposed by (53). The proposed technique showcased improved real time performance compared to existing methods. Experimental results highlighted the superiority of this approach over current deep learning techniques. Moreover, the author successfully addressed the needs of colorectal surgeons by providing satisfactory operational effectiveness and interpretable feedback. By incorporating Shapley additive explanations, the technique not only offers enhanced performance but also ensures interpretability, aligning with the requirements of the medical professionals in the field of colorectal surgery.
Upon investigation of prior research on the detection of gastrointestinal cancer using AI assistance, it became evident that this field will highly benefit from further exploration. While several AI models have been utilized to discover deformities in medical images, there remains a notable gap in the development of human-comprehensible models that can provide explanations for model predictions. Although there has been a recent surge of interest among researchers, only a limited number of studies have focused on creating AI models that offer interpretability, allowing healthcare professionals and stakeholders to understand and trust the predictions made by these models. In the context of gastrointestinal disease classification, the various shapes and sizes of a single lesion is a greater problem. Moreover, the single model extracts single type of features due to which classification accuracy reduces. Therefore, there is a clear need for more research efforts to develop AI models in gastrointestinal cancer detection that not only achieve high accuracy but also provide comprehensible explanations for their predictions.
3 Methodology
The proposed scheme jointly defines artificial intelligence (XAI) and presents an XAI-based model for gastrointestinal (GI) diagnosis. Figure 1 shows the proposed structure of XAI-based gastrointestinal cancer screening. The system was trained and evaluated using pathology results from the KvasirV2 dataset. A design was developed to improve the performance and accuracy of the system. This model incorporates predictions from multiple models and has the potential to increase power and improve overall classification. Additionally, the XAI process was used to uncover the determining factors associated with each category. This process can identify and describe key characteristics that influence the decision to classify the group. By integrating XAI into an integrated model and analyzing the decision, this approach aims to gain an understanding of the decision-making process of colon cancer testing, improving their transparency and interpretation.
3.1 Dataset
Datasets play a crucial role in the advancement of various computing domains, particularly in the field of deep learning applications. The availability and quality of datasets are critical since they must include enough examples, be sufficiently labeled, and show variety in the pictures. Several investigators and institutions have expanded the datasets for medical imaging so that it becomes easier to train and evaluate suggested models. This study made use of the Kvasir dataset initially introduced by (54) in 2017 which is composed of images that have been meticulously validated and annotated by medical experts. Each class contains a thousand images, thus showcasing pathological revelations, endoscopic approaches, and anatomical landmarks within the gastrointestinal tract.
However, for the purpose of this research, our focus was solely on the pathological findings class, which encompasses three distinct categories: Esophagitis, Polyps, and Ulcerative-Colitis, a chronic condition causing inflammation of the colon and rectum. To enhance the diversity and variety within the dataset, data augmentation techniques were applied to the original dataset. Specifically, rotation and zoom techniques were utilized to create variations of the existing images. This process involved rotating the images at different angles and applying zooming operations to produce new perspectives and scales.
By applying these data augmentation techniques, another dataset having 2000 images per class was generated. This increased dataset size provided a broader range of image variations and ensured a more comprehensive representation of the pathological findings within the gastrointestinal (GI) tract. The augmented dataset with its increased variety and enlarged sample size is crucial for training and evaluating the proposed models effectively. It enables the models to learn from a more diverse set of examples and improves their ability to generalize and make accurate predictions on unseen data. Figure 2 shows the sample images from the dataset. The dataset was divided into training and testing ratios, with 70% data used for training and remaining 30% for testing.
3.2 Convolutional neural network models
There are a total of three primary deep CNN that were implemented in the development of the ensemble model. InceptionV3, created by (55) in 2015. InceptionV3 is an upgraded version of GoogleNet (Inception V1) and comprises 42 layers. VGG16, created by (56) in 2014, was the third model used to create the ensemble model. It has 16 layers and uses the Softmax classifier. Finally, the InceptionResNetV2 was effectuated. InceptionResNetV2 is deep-CNN having Inception-architecture as its foundational basis though it makes use of residual connections instead of undergoing the filter concatenation phase. It comprises of 164 layers and was formed in 2016 by (57).
3.3 Ensemble learning
Ensemble models are a valuable technique in machine learning that combines multiple individual models to enhance the overall performance of a system. The fundamental concept behind ensemble modeling is to leverage the strengths of different models to compensate for their respective weaknesses, resulting in improved accuracy, robustness, and generalization capabilities. Ensemble models come in several varieties, such as bagging, boosting, and stacking.
Several models are separately trained on various subsets of the training data in bagging. Usually, the final forecast is determined by combining all the models’ predictions together using methods like majority voting or averaging. This method aids in lowering overfitting and boosting forecast stability. Boosting, on the other hand, entails repeated training models. Each new model focuses on the examples that were misclassified by the previous models, thereby progressively improving the overall performance. Boosting algorithms assign higher weights to difficult examples, allowing subsequent models to prioritize those instances during training. Stacking takes a different approach by utilizing the predictions of multiple models as input features for a meta-model. The meta-model is trained to learn how to combine these predictions effectively and make the final prediction. This approach can capture complex relationships between the base models’ outputs and potentially improve overall performance.
Ensemble models find applications in various domains of AI, including computer vision, natural language processing, and speech recognition. For instance, in image classification tasks, an ensemble of CNNs can be employed to enhance accuracy and robustness. Each CNN within the ensemble may specialize in different aspects of feature extraction or classification, leading to improved classification performance. Ensemble models are a powerful technique in machine learning that leverages the collective wisdom of multiple models. By combining diverse models, ensemble methods can mitigate individual model limitations and yield superior performance across a range of AI applications (58–60). The mathematical equations behind the ensemble modeling is as follows:
Where a model ft (y | x) is a model, moreover, and y is the estimated probability. Z is a normalization rule, ht is a model output variable.
This study focuses on the development of an ensemble model based on bagging techniques. This is executed through the synthesis of the predictions of three pretrained CNNs InceptionV3, InceptionResNetV2, and VGG16. The ensemble model’s architecture is depicted in Figure 3. Moreover, each of the three previously mentioned CNN models was applied to our enhanced KvasirV2 dataset which was separated into two parts 75% training and 25% validation. After individual training of the models, the average approach was used to develop the ensemble model through the combination of each model’s predictions. The average technique formulates an average of the predictions obtained from the three trained models resulting in the generation of the final prediction.
3.4 Explainable AI
The field of Explainable AI (XAI) is experiencing rapid growth, focusing on enhancing transparency and interpretability in machine learning algorithms. This advancement is of particular significance in the realm of medical imaging, as the outputs of machine learning models greatly influence patient care. XAI methods play a crucial role in enabling clinicians and radiologists to comprehend the rationale behind the model’s predictions, thereby instilling confidence in the accuracy of the model’s assessments. Moreover, XAI techniques aid in the identification of potential biases within the model, facilitating the prevention of misdiagnosis and promoting equitable healthcare outcomes (24, 25, 61).
XAI in medicine and healthcare have been classified in five categories by (36). Our goal is to improve the explain ability of medical imaging, which motivated us to investigate the XAI explanation using feature relevance approach. One such technique is the SHAP. The SHAP method developed by (37) is a model agnostic technique derived from cooperative game theory, enabling the interpretation of machine learning model outputs by quantifying the contribution of each feature. It provides a comprehensive framework that considers both global and local feature importance, accounting for feature interactions and ensuring fairness in assigning importance. The SHAP values align with desired axioms of feature attribution methods, including local accuracy, consistency, and missingness. Local accuracy ensures that the sum of SHAP values corresponds to the discrepancy between the model’s prediction and the expected output for a specific input. Consistency guarantees that fixing a feature’s value will not decrease its associated SHAP value. Missingness implies that irrelevant features have SHAP values close to zero. When you take into account all the many ways that features might combine, the Shapley value of a feature value indicates how much that feature contributed to the outcome (such as a reward or payout). It’s similar to calculating the relative contribution of each feature to the final result, accounting for all the various ways in which they may have cooperated. This aids in our comprehension of the elements that were crucial in reaching the desired outcome or making the final decision. The mathematical modeling for XAI is as:
Here, is the XAI model, is the simplified input such that ≈ (𝑥′) and ∈ (62) 𝑀 𝜙𝑖 ∈ 𝑅. Moreover, the function is defined in the manner shown in following equation to determine the impact of each attribute on model prediction.
Here, represents feature-sets, and is a subset of . is the trained model on and Nth feature. is the trained model without this feature. Moreover, represent the feature-value in th set.
Various applications have benefited from SHAP values, encompassing domains such as image recognition, natural language processing, and healthcare. For instance, in a study focusing on breast cancer detection, SHAP values were utilized to identify the most relevant regions in the images (63). Similarly, in another study concerning the detection of relevant regions in retinal images for predicting disease severity (64), SHAP values were employed to interpret the features of a deep neural network model.
4 Experimental results
In the initial phase of the experiments, an ensemble model was developed for the classification of gastrointestinal (GI) lesions. This involved training the InceptionV3, InceptionResNetV2, and VGG16 models individually on the KvasirV2 datasets. Subsequently, these models were combined to create the ensemble meta model. To adapt the primordial architectures of the 3 pre trained convolutional neural networks (CNNs), a global average pooling layer was added. This layer summarizes the spatial information from the previous layers and reduces the dimensionality of the feature maps. Following the pooling layer, a dropout layer with a dropout rate of 0.3 was applied to mitigate overfitting, enhancing the model’s generalization capabilities. The Adam optimization algorithm was employed for model optimization, using sparse-categorical-cross entropy as the loss-function. This combination allowed for effective training of the ensemble model by updating the model weights based on the calculated gradients. Each of the selected deep-CNNs was trained for 5 epochs, with a batch size of 32, to finetune the model parameters and improve its performance. After the training process, the softmax activation function was utilized for classification. This function assigned probabilities to each class, enabling the ensemble model to make predictions on the GI lesion classes.
By leveraging the strengths of multiple pre-trained CNN models through ensemble learning, the developed model aimed to enhance the accuracy and robustness of GI lesion classification. Table 1 shows the classification results obtained on individual models as well as Ensemble model. Moreover, for comparison among individual models as well as ensemble model, a graph is shown in Figure 4.
Based on Figure 5, the F1-score, recall, and precision for each class of esophagitis, polyps, and ulcerative colitis are displayed in the classification report, as shown in Figure 6. The confusion matrix derived from the classification results is shown in Figure 5. The confusion matrix, which displays the proportion of properly and erroneously identified samples for each class, offers a summary of the model’s performance.
The classification report is shown in Table 2 and contains the F1-score, recall, and accuracy metrics for the ulcerative colitis, polyps, and esophagitis classes. The model’s accuracy is gaged by the F1-score, which unifies recall and precision into a single number. The capacity of the model to accurately identify positive samples is reflected in recall, and the ability to correctly categorize positive predictions is reflected in accuracy. These metrics shed light on how well the model performs for a given class. The classification report and the confusion matrix provide useful data to assess the precision and potency of the created ensemble model in categorizing gastrointestinal lesions.
When compared to the individual models, the ensemble model’s findings show a notable improvement in overall accuracy. The three classes—ulcerative colitis, polyps, and esophagitis—showcase excellent accuracy, recall, and F1-score in the classification report. These metrics show that the model can effectively minimize false positives and false negatives while correctly detecting positive events. An overall accuracy of 93.17% and an F1-score of 97% for every class show that the ensemble model performs well in classifying gastrointestinal lesions. The model can properly detect both positive and negative examples, as suggested by the high F1-score, which also indicates that the model strikes a balance between precision and recall. The encouraging outcomes of the ensemble model point to its potential for further refinement and implementation in clinical settings, which is appropriate considering the significance of precise prediction in the context of gastrointestinal malignancies. The model is a useful tool in healthcare practice as it may help diagnose GI cancer because of its high F1-scores and overall accuracy.
We employed a blurring-based masker in conjunction with the SHAP partition explainer to gain an understanding of the deterministic elements that underlie the predictions of our ensemble model. We were able to explain the accurate forecasts by using this method to visualize the precise regions of the image that were important to the model’s predictions. Four photos from each class—ulcerative colitis, polyps, and esophagitis—that our model properly predicted were included in this study. We were able to generate a visual depiction of each class’s contributing attributes by utilizing the SHAP partition explainer. The deterministic characteristics and their significance for the ulcerative colitis, polyps, and esophagitis groups are shown in Figure 6.
These visualizations improve the interpretability and explain the ability of our model’s predictions by offering insightful information about the areas or patterns within the pictures that had a major impact on the ensemble model’s decision-making process. The chart shows the real image in Figure 6, with blue and red highlights in particular areas. The red color indicates elements that positively added to the prediction of a particular category, while the blue color represents parts that had an adverse contribution. By analyzing the fourth image in Figure 6 as an example, it can be observed that the red shades are predominantly concentrated around the region corresponding to the esophagitis pathology in the esophagitis class.
This suggests that these highlighted regions played a significant role in the model’s prediction for this category. However, when examining the subsequent two classes predicted by the model, we notice that both images exhibit mostly blue shades in the area associated with esophagitis pathology. This implies that these regions negatively influenced the model’s prediction for these classes. Overall, the model predicts and outputs the deterministic features of each tested image, highlighting the regions that contribute positively or adversely to the predicted categories. This provides valuable insights into the specific image characteristics that the model considers when making its predictions. Moreover, features visualization using t-sne is also shown in Figure 7.
The limited number of studies conducted on gastrointestinal cancer detection highlights the need for further research in this area. Existing studies have reported moderate to high accuracies using deep learning models such as InceptionResNetV2 and InceptionV3. For instance, one study (65) achieved an accuracy of 84.5% using InceptionResNetV2 with a dataset of 854 images, while another study (49) reported an accuracy of 90.1% using InceptionV3 with a test set of 341 endoscopic images.
In comparison, our optimized ensemble model, along with the individual models, demonstrates superior performance compared to these existing studies. The accuracy of our ensemble model is reported as 93.17% with an F1-score of 97% for each class. This indicates the effectiveness of our approach in accurately classifying gastrointestinal lesions. However, it is important to acknowledge the challenges faced in developing and evaluating deep learning models for gastrointestinal cancer due to the limited availability of publicly accessible datasets in this domain. This scarcity hinders the progress and thorough evaluation of deep learning models for gastrointestinal cancer detection. Moreover, the lack of explainability in deep learning models has contributed to the hesitation among healthcare professionals in adopting these models in clinical practices. To address this limitation, our proposed model incorporates the SHAP technique, which allows for the identification of deterministic features within the images associated with gastrointestinal pathologies. By providing explanations for the model decision-making process, our model enhances the interpretability and trustworthiness of the results. It has been observed that major misclassification occurs in the ulcerative-colitis and then polyp class. This occurs as both have similarity in their shape and size, the problem can be catered by applying the contours and highlighting the region, which will be done in future work. The limitation of the model is its reproducibility of results which is generally a deep learning issue, moreover the technique is not evaluated on a real time system therefore it should be trialed clinically before implementation of it.
The comparison of the proposed model with the latest techniques is shown in Table 3.
5 Conclusion
The use of technology in healthcare is often challenged by a lack of explanation. This research addresses this issue by examining SHAP (Shapley Additive exPlanations) technology in depth. Colon cancer pathology results can be used to extract preliminary characteristics thanks to SHAP. The use of SHAP in our research aims to enhance the comprehension and interpretation of the prediction model. We start our study by creating and improving augmented ensemble models. The averaging approach was used to merge three pre-trained CNN models: InceptionV3, InceptionResNetV2, and VGG16. Pathology findings from the KvasirV2 dataset, a helpful tool for diagnosing gastrointestinal disorders, were analyzed for this sample. Co-learning maximizes the model’s quality by increasing the model’s accuracy and efficiency. Because a pooled sample incorporates the unique strengths and capacities of each sample, cancer detection using it can be more robust and trustworthy. Furthermore, each disease’s characteristic traits were highlighted using the SHAP translator method. With the use of this technology, we can decipher certain details and regions of medical pictures, enabling the creation of prediction models. We may gain a better grasp of the decision-making mechanism and the underlying concepts of forecasting by extracting and visualizing these elements. Our results demonstrate the acceleration, quality, and use of descriptive intelligence (XAI) models for cancer detection, particularly in colon cancer. For future work, we will investigate other AI models, explainability methodologies, or applicability to other forms of cancer or disorders.
Data availability statement
The data that support the findings of this study are available from the first and corresponding authors upon reasonable request.
Author contributions
FB: Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Validation, Visualization, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research work was funded by Institutional Fund Projects under grant no. (IFPIP: 860–830-1443). The authors gratefully acknowledge technical and financial support provided by the Ministry of Education and King Abdulaziz University, DSR, Jeddah, Saudi Arabia.
Conflict of interest
The author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Lopes, J, Rodrigues, CM, Gaspar, MM, and Reis, CP. Melanoma management: from epidemiology to treatment and latest advances. Cancers. (2022) 14:4652. doi: 10.3390/cancers14194652
2. Arnold, M, Abnet, CC, Neale, RE, Vignat, J, Giovannucci, EL, McGlynn, KA, et al. Global burden of 5 major types of gastrointestinal cancer. Gastroenterology. (2020) 159:335–349.e15. e15. doi: 10.1053/j.gastro.2020.02.068
3. Islami, F, Goding Sauer, A, Miller, KD, Siegel, RL, Fedewa, SA, Jacobs, EJ, et al. Proportion and number of cancer cases and deaths attributable to potentially modifiable risk factors in the United States. CA Cancer J Clin. (2018) 68:31–54. doi: 10.3322/caac.21440
4. van den Brandt, PA, and Goldbohm, RA. Nutrition in the prevention of gastrointestinal cancer. Best Pract Res Clin Gastroenterol. (2006) 20:589–603. doi: 10.1016/j.bpg.2006.04.001
5. Matsuoka, T, and Yashiro, M. Precision medicine for gastrointestinal cancer: recent progress and future perspective. World J Gastrointest Oncol. (2020) 12:1–20. doi: 10.4251/wjgo.v12.i1.1
6. Allemani, C, Matsuda, T, Di Carlo, V, Harewood, R, Matz, M, Nikšić, M, et al. Global surveillance of trends in cancer survival 2000–14 (CONCORD-3): analysis of individual records for 37 513 025 patients diagnosed with one of 18 cancers from 322 population-based registries in 71 countries. Lancet. (2018) 391:1023–75. doi: 10.1016/S0140-6736(17)33326-3
7. Moghimi-Dehkordi, B, and Safaee, A. An overview of colorectal cancer survival rates and prognosis in Asia. World J Gastrointest Oncol. (2012) 4:71–5. doi: 10.4251/wjgo.v4.i4.71
8. Frenette, CT, and Strum, WB. Relative rates of missed diagnosis for colonoscopy, barium enema, and flexible sigmoidoscopy in 379 patients with colorectal cancer. J Gastrointest Cancer. (2007) 38:148–53. doi: 10.1007/s12029-008-9027-x
9. Grasgruber, P, Hrazdira, E, Sebera, M, and Kalina, T. Cancer incidence in Europe: an ecological analysis of nutritional and other environmental factors. Front Oncol. (2018) 8:151. doi: 10.3389/fonc.2018.00151
10. Gupta, J, Agrawal, T, Singh, P, and Diwakar, M. Optical biosensor for early diagnosis of Cancer. 2023 International Conference on Computer, Electronics & Electrical Engineering & their Applications (IC2E3); IEEE. (2023).
11. Ahmad, Z, Rahim, S, Zubair, M, and Abdul-Ghafar, J. Artificial intelligence (AI) in medicine, current applications and future role with special emphasis on its potential and promise in pathology: present and future impact, obstacles including costs and acceptance among pathologists, practical and philosophical considerations. A comprehensive review. Diagn Pathol. (2021) 16:1–16. doi: 10.1186/s13000-021-01085-4
12. Nasir, IM, Raza, M, Shah, JH, Khan, MA, Nam, Y-C, and Nam, Y. Improved shark smell optimization algorithm for human action recognition. Comput Mater Contin. (2023) 76:2667–84. doi: 10.32604/cmc.2023.035214
13. Nasir, IM, Raza, M, Ulyah, SM, Shah, JH, Fitriyani, NL, and Syafrudin, M. ENGA: elastic net-based genetic algorithm for human action recognition. Expert Syst Appl. (2023) 227:120311. doi: 10.1016/j.eswa.2023.120311
14. Nasir, IM, Raza, M, Shah, JH, Wang, S-H, Tariq, U, and Khan, MA. HAREDNet: a deep learning based architecture for autonomous video surveillance by recognizing human actions. Comput Electr Eng. (2022) 99:107805. doi: 10.1016/j.compeleceng.2022.107805
15. Tariq, J, Alfalou, A, Ijaz, A, Ali, H, Ashraf, I, Rahman, H, et al. Fast intra mode selection in HEVC using statistical model. Comput Mater Contin. (2022) 70:3903–18. doi: 10.32604/cmc.2022.019541
16. Nasir, IM, Rashid, M, Shah, JH, Sharif, M, Awan, MY, and Alkinani, MH. An optimized approach for breast cancer classification for histopathological images based on hybrid feature set. Curr Med Imaging. (2021) 17:136–47. doi: 10.2174/1573405616666200423085826
17. Mushtaq, I, Umer, M, Imran, M, Nasir, IM, Muhammad, G, and Shorfuzzaman, M. Customer prioritization for medical supply chain during COVID-19 pandemic. Comput Mater Contin. (2021) 70:59–72. doi: 10.32604/cmc.2022.019337
18. Nasir, IM, Raza, M, Shah, JH, Khan, MA, and Rehman, A. Human action recognition using machine learning in uncontrolled environment. 2021 1st International Conference on Artificial Intelligence and Data Analytics (CAIDA); IEEE. (2021).
19. Nasir, IM, Khan, MA, Yasmin, M, Shah, JH, Gabryel, M, Scherer, R, et al. Pearson correlation-based feature selection for document classification using balanced training. Sensors. (2020) 20:6793. doi: 10.3390/s20236793
20. Nasir, IM, Bibi, A, Shah, JH, Khan, MA, Sharif, M, Iqbal, K, et al. Deep learning-based classification of fruit diseases: an application for precision agriculture. Comput Mater Contin. (2021) 66:1949–62. doi: 10.32604/cmc.2020.012945
21. Nasir, IM, Khan, MA, Armghan, A, and Javed, MY. SCNN: a secure convolutional neural network using blockchain. 2020 2nd International Conference on Computer and Information Sciences (ICCIS); IEEE. (2020).
22. Khan, MA, Nasir, IM, Sharif, M, Alhaisoni, M, Kadry, S, Bukhari, SAC, et al. A blockchain based framework for stomach abnormalities recognition. Comput Mater Contin. (2021) 67:141–58. doi: 10.32604/cmc.2021.013217
23. Mashood Nasir, I, Attique Khan, M, Alhaisoni, M, Saba, T, Rehman, A, and Iqbal, T. A hybrid deep learning architecture for the classification of superhero fashion products: an application for medical-tech classification. Comput Mod Eng Sci. (2020) 124:1017–33. doi: 10.32604/cmes.2020.010943
24. Amann, J, Blasimme, A, Vayena, E, Frey, D, Madai, VI, and Consortium, PQ. Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med Inform Decis Mak. (2020) 20:1–9. doi: 10.1186/s12911-020-01332-6
25. Zhang, Y, Weng, Y, and Lund, J. Applications of explainable artificial intelligence in diagnosis and surgery. Diagnostics. (2022) 12:237. doi: 10.3390/diagnostics12020237
26. Dutta, S. An overview on the evolution and adoption of deep learning applications used in the industry. Wiley Interdiscip Rev. (2018) 8:e1257. doi: 10.1002/widm.1257
27. Tehsin, S, Rehman, S, Saeed, MOB, Riaz, F, Hassan, A, Abbas, M, et al. Self-organizing hierarchical particle swarm optimization of correlation filters for object recognition. IEEE Access. (2017) 5:24495–502. doi: 10.1109/ACCESS.2017.2762354
28. Tehsin, S, Rehman, S, Awan, AB, Chaudry, Q, Abbas, M, Young, R, et al. Improved maximum average correlation height filter with adaptive log base selection for object recognition. Optical Pattern Recognition XXVII; (2016).
29. Tehsin, S, Rehman, S, Bilal, A, Chaudry, Q, Saeed, O, Abbas, M, et al. Comparative analysis of zero aliasing logarithmic mapped optimal trade-off correlation filter. Pattern Recognition and Tracking XXVIII; SPIE. (2017).
30. Tehsin, S, Rehman, S, Riaz, F, Saeed, O, Hassan, A, Khan, M, et al. Fully invariant wavelet enhanced minimum average correlation energy filter for object recognition in cluttered and occluded environments. Pattern Recognition and Tracking XXVIII; SPIE. (2017).
31. Akbar, N, Tehsin, S, Bilal, A, Rubab, S, Rehman, S, and Young, R, editors. Detection of moving human using optimized correlation filters in homogeneous environments. Pattern Recognition and Tracking XXXI; SPIE. (2020).
32. Tehsin, S, Asfia, Y, Akbar, N, Riaz, F, Rehman, S, and Young, R. Selection of CPU scheduling dynamically through machine learning. Pattern Recognition and Tracking XXXI; SPIE. (2020).
33. Akbar, N, Tehsin, S, Ur Rehman, H, Rehman, S, and Young, R. Hardware design of correlation filters for target detection. Pattern Recognition and Tracking XXX; SPIE. (2019).
34. Asfia, Y, Tehsin, S, Shahzeen, A, and Khan, US. Visual person identification device using raspberry pi. The 25th conference of FRUCT association (2019).
35. Saad, SM, Bilal, A, Tehsin, S, and Rehman, S. Spoof detection for fake biometric images using feature-based techniques. SPIE Future Sensing Technologies; SPIE. (2020).
36. Yang, G, Ye, Q, and Xia, J. Unbox the black-box for the medical explainable AI via multi-modal and multi-Centre data fusion: a mini-review, two showcases and beyond. Inf Fusion. (2022) 77:29–52. doi: 10.1016/j.inffus.2021.07.016
37. Lundberg, SM, and Lee, S-I. A unified approach to interpreting model predictions. Adv Neural Inf Proces Syst. (2017) 30:4768–77. doi: 10.48550/arXiv.1705.07874
38. Nouman Noor, M, Nazir, M, Khan, SA, Song, O-Y, and Ashraf, I. Efficient gastrointestinal disease classification using pretrained deep convolutional neural network. Electronics. (2023) 12:1557. doi: 10.3390/electronics12071557
39. Nouman Noor, M, Nazir, M, Khan, SA, Ashraf, I, and Song, O-Y. Localization and classification of gastrointestinal tract disorders using explainable AI from endoscopic images. Appl Sci. (2023) 13:9031. doi: 10.3390/app13159031
40. Noor, MN, Ashraf, I, and Nazir, M. Analysis of GAN-based data augmentation for GI-tract disease classification In: H Ali, MH Rehmani, and Z Shah, editors. Advances in deep generative models for medical artificial intelligence. Berlin: Springer (2023). 43–64.
41. Noor, MN, Nazir, M, and Ashraf, I. Emerging trends and advances in the diagnosis of gastrointestinal diseases. BioScientific Rev. (2023) 5:118–43. doi: 10.32350/BSR.52.11
42. Noor, MN, Nazir, M, Ashraf, I, Almujally, NA, Aslam, M, and Fizzah, JS. GastroNet: a robust attention-based deep learning and cosine similarity feature selection framework for gastrointestinal disease classification from endoscopic images. CAAI transactions on intelligence. Technology. (2023) 2023:12231. doi: 10.1049/cit2.12231
43. Bertsimas, D, Margonis, GA, Tang, S, Koulouras, A, Antonescu, CR, Brennan, MF, et al. An interpretable AI model for recurrence prediction after surgery in gastrointestinal stromal tumour: an observational cohort study. Eclinicalmedicine. (2023) 64:102200. doi: 10.1016/j.eclinm.2023.102200
44. Auzine, MM, Khan, MH-M, Baichoo, S, Sahib, NG, Gao, X, and Bissoonauth-Daiboo, P. Classification of gastrointestinal Cancer through explainable AI and ensemble learning. 2023 Sixth International Conference of Women in Data Science at Prince Sultan University (WiDS PSU); IEEE. (2023).
45. Bang, CS, Lee, JJ, and Baik, GH. Computer-aided diagnosis of esophageal cancer and neoplasms in endoscopic images: a systematic review and meta-analysis of diagnostic test accuracy. Gastrointest Endosc. (2021) 93:1006–1015.e13. e13. doi: 10.1016/j.gie.2020.11.025
46. Janse, MH, Van der Sommen, F, Zinger, S, and Schoon, EJ. Early esophageal cancer detection using RF classifiers. Medical imaging 2016: computer-aided diagnosis; SPIE. (2016).
47. Lee, JH, Kim, YJ, Kim, YW, Park, S, Choi, Y-i, Kim, YJ, et al. Spotting malignancies from gastric endoscopic images using deep learning. Surg Endosc. (2019) 33:3790–7. doi: 10.1007/s00464-019-06677-2
48. Xiao, T, Renduo, S, Lianlian, W, and Honggang, Y. An automatic diagnosis system for chronic atrophic gastritis under white light endoscopy based on deep learning. Endoscopy. (2022) 54:S80. doi: 10.1055/s-0042-1744749
49. Li, L, Chen, Y, Shen, Z, Zhang, X, Sang, J, Ding, Y, et al. Convolutional neural network for the diagnosis of early gastric cancer based on magnifying narrow band imaging. Gastric Cancer. (2020) 23:126–32. doi: 10.1007/s10120-019-00992-2
50. Sakai, Y, Takemoto, S, Hori, K, Nishimura, M, Ikematsu, H, Yano, T, et al. Automatic detection of early gastric cancer in endoscopic images using a transferring convolutional neural network. 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); IEEE. (2018).
51. Obayya, M, Al-Wesabi, FN, Maashi, M, Mohamed, A, Hamza, MA, Drar, S, et al. Modified salp swarm algorithm with deep learning based gastrointestinal tract disease classification on endoscopic images. IEEE Access. (2023) 11:25959–67. doi: 10.1109/ACCESS.2023.3256084
52. Iakovidis, DK, Georgakopoulos, SV, Vasilakakis, M, Koulaouzidis, A, and Plagianakos, VP. Detecting and locating gastrointestinal anomalies using deep learning and iterative cluster unification. IEEE Trans Med Imaging. (2018) 37:2196–210. doi: 10.1109/TMI.2018.2837002
53. Wang, S, Yin, Y, Wang, D, Lv, Z, Wang, Y, and Jin, Y. An interpretable deep neural network for colorectal polyp diagnosis under colonoscopy. Knowl-Based Syst. (2021) 234:107568. doi: 10.1016/j.knosys.2021.107568
54. Pogorelov, K, Randel, KR, Griwodz, C, Eskeland, SL, de Lange, T, Johansen, D, et al. Kvasir: a multi-class image dataset for computer aided gastrointestinal disease detection. Proceedings of the 8th ACM on Multimedia Systems Conference (2017).
55. Szegedy, C, Vanhoucke, V, Ioffe, S, Shlens, J, and Wojna, Z. Rethinking the inception architecture for computer vision. Proceedings of the IEEE conference on computer vision and pattern recognition (2016).
56. Simonyan, K, and Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv. (2014) 2014:14091556. doi: 10.48550/arXiv.1409.1556
57. Szegedy, C, Ioffe, S, Vanhoucke, V, and Alemi, A. Inception-v4, inception-resnet and the impact of residual connections on learning. Proceedings of the AAAI conference on artificial intelligence (2017).
58. Nisbet, R, Elder, J, and Miner, GD. Handbook of statistical analysis and data mining applications. Cambridge, MA: Academic Press (2009).
59. Kotu, V, and Deshpande, B. Predictive analytics and data mining: Concepts and practice with rapidminer. Burlington, MA: Morgan Kaufmann (2014).
60. Ganaie, MA, Hu, M, Malik, A, Tanveer, M, and Suganthan, P. Ensemble deep learning: a review. Eng Appl Artif Intell. (2022) 115:105151. doi: 10.1016/j.engappai.2022.105151
61. Hulsen, T. Explainable artificial intelligence (XAI): concepts and challenges in healthcare. AI. (2023) 4:652–66. doi: 10.3390/ai4030034
62. Abou Jaoude, M, Sun, H, Pellerin, KR, Pavlova, M, Sarkis, RA, Cash, SS, et al. Expert-level automated sleep staging of long-term scalp electroencephalography recordings using deep learning. Sleep. (2020) 43:112. doi: 10.1093/sleep/zsaa112
63. Jansen, T, Geleijnse, G, Van Maaren, M, Hendriks, MP, Ten Teije, A, and Moncada-Torres, A. Machine learning explainability in breast cancer survival. Stud Health Technol Inform. (2020) 270:307–11. doi: 10.3233/SHTI200172
64. Tymchenko, B, Marchenko, P, and Spodarets, D. Deep learning approach to diabetic retinopathy detection. arXiv. (2020) 2020:200302261. doi: 10.48550/arXiv.2003.02261
Keywords: gastrointestinal cancer, explainable AI, SHAP, transfer learning, ensemble learning
Citation: Binzagr F (2024) Explainable AI-driven model for gastrointestinal cancer classification. Front. Med. 11:1349373. doi: 10.3389/fmed.2024.1349373
Edited by:
Vinayakumar Ravi, Prince Mohammad bin Fahd University, Saudi ArabiaReviewed by:
Prabhishek Singh, Bennett University, IndiaJani Anbarasi L., Vellore Institute of Technology, India
Copyright © 2024 Binzagr. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Faisal Binzagr, fbinzagr@kau.edu.sa