- 1Department of Computer Science, COMSATS Institute of Information Technology, Islamabad, Pakistan
- 2Department of Computer Science, Kohsar University Murree, Punjab, Pakistan
- 3Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, Riyadh, Saudi Arabia
- 4Department of AI and Software, Gachon University, Seongnam, Republic of Korea
Prostate cancer (PCa) is a major global concern, particularly for men, emphasizing the urgency of early detection to reduce mortality. As the second leading cause of cancer-related male deaths worldwide, precise and efficient diagnostic methods are crucial. Due to high and multiresolution MRI in PCa, computer-aided diagnostic (CAD) methods have emerged to assist radiologists in identifying anomalies. However, the rapid advancement of medical technology has led to the adoption of deep learning methods. These techniques enhance diagnostic efficiency, reduce observer variability, and consistently outperform traditional approaches. Resource constraints that can distinguish whether a cancer is aggressive or not is a significant problem in PCa treatment. This study aims to identify PCa using MRI images by combining deep learning and transfer learning (TL). Researchers have explored numerous CNN-based Deep Learning methods for classifying MRI images related to PCa. In this study, we have developed an approach for the classification of PCa using transfer learning on a limited number of images to achieve high performance and help radiologists instantly identify PCa. The proposed methodology adopts the EfficientNet architecture, pre-trained on the ImageNet dataset, and incorporates three branches for feature extraction from different MRI sequences. The extracted features are then combined, significantly enhancing the model’s ability to distinguish MRI images accurately. Our model demonstrated remarkable results in classifying prostate cancer, achieving an accuracy rate of 88.89%. Furthermore, comparative results indicate that our approach achieve higher accuracy than both traditional hand-crafted feature techniques and existing deep learning techniques in PCa classification. The proposed methodology can learn more distinctive features in prostate images and correctly identify cancer.
1 Introduction
A major challenge for medical science is cancer, which is the most widespread disease in humans around the globe. Cancer cells exhibit aggressive growth rates, and their precise diagnosis is pivotal to a patient’s survival. The most prevalent cancer diagnosed in men worldwide is PCa. Alarming statistics from the American Cancer Society predict approximately 288,300 new PCa cases in the United States by 2023, with an estimated 34,700 fatalities cancer society (1).
The conventional method for PCa classification relies on the Gleason Score (GS), as determined by analyzing biopsy samples. However, this conventional biopsy method has been found to exhibit reduced sensitivity in accurately identifying PCa. Furthermore, the Gleason classification method encounters challenges stemming from variations in interpretation, encompassing discrepancies among different observers (interobserver variability) and inconsistencies within assessments made by the same observer (intraobserver variability). These variations can predominantly be attributed to the heavy reliance on human interpretation within the Gleason classification method ŞCheck that all equations and special characters are displayed correctly.erbănescu et al. (2). Recent advancements in mpMRI have emerged as a pivotal tool for assessing the risk of PCa and improving the grading and classification of PCa Oberlin et al. (3); Bardis et al. (4). High-grade PCa is frequently characterized by the presence of more densely packed structures which can be effectively identified through the utilization of advanced MRI-based machine learning techniques. These approaches enable the detection and assessment of high-grade PCa, thereby enhancing diagnostic precision and treatment planning for patients. This study aims to create an automated method for classifying PCa, aiding radiologists’ assessments.
The implementation of quantitative assessments of mpMRI offers radiologists a valuable, noninvasive tool to enhance their clinical decision-making processes. Furthermore, it helps mitigate the discrepancies that can arise due to differences between readers. With the growing interest in the integration of artificial intelligence (AI) with medical practice, empowered by enhanced computational capabilities and the emergence of new AI methodologies, there has been a surge in studies introducing CAD approaches. These systems leverage machine-learning and deep-learning approaches to detect and classify tumors in medical imagery, with a pronounced impact on PCa detection and classification Abbasi et al. (5). This study aims to build an automated PCa classification method, addressing the limitations of traditional GS and improving the diagnostic potential offered by mp-MRI. Early-stage PCa identification is extremely important and beneficial for treatments. Identification of PCa using MRI images improves the rate of early diagnosis and assists in building a Computer Aided Diagnostic (CAD) system Jin et al. (6); Reda et al. (7). A pivotal objective within CAD systems is the development of objective and reproducible metrics for automated analysis Cem Birbiri et al. (8). The continuous refinement of PCa classification techniques holds significant importance, particularly in distinguishing between low and high-grade cancers.
To identify PCa, several approaches have been proposed. Most of them used traditional machine-learning approaches, to classify images and train classifiers Bardis et al. (4); Monni et al. (9); Abbasi et al. (5); Gillies et al. (10); Fehr et al. (11); Vignati et al. (12); Liu et al. (13); Ullah et al. (14); Shahzad et al. (15); Laghari et al. (16); Sobecki et al. (17); Giannini et al. (18); Wang et al. (19); Schelb et al. (20); Wildeboer et al. (21); Wibmer et al. (22). Many of these approaches used features of low radiomics focused on previous clinical reports, which may not fully leverage the entire information within the MRI images Källén et al. (23). Moreover, unsupervised approaches were previously used to acquire features that may contain unnecessary information or may exclude essential clues. Deep learning approaches recently acquired great performance and are widely used in classifying and identification tasks of both medical applications Esteva et al. (24); Albarqouni et al. (25); Yuan and Meng (26) and natural images LeCun et al. (27). They can train classifiers and learn features jointly. Because deep learning techniques have enormous potential and success, the authors use them to classify PCa. Deep learning-based architectures have yielded remarkable results because of their capability to autonomously acquire and represent features Tsehay et al. (28). Compared to conventional approaches, CNN-based models such as Alexnet demonstrated improved performance Kiraly et al. (29). A challenge associated with architectures like this is the substantial data needed for effective training such as in Chen et al. (30). Using transfer learning is an easier way to handle this problem. In order to extract features and identify data from one domain into another, transfer learning employs training experience as a sort of knowledge sharing Le et al. (31). Good performance can be accomplished with small training images by using the transfer learning technique Wildeboer et al. (32); Zhong et al. (33) that applies models of pre-trained images from other datasets. Furthermore, various MRI sequences of PCa present different concerns, and it is important to consider an effective way to incorporate different details. By combining details derived from multi-parametric images, a descriptive representation of PCa may be gained Cem Birbiri et al. (8).
A multi-parametric MRI transfer learning (mp-TL) system to identify PCa is presented in this study. To obtain features from various MRI sequences, the proposed transfer learning model has three branches: ADC and T2w (sagittal, trans-axial). The features extracted from these categories are combined in the model. For this study, we aim to utilize transfer learning techniques leveraging a family of networks of pre-trained EfficientNet models for the classification of prostate images. Compound scaling is used in the recently proposed Efficient-Nets architecture to balance the network’s three dimensions (Depth, Height, and Width). The proposed method demonstrates good performance in effectively classifying PCa images, contributing to enhanced diagnosis. The improvement in classifying PCa techniques is necessary to distinguish low and high-grade cancer. There is a need for an efficient deep learning-based architecture that efficiently classifies PCa images. In the proposed methodology, the important step for the classification of MRI images is the pre-processing stage. Pre-processing is used to process the PCa MRI images, and then the classification and feature extraction of PCa images is performed using deep CNN models. The contribution of this study is based on the classification of images and results in comparison with existing approaches. These are a few of this study’s major contributions.
• The proposed approach here harnesses transfer learning to jointly analyze multiple MRI sequences, rather than focusing solely on a single MRI sequence. This enables us to extract more discriminative features, leading to a substantial enhancement in PCa classification.
• To demonstrate the model’s effectiveness, the authors evaluate the PCa dataset utilizing a diverse range of Efficient-Net Models, encompassing B0, B5, and B7.
• Multi-view ensemble approach is used for classifying multi-parametric MRI images.
• An Efficient-Net model with fine-tuning and an additional Global Average Pooling (GAP) layer at the model’s end, serves as a crucial component. This not only extracts vital information but also forwards it to the activation function for further processing.
• The proposed approach’s effectiveness is highlighted through extensive experimentation conducted on the PCa dataset.
2 Related work
Numerous studies have been carried out by researchers to predict prostate MRI imaging. The literature on MRI image classification encompasses a range of both deep-learning and machine-learning techniques. There are various PCa MRI datasets that can be used for classification tasks, such as prostatex, ACRIN, and I2CVB. However, accessibility to these datasets is often limited or incomplete for many researchers. In contrast, Prostatex is a publicly accessible dataset specifically intended for research purposes.
For the MRI imaging classification of PCa, Chen et al. (34) suggested a deep-learning method focused on classification. A deep convolutional neural network, such as InceptionV3 and VGG-16 underwent pre-training on the ImageNet dataset. Subsequently, the multi-parametric magnetic resonance imaging dataset was fine-tuned. Xu et al. (35) employed residual networks for the identifying PCa. ResNets have demonstrated a capacity to learn both low-level and high-level features, making them well-suited for detecting subtle and intricate patterns in medical images, which are often indicative of diseases like PCa.Their study showcased the feasibility of training residual networks to acquire features that are valuable for identifying suspicious indicative of PCa.
In this study, Alkadi et al. (36), the authors employ a deep convolutional neural network to segment prostate lesions in T2W MRI images. They introduce a 3D sliding window technique for 3D context while maintaining computational efficiency. The approach distinguishes cancerous and non-cancerous tissues, with comparable results to multi-parametric systems, avoiding intricate alignment steps. This comprehensive study Viswanath et al. (37) assesses the performance of supervised classifiers in a multisite approach for detecting prostate cancer (PCa) extent using T2w MRI. The primary focus is on radiomic features extracted from high-resolution T2 images. The aim is to enhance the accuracy and timeliness of diagnoses in the context of medical imaging, particularly for PCa detection, where early and precise identification is critical for effective treatment.
Authors in this study Muhammad et al. (38) have highlighted the potential of utilizing a combination of multiple parameters, either as individual parameters or integrated multiple parameters within a machine learning framework, to enhance diagnostic capabilities. Their Schelb et al. (20) study highlights the effectiveness of training deep learning models to recognize and segment lesions in T2 and diffusion MRI data, significantly improving the clinical evaluation of MRI data. The UNet model was trained using cross-validation, incorporating split-sample techniques, and subsequently validated using an external test set. Singh et al. (39) suggest the use of deep neural networks for cribriform pattern classification. In this study, the authors introduce an automated image classification system employing deep learning and hand-crafted features to analyze prostate images. The focus is on detecting cribriform patterns, with results demonstrating diagnostic potential.
With notable advancements in computer vision, particularly in target recognition and identification through deep convolutional neural networks, the medical imaging research community is increasingly delving into the exploration of diverse CNN architectures. These architectures offer substantial potential for enhancing the accuracy of cancer detection systems. In this study, Yoo et al. (40) developed and introduced an automated pipeline based on CNN. This pipeline is designed to analyze images on a per-patient basis, aiming to detect clinically relevant PCa.
Bulten et al. (41) reported that the implementation of a semi-automatic labeling system eliminated the need for pathologists to manually annotate the images. A high degree of agreement with the reference norm has been obtained by the established framework. The deep learning method outperformed pathologists in different observation trials. Li et al. (42) clarify that for the diagnosis of disease, histology analysis is also seen as the gold standard. By reducing test time and inter-observer variability, computer-aided diagnostic software can theoretically further optimize existing pathology workflows. Previous cancer grading analyses have predominantly focused on the classification of predefined regions of significance or the handling of extensive volumes of fine-grained annotations.
Using a Genetic Algorithm, Namdar et al. (43) recommended fine-tuning a qualified CNN for enhanced PCa diagnosis, resulting in an improved AUC. Furthermore, Kwon et al. (44) proposed a radiomics-based method for prostate image identification. The purpose was to identify multi-parametric MRI for clinically important PCa. Lay et al. (45) stated because MR imaging has its limitations, researchers suggest a different PCa detection technique that can be most effective. The cancer detection approach trains random ferns on MR sequences in the absence of one or more of these MR sequences and then uses these random ferns to add the MR sequences.
An approach for evaluating the grade for PCa has been suggested in this paper by Abraham and Nair (46). In this method, features are extracted utilizing deep network autoencoders in conjunction with hand-crafted features, subsequently categorized with a softmax classifier. Song et al. (47) have shown that radiologists manually mark the regions of significance for PCa and measure the scores for each area. The authors developed a model on patch-based DCNN that utilizes a combination of MRI data to distinguish between cancerous and non-cancerous patients of PCa.
According to Lemaitre et al. (48), new magnetic resonance imaging (MRI) approaches have emerged to enhance diagnostic accuracy. However, factors like observer variability and the visibility and complexity of lesions can still impact diagnosis. In this respect, CAD-based applications are designed to support radiologists in their clinical practice. Taking account of all MRI modalities, the authors suggest a CAD method. The goal of this CAD scheme was to detect the prostate position of cancer. Liu et al. (49) stated that for the classification of PCa, deep learning architecture was developed using the 3D multipara-metric MRI data. The Xmas-Net model was used for extracting features in this study. Mehrtash et al. (50) have demonstrated that to better detect PCa Computer-assisted diagnosis of MRI PCa may be used as a method of clinical decision support to help interpretation by radiologists. CNN models are used to detect the probability of a patient being affected or not.Yang et al. (51) provides an integrated method for detecting PCa that can simultaneously image PCa and locate lesions based on characteristics of the deep convolutionary neural network and SVM.
3 Transfer learning
TL is a technique for transferring information across domains Orenstein and Beijbom (52). Deep learning is a challenging and time-intensive process, especially in medical imaging, where a substantial amount of training data is needed to understand certain patterns. To address the challenge of limited data, medical imaging datasets are utilized to fine-tune the weights of deep learning models that were previously trained for different computer vision applications, thus accelerating the training process. The strategy frequently used in various computer vision problems is fine-tuning transfer learning. For classification, the dense layers are well-tuned, while the top layers are frozen. The proposed methodology for classifying prostate images using transfer learning is shown in Figure 1.
4 Materials and methods
The prostate dataset was obtained from the Cancer Imaging Archive Nolan (53). The dataset consists of prostate MRI images which are labeled with the help of radiologists. After performing minor data preprocessing, we carefully selected 221 cases of prostate cancer for our dataset. Our dataset comprised 98 cases of low-grade and 123 cases of high-grade cancer. Every patient included in the study exhibited an initial screening result that raised suspicions regarding prostate cancer. Subsequently, each of these patients underwent a biopsy, from which a GS was determined. These cases are annotated with two-class labels distinguishing between low-grade (GS = 3 + 4, 3 + 3) and high-grade (GS = 4 + 4, 4 + 3, 5 + 3, 3 + 5) cancer. For each case, T2w (sagittal and transaxial) and ADC images were provided to conduct experiments.
In our experimental setup, we adopted a systematic approach to partitioning the dataset to ensure robust training and evaluation of our model. We performed a random selection process, wherein 80% of the dataset was utilized for various purposes, including training and validation, while the remaining 20% of the images were exclusively designated as the test set. Within this 80% portion of the dataset, we further allocated distinct proportions for training and validation. Approximately 50% of the dataset was allocated for training, which served as the foundation for our transfer learning process. The remaining 30% of this 80% portion was dedicated to the validation set. This set played a pivotal role in monitoring the model’s performance during training. By periodically evaluating the model’s predictions on this validation subset, we could make informed decisions regarding hyperparameter tuning and model adjustments, ultimately ensuring that our model’s generalization capabilities were optimized. Lastly, the 20% of the images that constituted the test set were kept entirely separate from the training and validation data. This segregation ensured that our model was assessed on entirely unseen data, mirroring real-world scenarios where it would be applied to make predictions. The test set served to evaluate the model’s performance, providing a reliable measure of its ability to generalize to new and previously unseen data. Through this well-structured data partitioning strategy, we aimed to achieve a robust and thorough assessment of our model’s capabilities, while also upholding the principles of fairness, rigor, and transparency in our experimental approach.
4.1 Proposed approach
This study presented a transfer learning model that utilizes multiparametric MRI for the classification of PCa into low-grade and high-grade. In Figure 2, the proposed model is mentioned. To learn features from multiparametric sequences (ADC, T2w), the authors make a transfer learning model with three branches and combine them to gain discriminative descriptors. A significant amount of training data is needed for deep convolution neural networks in medical imaging. When the available data is insufficient, deep CNNs often rely on pre-trained models. These models have been previously trained on extensive datasets, allowing for knowledge transfer, which is a fundamental aspect of TL.
Figure 3 describes the suggested model’s workflow. Transfer learning-based multi-parametric MRI model to automatic PCa identification is presented in this study. Various sequences of MRI reveal distinct aspects of PCa. T2-weighted (T2w) and ADC (Apparent Diffusion Coefficient) imaging modalities offer distinct insights, and their integration can significantly enhance the accuracy of PCa classification. To learn features from multi-parametric sequences T2w (sagittal and transaxial) and ADC, the authors make a transfer learning model with three branches of architecture to gain features separately for each modality and then combine them to gain one feature vector. We feed these sequences simultaneously in the network and their concatenation after the convolutional layer. Such a fusion approach allows the learning process to generate effective and discriminating PCa-related characteristics of multiple modalities mutually influenced by each other. To achieve better performance, we fine-tune our model by changing the top layers and defining the last layer classes to two nodes, as we identify PCa as a low-grade and high-grade form of cancer. After optimizing the MPTL methodology, we could perform the task of classifying prostate images. The performance of classifying PCa could be further improved by improving the ability to combine learned features.
4.2 Transfer learning using a single sequence to extract features
Due to the limited availability of prostate cancer data, we have opted for a transfer learning strategy instead of training an entire deep-learning neural network from the ground up. Specifically, we have harnessed the power of established deep learning architectures like ConvNets, AlexNet, and VGGNet, which have previously been trained on ImageNet datasets and are readily accessible as pre-trained networks. By implementing the EfficientNet architecture across multiple MRI series, we can extract essential features from ADC, T2w sagittal, and T2w transaxial images. This innovative approach allows us to transfer the knowledge acquired from ImageNet and effectively characterize PCa images.
4.3 A multisequence MRI-based feature fusion method
Different MRI modalities of PCa demonstrate different aspects. Various sequences of MRI disclose various PCa kinds. To provide separate and complementary data, T2w and ADC are recorded, and their combination can effectively increase the precision of PCa diagnosis. It is efficient to obtain the simultaneous information from MRI in deep learning method to optimize the relation between different MRIs. We feed these sequences simultaneously in the network and their concatenation after the convolutional layer. Such a fusion approach allows the learning process to generate effective and discriminating PCa-related characteristics of multiple modalities mutually influenced by each other. The performance is seen as the final joint characteristic after fully connected layers.
4.4 Developing a fine-tuned training strategy
In our approach utilizing the Efficient-Net architecture, we took several steps to enhance the classification of PCa into high and low-grade. We integrated fully connected layers into the network and fine-tuned it using our dataset. This fine-tuning process was pivotal in adapting the model to our specific classification task. To boost the feature extraction capabilities of our model, we introduced custom layers, including global average pooling within the classification layers. This addition helped in capturing more nuanced features from the medical images, which is crucial in accurately classifying cancer. What sets our approach apart is the use of Efficient-Net architectures, which come with distinct advantages. These models are not only faster in classification, being 6.1× faster compared to existing CNN models, but they are also significantly smaller, being 8.4× smaller. Importantly, their compact size doesn’t compromise their accuracy. Our Efficient-Net models leveraged TL based on architectures pre-trained on the extensive ImageNet dataset, known for its high accuracy and efficiency. This TL approach allowed us to benefit from the knowledge embedded in these pre-trained models, especially when our own dataset was limited. To further enhance the training process and ensure robustness, we employed data augmentation. This technique plays a crucial role in augmenting the dataset, increasing the diversity of training samples, and consequently, improving the model’s ability to generalize to unseen data. It is particularly effective in preventing overfitting, a common challenge in classification tasks. One notable aspect of our strategy is the use of a pre-trained model as a feature extractor. In this approach, the last fully connected layer is removed, and the remaining layers are treated as a fixed feature extractor. This significantly accelerates the training process. In essence, our approach combines the advantages of Efficient-Net architectures, Transfer Learning, data augmentation, and a pre-trained feature extractor to enhance the classification of prostate cancer. Figure 4 provides a visual representation of our model in action, demonstrating its potential in the field of medical image classification. The performance of the baseline Efficient-Net architecture is demonstrated in Figure 5.
4.5 Evaluation metrics
The output of input images is typically evaluated using the evaluation matrices listed below. The confusion matrix can be used to measure these matrices, including accuracy, precision, recall, and F1-score. The confusion matrix has four different types of parameters, where TP denotes a true positive, TN denotes a true negative, FP denotes a false positive, and FN denotes a false negative. as shown in Table 1 and equations are given below.
True Positives: The precise predicted positive values demonstrate that the predicted and the actual class value are both positive.
True Negatives: These are the accurately predicted negative values, showing that both the predicted and the actual class value are negative.
False Positives: When the predicted class is true but the actual class is false.
False Negatives: When the predicted class is no but the actual class is yes.
Accuracy: Accuracy is the most used performance metric, which may be calculated as the ratio of correctly expected observations to all observations. Having high accuracy will lead one to believe that our models outperform.
Precision: Precision is the ratio of accurately predicted positive observations to all predicted positive observations.
Recall: Recall is defined as the ratio of accurately predicted positive observations to all of the actual class observations.
F1 score: The weighted average of recall and accuracy is the Score. This score takes into consideration both false positives and false negatives. Although it is not as easy to immediately understand as accuracy, it is typically more beneficial than precision, especially if we have an uneven class distribution. If false positives and false negatives result in equal losses, accuracy performs better. It is simpler to include both accuracies and recall if the cost of false positives and negatives is significantly different.
4.6 Experimental settings
To get generalized results, the authors repeated the experiment several times, looking at different learning and test data combinations. To increase the robustness of the presented MPTL approach and to reduce overfitting, the authors used the data augmentation technique to increase the size of the training data of different image transformations. Before transferring images as input to the networks, the authors conducted some preprocessing steps. To fit the model, images are resized or reshaped from the original size to 244 × 244 for the family of Efficient-Net models for image classification. The Adam optimizer was used to further train the entire set of pre-trained Efficient-Net models.
The settings utilized to conduct the experiments are as follows.
• Experiment carried out using a Google Colab Pro platform with a GPU T4 P100 and 25 Gigabytes memory.
• Batch size of 16.
• Learning rate from le-1 to le-6 with weight decay of le-4.
• Epochs are set to 150.
• The Adam optimizer was used to further train the whole set of pre-trained Efficient-Net models Perez and Wang (54).
4.7 Experiments and results
We conducted an extensive comparative analysis to evaluate our proposed model alongside eight baseline methods. Initially, we employed a transfer learning model without fine-tuning, utilizing image features directly from ImageNet for experimentation. We then delved into three additional baseline experiments, finetuning single MRI sequences, specifically T2-weighted (T2w) and ADC (Apparent Diffusion Coefficient), for prostate cancer classification. Subsequently, we extended our experiments to three more baseline experiments, employing two MRI parameters as input. The comprehensive classification results for both our method and the eight baseline methods are meticulously detailed in Table 2.
To assess the classification performance of our proposed MPTL model, we carried out a comprehensive evaluation, comparing it with state-of-the-art prostate cancer classification methods, including both deep learning and machine learning-based approaches. These comparisons were conducted using our prostate cancer datasets, and we followed the experiment settings outlined in these reference papers to ensure a fair and equitable assessment. Table 3 meticulously presents the precision, recall, and accuracy metrics achieved by both our approach and the comparative methodologies. It’s noteworthy that deep learning-based techniques outperformed methods relying on traditional radiomics features or conventional machine learning approaches. This observation highlights the capability of deep learning-based techniques to capture more distinctive features for the identification of prostate cancer.
Our method performs better at classification than the preceding approaches. This is due to the fact that convergence issues and over-fitting issues with little data on PCa also hinder deep network training. In comparison, the image details in the transfer learning model using Efficient-Net were considered by our MPTL model. Therefore, as compared to previous classification techniques, our technique evaluates more precise parameters for PCa and achieves more efficiency.
This method classified the input image into cancer types with low and high grades. We elaborate on the experimental results performance to distinguish between the aggressive and non-aggressive forms of cancer. The Efficient-Net B7 architecture, which was trained on images of PCa, produces the greatest results. Table 3 shows the results of the proposed methods. In Şerbănescu et al. (2), authors apply the Google-Net approach for the identification of PCa classification for binary classification to distinguish low and high-grade forms of cancer and achieve 60.9 accuracies and performance. In Chen et al. (34) authors apply the VGG-16 approach for the identification of PCa classification for binary classification to distinguish the low and high-grade forms of cancer and achieve 83 accuracies and performance.
In Kwon et al. (44), authors apply the CART approach for the identification of PCa classification for binary classification to distinguish the low and high-grade forms of cancer and achieve 82.0 accuracies and performance. In Le et al. (31), authors apply the ResNet approach for the identification of PCa classification for binary classification to distinguish low and high-grade forms of cancer and achieve 82.09 accuracies and performance. In Muhammad et al. (55), authors apply the inceptionV3 approach for the identification of PCa classification for binary classification to distinguish the low and high-grade forms of cancer and achieve 80.09 accuracies and performance.
The comparison results of different methods with multi-parametric modalities are shown in Table 2. The results demonstrate that our proposed approach with a fusion of three modalities performs better results than single modalities and pair of modalities which depicts that our approach performance is better on multiple modalities.
The learning curve for accuracy and loss during training and validation is depicted in Figures 6, 7. Our approach also shows better identification performance compared with the other machine learning approaches with extraction features from a single MRI sequence, showing that the methods based on deep learning will learn more high-level discriminative features. ROC curves of PCa classification are shown in Figure 8. The results demonstrate the performance of the model to identify input images is classified as low and high-grade forms of cancer that are aggressive and non-aggressive forms of cancer.
4.8 Cross dataset validation
To comprehensively evaluate the generalization performance of the proposed Multi-Parameter Transfer Learning (MPTL) model, an extensive experiment was conducted across diverse datasets. The primary objective of this experiment was to showcase the practicality and effectiveness of the MPTL framework in real-world scenarios where data sources may vary significantly.
For this purpose, we strategically leveraged two distinct datasets, both of which are publicly available to ensure transparency and reproducibility in our research. The first dataset, sourced from the Cancer Imaging Archive Nolan (53), is a substantial collection of medical images related to prostate cancer. The second dataset, referred to as I2CVB G. Lemaitre et al. (56), provides an additional set of prostate MRI images for comparative analysis. It’s worth noting that these datasets bring a wealth of diversity to the experiment. They exhibit variations in terms of image characteristics such as shapes, angles, sizes, resolutions, and formats. This diversity mirrors the real-world scenario where medical imaging data can originate from various sources and possess inherent dissimilarities. After minor cleaning, our training process was conducted on a robust training set comprising 5096 images from the first dataset. Subsequently, we rigorously assessed the model’s performance on an independent test set comprising 1371 images sourced from the I2CVB dataset. This demarcation of training and testing datasets enabled us to simulate a real-world scenario where a model is required to adapt and generalize across distinct data sources.
The results obtained in this cross-dataset experiment are highly encouraging. The proposed MPTL framework exhibited remarkable performance, further emphasizing its versatility and effectiveness in handling diverse data sources. Specifically, our model achieved an accuracy rate of 86.65%, indicating its capability to make correct classifications. The precision rate, measuring the model’s ability to correctly classify positive cases, stood at an impressive 83.36%. Furthermore, the recall rate, signifying the model’s capacity to identify all relevant instances, reached an impressive 89.18%. Lastly, the F1-score, which strikes a balance between precision and recall, demonstrated a robust performance at 86.13%. These outcomes underscore the generalization power of the MPTL framework for the classification of prostate MRI images. The model’s consistent and high-quality performance across datasets with diverse characteristics reinforces its potential utility in real-world medical applications, where data heterogeneity is often encountered.
5 Conclusion
In terms of replacing manual cancer assessment by radiologists using MRI images, CAD plays a critical role. There are, however, numerous risks and a high level of complexity involved in this task, along with expert-level opinions. The manual extraction of handcrafted features and subsequent classification not only consumes time but also introduces a higher likelihood of errors. To streamline the assessment process for radiologists and mitigate diagnostic errors, the necessity for an automated decision-making classification model becomes evident. In this paper, we introduce an innovative MPTL model for the automatic classification of PCa. Our model leverages knowledge from ImageNet to aid in the feature learning process from multi-parametric MRI (mp-MRI) sequences. These transferred features are combined to enhance the accuracy of PCa classification. A refined fine-tuning method including global average pooling is further applied to enhance PCa classification. As a result, the learned features exhibit significantly enhanced discriminative capabilities. Through an extensive series of comparative studies, we have highlighted the exceptional performance of our model in direct comparison to the prevailing state-of-the-art cancer classification methods. Our empirical results unequivocally establish the efficacy of our proposed approach in achieving high-precision PCa classification. Our findings highlight the potential benefits of transfer learning techniques from natural images to the medical domain, potentially offering valuable solutions in scenarios where the availability of annotated training datasets is limited for various practical considerations.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author contributions
Conceptualization, MM, MFM, and SA; methodology, MM, MSA, and MA; software, MM, KA, and MSA; validation, MFM, SA, and KA; formal analysis, MA and MSA; investigation, MSA, and SA; resources, MM, MFM and SA; data curation, KA, MSA, and MFM; writing—original draft preparation, MM, MFM, and SA; writing—review and editing, MSA, MA, KA, and MFM; visualization, MM, MFM, and SA; supervision, MFM and SA; project administration, MA and KA; funding acquisition, MA. All authors contributed to the article and approved the submitted version.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This Research is funded by Researchers Supporting Project Number (RSPD2023R947), King Saud University, Riyadh, Saudi Arabia.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
2. Şerbănescu M-S, Manea NC, Streba L, Belciug S, Ples¸ea IE, Pirici I, et al. Automated Gleason grading of prostate cancer using transfer learning from general-purpose deep-learning networks. Romanian J Morphology Embryology (2020) 61:149. doi: 10.47162/RJME.61.1.17
3. Oberlin DT, Casalino DD, Miller FH, Meeks JJ. Dramatic increase in the utilization of multiparametric magnetic resonance imaging for detection and management of prostate cancer. Abdominal Radiol (2017) 42:1255–8. doi: 10.1007/s00261-016-0975-5
4. Bardis MD, Houshyar R, Chang PD, Ushinsky A, Glavis-Bloom J, Chahine C, et al. Applications of artificial intelligence to prostate multiparametric MRI (mpMRI): Current and emerging trends. Cancers (2020) 12:1204. doi: 10.3390/cancers12051204
5. Abbasi AA, Hussain L, Awan IA, Abbasi I, Majid A, Nadeem MSA, et al. Detecting prostate cancer using deep learning convolution neural network with transfer learning approach. Cogn Neurodynamics (2020) 14:523–33. doi: 10.1007/s11571-020-09587-5
6. Jin J, Zhang L, Leng E, Metzger GJ, Koopmeiners JS. Detection of prostate cancer with multiparametric MRI utilizing the anatomic structure of the prostate. Stat Med (2018) 37:3214–29. doi: 10.1002/sim.7810
7. Reda I, Shalaby A, Khalifa F, Elmogy M, Aboulfotouh A, Abou El-Ghar M, et al. Computeraided diagnostic tool for early detection of prostate cancer, in: 2016 IEEE International Conference on Image Processing (ICIP) (IEEE). (2016) 2668–72.
8. Cem Birbiri U, Hamidinekoo A, Grall A, Malcolm P, Zwiggelaar R. Investigating the performance of generative adversarial networks for prostate tissue detection and segmentation. J Imaging (2020) 6:83. doi: 10.3390/jimaging6090083
9. Monni F, Fontanella P, Grasso A, Wiklund P, Ou Y-C, Randazzo M, et al. Magnetic resonance imaging in prostate cancer detection and management: a systematic review. Minerva urologica e nefrologica= Ital J Urol Nephrol (2017) 69:567–78. doi: 10.23736/S0393-2249.17.02819-3
10. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology (2016) 278:563–77. doi: 10.1148/radiol.2015151169
11. Fehr D, Veeraraghavan H, Wibmer A, Gondo T, Matsumoto K, Vargas HA, et al. Automatic classification of prostate cancer Gleason scores from multiparametric magnetic resonance images. Proc Natl Acad Sci (2015) 112:E6265–73. doi: 10.1073/pnas.1505935112
12. Vignati A, Mazzetti S, Giannini V, Russo F, Bollito E, Porpiglia F, et al. Texture features on T2-weighted magnetic resonance imaging: new potential biomarkers for prostate cancer aggressiveness. Phys Med Biol (2015) 60:2685. doi: 10.1088/0031-9155/60/7/2685
13. Liu L, Tian Z, Zhang Z, Fei B. Computer-aided detection of prostate cancer with MRI: technology and applications. Acad Radiol (2016) 23:1024–46. doi: 10.1016/j.acra.2016.03.010
14. Ullah N, Khan MS, Khan JA, Choi A, Anwar MS. A robust end-to-end deep learning-based approach for effective and reliable BTD using MR images. Sensors (2022) 22:7575. doi: 10.3390/s22197575
15. Shahzad M, Tahir MA, Alhussein M, Mobin A, Shams Malick RA, Anwar MS. NeuPD—A neural network-based approach to predict antineoplastic drug response. Diagnostics (2023) 13:2043. doi: 10.3390/diagnostics13122043
16. Laghari AA, Sun Y, Alhussein M, Aurangzeb K, Anwar MS, Rashid M. Deep residual-dense network based on bidirectional recurrent neural network for atrial fibrillation detection. Sci Rep (2023) 13:15109. doi: 10.1038/s41598-023-40343-x
17. Sobecki P, Życka-Malesa D, Mykhalevych I, Sklinda K, Przelaskowski A. MRI imaging texture features in prostate lesions classification, in: EMBEC & NBC 2017: Joint Conference of the European Medical and Biological Engineering Conference (EMBEC) and the Nordic-Baltic Conference on Biomedical Engineering and Medical Physics (NBC), Tampere, Finland, June 2017 (Springer). (2018) 827–30.
18. Giannini V, Vignati A, Mirasole S, Mazzetti S, Russo F, Stasi M, et al. MR-T2-weighted signal intensity: a new imaging biomarker of prostate cancer aggressiveness. Comput Methods Biomechanics Biomed Engineering: Imaging Visualization (2016) 4:130–4. doi: 10.1080/21681163.2014.910476
19. Wang X, Yang W, Weinreb J, Han J, Li Q, Kong X, et al. Searching for prostate cancer by fully automated magnetic resonance imaging classification: deep learning versus non-deep learning. Sci Rep (2017) 7:15415. doi: 10.1038/s41598-017-15720-y
20. Schelb P, Kohl S, Radtke JP, Wiesenfarth M, Kickingereder P, Bickelhaupt S, et al. Classification of cancer at prostate MRI: deep learning versus clinical PI-RADS assessment. Radiology (2019) 293:607–17. doi: 10.1148/radiol.2019190938
21. Wildeboer RR, van Sloun RJG, Huang P, Wijkstra H, Mischi M. 3-D multi-parametric contrast-enhanced ultrasound for the prediction of prostate cancer. Ultrasound Med Biol (2019) 45:2713–24. doi: 10.1016/j.ultrasmedbio.2019.05.017
22. Wibmer A, Hricak H, Gondo T, Matsumoto K, Veeraraghavan H, Fehr D, et al. Haralick texture analysis of prostate MRI: utility for differentiating non-cancerous prostate from prostate cancer and differentiating prostate cancers with different Gleason scores. Eur Radiol (2015) 25:2840–50. doi: 10.1007/s00330-015-3701-8
23. Källén H, Molin J, Heyden A, Lundström C, Åström K. Towards grading gleason score using generically trained deep convolutional neural networks, in: 2016 IEEE 13th International Symposium on Biomedical Imaging (ISBI) (IEEE). (2016) 1163–7.
24. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. nature (2017) 542:115–8. doi: 10.1038/nature21056
25. Albarqouni S, Baur C, Achilles F, Belagiannis V, Demirci S, Navab N. Aggnet: deep learning from crowds for mitosis detection in breast cancer histology images. IEEE Trans Med Imaging (2016) 35:1313–21. doi: 10.1109/TMI.2016.2528120
26. Yuan Y, Meng MQ-H. Deep learning for polyp recognition in wireless capsule endoscopy images. Med Phys (2017) 44:1379–89. doi: 10.1002/mp.12147
28. Tsehay Y, Lay N, Wang X, Kwak JT, Turkbey B, Choyke P, et al. Biopsy-guided learning with deep convolutional neural networks for Prostate Cancer detection on multiparametric MRI, in: 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017) (IEEE). (2017) 642–5.
29. Kiraly AP, Nader CA, Tuysuzoglu A, Grimm R, Kiefer B, El-Zehiry N, et al. Deep convolutional encoder-decoders for prostate cancer detection and classification, in: Medical Image Computing and Computer Assisted Intervention- MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, September 11-13, 2017, Proceedings, Part III 20 (Springer). (2017) 489–97.
30. Chen Q, Xu X, Hu S, Li X, Zou Q, Li Y. A transfer learning approach for classification of clinical significant prostate cancers from mpMRI scans. in: Medical Imaging 2017: Computer-Aided Diagnosis (SPIE). (2017) 10134:1154–7.
31. Le MH, Chen J, Wang L, Wang Z, Liu W, Cheng K-TT, et al. Automated diagnosis of prostate cancer in multi-parametric MRI based on multimodal convolutional neural networks. Phys Med Biol (2017) 62:6497. doi: 10.1088/1361-6560/aa7731
32. Wildeboer RR, Mannaerts CK, van Sloun RJG, Budäus L, Tilki D, Wijkstra H, et al. Automated multiparametric localization of prostate cancer based on B-mode, shear-wave elastography, and contrast-enhanced ultrasound radiomics. Eur Radiol (2020) 30:806–15. doi: 10.1007/s00330-019-06436-w
33. Zhong X, Cao R, Shakeri S, Scalzo F, Lee Y, Enzmann DR, et al. Deep transfer learning-based prostate cancer classification using 3 Tesla multi-parametric MRI. Abdominal Radiol (2019) 44:2030–9. doi: 10.1007/s00261-018-1824-5
34. Chen Q, Hu S, Long P, Lu F, Shi Y, Li Y. A transfer learning approach for Malignant prostate lesion detection on multiparametric MRI. Technol Cancer Res Treat (2019) 18:1533033819858363. doi: 10.1177/1533033819858363
35. Xu H, Baxter JSH, Akin O, Cantor-Rivera D. Prostate cancer detection using residual networks. Int J Comput assisted Radiol Surg (2019) 14:1647–50. doi: 10.1007/s11548-019-01967-5
36. Alkadi R, Taher F, El-Baz A, Werghi N. A deep learning-based approach for the detection and localization of prostate cancer in T2 magnetic resonance images. J digital Imaging (2019) 32:793–807. doi: 10.1007/s10278-018-0160-1
37. Viswanath SE, Chirra PV, Yim MC, Rofsky NM, Purysko AS, Rosen MA, et al. Comparing radiomic classifiers and classifier ensembles for detection of peripheral zone prostate tumors on T2-weighted MRI: a multi-site study. BMC Med Imaging (2019) 19:1–12. doi: 10.1186/s12880-019-0308-6
38. Muhammad S, Dailey MN, Farooq M, Majeed MF, Ekpanyapong M. Spec-Net and Spec-CGAN: Deep learning models for specularity removal from faces. Image Vision Computing (2020) 93:103823. doi: 10.1016/j.imavis.2019.11.001
39. Singh M, Kalaw EM, Jie W, Al-Shabi M, Wong CF, Giron DM, et al. Cribriform pattern detection in prostate histopathological images using deep learning models. arXiv preprint arXiv:1910.04030 (2019).
40. Yoo S, Gujrathi I, Haider MA, Khalvati F. Prostate cancer detection using deep convolutional neural networks. Sci Rep (2019) 9:1–10. doi: 10.1038/s41598-019-55972-4
41. Bulten W, Pinckaers H, van Boven H, Vink R, de Bel T, van Ginneken B, et al. Automated deep-learning system for Gleason grading of prostate cancer using biopsies: a diagnostic study. Lancet Oncol (2020) 21:233–41. doi: 10.1016/S1470-2045(19)30739-9
42. Li J, Li W, Gertych A, Knudsen BS, Speier W, Arnold CW. An attention-based multi-resolution model for prostate whole slide imageclassification and localization. arXiv preprint arXiv:1905.13208 (2019).
43. Namdar K, Gujrathi I, Haider MA, Khalvati F. Evolution-based fine-tuning of CNNs for prostate cancer detection. arXiv preprint arXiv:1911.01477 (2019). doi: 10.48550/arXiv.1911.01477
44. Kwon D, Reis IM, Breto AL, Tschudi Y, Gautney N, Zavala-Romero O, et al. Classification of suspicious lesions on prostate multiparametric MRI using machine learning. J Med Imaging (2018) 5:034502–2. doi: 10.1117/1.JMI.5.3.034502
45. Lay N, Tsehay Y, Sumathipala Y, Cheng R, Gaur S, Smith C, et al. A decomposable model for the detection of prostate cancer in multi-parametric MRI, in: Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, September 16-20, 2018, Proceedings, Part II 11 (Springer). (2018) 930–9.
46. Abraham B, Nair MS. Computer-aided classification of prostate cancer grade groups from MRI images using texture features and stacked sparse autoencoder. Computerized Med Imaging Graphics (2018) 69:60–8. doi: 10.1016/j.compmedimag.2018.08.006
47. Song Y, Zhang Y-D, Yan X, Liu H, Zhou M, Hu B, et al. Computer-aided diagnosis of prostate cancer using a deep convolutional neural network from multiparametric MRI. J Magnetic Resonance Imaging (2018) 48:1570–7. doi: 10.1002/jmri.26047
48. Lemaitre G, Martí R, Rastgoo M, Mériaudeau F. Computer-aided detection for prostate cancer detection based on multi-parametric magnetic resonance imaging, in: 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) (IEEE). (2017) 3138–41.
49. Liu S, Zheng H, Feng Y, Li W. Prostate cancer diagnosis using deep learning with 3D multiparametric MRI. in: Medical imaging 2017: computer-aided diagnosis (SPIE). (2017) 10134:581–4. doi: 10.1117/12.2277121
50. Mehrtash A, Sedghi A, Ghafoorian M, Taghipour M, Tempany CM, Wells WM III, et al. Classification of clinical significance of MRI prostate findings using 3D convolutional neural networks. In: Medical imaging 2017: computer-aided diagnosis (SPIE). (2017) 10134:589–92. doi: 10.1117/12.2277123
51. Yang X, Liu C, Wang Z, Yang J, Le Min H, Wang L, et al. Co-trained convolutional neural networks for automated detection of prostate cancer in multi-parametric MRI. Med image Anal (2017) 42:212–27. doi: 10.1016/j.media.2017.08.006
52. Orenstein EC, Beijbom O. Transfer learning and deep feature extraction for planktonic image data sets, in: 2017 IEEE winter conference on applications of computer vision (WACV) (IEEE). (2017) 1082–8.
53. Nolan T. Spie-aapm-nci prostatex challenges (prostatex). (2017), Dataset. doi: 10.7937/K9TCIA.2017.MURS5CL.
54. Perez L, Wang J. The effectiveness of data augmentation in image classification using deep learning. arXiv preprint arXiv:1712.04621 (2017). doi: 110.3389/fmed.2021.629134
55. Muhammad S, Dailey MN, Sato I, Majeed MF. Handling specularity in intrinsic image decomposition, in: Image Analysis and Recognition: 15th International Conference, ICIAR 2018, Povoa´ de Varzim, Portugal, June 27–29, 2018, Proceedings 15 (Springer). (2018) 107–15.
Keywords: transfer learning, convolutional neural network, deep learning, PCA, MRI images
Citation: Mehmood M, Abbasi SH, Aurangzeb K, Majeed MF, Anwar MS and Alhussein M (2023) A classifier model for prostate cancer diagnosis using CNNs and transfer learning with multi-parametric MRI. Front. Oncol. 13:1225490. doi: 10.3389/fonc.2023.1225490
Received: 19 May 2023; Accepted: 16 October 2023;
Published: 09 November 2023.
Edited by:
Fabio Grizzi, Humanitas Research Hospital, ItalyReviewed by:
Shady Saikali, AdventHealth, United StatesJeffrey Tuan, National Cancer Centre Singapore, Singapore
Copyright © 2023 Mehmood, Abbasi, Aurangzeb, Majeed, Anwar and Alhussein. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Muhammad Shahid Anwar, shahidanwar786@gachon.ac.kr; Muhammad Faran Majeed, m.faran.majeed@kum.edu.pk