Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 06 July 2022
Sec. Genitourinary Oncology
This article is part of the Research Topic The Application of Artificial Intelligence In Diagnosis, Treatment and Prognosis In Urologic Oncology View all 11 articles

Machine Learning-Based Models Enhance the Prediction of Prostate Cancer

Sunmeng ChenSunmeng Chen1Tengteng JianTengteng Jian1Changliang ChiChangliang Chi1Yi LiangYi Liang2Xiao LiangXiao Liang2Ying YuYing Yu1Fengming JiangFengming Jiang1Ji Lu*Ji Lu1*
  • 1Department of Urology, The First Hospital of Jilin University, Changchun, China
  • 2School of Business and Management, Jilin University, Changchun, China

Purpose: PSA is currently the most commonly used screening indicator for prostate cancer. However, it has limited specificity for the diagnosis of prostate cancer. We aim to construct machine learning-based models and enhance the prediction of prostate cancer.

Methods: The data of 551 patients who underwent prostate biopsy were retrospectively retrieved and divided into training and test datasets in a 3:1 ratio. We constructed five PCa prediction models with four supervised machine learning algorithms, including tPSA univariate logistic regression (LR), multivariate LR, decision tree (DT), random forest (RF), and support vector machine (SVM). The five prediction models were compared based on model performance metrics, such as the area under the receiver operating characteristic curve (AUC), accuracy, sensitivity, specificity, calibration curve, and clinical decision curve analysis (DCA).

Results: All five models had good calibration in the training dataset. In the training dataset, the RF, DT, and multivariate LR models showed better discrimination, with AUCs of 1.0, 0.922 and 0.91, respectively, than the tPSA univariate LR and SVM models. In the test dataset, the multivariate LR model exhibited the best discrimination (AUC=0.918). The multivariate LR model and SVM model had better extrapolation and generalizability, with little change in performance between the training and test datasets. Compared with the DCA curves of the tPSA LR model, the other four models exhibited better net clinical benefits.

Conclusion: The results of the current retrospective study suggest that machine learning techniques can predict prostate cancer with significantly better AUC, accuracy, and net clinical benefits.

Introduction

Prostate cancer (PCa) is the second leading malignancy in men and the fifth leading cause of cancer mortality in men worldwide (1). Although PSA is still the most commonly used screening tool for prostate cancer, it has been controversial in recent decades (2). It is suggested that PSA screening improves the detection rate of localized and less aggressive prostate cancers but also reduces the proportions of advanced PCa and PCa-specific mortality (35). However, due to the obvious overlap of PSA levels in various conditions, such as benign prostatic hyperplasia, prostatitis, and prostate cancer, the specificity of PSA screening is low, which leads to a plethora of unrelated diseases for prostate biopsy (2). These unnecessary prostate biopsies result in not only a significant waste of medical resources but also an increased incidence of sepsis, which can be life-threatening to patients (6). Therefore, there is a need for a new convenient method to improve the diagnostic ability of PCa.

Machine learning is a branch of artificial intelligence (AI) in which machines are programmed to learn patterns from data, and the learning itself is based on a set of mathematical rules and statistical assumptions. It is widely used in biology because of its enormous advantages in dealing with large datasets (7, 8). It has also been rapidly developed and applied in the medical field, especially in the construction of predictive models (9). Therefore, using machine learning to construct PCa prediction models would be a feasible and promising approach.

In this study, we constructed generalizable machine learning predictive models to improve the accuracy of PCa risk assessment by using objective parameters present in electronic medical records and then evaluated their performance.

Methods

Data Sources

A total of 789 male patients in the First Hospital of Jilin University who underwent transrectal ultrasound-guided prostate biopsy from January 2013 to January 2021 were included. Indications for prostate biopsy included serum tPSA >4 ng/ml, abnormal digital rectal exam (DRE), or imaging findings suggestive of suspected prostate cancer. All patients underwent systematic biopsy with 10-12 cores. Patients with one of the following criteria were excluded from the study: taking medications that could affect serum PSA levels, unclear results of the prostate biopsy, and significant abnormal values or missing data. A final total of 551 patients were included in the study. All patient data were collected through electronic medical records, including age, BMI, hypertension, diabetes, total PSA (tPSA), free PSA (fPSA), the ratio of serum fPSA to tPSA (f/tPSA), prostate volume (PV), PSA density (PSAD), neutrophil-to-lymphocyte ratio (NLR), and pathology reports of prostate biopsy. All examinations were completed within one week before prostate biopsy. PSAD is the ratio of tPSA to PV. The calculation of PV was calculated by the following formula: maximal transverse diameter × maximal anterior-posterior diameter × maximal superior-inferior diameter × 0.52.

Model Development

R software (version 4.1.4, https://www.rproject.org/) was used to develop machine learning models. A total of five prediction models were constructed by dividing the data into a training dataset and a test dataset at a ratio of 3:1. The pathologic type of PCa or other benign disease was used as the dichotomous variable, and other variables were all used as continuous variables. For logistic regression (LR) model construction, the best predictive variables were first screened in the training dataset using stepwise regression, association plots between the variables were made to understand the magnitude of the association between the variables, and the presence or absence of collinearity between variables was judged according to the variance inflation factor (VIF). Then, LR models were constructed using the “lrm” function in the “rms” package. For the decision tree (DT) model, we used the “rpart” package for training, using two hyperparameters, the complexity parameters cp and spilt. The initial cp value was set to 0.001, and then the best cp value was found and pruned based on the best cp value. The input variables were obtained through the selection of important features, and the best DT model was then output. The random forest (RF) model screened the optimal input variables by significant feature selection. The RF model was trained using the “randomForest” package in R software, using two hyperparameters, ntree and mtry, which were set at 500 and 6, respectively. The support vector machine (SVM) model was filtered by the “caret” package for important features. Training was performed using the “e1071” package, using a Gaussian kernel function and setting the two hyperparameters, cost and gamma, to 1 and 0.1, respectively.

Model Performance Evaluation

The performance of the developed models was validated using a test dataset in a process that was completely independent of the algorithm training. The performance of the five models was then evaluated by comparing four metrics: the receiver operating characteristic (ROC) curve and its corresponding area under the curve (AUC) and the accuracy, sensitivity, and specificity. The calibration curve was used to evaluate the calibration of the model, the Brier score was used to assess the calibration, the Hosmer–Lemeshow goodness-of-fit test was used to judge whether there was a significant difference between the observed and predicted values, and clinical decision curve analysis (DCA) was used to assess the net benefit of the model.

Data Analysis

For comparative analysis between two samples, Student’s t test was used for normally distributed continuous variables, and the Mann–Whitney U test was used for categorical variables with nonnormal continuous variables. Continuous variables in the data were expressed as medians and IQRs or means and SDs, categorical variables were expressed as frequencies and percentages, and the bilateral significance level for the left-right test was set at 5% (p<0.05).

Results

Baseline Patient Characteristics

Table 1 shows the baseline characteristics of the patients. A total of 302 (54.8%) of the 551 patients were diagnosed with PCa. The PCa detection rate in patients ≥65 years old was higher than patients <65. The mean levels of tPSA, fPSA, and PSAD were significantly higher in the PCa group than in the non-PCa group. When tPSA>4 ng/ml, the PCa detection rate increased with increasing tPSA. In the subgroups of 4<tPSA<10, 10≤tPSA<20, 20≤tPSA<100, and tPSA≥100, the detection rates were 18.6%, 26.2%, 54.2%, and 97.2%, respectively. The mean neutrophil count, PV, and BMI were lower in the PCa group than in the non-PCa group. No significant differences in other variables were found between two groups.

TABLE 1
www.frontiersin.org

Table 1 Characteristics of Patients, stratified by biopsy outcomes.

LR Algorithm-Based PCa Prediction Model

First, the tPSA univariate LR model was constructed in the training dataset by including only one single factor, tPSA. This model showed that tPSA was positively correlated with the diagnosis of PCa (coefficient=0.034). Then, a multivariate LR model was constructed by including all variables through a stepwise regression method. When the VIC value reached the minimum value, a total of seven best predictive variables were selected, including age, tPSA, fPSA, PV, NLR, peripheral blood neutrophil count and lymphocyte count. The two LR models are shown in Supplementary Table 1. In the multivariate LR model, age, tPSA, and fPSA were positively correlated with PCa, while PV and neutrophil count were negatively correlated with PCa (Supplementary Figure 1). In addition, there was a significant association between peripheral blood neutrophil count and NLR, tPSA and fPSA, suggesting their respective possible collinearity. The variance inflation factor (VIF) was subsequently calculated for verification, and all VIF values were less than 5, indicating no collinearity between any of the variables.

DT Algorithm-Based PCa Prediction Model

The optimal cp value for the DT model was 0.008. Based on the corresponding ranking of important features, the final input variables for the DT model were age, tPSA, fPSA, PV, PSAD, NLR, f/tPSA, and biopsy results (Supplementary Figure 2A). The process and results of model classification are shown in Supplementary Figure 2B. This model correctly classified 87.7% (363/414) of the cases in the training dataset.

RF Algorithm-Based PCa Prediction Model

After ranking the important features of the RF model, seven features with the highest predictive accuracy were selected as the input features, including age, tPSA, fPSA, PV, PSAD, NLR, and peripheral blood neutrophil count (Supplementary Figure 3A). The error of the model gradually decreased as the number of decision trees increased, and the minimum error value of the RF model was reached when the number of decision trees was 313 (Supplementary Figure 3B).

SVM Algorithm-Based PCa Prediction Model

The best SVM model was screened by the “rfe” function in the caret package using a 10-fold cross-validation method. The number of variables was screened one by one from 1 to 12, and the best model was obtained when the number of variables was 5. The input variables at this time were age, PSAD, tPSA, fPSA, and PV.

Performance of the Developed Models

In the training dataset, the RF and DT models performed particularly well in differentiation with the AUCs of 1.0 and 0.922, respectively (Table 2). The model with the lowest AUC was the tPSA univariate LR model (0.842). The calibration in all five models was very good, which suggested that the predicted values of the models were in high agreement with the actual values (Figure 1A). The Brier scores of the multivariate LR, DT, RF, SVM and tPSA LR models were 0.119, 0.122, 0.121, 0.118 and 0.154, respectively. The p values for the Hosmer–Lemeshow test were all greater than 0.05, indicating that there was no statistical bias in the near-perfect fit between the predicted and actual values. Therefore, the models we constructed were valid and reliable.

TABLE 2
www.frontiersin.org

Table 2 Diagnostic performance of different machine learning models.

FIGURE 1
www.frontiersin.org

Figure 1 Performance of machine learning models. (A) Calibration curves of five prediction models in the training dataset. The predicted probabilities are plotted on the X-axis, and the actual probabilities are plotted on the Y-axis. (B) The ROC curves of the tPSA LR, multivariate LR, DT, RF, and SVM models in the test dataset. (C) Clinical decision curve analysis (DCA) of the five models in the test dataset.

All characteristics of samples were comparable in the training and test datasets (Supplementary Table 2). In the test dataset, the specificity of the tPSA LR model was highest, reaching 93.2%; however, the sensitivity and accuracy were relatively low, at 63.9% and 77.1%, respectively (Table 2). The sensitivity and accuracy of the multivariate LR model were improved significantly in the test dataset compared with the training dataset, although there was a slight decrease in specificity. The diagnostic performance of the SVM model was not outstanding in either the training or the test dataset, but most outcomes were improved in the test dataset. Thus, the multivariate LR model exhibited the best discrimination, and the extrapolation and generalization abilities of the multivariate LR and SVM models were relatively strong. In contrast, although the DT and RF models performed well in the training dataset, their performances in the test dataset decreased significantly. The corresponding ROC curves for the five models are shown in Figure 1B. To further evaluate the potential clinical benefits of these prediction models, we performed DCA curves using the test dataset (Figure 1C). All models demonstrated significant net benefits. Compared with the tPSA LR model, the other four models showed significantly higher net clinical benefits.

To evaluate the performance of constructed models in the subgroups of PSA of 4-10 ng/ml and 4-20 ng/ml, we used these two subgroups as the test datasets, and evaluated their AUCs, sensitivities, specificities, and accuracies (Supplementary Tables 3, 4). The diagnostic performances of the RF algorithm-based machine learning prediction model showed best in both subgroups. The AUCs were 0.856 and 0.94 respectively. Although the sensitivity decreased compared with it in the training dataset, the specificity and accuracy were still relatively high. The performance of other algorithms-based developed models were not outstanding.

Discussion

In the past decade, PSA has been widely used as the most valuable diagnostic and prognostic marker for PCa (10). However, some studies have shown that less than 30% of men with PSA levels in the gray zone (4-10 ng/ml) have pathologically confirmed PCa, indicating that a large proportion of patients have undergone unnecessary biopsies and been overdiagnosed (11). In our study, the detection rate of PCa with tPSA in the gray zone was only 18.6%. Similar detection rates were reported in other studies (1215). Even if PSA is not in the gray zone, for example, between 10 and 20 ng/ml, the detection rate in our study was only 26.2%. In recent years, DRE, PSAD, PSAV, 4Kscore, f/tPSA ratio, prostate health index (PHI) and age-specific PSA have been proposed as predictors for PCa (3, 1618). However, it may be difficult to achieve good predictive results with any single factor.

In this study, we constructed prediction models of PCa based on machine learning algorithms. Four algorithms were used, and a total of five models were constructed. First, we constructed a univariate logistic regression model using tPSA. As shown in Table 2, in the test dataset, although the specificity of the tPSA LR model reached 0.932, its sensitivity was decreased to only 0.639. The AUC of this model was only 0.846, which was significantly lower than that of the other models, showing the limitations of using tPSA alone as a predictor of PCa. Among them, the multivariate LR model had higher specificity, sensitivity, accuracy, and AUC in the test dataset, showing good predictive ability. Therefore, the shortcomings of low sensitivity and accuracy of the tPSA LR model were complemented very well by the inclusion of more variables. In our study, the multivariate LR model also had outstanding extrapolation and generalization ability due to the small number of changes between the training and test datasets. As a model similar to traditional statistical analysis methods, the results of the LR model had strong interpretability, which could help clinicians predict PCa based on relevant factors. The output of the DT model was similar to the clinical pathway. It is a clinician- and patient-friendly model and has strong clinical operability. Anyone can follow the predicted model from the root node to the leaf node to make decisions. However, in our study, the performance of the constructed DT model decreased significantly when it was validated in the test dataset and had very low specificity. The RF and SVM models had average diagnostic performance in our study, and their “black box” style reduced the clinical interpretability slightly.

The PCa detection rate in patients with PSA<20 ng/ml was relatively low in our study. Thus it is important to predict PCa in this population. The RF model performed best in both subgroups PSA of 4-10 ng/ml and 4-20 ng/ml. This suggested that the machine learning models we constructed based on overall population might also be applicable in patients with PSA ranged 4-10 or 4-20 ng/ml. In the training dataset, the RF model outperformed other models with the AUC of 1.0, while its performance decreased significantly in the test dataset. But in these two PSA subgroups, about 72% of samples overlapped with the samples in the training dataset, and were involved in the construction of models. That may lead to the outperformance of RF model rather than other models in PSA subgroups. Future study should focus on this population and develop more accurate machine learning models.

In recent years, some studies on the prediction of PCa by machine learning models have been published. In the study of Peter Ka-Fung Chiu et al., four variables, PSA, DRE, PV and transrectal ultrasound findings, were included, and SVM, LR, and RF models were constructed. All models were shown to have better prediction for PCa and clinically significant PCa than PSA and PSAD alone (19). Similarly, Nitta et al. suggested that compared to the AUCs of the PSA level, PSAD, and PSAV alone, the AUCs of artificial neural network (ANN), RF, SVM machine learning models were all improved when age, PSA level, PV, and white blood cell count in urinalysis were incorporated (20).In a study including patients with tPSA<10 ng/ml, a PSA-based machine learning model was constructed based on dense neural network with an AUC of 0.72, which was improved compared to PSA alone, age, fPSA and f/tPSA alone (21). In another study, multiparametric MRI (mpMRI) combined with other characteristics of patients was included to construct a machine learning model. The SVM and RF yielded similar diagnostic accuracy and net benefit and spared more biopsies at 95% sensitivity for the detection of clinically significant PCa compared with logistic regression (15).

Recently, a meta-analysis showed that the performance of selectMDx test in urine was comparable to that of mpMRI with regards to PCa detection. The AUC of selectMDx only was 0.854, and the AUC of one or both positive finding with selectMDx and/or mpMRI could reach 0.909. However, the multivariate LR model in our study still shows strength with AUC of 0.918 in test dataset (22). The integrative machine learning model was constructed to predict negative prostate biopsy utilizing both radiomics and clinical features. Although that model got high performance with negative predictive value of 98.3%, the AUC, sensitivity, and specificity were 0.798, 83.3%, and 75.2%, respectively, which were relatively lower compared with those in our models (23). Considering that all variables in our models are objective indicators, reducing possible errors of manual evaluation, we believe that the advantages of our models are more obvious.

Although machine learning-based models for PCa prediction have been constructed and validated, there are several limitations in our study. First, our study is retrospective in nature and may be potentially biased, and the sample size may not be adequate for some machine learning algorithms. Second, both the training and test datasets were from the same hospital, so further external validation at other centers is needed to confirm the findings. Third, some important factors or variables were not included in our models. For example, it has been shown that mpMRI provides more imaging information than conventional ultrasound, not only improving the detection of PCa but also helping to distinguish clinically significant PCa (24). Limited by insufficient data, mpMRI and its PI-RADS data were not included in our study. It is hoped that incorporating mpMRI into machine learning models may help to further improve the diagnostic performance of models in the future.

Conclusion

In conclusion, by retrieving electronic medical records, we developed, validated, and compared machine learning models to predict PCa in the biopsy population. All models showed clinical benefits based on DCA. Multivariate LR, DT, RF, and SVM models were better than tPSA univariate LR. Among these models, multivariate LR performed best, with an AUC of 0.918 in the test dataset. Constructing machine learning-based models and predicting PCa is feasible. This could enhance the detection of PCa and help to avoid unnecessary prostate biopsy.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author Contributions

SC: Data analysis, Manuscript writing/editing. TJ: Data collection. CC: Data collection. YL: Data analysis. XL: Data analysis. YY: Data collection. FJ: Project development, supervision. JL: Project development, supervision, manuscript writing/editing. All authors contributed to the article and approved the submitted version.

Funding

This study was funded by Science and Technology Development Project of Jilin Province, China (20200201315JC).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.941349/full#supplementary-material

References

1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global Cancer Statistics 2020: Globocan Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA: Cancer J Clin (2021) 71(3):209–49. doi: 10.3322/caac.21660

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Eldred-Evans D, Tam H, Sokhi H, Padhani AR, Winkler M, Ahmed HU. Rethinking Prostate Cancer Screening: Could Mri Be an Alternative Screening Test? Nat Rev Urol (2020) 17(9):526–39. doi: 10.1038/s41585-020-0356-2

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Catalona WJ. Prostate Cancer Screening. Med Clinics North America (2018) 102(2):199–214. doi: 10.1016/j.mcna.2017.11.001

CrossRef Full Text | Google Scholar

4. Tsodikov A, Gulati R, Heijnsdijk EAM, Pinsky PF, Moss SM, Qiu S, et al. Reconciling the Effects of Screening on Prostate Cancer Mortality in the Erspc and Plco Trials. Ann Internal Med (2017) 167(7):449–55. doi: 10.7326/m16-2586

CrossRef Full Text | Google Scholar

5. Fenton JJ, Weyrich MS, Durbin S, Liu Y, Bang H, Melnikow J. Prostate-Specific Antigen-Based Screening for Prostate Cancer: Evidence Report and Systematic Review for the Us Preventive Services Task Force. Jama (2018) 319(18):1914–31. doi: 10.1001/jama.2018.3712

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Tan GH, Nason G, Ajib K, Woon DTS, Herrera-Caceres J, Alhunaidi O, et al. Smarter Screening for Prostate Cancer. World J Urol (2019) 37(6):991–9. doi: 10.1007/s00345-019-02719-5

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Greener JG, Kandathil SM, Moffat L, Jones DT. A Guide to Machine Learning for Biologists. Nat Rev Mol Cell Biol (2022) 23(1):40–55. doi: 10.1038/s41580-021-00407-0

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Camacho DM, Collins KM, Powers RK, Costello JC, Collins JJ. Next-Generation Machine Learning for Biological Networks. Cell (2018) 173(7):1581–92. doi: 10.1016/j.cell.2018.05.015

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Schwalbe N, Wahl B. Artificial Intelligence and the Future of Global Health. Lancet (2020) 395(10236):1579–86. doi: 10.1016/s0140-6736(20)30226-9

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Matlaga BR, Eskew LA, McCullough DL. Prostate Biopsy: Indications and Technique. J Urol (2003) 169(1):12–9. doi: 10.1097/01.ju.0000041828.84343.53

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Catalona WJ, Smith DS, Ratliff TL, Dodds KM, Coplen DE, Yuan JJ, et al. Measurement of Prostate-Specific Antigen in Serum as a Screening Test for Prostate Cancer. New Engl J Med (1991) 324(17):1156–61. doi: 10.1056/nejm199104253241702

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Lee A, Chia SJ. Contemporary Outcomes in the Detection Of Prostate Cancer Using Transrectal Ultrasound-Guided 12-Core Biopsy in Singaporean Men With Elevated Prostate Specific Antigen and/or Abnormal Digital Rectal Examination. Asian J Urol (2015) 2(4):187–93. doi: 10.1016/j.ajur.2015.08.003

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Seo HK, Chung MK, Ryu SB, Lee KH. Detection Rate of Prostate Cancer According to Prostate-Specific Antigen and Digital Rectal Examination in Korean Men: A Nationwide Multicenter Study. Urology (2007) 70(6):1109–12. doi: 10.1016/j.urology.2007.07.052

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Matsumoto K, Satoh T, Egawa S, Shimura S, Kuwao S, Baba S. Efficacy and Morbidity of Transrectal Ultrasound-Guided 12-Core Biopsy for Detection of Prostate Cancer in Japanese Men. Int J Urol (2005) 12(4):353–60. doi: 10.1111/j.1442-2042.2005.01058.x

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Yu S, Tao J, Dong B, Fan Y, Du H, Deng H, et al. Development and Head-To-Head Comparison of Machine-Learning Models to Identify Patients Requiring Prostate Biopsy. BMC Urol (2021) 21(1):80. doi: 10.1186/s12894-021-00849-w

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Moradi A, Srinivasan S, Clements J, Batra J. Beyond the Biomarker Role: Prostate-Specific Antigen (Psa) in the Prostate Cancer Microenvironment. Cancer Metastasis Rev (2019) 38(3):333–46. doi: 10.1007/s10555-019-09815-3

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Catalona WJ, Partin AW, Slawin KM, Brawer MK, Flanigan RC, Patel A, et al. Use of the Percentage of Free Prostate-Specific Antigen to Enhance Differentiation of Prostate Cancer From Benign Prostatic Disease: A Prospective Multicenter Clinical Trial. Jama (1998) 279(19):1542–7. doi: 10.1001/jama.279.19.1542

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Carter HB, Ferrucci L, Kettermann A, Landis P, Wright EJ, Epstein JI, et al. Detection of Life-Threatening Prostate Cancer With Prostate-Specific Antigen Velocity During a Window of Curability. J Natl Cancer Inst (2006) 98(21):1521–7. doi: 10.1093/jnci/djj410

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Chiu PK, Shen X, Wang G, Ho CL, Leung CH, Ng CF, et al. Enhancement of Prostate Cancer Diagnosis by Machine Learning Techniques: An Algorithm Development and Validation Study. Prostate Cancer Prostatic Dis (2021). doi: 10.1038/s41391-021-00429-x

CrossRef Full Text | Google Scholar

20. Nitta S, Tsutsumi M, Sakka S, Endo T, Hashimoto K, Hasegawa M, et al. Machine Learning Methods Can More Efficiently Predict Prostate Cancer Compared With Prostate-Specific Antigen Density and Prostate-Specific Antigen Velocity. Prostate Int (2019) 7(3):114–8. doi: 10.1016/j.prnil.2019.01.001

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Perera M, Mirchandani R, Papa N, Breemer G, Effeindzourou A, Smith L, et al. Psa-Based Machine Learning Model Improves Prostate Cancer Risk Stratification in a Screening Population. World J Urol (2021) 39(6):1897–902. doi: 10.1007/s00345-020-03392-9

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Sari Motlagh R, Yanagisawa T, Kawada T, Laukhtina E, Rajwa P, Aydh A, et al. Accuracy of Selectmdx Compared to Mpmri in the Diagnosis of Prostate Cancer: A Systematic Review and Diagnostic Meta-Analysis. Prostate Cancer Prostatic Dis (2022) 25(2):187–98. doi: 10.1038/s41391-022-00538-1

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Zheng H, Miao Q, Liu Y, Raman SS, Scalzo F, Sung K. Integrative Machine Learning Prediction of Prostate Biopsy Results From Negative Multiparametric Mri. J Magn Reson Imaging (2022) 55(1):100–10. doi: 10.1002/jmri.27793

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Stabile A, Giganti F, Rosenkrantz AB, Taneja SS, Villeirs G, Gill IS, et al. Multiparametric Mri for Prostate Cancer Diagnosis: Current Status and Future Directions. Nat Rev Urol (2020) 17(1):41–61. doi: 10.1038/s41585-019-0212-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: prostate cancer, machine learning, prostate-specific antigen, prostate biopsy, prediction models

Citation: Chen S, Jian T, Chi C, Liang Y, Liang X, Yu Y, Jiang F and Lu J (2022) Machine Learning-Based Models Enhance the Prediction of Prostate Cancer. Front. Oncol. 12:941349. doi: 10.3389/fonc.2022.941349

Received: 11 May 2022; Accepted: 13 June 2022;
Published: 06 July 2022.

Edited by:

Jian Lu, Peking University Third Hospital, China

Reviewed by:

Clara Cerrato, University of California, San Diego, United States
Yishuo Wu, Fudan University, China

Copyright © 2022 Chen, Jian, Chi, Liang, Liang, Yu, Jiang and Lu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Ji Lu, lu_ji@jlu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.