Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 03 May 2019
Sec. Medicine and Public Health

Pancreatic Cancer Prediction Through an Artificial Neural Network

  • 1Department of Therapeutic Radiology, School of Medicine, Yale University, New Haven, CT, United States
  • 2Department of Internal Medicine, School of Medicine, Yale University, New Haven, CT, United States

Early detection of pancreatic cancer is challenging because cancer-specific symptoms occur only at an advanced stage, and a reliable screening tool to identify high-risk patients is lacking. To address this challenge, an artificial neural network (ANN) was developed, trained, and tested using the health data of 800,114 respondents captured in the National Health Interview Survey (NHIS) and Pancreatic, Lung, Colorectal, and Ovarian cancer (PLCO) datasets, together containing 898 patients diagnosed with pancreatic cancer. Prediction of pancreatic cancer risk was assessed at an individual level by incorporating 18 features into the neural network. The established ANN model achieved a sensitivity of 87.3 and 80.7%, a specificity of 80.8 and 80.7%, and an area under the receiver operating characteristic curve of 0.86 and 0.85 for the training and testing cohorts, respectively. These results indicate that our ANN can be used to predict pancreatic cancer risk with high discriminatory power and may provide a novel approach to identify patients at higher risk for pancreatic cancer who may benefit from more tailored screening and intervention.

Introduction

Pancreatic cancer (PC) remains the fourth leading cause of cancer-related death in both men and women in the United States (Klein et al., 2013; American Cancer Society, 2017) despite its low incidence rate (Pannala et al., 2009). In 2017, a total of 53,670 new PC cases (3.18% of all new cancer cases) and a total of 43,093 associated deaths (7.17% of all cancer deaths) were recorded in the United States (American Cancer Society, 2017). The age-adjusted cancer-related death rate is increasing for PC, and it is predicted that PC will become the second most common cause of cancer-related deaths by 2030(Klein et al., 2013; Boursi et al., 2017). PC has a high mortality rate in part because cancer-specific symptoms in most patients (>80%) occur only at an advanced stage (Pannala et al., 2009; Klein et al., 2013,; Boursi et al., 2017).

According to the 2017 American Cancer Society (ACS) statistics, the recent 5-years survival rate for all stages of PC is 8.5% (American Cancer Society, 2017). The 5-years survival rates for patients with early-stage diagnosis can be as high as 20% (Winter et al., 2006; Howlader, 2011; Klein et al., 2013). However, only a small portion of patients (<15%) have surgically resectable disease at the time of diagnosis (Pannala et al., 2009). Furthermore, identification of individuals at high risk for PC or with early-stage disease is difficult due to the lack of a reliable screening tools, the absence of sensitive and specific biomarkers, and the low prevalence (Pannala et al., 2009; Yu et al., 2016; Boursi et al., 2017).

Recently, numerous studies have been focused on early detection of PC through the identification and validation of promising biomarkers (Grønborg et al., 2004; Gold et al., 2010; Klein et al., 2013). Further, the ability to detect pre-cancerous changes in the pancreas among high-risk individuals via Doppler ultrasound (US), endoscopic ultrasound (EUS), magnetic resonance imaging (MRI), computed tomography (CT) scan, or positron emission tomography (PET) has also been demonstrated in several clinical studies (Canto et al., 2004, 2006; Poley et al., 2009; Verna et al., 2010; Klein et al., 2013). Pancreatic tumors as small as 0.5 cm can be identified with diagnostic imaging, such as CT, MRI, or EUS. However, despite the high sensitivity of these techniques (Klein et al., 2013; Boursi et al., 2017), it is not practical or economically feasible to perform widespread PC screening in the general population due to the relatively low incidence rate (Klein et al., 2013; Boursi et al., 2017). However, these techniques can be used more efficiently and cost-effectively if employed in a high-risk subset of the population. For example, screening protocols are applied in patients with germline mutations associated with PC and patients with familial PC (Boursi et al., 2017). However, only 10–20% of all PC cases can be attributed to familial PC (Boursi et al., 2017).

Various epidemiologic and clinical characteristics are associated with occurrence of PC, including family history of PC (Permuth-Wey and Egan, 2009), inherited genetic variation/influence (Lichtenstein et al., 2000; Klein et al., 2013), anthropometric variables [e.g., body mass index (BMI)] (Pannala et al., 2008; Arslan et al., 2010; Hart et al., 2011; Klein et al., 2013; Association, 2014), lifestyle (e.g., smoking, drinking alcohol) (Iodice et al., 2008; Michaud et al., 2010; Lucenteforte et al., 2011; Klein et al., 2013), and medical comorbidities (e.g., pancreatitis, diabetes) (Lowenfels et al., 1993; Pannala et al., 2009; Ben et al., 2011; Klein et al., 2013; Boursi et al., 2017). New onset diabetes is considered one of the strongest predictors of PC, and numerous epidemiologic studies reported that the association between newly diagnosed PC and diabetes mellitus was ~50% (American Cancer Society, 2017; Boursi et al., 2017). Chari et al. (2005) showed that the 3-years cumulative incidence of PC among patients with new onset diabetes is 8 times higher than expected (Boursi et al., 2017). Hence, it can be stated that diabetes associated with PC may be a paraneoplastic phenomenon caused by the cancer (Pelaez-Luna et al., 2007; Sah et al., 2013; Boursi et al., 2017). Smoking also increases the risk of PC by a factor of two (American Cancer Society, 2017). Even the use of smokeless tobacco increases PC risk (American Cancer Society, 2017). Family history of PC is also considered a risk factor (American Cancer Society, 2017).

To our knowledge, no established screening strategy has been introduced for sporadic PC. The non-invasive precursor lesions known as pancreatic intraepithelial neoplasia (PanIN) progress from PanIN1 to PanIN3 and into PC within an undefined timeline (Hruban et al., 2000; Pannala et al., 2009; Yu et al., 2016). Brat et al. (1998) reported the presence of PanINs 1.4–10 years before the appearance of PC clinically. In another study, 114 CT scans in 45 patients (done either at or before PC diagnosis) were reviewed to estimate the timeline for progression of PC (Pannala et al., 2009). Multiple studies indicate that the radiographic features of unresectability and the onset of symptoms of the cancer appeared simultaneously (Gangi et al., 2004; Pelaez-Luna et al., 2007; Pannala et al., 2009). Pannala et al. (2009) stated that PC remains resectable when asymptomatic and thus is unlikely to be detected. It is estimated that symptoms manifest about 6 months after PC becomes unresectable (Pannala et al., 2009). Therefore, identifying those at high risk yet asymptomatic is very important to find PC while it is still resectable.

The artificial neural network (ANN), which is based on the brain's neural structure (Rosenblatt, 1958), raised the interest of scientific community worldwide in the field of medicine due to its potential for diagnostic and prognostic applications (Smith et al., 1988; Salim, 2004; Kamruzzaman et al., 2010; Patil and Mudholkar, 2012). It has been used in heart disease (Kamruzzaman et al., 2010), predicting headache, pre-diagnosis of hypertension (Sumathi and Santhakumaran, 2011), kidney stone diseases (Kumar and Abhishek, 2012), classifying breast masses to identify breast cancer (Das and Bhattacharya, 2008; Pandey et al., 2012), dermatologist-level classification of skin diseases/cancer (Bakpo and Kabari, 2011; Esteva et al., 2017), prediction of skin cancer and blood cancer (Payandeh et al., 2009; Esteva et al., 2017; Roffman et al., 2018a), and diagnosis of PC (Sanoob et al., 2016). As an example of the workflow in these applications, classification of skin cancer was performed via a single convolutional neural network, which was trained with a dataset of 129,450 clinical images (Esteva et al., 2017). In another study, an ANN model was created to diagnose PC based on a data set of symptoms (Sanoob et al., 2016). A total sample of 120 patients (i.e., 90 training samples and 30 testing samples) with 11 possible symptoms and 3 outcomes were considered for this model (Sanoob et al., 2016). The authors claimed that the ANN model has advantages over typical strategies for disease diagnosis (Sanoob et al., 2016).

Roffman et al. (2018a) took a novel approach to predict non-melanoma skin cancer by using personal health data (e.g., gender, race, Hispanic ethnicity, hypertension, heart disease, exercise habits, history of stroke, etc.) commonly available in electronic medical record (EMR) systems. The area under the conventional receiver operating characteristic (ROC) curve was 0.81 and 0.81 for training and validation, respectively (Roffman et al., 2018a). This study suggests that the ANN can be a convenient and cost-effective method in evaluating cancer risk for individuals (Roffman et al., 2018a). Likewise, the goal of this study is to develop an ANN to calculate risk for PC in the general population and to identify a high-risk population in a cost-effective manner by utilizing easily available personal health data.

Materials and Methods

Two Data Sources

The National Health Interview Survey (NHIS) (Blewett et al., 2017) was established in 1957 to monitor the overall health status of the United States through personal household interviews on a broad range of health topics. Numerous epidemiologic studies have been conducted using NHIS (Blewett et al., 2017; Roffman et al., 2018a). The NHIS datasets of 1997 to 2017 (Blewett et al., 2017) were used in this study. The target study population consisted of people with onset of pancreatic cancer <4 years prior to the survey date. Considering the time dependency of input features to the model, this 4-years cutoff on the pancreatic cancer group was selected after careful testing of different cutoffs on model performance to strike a balance between sample size and the predictive power of our model. After applying this cutoff, we have 645,217 respondents, 131 of whom had PC.

The Prostate, Lung, Colorectal, and Ovarian (PLCO) trial (NCI, 2018) is a randomized, controlled trial investigating whether certain screening exams reduce mortality from prostate, lung, colorectal and ovarian cancer. Between November 1993 and July 2001, 154,897 participants were enrolled, 767 of whom developed PC during 13 years of follow up. For this study, PC status, personal health data, family history, socio-behavior, lifestyle and dietary data have been extracted from PLCO datasets via an in-house Matlab code.

Primary Outcome

The primary outcome of interest includes (1) the accuracy of model prediction for PC; and (2) the feasibility of individualized cancer risk stratification for tailored intervention.

Predictors

A total of 18 personal health features were selected for use in the ANN for PC risk prediction based on literature review, biological plausibility, and clinical judgment. The details of these personal health features are given in Table 1. Some features are converted to binary format [one-hot encoding (Harris and Harris, 2014)] and the others are rescaled to fall between 0 and 1 (Roffman et al., 2018a). All these features were available in the NHIS dataset and most of them were also in the PLCO dataset.

TABLE 1
www.frontiersin.org

Table 1. Description of the personal health features from NHIS and PLCO datasets used in the ANN.

Sample Size Considerations

All the data in the NHIS dataset from 1997 to 2017 and PLCO dataset were used to maximize the power and generalizability of the results. To investigate the performance of ANN on different datasets, three datasets were built:

1. DS1 = NHIS dataset (645,217 participants, with 131 PC cases)

2. DS2 = PLCO dataset (154,897 participants with 767 PC cases)

3. DS3 = NHIS dataset + PLCO dataset (800,114 participants with 898 PC cases)

After constructing and randomizing these three datasets, we used a train/validate/test scheme. The ANN was trained on 70% (training dataset) of the data using 10-fold cross-validation, while the remaining 30% was withheld for further testing (testing dataset). Cancer risk, sensitivity, and specificity were calculated for both training and testing datasets.

Missing Data

Some entries for some respondents were missing because they did not respond, or the question was not applicable. The details of these missing data are given in Table 1. To address these missing data, we used the idea of one-hot encoding (Harris and Harris, 2014). Essentially, for each feature we create a binary variable indicating whether a respondent has a value for that feature. Then the missing value is set to −1, outside of the range of the “real” data.

Statistical Analysis

Given the binary outcome, we developed our prediction model using the logistic activation function. The model was developed, and all analyses were performed using an in-house Matlab code.

Artificial Neural Network (ANN)

In our group, besides PC, we have also investigated a variety of other cancer types, such as lung cancer (Hart et al., 2018), prostate cancer (Roffman et al., 2018b), endometrial cancer (Hart et al., 2019), and colorectal cancer (Nartowt et al., 2019a,b) using ANN, Support Vector Machine, Decision Tree, Naive Bayes, Linear Discriminant Analysis, and Logistic Regression. Our results indicated that in general, ANN achieves the best performance as compared to other algorithms in terms of sensitivity, specificity, and AUC. Therefore, we used ANN in the present work. A schematic of an ANN model is shown in Figure 1. Our ANN had, in addition to the input and output layers, two hidden layers (each consisting of 12 neurons). The input features (between 0 to 1) and output (0 or 1) were split into 70/30 for training and testing datasets while keeping the ratio of the number of cancer cases to non-cancer cases constant. Within the training dataset, 10-fold stratified cross validation was used to evaluate the performance of models trained on the different datasets. Once the best model was chosen, we trained it on the full training dataset and then evaluated it on the test dataset. We used a logistic activation function and the sum of squared errors cost function. We trained our model using the standard backpropagation algorithm with simple gradient descent (http://ufldl.stanford.edu/tutorial/supervised/MultiLayerNeuralNetworks/), except that we used momentum to speed up the convergence. We batch trained our model (using the whole dataset at once) instead of online training (Roffman et al., 2018a). We ran the training for 5,000 iterations. The output of the ANN is a fractional number between 0 and 1. A higher output value means higher risk of PC. This fractional value can be transformed into cancer status (Yes or No) by choosing a threshold value above which the ANN will give a positive prediction for the cancer status (YES) or otherwise a “NO” for non-cancer. A variety of threshold values are tested to compute sensitivity and specificity after completion of the training. The selected threshold value from the training dataset is used to compute the sensitivity and specificity for the testing set.

FIGURE 1
www.frontiersin.org

Figure 1. A schematic of an ANN. Each box and line represents a neuron and a weight, respectively. The number of weights grows rapidly with the number of neurons.

Model Performance Evaluation

The models trained on different datasets were evaluated based on the mean of the performance on the validation datasets. Specifically, we used the area under the ROC curve (AUC) as the measure of performance. This was chosen because in order to stratify the population into risk groups we want to have good discrimination (Metz, 1978).

Once the best model was selected, its performance on both the training and testing datasets was evaluated, testing the ability of the risk score to differentiate between the individuals with onset of PC and non-PC individuals. In addition to the AUC, the agreement between the predicted probabilities from the model and the observed outcomes are reflected from the training of the model.

Risk Stratification

A risk stratification scheme was tested to demonstrate the potential application of our ANN model in the clinic. The scheme was designed to divide the population into three categories: low, medium, and high risk. These boundaries were conservatively selected using the training dataset, such that no more than 1% of respondents without cancer and with cancer would be categorized as high and low risk, respectively. However, the medium-high risk boundary could be selected to stratify more respondents with cancer in the high-risk category in case of low cost and/or potential harms in screening non-cancerous respondents. With these boundaries selected from the training data, the stratification scheme is then applied to the testing dataset to demonstrate the potential clinical application of the model. Per this risk stratification scheme, high-risk individuals could be screened immediately. The medium-risk and low-risk individuals could receive their standard regular and less frequent screenings, respectively.

Results

Model Selection

The performance of the model was assessed by calculating the AUC of the ROC plots for all three datasets (i.e., DS1, DS2, and DS3). For DS1, the AUC of the ROC plot is 0.75 ± 0.06 for the training sets, and 0.71 ± 0.11 for the testing sets (Figure 2A), while for DS2, these values are 0.64 ± 0.01 for training and 0.62 ± 0.04 for testing (Figure 2B). Similarly, the AUCs for DS3 are 0.86 ± 0.01 and 0.85 ± 0.02 for the training and testing sets, respectively (see Figure 2C). The best performance of the model was observed for DS3.

FIGURE 2
www.frontiersin.org

Figure 2. Receiver operating characteristic (ROC) plots for the training and validation datasets of (A) DS1, (B) DS2, and (C) DS3.

Final Model Performance

Having selected the DS3 model, we train it on the full training dataset and evaluated it on the testing dataset. The sensitivity and specificity for both training and testing are plotted as functions of the threshold risk to study their trends (Figure 3A). Selecting the threshold risk that maximizes the sum of the sensitivity and specificity, we get specific values plotted in Figure 3B. The positive predictive value (PPV) and negative predictive value (NPV) are plotted as a function of threshold value shown in Figure 4. For the presented values of sensitivity and specificity of the DS3 training dataset, PPV and NPV values are 0.1% (95% confidence interval (CI): 0.09–0.100%) and 99.997% (95% CI: 99.996–99.997%), respectively. Similarly, for the DS3 testing dataset, 0.089% (95% CI: 0.084–0.095%) and 99.995% (95% CI: 99.993–99.996%) are PPV and NPV values, respectively for the presented values of sensitivity and specificity.

FIGURE 3
www.frontiersin.org

Figure 3. (A) Sensitivity and specificity from final training as functions of the cutoff values for DS3, and (B) representation of sensitivity and specificity values for DS3.

FIGURE 4
www.frontiersin.org

Figure 4. Positive predictive value (PPV) and negative predictive value (NPV) for training dataset of DS3.

Risk Stratification

Running through the DS3 dataset, the outputs of the ANN were categorized as low-, medium- and high-risk. The categorized fraction of the respondents with and without PC varied at different risk levels. It was clear from Figure 5 that most of non-cancer respondents were categorized in either low or medium risk while most of the respondents with cancers were either categorized as medium or high-risk. Risk stratification results for the testing datasets were summarized in Table 2.

FIGURE 5
www.frontiersin.org

Figure 5. Population with Cancer (blue) and without cancer (orange) as functions of high-risk (solid) and low-risk (dashed). Assuming a 1% miss classification rate in the low- and high-risk categories (black line), individuals can be divided into 3 categories according to their cancer risk: high (red), medium (yellow), and low (green) for DS3.

TABLE 2
www.frontiersin.org

Table 2. Risk Stratification for DS3 (NHIS and PLCO datasets combined).

Discussions

In this study, risk of PC is predicted and stratified based on basic personal health data (NHIS and PLCO datasets) using a multi-parameterized ANN model. The model performance was evaluated by training and testing it on different datasets to determine its optimum performance. The best performance of the model was observed for DS3 with an AUC of 0.86 [CI 0.85–0.86 ±1 standard deviation (SD)] and 0.85 (CI 0.85–0.87 ±1 SD) for training and testing, respectively (Figure 2C). The best observed values for sensitivity and specificity for the training (testing) datasets of DS3 are 87.3% (80.7%) and 80.8% (80.7%), respectively. In 2017, the number of new cases of PC was 12.6 per 100,000 men and women per year (American Cancer Society, 2017). With our NPV value from the testing dataset being 99.995%, when our model predicts someone does not have cancer it is only wrong 0.005% of the time (5 per 100,000). For the DS3 testing dataset our PPV value is 0.09% (90 per 100,000). The group our ANN flags as having cancer is enriched more than 7-fold over the general population.

Because of the low number of PC cases for NHIS datasets (DS1), the model overfit and did not perform very well which is evident from the standard deviation in the validation AUC. The model also did not perform well for DS2 because the PLCO data consists of an enriched population of high-risk individuals with a higher median age. Also, there were a number of input features (e.g., alcohol use) that were completely absent in the PLCO datasets. Therefore, the model lost diversity and predictive power and relatively lower AUC values were observed. By combining NHIS with PLCO datasets, AUC value increased to 0.85, indicating a significant improvement in the discriminatory power of the model.

Currently, contrast-enhanced US, EUS, MRI, CT, and PET are the most promising modalities for PC screening (Verna et al., 2010; Klein et al., 2013). Each of these techniques has its advantages and limitations in screening for PC, but these techniques are often applied after the appearance of symptoms, which may be fatally too late in most cases. However, our ANN is focused on the early prediction and stratification of PC risk before symptoms appear. The results show that without any screening tests, the ANN produced very good predictions for PC. By comparing our results with already established screening modalities (i.e., EUS and MRI), PC risk was estimated with a high sensitivity and decent specificity. We stress that only personal health data (the type that is readily available in the EMR system) was used to reach this level of sensitivity and specificity.

The ANN can also be used to categorize the general public into low, medium, or high risk for PC based on easily obtainable personal health data in NHIS format. Reliable identification of high-risk patients who may benefit from tailored screening may improve a probability to detect PC at early stages. According to our testing results for the model, only 3 (1.9%) of respondents with cancer are incorrectly classified as low-risk, while only 2,394 (1%) of respondents in the total stratified population without cancer are false-positively categorized as high-risk (Table 2). With an AUC of 0.85, our model can effectively discriminate between respondents with and without PC (Figure 2).

Recently, a clinical prediction model has been used to assess PC risk with pre-diabetic and new onset diabetic patients (Boursi et al., 2017, 2018). For pre-diabetic study, a total number of 138,232 patients with new onset impaired fasting glucose (IFG) were selected where 245 individuals were diagnosed with pancreatic ductal adenocarcinoma within 3 years of IFG diagnosis. The prediction model included age, BMI, PPIs, total cholesterol, LDL, ALT and alkaline phosphatase. The reported AUC of the model was 0.71 (95% CI 0.67–0.75) (Boursi et al., 2018). By analyzing 109,385 onset diabetic patients including 390 PC cases, their model produced AUC of 0.82 (95% CI, 0.75–0.89) (Boursi et al., 2017). However, a comprehensive list of PC risk factors (54 in total) were used, e.g., age, BMI, change in BMI, smoking, use of proton pump inhibitors, and anti-diabetic medications, as well as levels of hemoglobin A1C, cholesterol, hemoglobin, creatinine, and alkaline phosphatase. This set of data requires specialized equipment to collect and may not be reportable by all members of the general public. In contrast, our ANN works on personal health data that are easily reportable by the general public while maintaining an AUC of 0.85.

Cai et al. (2011) developed a PC risk stratification prediction rule by studying 138 patients with chronic pancreatitis. A scoring method based logistic regression was used to develop the prediction rule. Hsieh et al. (2018) predicted PC in the patients with type 2 diabetes using logistic regression and artificial neural network models. In another study, Wang et al. (2007) predicted familial PC risk through a Mendelian model (i.e., PancPRO) that was built by extending the Bayesian modeling framework. The AUCs achieved by these models were 0.72 (Cai et al., 2011), 0.73 (Hsieh et al., 2018), and 0.75 (Wang et al., 2007), respectively. With lower AUCs as compared to the current study and being designed for specific conditions, these studies may not be widely used for the general public. In another study, a weighted Bayesian network was used for prediction of PC by combining PubMed knowledge and electronic health record (EHR) data (Zhao and Weng, 2011). A total of 20 common risk factors (i.e., age, gender, smoking, and/or alcohol use, weight loss, vomiting, nausea, fatigue, appetite loss, jaundice, abdominal pain, diabetes, depression, AST, ALT, albumin, alkaline phosphatase, GGT, glucose, bilirubin, CEA, and CA 19-9) associated with PC were used with PubMed knowledge to weigh the risk factors. Their network produced an AUC of 0.91 (95% CI, 0.869–0.951). Although these results are promising, the weighting has been calculated separately for each risk factor. If more risk factors are added, the prediction results will be different due to added weightings from PubMed knowledge. Secondly, in these studies, most features are clinical and hence not readily available. Our ANN's weights were fit on the training dataset and if more risk factors are added, updating the weights to include the new factors can be done quickly by re-fitting the ANN.

Nakatochi et al. (2018) presented a PC risk prediction model in the general population in Japan with AUC of 0.63. However, their model was based on data including directly determined or imputed single nucleotide polymorphisms (SNPs) genotypes. While our ANN model performed considerably well to predict PC on the basis of commonly available data in the EMR, inclusion of personal high-risk features for PC (e.g., pancreatic cysts, family history etc.) could potentially improve the performance of the model. Our approach is also distinct from previous studies because it is based on survey data representative of the general population. The previous studies are based on either one or more clinical conditions or smaller sample sizes. Furthermore, the developed ANN may be very helpful to primary care physicians due to its ability to stratify people into various risk categories. Higher risk people could be referred to a diagnostic department for more tailored and intensive assessments. We envisage that this model can be integrated into an EMR system or be available on websites and portable devices, such as mobile phones and tablets. This will be very helpful for the clinicians to calculate the PC risk of their patients immediately after entering their data. More importantly, with the tool embedded in the clinical workflow, pancreatic cancer could be detected at an early stage, hence improving the survival rate in the long run.

Conclusion

We reported an ANN that can be used to predict pancreatic cancer with a sensitivity of 80.7%, a specificity of 80.7%, and an AUC of 0.85 based solely on personal health data. In addition, the developed ANN was able to stratify people into low, medium and high cancer risk for more tailored screening and risk management. Compared to current screening techniques, this ANN is non-invasive, cost-effective, and easy to implement with readily available personal health data. More data and testing would be needed to further improve the performance of the ANN in order to facilitate its application in the clinic.

Author Contributions

WM analyzed data, produced results, and wrote technical details. GH provided first version of working code. GH, BN, JF, KJ, and YL provided consultation, produced technical details and reviewed the manuscript. JD generated research ideas and reviewed manuscript.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

Research reported in this publication was supported by the National Institute of Biomedical Imaging and Bioengineering of the National Institutes of Health under Award Number R01EB022589. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

References

American Cancer Society (2017). Cancer Facts & Figures 2017. Atlanta, GA: American Cancer Society.

Arslan, A. A., Helzlsouer, K. J., Kooperberg, C., Shu, X.-O., Steplowski, E., Bueno-De-Mesquita, H. B., et al. (2010). Anthropometric measures, body mass index, and pancreatic cancer: a pooled analysis from the Pancreatic Cancer Cohort Consortium (PanScan). Arch. Intern. Med. 170, 791–802. doi: 10.1001/archinternmed.2010.63

PubMed Abstract | CrossRef Full Text | Google Scholar

Association, A. D. (2014). Diagnosis and classification of diabetes mellitus. Diabetes Care. 37, S81–S90. doi: 10.2337/dc10-S062

CrossRef Full Text | Google Scholar

Bakpo, F., and Kabari, L. (2011). “Diagnosing skin diseases using an artificial neural network,” in Artificial Neural Networks-Methodological Advances and Biomedical Applications, ed K. Suzuki (InTech), 253–270.

Google Scholar

Ben, Q., Xu, M., Ning, X., Liu, J., Hong, S., Huang, W., et al. (2011). Diabetes mellitus and risk of pancreatic cancer: a meta-analysis of cohort studies. Eur. J. Cancer. 47, 1928–1937. doi: 10.1016/j.ejca.2011.03.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Blewett, L. A., Rivera Drew, J. A., Griffin, R., King, M. L., and Williams, K. C. W. (2017). IPUMS Health Surveys: National Health Interview Survey, Version 6.2 [dataset]. Minneapolis, MN: University of Minnesota.

Boursi, B., Finkelman, B., Giantonio, B. J., Haynes, K., Rustgi, A. K., Rhim, A. D., et al. (2017). A clinical prediction model to assess risk for pancreatic cancer among patients with new-onset diabetes. Gastroenterology. 152, 840–850.e843. doi: 10.1053/j.gastro.2016.11.046

PubMed Abstract | CrossRef Full Text | Google Scholar

Boursi, S. B., Finkelman, B., Giantonio, B. J., Haynes, K., Rustgi, A. K., Rhim, A., et al. (2018). A clinical prediction model to assess risk for pancreatic cancer among patients with pre-diabetes. J. Clin. Oncol. 36(15_Suppl.). doi: 10.1200/JCO.2018.36.15_suppl.e16226

CrossRef Full Text | Google Scholar

Brat, D. J., Lillemoe, K. D., Yeo, C. J., Warfield, P. B., and Hruban, R. H. (1998). Progression of pancreatic intraductal neoplasias to infiltrating adenocarcinoma of the pancreas. Am. J. Surg. Pathol. 22, 163–169.

PubMed Abstract | Google Scholar

Cai, Q. C., Chen, Y., Xiao, Y., Zhu, W., Xu, Q. F., Zhong, L., et al. (2011). A prediction rule for estimating pancreatic cancer risk in chronic pancreatitis patients with focal pancreatic mass lesions with prior negative EUS-FNA cytology. Scand. J. Gastroenterol. 46, 464–470. doi: 10.3109/00365521.2010.539256

PubMed Abstract | CrossRef Full Text | Google Scholar

Canto, M. I., Goggins, M., Hruban, R. H., Petersen, G. M., Giardiello, F. M., Yeo, C., et al. (2006). Screening for early pancreatic neoplasia in high-risk individuals: a prospective controlled study. Clin. Gastroenterol. Hepatol. 4, 766–781. doi: 10.1016/j.cgh.2006.02.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Canto, M. I., Goggins, M., Yeo, C. J., Griffin, C., Axilbund, J. E., Brune, K., et al. (2004). Screening for pancreatic neoplasia in high-risk individuals: an EUS-based approach. Clin. Gastroenterol. Hepatol. 2, 606–621.

PubMed Abstract | Google Scholar

Chari, S. T., Leibson, C. L., Rabe, K. G., Ransom, J., De Andrade, M., and Petersen, G. M. (2005). Probability of pancreatic cancer following diabetes: a population-based study. Gastroenterology. 129, 504–511. doi: 10.1016/j.gastro.2005.05.007

PubMed Abstract | CrossRef Full Text | Google Scholar

Das, A., and Bhattacharya, M. (2008). “GA based neuro fuzzy techniques for breast cancer identification,” in Machine Vision and Image Processing Conference, 2008. IMVIP'08. International (IEEE), 136–141.

Google Scholar

Esteva, A., Kuprel, B., Novoa, R. A., Ko, J., Swetter, S. M., Blau, H. M., et al. (2017). Dermatologist-level classification of skin cancer with deep neural networks. Nature. 542:115. doi: 10.1038/nature21056

PubMed Abstract | CrossRef Full Text | Google Scholar

Gangi, S., Fletcher, J. G., Nathan, M. A., Christensen, J. A., Harmsen, W. S., Crownhart, B. S., et al. (2004). Time interval between abnormalities seen on CT and the clinical diagnosis of pancreatic cancer: retrospective review of CT scans obtained before diagnosis. Am. J. Roentgenol. 182, 897–903. doi: 10.2214/ajr.182.4.1820897

PubMed Abstract | CrossRef Full Text | Google Scholar

Gold, D. V., Goggins, M., Modrak, D. E., Newsome, G., Liu, M., Shi, C., et al. (2010). Detection of early-stage pancreatic adenocarcinoma. Cancer Epidemiol. Prev. Biomarkers. 19, 2786–2794. doi: 10.1158/1055-9965.EPI-10-0667

PubMed Abstract | CrossRef Full Text | Google Scholar

Grønborg, M., Bunkenborg, J., Kristiansen, T. Z., Jensen, O. N., Yeo, C. J., Hruban, R. H., et al. (2004). Comprehensive proteomic analysis of human pancreatic juice. J. Proteome Res. 3, 1042–1055. doi: 10.1021/pr0499085

PubMed Abstract | CrossRef Full Text | Google Scholar

Harris, D., and Harris, S. L. (2014). Digital Design and Computer Architecture. Waltham, MA: Morgan Kaufmann.

Google Scholar

Hart, G. R., Nartowt, B. J., Muhammad, W., Liang, Y., Huang, G. S., and Deng, J. (2019), Endometrial cancer risk prediction stratification using personal health data (to be submitted).

Hart, G. R., Roffman, D. A., Decker, R., and Deng, J. (2018). A multi-parameterized artificial neural network for lung cancer risk prediction. PLoS ONE. 13:e0205264. doi: 10.1371/journal.pone.0205264

PubMed Abstract | CrossRef Full Text | Google Scholar

Hart, P. A., Kamada, P., Rabe, K. G., Srinivasan, S., Basu, A., Aggarwal, G., et al. (2011). Weight loss precedes cancer specific symptoms in pancreatic cancer associated diabetes mellitus. Pancreas. 40:768. doi: 10.1097/MPA.0b013e318220816a

PubMed Abstract | CrossRef Full Text | Google Scholar

Howlader, N. (2011). SEER Cancer Statistics Review, 1975–2008. Available online at: http://seer.cancer.gov/csr/1975_2008/ (accessed September 15, 2018).

Hruban, R. H., Goggins, M., Parsons, J., and Kern, S. E. (2000). Progression model for pancreatic cancer. Clin. Cancer Res. 6, 2969–2972.

PubMed Abstract | Google Scholar

Hsieh, M. H., Sun, L. M., Lin, C. L., Hsieh, M. J., Hsu, C. Y., and Kao, C. H. (2018). Development of a prediction model for pancreatic cancer in patients with type 2 diabetes using logistic regression and artificial neural network models. Cancer Manage. Res. 10:6317–6324. doi: 10.2147/CMAR.S180791

PubMed Abstract | CrossRef Full Text | Google Scholar

Iodice, S., Gandini, S., Maisonneuve, P., and Lowenfels, A. B. (2008). Tobacco and the risk of pancreatic cancer: a review and meta-analysis. Langenbecks Arch. Surg. 393, 535–545. doi: 10.1007/s00423-007-0266-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Kamruzzaman, S., Hasan, A. R., Siddiquee, A. B., Mazumder, M., and Hoque, E. (2010). Medical diagnosis using neural network. arXiv preprint arXiv:1009.4572.

Klein, A. P., Lindström, S., Mendelsohn, J. B., Steplowski, E., Arslan, A. A., Bueno-De-Mesquita, H. B., et al. (2013). An absolute risk model to identify individuals at elevated risk for pancreatic cancer in the general population. PLoS ONE. 8:e72311. doi: 10.1371/journal.pone.0072311

PubMed Abstract | CrossRef Full Text | Google Scholar

Kumar, K., and Abhishek, B. (2012). Artificial neural networks for diagnosis of kidney stones disease. I. J. Infor. Technol. Comput. Sci. 7, 20–25. doi: 10.5815/ijitcs.2012.07.03

CrossRef Full Text | Google Scholar

Lichtenstein, P., Holm, N. V., Verkasalo, P. K., Iliadou, A., Kaprio, J., Koskenvuo, M., et al. (2000). Environmental and heritable factors in the causation of cancer—analyses of cohorts of twins from Sweden, Denmark, and Finland. N. Engl. J. Med. 343, 78–85. doi: 10.1056/NEJM200007133430201

PubMed Abstract | CrossRef Full Text | Google Scholar

Lowenfels, A. B., Maisonneuve, P., Cavallini, G., Ammann, R. W., Lankisch, P. G., Andersen, J. R., et al. (1993). Pancreatitis and the risk of pancreatic cancer. N. Engl. J. Med. 328, 1433–1437.

PubMed Abstract | Google Scholar

Lucenteforte, E., La Vecchia, C., Silverman, D., Petersen, G., Bracci, P., Ji, B. A., et al. (2011). Alcohol consumption and pancreatic cancer: a pooled analysis in the International Pancreatic Cancer Case–Control Consortium (PanC4). Ann. Oncol. 23, 374–382. doi: 10.1093/annonc/mdr120

PubMed Abstract | CrossRef Full Text | Google Scholar

Metz, C. E. (1978). Basic principles of ROC analysis. Semin. Nucl. Med. 88, 283–298.

Michaud, D. S., Vrieling, A., Jiao, L., Mendelsohn, J. B., Steplowski, E., Lynch, S. M., et al. (2010). Alcohol intake and pancreatic cancer: a pooled analysis from the pancreatic cancer cohort consortium (PanScan). Cancer Causes Control. 21, 1213–1225. doi: 10.1007/s10552-010-9548-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Nakatochi, M., Lin, Y., Ito, H., Hara, K., Kinoshita, F., Kobayashi, Y., et al. (2018). Prediction model for pancreatic cancer risk in the general Japanese population. PLoS ONE. 13:e0203386. doi: 10.1371/journal.pone.0203386

PubMed Abstract | CrossRef Full Text | Google Scholar

Nartowt, B. J. G. R. H, Muhammad, W., Liang, Y., and Deng, J. (2019a). Supervised machine learning algorithms in scoring colorectal cancer risk in a cross-sectional study and in a longitudinal study-An externally-validated neural network model (under review).

Google Scholar

Nartowt, B. J., Hart, G. R., Muhammad, W., Liang, Y., and Deng, J. (2019b). Robust machine learning for colorectal cancer risk prediction and stratification (to be submitted).

NCI (2018). Cancer Data Access System (CDAS): Prostate, Lung, Colorectal and Ovarian (PLCO).

Pandey, B., Jain, T., Kothari, V., and Grover, T. (2012). Evolutionary modular neural network approach for breast cancer diagnosis. Int. J. Comp. Sci. Issues. 9, 219–225.

Google Scholar

Pannala, R., Basu, A., Petersen, G. M., and Chari, S. T. (2009). New-onset diabetes: a potential clue to the early diagnosis of pancreatic cancer. Lancet Oncol. 10, 88–95. doi: 10.1016/S1470-2045(08)70337-1

PubMed Abstract | CrossRef Full Text | Google Scholar

Pannala, R., Leirness, J. B., Bamlet, W. R., Basu, A., Petersen, G. M., and Chari, S. T. (2008). Prevalence and clinical profile of pancreatic cancer–associated diabetes mellitus. Gastroenterology. 134, 981–987. doi: 10.1053/j.gastro.2008.01.039

PubMed Abstract | CrossRef Full Text | Google Scholar

Patil, S. M., and Mudholkar, R. (2012). An osteoarthritis classifier using back propagation neural network. Int. J. Adv. Eng. Technol. 4:292.

Google Scholar

Payandeh, M., Aeinfar, M., Aeinfar, V., and Hayati, M. (2009). A new method for diagnosis and predicting blood disorder and cancer using artificial intelligence (artificial neural networks). Int. J. Hematol. Oncol. Stem Cell Res. 3, 25–33.

Google Scholar

Pelaez-Luna, M., Takahashi, N., Fletcher, J. G., and Chari, S. T. (2007). Resectability of presymptomatic pancreatic cancer and its relationship to onset of diabetes: a retrospective review of CT scans and fasting glucose values prior to diagnosis. Am. J. Gastroenterol. 102:2157. doi: 10.1111/j.1572-0241.2007.01480.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Permuth-Wey, J., and Egan, K. M. (2009). Family history is a significant risk factor for pancreatic cancer: results from a systematic review and meta-analysis. Fam. Cancer. 8, 109–117. doi: 10.1007/s10689-008-9214-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Poley, J. W., Kluijt, I., Gouma, D. J., Harinck, F., Wagner, A., Aalfs, C., et al. (2009). The yield of first-time endoscopic ultrasonography in screening individuals at a high risk of developing pancreatic cancer. Am. J. Gastroenterol. 104:2175–2181. doi: 10.1038/ajg.2009.276

PubMed Abstract | CrossRef Full Text | Google Scholar

Roffman, D., Hart, G., Girardi, M., Ko, C. J., and Deng, J. (2018a). Predicting non-melanoma skin cancer via a multi-parameterized artificial neural network. Sci. Rep. 8:1701. doi: 10.1038/s41598-018-19907-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Roffman, D. A., Hart, G. R., Leapman, M. S., Yu, J. B., Guo, F. L., Ali, I., et al. (2018b). Development and validation of a multiparameterized artificial neural network for prostate cancer risk prediction and stratification. JCO Clin. Cancer Inform. 2, 1–10. doi: 10.1200/CCI.17.00119

PubMed Abstract | CrossRef Full Text | Google Scholar

Rosenblatt, F. (1958). The perceptron: a probabilistic model for information storage and organization in the brain. Psychol. Rev. 65:386.

PubMed Abstract | Google Scholar

Sah, R. P., Nagpal, S. J. S., Mukhopadhyay, D., and Chari, S. T. (2013). New insights into pancreatic cancer-induced paraneoplastic diabetes. Nat. Rev. Gastroenterol. Hepatol. 10, 423–433. doi: 10.1038/nrgastro.2013.49

PubMed Abstract | CrossRef Full Text | Google Scholar

Salim, N. (2004). Medical Diagnosis Using Neural Networks. Sintok: Faculty of Information Technology, University Utara Malaysia.

Google Scholar

Sanoob, M., Madhu, A., Ajesh, K., and Varghese, S. M. (2016). Artificial neural network for diagnosis of pancreatic cancer. Int. J. Cybernet. Inform. 5, 41–49. doi: 10.5121/ijci.2016.5205

CrossRef Full Text | Google Scholar

Smith, J. W., Everhart, J., Dickson, W., Knowler, W., and Johannes, R. (1988). “Using the ADAP learning algorithm to forecast the onset of diabetes mellitus,” in Proceedings of the Annual Symposium on Computer Application in Medical Care (American Medical Informatics Association), 261.

Google Scholar

Sumathi, B., and Santhakumaran, A. (2011). Pre-diagnosis of hypertension using artificial neural network. Glob. J. Comput. Sci. Technol. 11.

Google Scholar

Verna, E. C., Hwang, C., Stevens, P. D., Rotterdam, H., Stavropoulos, S. N., Sy, C. D., et al. (2010). Pancreatic cancer screening in a prospective cohortof high-risk patients: a comprehensive strategy of imaging and genetics. Clin. Cancer Res. 16, 5028–5037. doi: 10.1158/1078-0432.CCR-09-3209

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, W., Chen, S., Brune, K. A., Hruban, R. H., Parmigiani, G., and Klein, A. P. (2007). PancPRO: risk assessment for individuals with a family history of pancreatic cancer. J. Clin. Oncol. 25:1417–1422. doi: 10.1200/JCO.2006.09.2452

PubMed Abstract | CrossRef Full Text | Google Scholar

Winter, J. M., Cameron, J. L., Campbell, K. A., Arnold, M. A., Chang, D. C., Coleman, J., et al. (2006). 1423 pancreaticoduodenectomies for pancreatic cancer: a single-institution experience. J. Gastrointest. Surg. 10, 1199–1211. doi: 10.1016/j.gassur.2006.08.018

PubMed Abstract | CrossRef Full Text | Google Scholar

Yu, A., Woo, S. M., Joo, J., Yang, H.-R., Lee, W. J., Park, S.-J., et al. (2016). Development and validation of a prediction model to estimate individual risk of pancreatic cancer. PLoS ONE. 11:e0146473. doi: 10.1371/journal.pone.0146473

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, D., and Weng, C. (2011). Combining PubMed knowledge and EHR data to develop a weighted bayesian network for pancreatic cancer prediction. J. Biomed. Inform. 44, 859–868. doi: 10.1016/j.jbi.2011.05.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: pancreatic cancer, cancer risk, cancer prediction, artificial neural network, big data

Citation: Muhammad W, Hart GR, Nartowt B, Farrell JJ, Johung K, Liang Y and Deng J (2019) Pancreatic Cancer Prediction Through an Artificial Neural Network. Front. Artif. Intell. 2:2. doi: 10.3389/frai.2019.00002

Received: 08 January 2019; Accepted: 15 April 2019;
Published: 03 May 2019.

Edited by:

Nisha S. Sipes, National Institute of Environmental Health Sciences (NIEHS), United States

Reviewed by:

Dokyoon Kim, University of Pennsylvania, United States
Hansapani Rodrigo, University of Texas Rio Grande Valley Edinburg, United States

Copyright © 2019 Muhammad, Hart, Nartowt, Farrell, Johung, Liang and Deng. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jun Deng, jun.deng@yale.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.