Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 11 November 2021
Sec. Head and Neck Cancer
This article is part of the Research Topic Prognosis Prediction and Risk Stratification in Head and Neck Cancer View all 59 articles

Dose Prediction Using a Three-Dimensional Convolutional Neural Network for Nasopharyngeal Carcinoma With Tomotherapy

Yaoying Liu,Yaoying Liu1,2Zhaocai ChenZhaocai Chen3Jinyuan WangJinyuan Wang1Xiaoshen WangXiaoshen Wang1Baolin QuBaolin Qu1Lin MaLin Ma1Wei ZhaoWei Zhao2Gaolong Zhang*Gaolong Zhang2*Shouping Xu*Shouping Xu1*
  • 1Department of Radiation Oncology, the First Medical Center of the People's Liberation Army General Hospital, Beijing, China
  • 2School of Physics, Beihang University, Beijing, China
  • 3Manteia Technologies Co., Ltd, Xiamen, China

Purpose: This study focused on predicting 3D dose distribution at high precision and generated the prediction methods for nasopharyngeal carcinoma patients (NPC) treated with Tomotherapy based on the patient-specific gap between organs at risk (OARs) and planning target volumes (PTVs).

Methods: A convolutional neural network (CNN) is trained using the CT and contour masks as the input and dose distributions as output. The CNN is based on the “3D Dense-U-Net”, which combines the U-Net and the Dense-Net. To evaluate the model, we retrospectively used 124 NPC patients treated with Tomotherapy, in which 96 and 28 patients were randomly split and used for model training and test, respectively. We performed comparison studies using different training matrix shapes and dimensions for the CNN models, i.e., 128 ×128 ×48 (for Model I), 128 ×128 ×16 (for Model II), and 2D Dense U-Net (for Model III). The performance of these models was quantitatively evaluated using clinically relevant metrics and statistical analysis.

Results: We found a more considerable height of the training patch size yields a better model outcome. The study calculated the corresponding errors by comparing the predicted dose with the ground truth. The mean deviations from the mean and maximum doses of PTVs and OARs were 2.42 and 2.93%. Error for the maximum dose of right optic nerves in Model I was 4.87 ± 6.88%, compared with 7.9 ± 6.8% in Model II (p=0.08) and 13.85 ± 10.97% in Model III (p<0.01); the Model I performed the best. The gamma passing rates of PTV60 for 3%/3 mm criteria was 83.6 ± 5.2% in Model I, compared with 75.9 ± 5.5% in Model II (p<0.001) and 77.2 ± 7.3% in Model III (p<0.01); the Model I also gave the best outcome. The prediction error of D95 for PTV60 was 0.64 ± 0.68% in Model I, compared with 2.04 ± 1.38% in Model II (p<0.01) and 1.05 ± 0.96% in Model III (p=0.01); the Model I was also the best one.

Conclusions: It is significant to train the dose prediction model by exploiting deep-learning techniques with various clinical logic concepts. Increasing the height (Y direction) of training patch size can improve the dose prediction accuracy of tiny OARs and the whole body. Our dose prediction network model provides a clinically acceptable result and a training strategy for a dose prediction model. It should be helpful to build automatic Tomotherapy planning.

Introduction

Radiotherapy (RT) Plan optimization is a time-consuming process in routine clinical practice. It may cost several hours to constrain the dose distribution to meet the optimal clinical criteria. The plan quality, which the total voxel information can guide the RT plan optimization and ensure, depends on the medical dosimetrist or the medical physicist’s clinical experience and skills. It can minimize the uncertainty of the planning outcome due to different planners handling the planning process (13).

Recently, artificial intelligence (AI) and deep learning (DL) methods have been extensively involved in radiotherapy workflow, such as dose prediction (47). The DL-based methods perform well in automatic feature extraction and mapping transformation (5, 8). The dose prediction model can make an end-to-end mapping transformation between patients’ anatomical and dose distribution information with organs-at-risk (OARs) constraints (912). Compared with using the conventional treatment planning system (TPS), using the DL model to generate predicted dose distribution reduces planning time significantly (1316).

Tomotherapy is a superior RT modality for treating advanced cancers, such as head and neck cancer. Compared to conventional RT treatment, Tomotherapy plan optimization is a time-consuming process. To make a plan with desirable quality, the planner needs to adjust the dose-volume histogram (DVH) limitation and plan criteria to update the plan weights iteratively. In this context, the total voxel information becomes a crucial consideration in dose prediction. It can guide Tomotherapy plan optimization, reducing the iteration times by lessening TPS optimization’s adjustment steps and minimizing the planning outcome uncertainty caused by anthropogenic factors. Different planners may handle the planning process.

Due to the complex anatomy, it is highly challenging to make a plan that can precisely deliver the prescribed dose to the target for the head and neck cancer patients (17, 18). They carry great essential functions for humans, and they need to be protected from unnecessary doses to guarantee which could still function well after the treatment (safe during the treatment). It results in more difficulty in achieving the desirable dose for planning target volumes (PTVs).

This study aims to establish the underlying relationship between anatomical and dose distribution information for nasopharyngeal carcinoma (NPC) patients treated with Tomotherapy using deep-learning approaches. Since few studies have been performed to investigate dose prediction for NPC, this study should be potentially exciting and valuable as guidance or reference for future RT planning.

Materials and Methods

Data Collection and Preparation

One hundred twenty-four NPC patients were treated with Tomotherapy, and our study collected their data. PTVs, the OARs, and the external contour (Body) were labeled as the contoured structures. We added a 3 mm margin around the gross tumor volume of the nasopharynx (GTVnx) and clinical target volume (CTV) to create the planning GTVnx (pGTVnx) and PTV, respectively. The PTVs include PTV60 (a prescription dose of 60 Gy) and PTV54 (54 Gy). The OARs included Brainstem, Spinal-cord, Eyes, Lens, Larynx-esophagus-trachea (L-E-T), Optic-nerves, Oral-cavity, Parotid-glands (PGs), Pituitary, Thyroid, Submaxillary-glands (SMGs). The study collected Digital Imaging and Communications in Medicine (DICOM) files for each case, including CT series, RT Plan, RT Structure, and RT Dose files. All cases corresponding DICOM files involved in our study have been done for particular quality assurance (QA) and delivered.

The collected cases have good consistency: have all PTV60 (with prescription dose of 60 Gy) and have the same types of OARs. We did the data preprocessing before the model training. It ensures the CNN network could load and correctly process the mapping transformation between the patient’s anatomical and dose distribution information. We extracted the 3D CT matrix from CT DICOM files, and the voxel values were normalized for each case. The normalized CT matrix holds a zero mean value and one as the variance. The study converted the region of interest (ROI) information to a binary mask, which means the pixels inside the contouring area with a value of 1 and pixels outside the contouring area with 0. The spacing and matrix shapes of the ROI contouring mask were adjusted equal to the corresponding CT matrix. We obtained the dose array from RT Dose files, with dose values (from 0 to 74 Gy) directly recorded in the dose matrix. All data preprocessing had been done by Python codes. NumPy, pydicom, and other python packages were used to conserve the raw data to the “npy” format.

3D Neural Network

The 3D Dense-U-Net was built as the neural network architecture (Figure 1). “U-Net” is a famous well-behaved CNN network specializing in end-to-end matrix mapping (19). The U-Net architecture consists of down-sampling and up-sampling blocks concatenated across the bottleneck symmetrically, thus allowing the model to extract features for high, middle, and low level (20). The Dense-U-Net structure preserves the up-sampling and down-sampling portions and adds the densely connected layers within each hierarchical level to create the “Dense structure” (21). Every hierarchical level of Dense-U-Net preserves all features from previous layers. It allows the features to be reused and propagated along with successive layers. The 3D Dense-U-Net is the 3D version of the Dense U-Net model. Compared to the 2D Dense U-Net, the 3D one can directly process the input 3D matrix’s information and capture features along the Y direction.

FIGURE 1
www.frontiersin.org

Figure 1 3D Dense-U-NET structure.

Training and Testing

For the 124 nasopharyngeal carcinoma cases, we randomly chose 96 cases for training and 28 cases for testing. The model input matrix contained 21 channels. The first channel is for CT image information, and the 2–21 channels contain ROIs contour information, which includes pGTVnx, PTV60, PTV54, Body, Brainstem, Spinal-cord, Eye-L, Eye-R, Lens-L, Lens-R, L-E-T, Optic-nerve-L, Optic-nerve-R, Oral-cavity, Parotid-L, Parotid-R, Pituitary, Thyroid, SMG-L, and SMG-R. In this study, the ground truth is the dose distribution from the collected RT DOSE DICOM files. Due to the GPU memory limitation, we specified the patch-training strategy. The 128×128×48 shape matrix for training was randomly selected from the 3D dose matrix. The 3D Dense-U-Net model was built up by connection of Dense Block. Every Dense Block includes a Relu activation process, followed by convolution (kernel size 3×3×3), batch normalization, and concatenation with the previous layer. We used zero paddings in each convolution, and each convolution layer had 12 channels. The 3D Dense-U-NET model went through four times down-sampled by max-pooling (kernel size 2×2×2) and symmetrically with four times up-sampled by deconvolution (kernel size 2×2×2, channel =80). The down-sampling process reduced the initial input matrix size from 128×128×48(128➔64➔32➔16➔8) to 8×8×3. It allows the network to be able to extract features both locally and globally; the up-sampling restored the matrix size from 8×8×3 to 128×128×48. The final hierarchical layer of convolution forms a single channel matrix and becomes the output matrix. We used the Adam optimizer (22) with the MAE loss function (1nΣi=1n|f(x)y|) and settled the batch size as 4. The learning rate decayed from 10−4 to 10−6 during CNN network training. When the loss values and learning rate stabilized, the process stopped training. And an Nvidia RTX 3090 GPU accelerated the entire training and testing process in this study. The deep learning framework was TensorFlow and Keras.

This study used 28 untrained cases for the model testing. The CT images and ROI contours were used as the model input data, and dose distribution was the model output (Figure 2). The matrix height (Y direction) of the testing case patch was 64. We concatenated the full-body dose distribution after the model generated the predicted dose distribution for each testing case patch.

FIGURE 2
www.frontiersin.org

Figure 2 Dose difference between the predicted dose distribution and the ground truth for Model I, II, or III. The deep red color shows the dose difference beyond 4 Gy.

We trained two comparative models with different Y lengths (height) to verify whether the 3D model with a large-height training patch could extract more interrelation information from different OAR-PTV distances in the Y direction. From our statistics results, the distance between specific OARs to PTV varies a lot among different patients. For example, the optic nerves’ distances to PTV ranged from 0 to 30 mm, which already equals 10 slices thickness of a CT scan with 3 mm thickness. Model I used the above model training method, and the shape of the training matrix was 128×128×48. Model II reduced the height of the training matrix to 128×128×16 shape. Training Model II aimed to verify whether the increase of height of the training matrix would be helpful to modulate the model to provide more accurate dose prediction for OARs. If the maximum distance from the optic nerve to PTV was 10 slices, and the height of the training matrix was just 16 slice distances, the training matrix may not be able to find enough spatial relationship from optic-organ to PTV. Increasing the height of the training matrix may allow the model to explore a more spatial relationship between the optic nerves and PTV, therefore, to generate more accurate OAR dose prediction. Model III used 2D Dense U-net. It is a simplified change from 3D Dense-U-net to a 2D version. Model III was designed to eliminate the “Y-direction distance” influence in the learning process. Model III can be seen as a comparison experiment to verify if the OAR-PTV distance in the Y direction would be a factor affecting the DL model output.

Quantitative Evaluation

Percentage of errors (δDi), p-value, and gamma passing rate were calculated to evaluate our three models’ accuracy. The formula of the percentage of errors was:

δDi=|DiGroundtruthDipredicted|prescription dose

We calculated δDi of D98, D95, D50, D2 for PTV60, and Dmean, Dmax for all ROIs. All corresponding δDi for 28 test patients were counted and formed mean and standard deviation (Mean ± SD) for each ROI. The p-value of the two models’ δDi was calculated using a T-test; when the p-value<0.05, the prediction results have no statistical correlation. The gamma passing rates with the 3%/3 mm criteria and 10% threshold for the three approaches were calculated by 3D Slicer 4.10.2 [National Institutes of Health (NIH), USA] software.

Results

The mean deviations from the mean and maximum dose of PTVs and OARs were 2.42 and 2.93%, respectively. Error for the maximum dose of optic nerves-R in Model I was 4.87 ± 6.88%, compared with 7.9 ± 6.8% in Model II (p=0.08) and 13.85 ± 10.97% in Model III (p<0.01); Model I showed well. The gamma passing rate of PTV60 for 3%/3 mm criteria was 83.6 ± 5.2% in Model I, compared with 75.9 ± 5.5% in Model II (p<0.001) and 77.2 ± 7.3% in Model III (p<0.01); Model I also did the best job. The prediction error of D95 for PTV60 was 0.64 ± 0.68% in Model I, compared with 2.04 ± 1.38% in Model II (p<0.01) and 1.05 ± 0.96% in Model III (p=0.01); Model I still performed well. The details of prediction errors are presented in Table 1 and Table 2.

TABLE 1
www.frontiersin.org

Table 1 Mean and standard deviation (Mean ± SD) of maximum and mean values between the predicted dose and the ground truth received on PTVs and OARs relative to the prescription dose.

TABLE 2
www.frontiersin.org

Table 2 Means and standard deviations (Mean ± SD) of absolute differences for clinical DVH metrics between the predicted and ground truth doses.

To compare the three models’ accuracy intuitively, we randomly selected a test patient. We showed the dose difference between the predicted dose and the ground truth in Figure 2 and the DVH plots of ROIs in Figure 3. Figures of the dose differences and DVH plots showed that Model I has the best prediction among the three models and an advantage in predicting the optic organs’ dose.

FIGURE 3
www.frontiersin.org

Figure 3 DVH plots for a test patient, such as (A) DVH plot of PTVs, (B) DVH plot of optic organs, (C, D) DVH plots of the other OARs. DVH, Ground-truth (Solid line), Model I (Dashed line), Model II (Dashed and dotted line), Model III (Dotted line).

Discussion

Precise automatic dose prediction can significantly improve clinical planning efficiency and safety (23). 3D dose prediction results can refer to current RT plan optimization in TPS (24, 25). Here, we built CNN-based dose prediction on the previous approved delivered plans. Since, in daily clinical practice, different medical physicists handled the planning process, which provided a source of uncertainty of the RT planning outcome. Using CNN-based dose prediction results guiding plan optimization can reduce the uncertainty of the planning outcomes and improve the plan optimization speed (26). A few fluence-prediction-based auto-planning researches have been done in the past few years. They mentioned that dose distribution could be predicted utilizing a fluence map as well. Furthermore, this enlightens us to get the dose prediction based on an auto-planning system (27, 28). Dose prediction studies can be the basis for much RT-relevant research and technology development.

NPC cases with Tomotherapy have great value in deep-learning dose prediction research. As we know, NPC patients with Tomotherapy are relatively rare in clinical RT practice. And in the past, studies about dose prediction of NPC patients with Tomotherapy were also not too many. Our study for dose prediction found that using a 3D CNN network for training could provide a better outcome than using a 2D CNN network, and the dose prediction accuracy has reached the clinical standard (the mean deviations for the mean and maximum doses of PTVs and OARs were 2.42 and 2.93%, respectively). It can refer to future dose prediction of NPC patients with Tomotherapy, even though this method still needs more research to improve its accuracy.

Our dose prediction model performed well in OARs and PTV areas but didn’t work well in the outside area of OARs and PTVs. Although the outcome accuracy in this study met the clinical requirements, the evaluating indicators included the deviations for the mean and maximum doses for ROIs, the gamma passing rates for PTV, and the DVH plots. But there are still some problems, such as the passing rate for Body was 70.2 ± 9.8%, which was relatively poor. That means further research should focus on how to predict accurate doses in no-contoured areas. Future studies recommend inputting more features such as the help region and the outward expansion area or controlling training data’s consistency, such as only using the designed plan from the same planner.

Besides building a dose prediction model, there is another critical factor that needs to be thought over (1). Training with the CNN network should follow the clinical logic concept (2). Training strategy should not directly duplicate from other studies, considering the dataset’s features should be ahead of training.

In this study, the network structure is similar to some medical imaging segmentation networks. Previous studies showed that the U-net could perform very well in dose prediction and CT image segmentation tasks. But the training strategy should be suitable for the specific prediction tasks. For example, the 2D U-Net can perform pretty well in the task of CT image segmentation (19, 29, 30). Slice by slice segmentation prediction is similar to the clinical logic flow. As is well known, clinical staff always creates the contouring slice by slice, which affirms that each single CT slice should contain enough segmentation information. But as shown in the study, directly using the 2D network to predict dose distribution slice by slice cannot give us the wanted outcome, which may be due to the loss of Y-direction information as shown in the results, the OARs (such as Spinal-cord), which were close to PTV in the Y direction. It didn’t show the different results using the 2D or 3D network to predict the dose distribution. But for OARs far from PTV in the Y direction, such as the optic organs, the dose-prediction results of the 2D network brought out lacks ability. The reason for the outcome difference could be that the algorithm logic is different from the clinical logic. When a medical physicist or dosimetrist makes a treatment plan, the staff should consider the relationship of the relative location between OAR and PTVs. We can quickly understand that it is difficult to avoid unnecessary doses for NPC patients if the optic organs are close to PTV. On the other hand, if the organs are far from PTV, they would be protected from radiation more efficiently. So, using the 3D network for training can allow the model to get the relative location between OARs and PTV. This action conforms to the clinical logic flow. Thus, a good outcome could meet.

Meanwhile, it is necessary to formulate the training strategy by considering the dataset’s features. Some deep-learning-based dose prediction studies have been made for cervical carcinoma. The studies used a general 3D-model-patch-training strategy with 16 pixels height matrix to train (shape of n×n×16) or directly used a 2D network for data training (5, 31, 32). From some dose prediction studies proposed, 2D network training is good enough to provide excellent results of dose prediction. But in transplanting the patch-training strategy to this project, using (n×n×16) shape matrix to train or using the 2D network gives different results. We found that the results were not so good. Reviewing the patient’s anatomic structure, we finally uncovered the dependency between the dose prediction results and the patient’s anatomic information. Using an (n×n×16) shape training matrix/patch, we got an ideal dose prediction for the patients whose PTV was close to the eyes. But for the patients whose PTV was far from the eyes, it resulted in a wrong prediction. The statistics results showed that the optic nerve’s dose delivered was negatively correlated with the distance from PTV to it. For all patients involved in the study, the maximum dose for optic nerves ranged from 9.7 to 71.4 Gy; the distance from the optic nerves to PTV ranged from 0 to 30 mm. The deep-learning model needs to know the spatial relationship between OARs and PTV. Since the predicted doses of optic nerves were highly related to its distance to PTV, using (n×n×16) shape matrix for training, it wouldn’t get an accurate dose prediction for the cases with sizeable PTV-eye distance. We believe that the (n×n×16) shape training matrix fitted better to extract anatomic information in pelvic cancer because the pelvic tissues were generally compact to PTV. For NPC patients, the PTV-eye distance varies from 0 to several centimeters. If the training patch’s height is small such as 16 pixels, it may be difficult for the deep-learning network to find the PTV-eye spatial relationship. PTVs have usually more than 70 slices thickness height for NPC patients. Suppose the training patch’s matrix with a considerable height, such as height = 48 pixels, and the model could extract more features of the spatial relationship among the PTV-eye voxels.

Clinical and actual treatment logic concept includes a lot of information, which are greatly important. We could utilize them to optimize the deep-learning network performance relevant to the RT aspect. The training matrix should be considered the network’s field of view from which the model could find the transformation relationship. Training with the 3D Dense-U-NET could predict each pixel’s dose value by considering the full input matrix. Increasing the input matrix height (Y direction) would be a strategy realized the extraction combination features of model training and clinical logic concepts. Increasing the height of the input matrix (increases the local sense of field) can make the DL model find more spatial features and relationships correlated to PTV-OAR distance, which provides a more accurate outcome for dose prediction.

The deep-learning-based dose prediction method still has many problems that need to be solved. Firstly, previous research never focused on excavating the data’s internal features and comparing the data differences. The anatomical information holds tremendous differences among different patients. Secondly, we can’t directly use the previous researchers’ method for deep learning, for different tumor types and treatment techniques have specific dose prediction methods. According to the tumor type and treatment mode, developing a specific dose prediction method can be a better way to improve dose prediction efficiency and accuracy. Our research was focused on adding the clinical logic concept with the deep-learning method together. Therefore, we developed a more reasonable deep-learning model training strategy.

A deep-learning-based study focuses on the relevant software and hardware, the clinical logic concepts, and the collected data characteristic. Combining computer technology, clinical logic flow, and data characteristics would be an ideal pathway to develop an excellent-performance dose prediction model.

Conclusions

In this study, we successfully developed an accurate dose prediction model using a 3D convolutional neural network. It proves well for NPC patients with Tomotherapy. It also tells that exploring the spatial features between OARs and PTV is necessary for dose prediction. We found that a 3D DL model with a larger Y-dimension training matrix increases the accuracy of dose prediction outcomes. With this extra consideration, our accuracy improvement method of dose prediction is good enough to be considered a milestone for the automatic planning process with Tomotherapy and other RT techniques. The predicted results could be used as a reference or guidance for systematic clinical RT planning.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding authors.

Ethics Statement

This study was approved by the Ethics Committee of the Chinese PLA General Hospital (approved no. S2016-122-01). Written informed consent to participate in this study was provided by the participant’s legal guardian.

Author Contributions

YL: Experiment design and code implementation; article writing. ZC: Technical support. JW: Data collection. XW: Technical support. BQ: Data collection. LM: Data collection. WZ: Technical support, article modification. SX: Experiment design, article modification. GZ: Article modification. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the Medical Big Data AI R&D Project (2019MBD-043).

Conflict of Interest

Manteia Technologies Co., Ltd employed author ZC.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Barragan-Montero AM, Nguyen D, Lu W, Lin MH, Norouzi-Kandalan R, Geets X, et al. Three-Dimensional Dose Prediction for Lung IMRT Patients With Deep Neural Networks: Robust Learning From Heterogeneous Beam Configurations. Med Phys (2019) 46(8):3679–91. doi: 10.1002/mp.13597

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Chen X, Men K, Li Y, Yi J, Dai J. A Feasibility Study on an Automated Method to Generate Patient-Specific Dose Distributions for Radiotherapy Using Deep Learning. Med Phys (2019) 46(1):56–64. doi: 10.1002/mp.13262

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Kajikawa T, Kadoya N, Ito K, Takayama Y, Chiba T, Tomori S, et al. A Convolutional Neural Network Approach for IMRT Dose Distribution Prediction in Prostate Cancer Patients. J Radiat Res (2019) 60(5):685–93. doi: 10.1093/jrr/rrz051

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Nguyen D, Jia X, Sher D, Lin M-H, Iqbal Z, Liu H, et al. Three-Dimensional Radiotherapy Dose Prediction on Head and Neck Cancer Patients With a Hierarchically Densely Connected U-Net Deep Learning Architecture. Phys Med Biol (2019) 64(6):065020. doi: 10.1088/1361-6560/ab039b

CrossRef Full Text | Google Scholar

5. Ma M, Kovalchuk N, Buyyounouski MK, Xing L, Yang Y. Incorporating Dosimetric Features Into the Prediction of 3D VMAT Dose Distributions Using Deep Convolutional Neural Network. Phys Med Biol (2019) 64(12):125017. doi: 10.1088/1361-6560/ab2146

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Boldrini L, Bibault J-E, Masciocchi C, Shen Y, Bittner M-I. Deep Learning: A Review for the Radiation Oncologist. Front Oncol (2019) 9:977. doi: 10.3389/fonc.2019.00977

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Kandalan RN, Nguyen D, Rezaeian NH, Barragán-Montero AM, Breedveld S, Namuduri K, et al. Dose Prediction With Deep Learning for Prostate Cancer Radiation Therapy: Model Adaptation to Different Treatment Planning Practices. Radiother Oncol (2020) 153:228–35. doi: 10.1016/j.radonc.2020.10.027

PubMed Abstract | CrossRef Full Text | Google Scholar

8. LeCun Y, Bengio Y, Hinton G. Deep Learning. Nature (2015) 521(7553):436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Guerreiro F, Seravalli E, Janssens G, Maduro J, Knopf A, Langendijk J, et al. Deep Learning Prediction of Proton and Photon Dose Distributions for Paediatric Abdominal Tumours. Radiother Oncol (2021) 156:36–42. doi: 10.1016/j.radonc.2020.11.026

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Gronberg MP, Gay SS, Netherton TJ, Rhee DJ, Court LE, Cardenas CE. Dose Prediction for Head and Neck Radiotherapy Using a Three-Dimensional Dense Dilated U-Net Architecture. Med Phys (2021) 48(9):5567–73. doi: 10.1002/mp.14827

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Zimmermann L, Faustmann E, Ramsl C, Georg D, Heilemann G. Dose prediction for radiation therapy using feature-based losses and One Cycle Learning. Med Phys (2021) 48(9):5562–6. doi: 10.1002/mp.14774

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Kearney V, Chan JW, Wang T, Perry A, Descovich M, Morin O, et al. DoseGAN: A Generative Adversarial Network for Synthetic Dose Prediction Using Attention-Gated Discrimination and Generation. Sci Rep (2020) 10(1):1–8. doi: 10.1038/s41598-020-68062-7

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Liang B, Tian Y, Chen X, Yan H, Yan L, Zhang T, et al. Prediction of Radiation Pneumonitis With Dose Distribution: A Convolutional Neural Network (CNN) Based Model. Front Oncol (2020) 9:1500. doi: 10.3389/fonc.2019.01500

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Hu J, Song Y, Wang Q, Bai S, Yi Z. Incorporating Historical Sub-Optimal Deep Neural Networks for Dose Prediction in Radiotherapy. Med Image Anal (2021) 67:101886. doi: 10.1016/j.media.2020.101886

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Bakx N, Bluemink H, Hagelaar E, van der Sangen M, Theuws J, Hurkmans C. Development and Evaluation of Radiotherapy Deep Learning Dose Prediction Models for Breast Cancer. Phys Imaging Radiat Oncol (2021) 17:65–70. doi: 10.1016/j.phro.2021.01.006

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Barragán-Montero AM, Thomas M, Defraene G, Michiels S, Haustermans K, Lee JA, et al. Deep Learning Dose Prediction for IMRT of Esophageal Cancer: The Effect of Data Quality and Quantity on Model Performance. Physica Med (2021) 83:52–63. doi: 10.1016/j.ejmp.2021.02.026

CrossRef Full Text | Google Scholar

17. Caudell JJ, Torres-Roca JF, Gillies RJ, Enderling H, Kim S, Rishi A, et al. The Future of Personalised Radiotherapy for Head and Neck Cancer. Lancet Oncol (2017) 18(5):e266–73. doi: 10.1016/S1470-2045(17)30252-8

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Castelli J, Simon A, Lafond C, Perichon N, Rigaud B, Chajon E, et al. Adaptive Radiotherapy for Head and Neck Cancer. Acta Oncol (2018) 57(10):1284–92. doi: 10.1080/0284186X.2018.1505053

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Jin J, Zhu H, Zhang J, Ai Y, Zhang J, Teng Y, et al. Multiple U-Net-Based Automatic Segmentations and Radiomics Feature Stability on Ultrasound Images for Patients With Ovarian Cancer. Front Oncol (2021) 10:614201. doi: 10.3389/fonc.2020.614201

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Munich, Germany: Springer (2015).

Google Scholar

21. Huang G, Liu Z, van der Maaten L, Weinberger KQ. Densely Connected Convolutional Networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI, United States (2017).

Google Scholar

22. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. In: International Conference on Learning Representations 2015 (ICLR 2015) . San Diego, California, United States (2014).

Google Scholar

23. Ma J, Nguyen D, Bai T, Folkerts M, Jia X, Lu W, et al. A Feasibility Study on Deep Learning–Based Individualized 3d Dose Distribution Prediction. Med Phys (2021) 48(8):4438–47. doi: 10.1002/mp.15025

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Bai X, Wang B, Wang S, Wu Z, Gou C, Hou Q. Radiotherapy Dose Distribution Prediction for Breast Cancer Using Deformable Image Registration. Biomed Eng Online (2020) 19:1–20. doi: 10.1186/s12938-020-00783-2

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Norouzi Kandalan R, Nguyen D, Hassan Rezaeian N, Barragan-Montero AM, Breedveld S, Namuduri K, et al. Dose Prediction With Deep Learning for Prostate Cancer Radiation Therapy: Model Adaptation to Different Treatment Planning Practices. Radiother Oncol (2020) 153:228–35. doi: 10.1016/j.radonc.2020.10.027

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Sun Y, Yang Y, Qian J, Tian Y. Evaluation and Prediction of Pelvic Dose in Postoperative IMRT for Cervical Cancer. Chin J Radiat Oncol (2020) 6(2020):136–40.

Google Scholar

27. Zhong Y, Yu L, Zhao J, Fang Y, Yang Y, Wu Z, et al. Clinical Implementation of Automated Treatment Planning for Rectum Intensity-Modulated Radiotherapy Using Voxel-Based Dose Prediction and Post-Optimization Strategies. Front Oncol (2021) 11:697995. doi: 10.3389/fonc.2021.697995

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Nilsson V, Gruselius H, Zhang T, De Kerf G, Claessens M. Probabilistic Dose Prediction Using Mixture Density Networks for Automated Radiation Therapy Treatment Planning. Phys Med Biol (2021) 66(5):055003. doi: 10.1088/1361-6560/abdd8a

PubMed Abstract | CrossRef Full Text | Google Scholar

29. He K, Liu X, Shahzad R, Reimer R, Thiele F, Niehoff J, et al. Advanced Deep Learning Approach to Automatically Segment Malignant Tumors and Ablation Zone in the Liver With Contrast-Enhanced CT. Front Oncol (2021) 11:669437. doi: 10.3389/fonc.2021.669437

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Liu X, Li K-W, Yang R, Geng L-S. Review of Deep Learning Based Automatic Segmentation for Lung Cancer Radiotherapy. Front Oncol (2021) 11:717039. doi: 10.3389/fonc.2021.717039

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Kajikawa T, Kadoya N, Ito K, Takayama Y, Chiba T, Tomori S, et al. Automated Prediction of Dosimetric Eligibility of Patients With Prostate Cancer Undergoing Intensity-Modulated Radiation Therapy Using a Convolutional Neural Network. Radiol Phys Technol (2018) 11(3):320–7. doi: 10.1007/s12194-018-0472-3

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Ma M, Buyyounouski K, Vasudevan M,V, Xing L, Yang Y. Dose Distribution Prediction in Isodose Feature-Preserving Voxelization Domain Using Deep Convolutional Neural Network. Med Phys (2019) 46(7):2978–87. doi: 10.1002/mp.13618

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: dose prediction, deep learning, Tomotherapy, nasopharyngeal carcinoma, radiotherapy plan

Citation: Liu Y, Chen Z, Wang J, Wang X, Qu B, Ma L, Zhao W, Zhang G and Xu S (2021) Dose Prediction Using a Three-Dimensional Convolutional Neural Network for Nasopharyngeal Carcinoma With Tomotherapy. Front. Oncol. 11:752007. doi: 10.3389/fonc.2021.752007

Received: 02 August 2021; Accepted: 21 October 2021;
Published: 11 November 2021.

Edited by:

Heming Lu, People’s Hospital of Guangxi Zhuang Autonomous Region, China

Reviewed by:

Yinglin Peng, Sun Yat-sen University Cancer Center (SYSUCC), China
Changsheng Ma, Shandong University, China

Copyright © 2021 Liu, Chen, Wang, Wang, Qu, Ma, Zhao, Zhang and Xu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shouping Xu, c2hvdXBpbmdfeHVAeWFob28uY29t; Gaolong Zhang, emdsQGJ1YWEuZWR1LmNu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.