
95% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Oncol. , 26 March 2025
Sec. Radiation Oncology
Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1509449
Purpose: This study introduces a deep learning (DL) model that leverages doses calculated from both a treatment planning system (TPS) and independent dose verification software using Monte Carlo (MC) simulations, aiming to predict the gamma passing rate (GPR) in VMAT patient-specific QA more accurately.
Materials and method: We utilized data from 710 clinical VMAT plans measured with an ArcCHECK phantom. These plans were recalculated on an ArcCHECK phantom image using Pinnacle TPS and MC algorithms, and the planar dose distributions corresponding to the detector element surfaces were utilized as input for the DL model. A convolutional neural network (CNN) comprising four layers was employed for model training. The model’s performance was evaluated through multiple predictive error metrics and receiver operator characteristic (ROC) curves for various gamma criteria.
Results: The mean absolute errors (MAE) between measured GPR and predicted GPR are 1.1%, 1.9%, 1.7%, and 2.6% for the 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm gamma criteria, respectively. The correlation coefficients between predicted GPR and measured GPR are 0.69, 0.72, 0.68, and 0.71 for each gamma criterion. The AUC (Area Under the Curve) values based on ROC curve for the four gamma criteria are 0.90, 0.92, 0.93, and 0.89, indicating high classification performance.
Conclusion: This DL-based approach showcases significant potential in enhancing the efficiency and accuracy of VMAT patient-specific QA. This approach promises to be a useful tool for reducing the workload of patient-specific quality assurance.
In the realm of modern radiation therapy, intensity-modulated radiation therapy (IMRT) and volumetric modulated arc therapy (VMAT) have become the most widely used techniques due to their superior dose conformity to the target volume (1). Even though these techniques are sophisticated, they are fundamentally complex and require patient-specific quality assurance (PSQA) to guarantee dose delivery accuracy (2). Traditionally, measurement-based quality assurance has been used to do the verification utilizing ionization chambers, films, or multi-dimensional detector arrays.
Gamma analysis, which examines the discrepancy between the measured and computed dose distributions, is commonly used in the verification process. For a given set of doses and distance-to-agreement (DTA) criteria, the gamma index is calculated using percentage dose difference and DTA at each detector. Traditional measurement-based methods, while effective, are labor-intensive and time-consuming, particularly when the initial results don’t meet acceptance criteria (3, 4). Traditional methods also impede the implementation of online adaptive radiotherapy, necessitating a rapid and real-time process for treatment planning and quality assurance (5, 6).
Recently, machine learning (ML) has been introduced for prediction of GPR. The complexity of IMRT and VMAT plans, characterized by factors such as the modulation complexity score (MCS), leaf motion constraints, and gantry speed etc., has been scrutinized for its impact on treatment accuracy and GPR (7–11). With ML, researchers have successfully harnessed these and treatment plan parameters to develop predictive models (12–21). These models are capable of directly forecasting the pass rate of radiation therapy plans, marking a significant advancement in PSQA. However, these models still face challenges. One issue is that these models often require manual selection of parameters, and inappropriate parameter choices can significantly affect prediction results.
Deep learning has been leveraged to enhance the accuracy and efficiency of predicting the GPR in treatment plans without defining various features like ML, otherwise extracting features from images automatically (22–25). Interian et al. and Tomori et al. both utilized convolutional neural networks to predict the GPR based on the 2D dose information (26, 27). Huang et al. introduced a virtual PSQA method for IMRT, employing UNet++ trained to forecast three key outputs (gamma pass rates, dose differences, and classification outcomes) to ascertain the success or failure of the QA process (28). Additionally, researchers have utilized fluence maps and 3D dose distributions to predict the verification pass rates or gamma distributions of radiation therapy plans (29–31). This kind of approaches involve analyzing the intricate patterns in dose distribution across three dimensions, providing a comprehensive view that enhances the accuracy of predicting treatment plan quality and effectiveness.
Independent calculation-based dose verification is another potential method recommended by the AAPM report TG219 (32). As a Supplementary Method, calculation-based dose verification is characterized by its low manpower consumption and high level of automation. However, it should be noted that this calculation method cannot replace the measurement-based patient-specific QA now. They serve as a supplementary check, particularly as certain machine delivery errors may not be detectable solely through software calculations (33, 34).
Moreover, the application of Monte Carlo methods, known for their high accuracy in dose calculation, has been integrated into predictive models (35–40). Monte Carlo methods provide highly accurate dose calculations by simulating the interaction of radiation with matter at the particle level. The method takes into account complex variables such as tissue heterogeneity, irregular geometries, and detailed dose deposition patterns. Additionally, Monte Carlo simulations can model the probabilistic nature of radiation transport, allowing for more precise dose distributions compared to traditional deterministic methods. These factors make the Monte Carlo methods suitable for PSQA. Some third-party radiotherapy treatment plan verification software, such as SunCHECK, Mobius3D, ArcherQA, and RadCalc, complement plan verification efforts and play an important role in ensuring the precision and safety of radiation therapy. These software applications commonly employ the highly accurate Monte Carlo-based dose calculation methods for dose computation. The combination of advanced dose calculation algorithms with AI methodologies offers a promising avenue for enhancing the precision and efficiency of plan verification processes.
In this context, the purpose of this study is to investigate the feasibility and performance of DL integrated with doses derived from TPS and independent verification software using Monte Carlo dose calculations in the PSQA. This approach is expected to offer a novel tool for ensuring the accuracy and safety of treatments.
The overall study design for predicting GPR is shown in Figure 1. Initially, patient plans were optimized within the TPS. Subsequently, the RT plan was transferred to the accelerator for dose measurement. Doses were recalculated for the clinical plans on a QA phantom using both the TPS and plan verification software ArcherQA (Wisdom Technology Company Limited, Hefei, China) with the MC algorithm. Gamma analysis was performed by comparing the TPS-calculated dose on the QA phantom against the measured dose, obtaining GPR with different criteria. A Python script was utilized to extract cylindrical dose data corresponding to the detector array from the TPS and ArcherQA exports. The cylindrical dose, along with the GPR were utilized for training convolutional neural network models and predictive analysis.
We retrospectively selected 710 clinical VMAT plans treated with a 6-MV photon beam delivered by an Elekta Versa HD accelerator (Elekta AB, Stockholm, Sweden) in flattering filter-free mode. The treatment sites encompassed the head and neck (209), thorax (198), abdomen (205), and pelvis (98). The distribution of disease sites across the training, validation, and testing datasets is detailed in Supplementary Appendix A. For these plans, the VMAT optimization was carried out using Pinnacle TPS (version 16.2, Philips Healthcare, Eindhoven, Netherlands). Within this system, the dose calculation was performed using an adaptive convolve dose engine.
The ArcCHECK dosimetry system (Sun Nuclear Corporation, Melbourne, USA) was used to perform verification measurements. ArcCHECK is a helical 3D detector array comprising 1386 diodes arranged within the cylindrical wall of a phantom. Each patient plan was recalculated on ArcCHECK with a dose grid resolution of 0.2 cm in each dimension. The GPR was calculated with SNC patient software (version 6.7.2, Sun Nuclear Corporation, Melbourne, USA) for evaluating dose discrepancies between the TPS-calculated and phantom-measured values. For the gamma analysis (3D mode), we applied criteria of 3%/3 mm, 3%/2 mm, 2%/3 mm, and 2%/2 mm, alongside a 10% dose threshold, using absolute dose mode and global normalization (41, 42).
All the DICOM files of verification plans, including the RT plan, RT structure, RT dose, and CT image, were imported into ArcherQA to recalculate the dose distribution using the Monte Carlo algorithm. ArcherQA has already been clinically implemented for all our machines for independent dose verification. The details for beam modeling was introduced before (43). The modeling was also commissioned and validated with phantom measurement results. The uncertainty was 1% for transportation used in Monte Carlo simulations. The grid for Monte Carlo calculation were identical to the settings in the validation plan within the TPS.
The calculated 3D dose was exported from the TPS and ArcherQA, respectively, to our in-house cylindrical dose generator. This Python-based tool extracts the dose distribution on the cylindrical surface where the detector array resides. The extracted dose distribution, with a resolution of 220×673, has been normalized to the maximum dose value. Finally, the dose distribution data is saved in the form of an h5 file, which is then used for the training and testing of the model.
In this study, a convolutional neural network (CNN) was used to predict the GPR. Figure 2 illustrates the model’s architecture, which comprises four convolutional layers, four max-pooling layers, four activation layers, a flatten layer, and three fully connected layer. All activation layers utilized the rectified linear unit (ReLU) function. ReLU helps improve the efficiency of learning by eliminating output values below zero. Additionally, dropout layer with a 0.25 drop rate before the first fully connected layer was employed to enhance network robustness, mitigating overfitting by randomly removing neurons (44). The model’s input consists of dual-channel unwrapped dose on the cylindrical surface, with doses originating from both the TPS and ArcherQA. The model’s output is gamma pass rates for four different criteria.
Our PSQA dataset consisted of 710 case records. For the dataset splitting, we allocated 426 cases for training, 142 for validation, and the remaining 142 for testing. During model training, we used data augmentation strategies, such as translations and flipping, to artificially increase the size and diversity of the dataset. The neural network architecture is based on PyTorch and was trained on an NVIDIA GeForce RTX 4080 GPU. During training, the L1 Loss (Mean Absolute Error) was used as our loss function. The optimization process utilized the Adam optimizer with a base learning rate (lr) set to 0.0001 and adaptive moment estimation parameters (betas) configured as (0.9, 0.999). To enhance training dynamics, a learning rate scheduler was employed, gradually reducing the learning rate by a factor of 0.98 every 5 epochs. The deep learning models were trained for a maximum of 300 epochs. Early stopping was employed to prevent overfitting by halting training when the validation loss started to increase.
In pursuit of a dependable and consistent model, a 5-fold cross-validation approach was used to train the model. Through cross-validation, optimal model parameters were determined and then used to generate predicted GPR for the 142 test cases. Subsequently, a GPR comparison was performed between the predicted GPR and the measured GPR, taking into account metrics of mean error (ME), mean absolute error (MAE), and root-mean-square error (RMSE) as described by Equations 1–3.
where N represents the total count of test cases, p(i) denotes the GPR value predicted for the i-th case, and m(i) signifies the corresponding measured GPR value.
The performance across various criteria was also evaluated using the ROC curve. This curve represents the relationship between the true positive rate (TPR) and the false positive rate (FPR), as described by Equations 4, 5. To quantify the classifier’s performance, the AUC (Area Under the Curve) was calculated. Typically, AUC values range from 0.5 to 1, with 0.5 indicating random classification, and a value close to one signifying an excellent classifier.
where, N represents the number of a specific value, and definitions for TP (true positive), TN (true negative), FP (false positive), FN (false negative) are provided in Table 1. The “pass” threshold was defined as achieving a GPR greater than or equal to 95% at 3%/3 mm, greater than or equal to 90% at 3%/2 mm, 2%/3 mm, and 2%/2 mm (41, 42). Otherwise, it was classified as “fail”.
The entire training process for the model takes around 110 minutes, and the detailed training and validation loss curves over epochs can be found in Supplementary Appendix B. Figure 3 illustrates the correlation between measured and predicted GPR values for the 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm gamma criteria. The black diagonal line in the figure represents a perfect prediction where the measured values equal the predicted values. Table 2 presents mean values and standard deviations (SD) for both measured and predicted GPR values of the test cases.
Figure 3. Plot of measured and predicted GPR at 3%/3 mm criterion, 3%/2 mm criterion, 2%/3 mm criterion, and 2%/2 mm criterion.
Figure 4 shows the distribution of prediction errors for GPR values across four distinct criteria: 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm. Specific statistical error metrics such as ME, MAE, and RMSE can be found in Tables 3 and 4. The Pearson’s correlation coefficient (CC) between predicted GPR and measured GPR was calculated to be 0.69, 0.72, 0.68, and 0.71 for the 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm criteria, respectively. The results demonstrate a robust correlation between the predicted and measured gamma pass rates.
Figure 4. The prediction error distribution for GPR values across four different criteria: 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm. The black bars represent the interquartile range (25% - 75%), the lines extending from the bars indicate the 1.5× interquartile range, and the white circles show the median prediction error for each category. The shape of each violin plot illustrates the probability density of the data at different error levels, with wider sections representing a higher probability of data points falling at a particular error percentage.
Table 4. Comparison of GPR (%) error statistics in the test dataset among different model inputs under various gamma criteria.
The ROC curves for different GPR value classification criteria are illustrated in Figure 5, with the AUC values being 0.90, 0.92, 0.93, and 0.95 for the 3%/3mm, 3%/2mm, 2%/3mm, and 2%/2mm criteria, respectively. The ROC curve and the corresponding AUC values serve as key indicators of the model’s classification performance. An AUC value closer to 1 indicates excellent predictive accuracy, suggesting that the model is effectively distinguishing between true positive and true negative cases across the different criteria. These AUC values indicate that each criterion provides excellent predictive accuracy and high classification performance.
Figure 5. ROC curves for four different criteria of predicted GPR. Each curve represents the trade-off between the true positive rate (sensitivity) and false positive rate (1-specificity) for the 3%/3mm (blue line, AUC = 0.90), 3%/2mm (orange line, AUC = 0.92), 2%/3mm (green line, AUC = 0.93), and 2%/2mm (red line, AUC = 0.89) criteria.
In this study, we present an exploration of PSQA using DL for predicting GPR in radiation therapy. Traditional PSQA, primarily relies on measurement-based methods, which, although accurate, can be time-consuming and resource-intensive. The model developed in this study can rapidly provide planners with GPR results, allowing for timely actions like replanning when predicted GPR falls below acceptance criteria. High-accuracy predictions enhance the credibility of the virtual QA method and support decision-making. Our prediction model can be integrated into an automated software platform. When new plan data is received, the platform automatically extracts the dual-channel dose, calls the model for prediction, and generates a report. Compared to independent Monte Carlo calculations, this process only requires an additional few minutes of time, which fully meets the clinical application time requirements.
The International Commission on Radiation Units and Measurements (ICRU) Report 83 (45) and the American Association of Physicists in Medicine (AAPM) Task Group 219 Report (32) endorse the use of independent dose calculation methods as an additional verification step for treatment plans generated by TPS. Utilizing random sampling to simulate particle transport and interactions, the MC method not only ensures a more accurate representation of radiation transport and scattering in various media but also excels in dose calculations for small field sizes where traditional convolution methods may falter in precision. Consequently, these attributes have made MC algorithms commonly employed as independent dose calculation methods. Motivated by this, we explored the integration of advanced MC dose calculation algorithms and AI methodologies to achieve more accurate predictions of GPR. In this study, a commercial MC-based dose engine (ArcherQA) was used for independent dose calculations. In the model construction, we utilized a dual-channel dose as input, incorporating dose distributions from both the TPS and ArcherQA, to enhance the model’s performance. The model with dual-channel dose input exhibits smaller MAE and RMSE, as detailed in Table 4 for specific comparative results. One possible reason is that the doses calculated by Monte Carlo algorithms are closer to the measured values, especially in the calculation of small field doses (39). However, due to the inherent lack of interpretability common in deep learning algorithms, a direct and reliable cause still requires further exploration.
GPR prediction has been explored by various researchers in prior studies. Notably, the AI-driven regression method exhibited favorable predictive performance. Hirashima et al. utilized plan complexity and dosiomics features as input data to predict the GPR value of a helical diode array, achieving correlation coefficients ranging from 0.45 to 0.61 and MAE ranging from 2.7% to 3.2% for a 3%/2mm gamma criteria (17). Matsuura et al. utilized deep learning to predict the GPR of a 3D detector array-based quality assurance for volumetric modulated arc therapy in prostate cancer (30). They achieved a correlation coefficient of 0.7 and a MAE of 2.5% for a gamma criterion of 3%/2mm. In the course of this study, we acquired a correlation coefficient of 0.72 and a MAE of 1.9% using the predictive model under a 3%/2mm gamma criterion. While direct comparisons of these results are challenging due to variations in prediction methods, such as treatment site and measurement device, our study demonstrated the DL model utilizing dual-channel dose input for GPR prediction outperforms or equals other ML or DL-based algorithms.
In the process of radiotherapy, many factors can lead to dose deviations. Our DL model is adept at detecting systemic dose calculation errors and discrepancies in fluence/geometric modeling of the treatment machine. It can also identify dose deviations resulting from suboptimal field sizes, such as those that are too small, too large, or significantly off-center. However, it does not track errors in data transfer to the treatment unit nor detect actual errors in treatment delivery if the machine fails to interlock itself, as these would require real-time monitoring or separate verification systems.
In spite of its efficacy, this approach faces certain challenges. Firstly, the model’s performance is inherently tied to the quality and diversity of the training data. Although our dataset encompasses multiple sites, it is limited to 710 clinical VMAT plans from a single machine, requiring further investigation into the model’s generalizability across different machines, treatment modalities, and patient populations. Before clinical implementation, the model requires training on a significantly larger dataset to ensure its robustness and reliability. During the initial phase of clinical use, the model’s predictions will be used in parallel with existing clinical methods to cross-validate results and ensure consistency. Another consideration is the model’s dependency on the accuracy of input data from TPS, ArcherQA and ArcCHECK. Any discrepancies or errors in these systems could negatively impact the model’s predictions. Therefore, continuous verification of these systems is essential if the model is to be applied in clinical settings.
Our approach offers several key advantages compared to similar techniques. Firstly, compared to machine learning-based methods, our method does not rely on the manual extraction of feature parameters, which reduces human bias and the risk of errors in feature selection. In comparison to third-party verification methods based solely on Monte Carlo simulations, our approach offers higher reliability. By incorporating dual-channel dose into the predictive model, it provides a more comprehensive assessment of the treatment plan’s quality. Furthermore, when compared to other deep learning-based models, our approach has demonstrated higher accuracy in predicting GPR, making it a more reliable tool for PSQA.
Our model introduces an innovative method to support and streamline the PSQA process. In the near future, if our model is applied clinically, it may first be used to assist in checking or screening low pass-rate verification plans. It would serve as a supplement rather than a replacement for traditional PSQA. The potential application of our deep learning model is to identify treatment plans with low gamma passing rates. This selective focus would enable clinical physicists to concentrate their efforts on plans that are most likely to require adjustments. Additionally, integrating this method with emerging technologies like magnetic resonance imaging-guided radiation therapy could provide novel insights into online adaptive radiation therapy strategies.
We proposed a deep learning-based prediction model for VMAT patient-specific QA, using extracted doses from both TPS and MC calculations as input to the model. The dual-channel dose distribution, associated with cylindrically arranged detector elements, captures plan characteristics and holds predictive potential for the GPR in VMAT plans. This method is expected to be a valuable tool capable of reducing the workload associated with PSQA.
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
JM: Writing – original draft, Writing – review & editing. YX: Funding acquisition, Software, Validation, Writing – review & editing. KM: Methodology, Writing – review & editing. JD: Funding acquisition, Methodology, Supervision, Writing – review & editing.
The author(s) declare that financial support was received for the research and/or publication of this article. This work is supported by the National High Level Hospital Clinical Research Funding (2022-CICAMS-80102022203), CAMS Innovation Fund for Medical Sciences (CIFMS, 2023-I2M-C&T-B-095) and the Beijing Hope Run Special Fund of Cancer Foundation of China (No. LC2022B25).
We thank Dr. X. George Xu, Dr. Xi Pei and Dr. Bo Cheng from University of Science and Technology of China and Anhui Wisdom Technology Company Limited for support in Monte Carlo software ArcherQA.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declare that no Generative AI was used in the creation of this manuscript.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1509449/full#supplementary-material
1. Gorayski P, Fitzgerald R, Barry T, Burmeister E, Foote M. Volumetric modulated arc therapy versus step-and-shoot intensity modulated radiation therapy in the treatment of large nerve perineural spread to the skull base: a comparative dosimetric planning study. J Med Radiat Sci. (2014) 61:85–90. doi: 10.1002/jmrs.45
2. Chiavassa S, Bessieres I, Edouard M, Mathot M, Moignier A. Complexity metrics for IMRT and VMAT plans: a review of current literature and applications. Br J Radiol. (2019) 92:20190270. doi: 10.1259/bjr.20190270
3. Miles EA, Clark CH, Urbano MT, Bidmead M, Dearnaley DP, Harrington KJ, et al. The impact of introducing intensity modulated radiotherapy into routine clinical practice. Radiother Oncol. (2005) 77:241–6. doi: 10.1016/j.radonc.2005.10.011
4. Nelms BE, Chan MF, Jarry G, Lemire M, Lowden J, Hampton C, et al. Evaluating IMRT and VMAT dose accuracy: practical examples of failure to detect systematic errors when applying a commonly used metric and action levels. Med Phys. (2013) 40:111722. doi: 10.1118/1.4826166
5. Bohoudi O, Bruynzeel AME, Senan S, Cuijpers JP, Slotman BJ, Lagerwaard FJ, et al. Fast and robust online adaptive planning in stereotactic MR-guided adaptive radiation therapy (SMART) for pancreatic cancer. Radiother Oncol. (2017) 125:439–44. doi: 10.1016/j.radonc.2017.07.028
6. de Jong R, Crama KF, Visser J, van Wieringen N, Wiersma J, Geijsen ED, et al. Online adaptive radiotherapy compared to plan selection for rectal cancer: quantifying the benefit. Radiat Oncol. (2020) 15:162. doi: 10.1186/s13014-020-01597-1
7. Shen L, Chen S, Zhu X, Han C, Zheng X, Deng Z, et al. Multidimensional correlation among plan complexity, quality and deliverability parameters for volumetric-modulated arc therapy using canonical correlation analysis. J Radiat Res. (2018) 59:207–15. doi: 10.1093/jrr/rrx100
8. McNiven AL, Sharpe MB, Purdie TG. A new metric for assessing IMRT modulation complexity and plan deliverability. Med Phys. (2010) 37:505–15. doi: 10.1118/1.3276775
9. McDonald DG, Jacqmin DJ, Mart CJ, Koch NC, Peng JL, Ashenafi MS, et al. Validation of a modern second-check dosimetry system using a novel verification phantom. J Appl Clin Med Phys. (2017) 18:170–7. doi: 10.1002/acm2.12025
10. Masi L, Doro R, Favuzza V, Cipressi S, Livi L. Impact of plan parameters on the dosimetric accuracy of volumetric modulated arc therapy. Med Phys. (2013) 40:071718. doi: 10.1118/1.4810969
11. Lambri N, Hernandez V, Sáez J, Pelizzoli M, Parabicoli S, Tomatis S, et al. Multicentric evaluation of a machine learning model to streamline the radiotherapy patient specific quality assurance process. Phys Med. (2023) 110:102593. doi: 10.1016/j.ejmp.2023.102593
12. Valdes G, Scheuermann R, Hung CY, Olszanski A, Bellerive M, Solberg TD. A mathematical framework for virtual IMRT QA using machine learning. Med Phys. (2016) 43:4323. doi: 10.1118/1.4953835
13. Valdes G, Chan MF, Lim SB, Scheuermann R, Deasy JO, Solberg TD. IMRT QA using machine learning: A multi-institutional validation. J Appl Clin Med Phys. (2017) 18:279–84. doi: 10.1002/acm2.12161
14. Granville DA, Sutherland JG, Belec JG, La Russa DJ. Predicting VMAT patient-specific QA results using a support vector classifier trained on treatment plan characteristics and linac QC metrics. Phys Med Biol. (2019) 64:095017. doi: 10.1088/1361-6560/ab142e
15. Lam D, Zhang X, Li H, Deshan Y, Schott B, Zhao T, et al. Predicting gamma passing rates for portal dosimetry-based IMRT QA using machine learning. Med Phys. (2019) 46:4666–75. doi: 10.1002/mp.13752
16. Li J, Wang L, Zhang X, Liu L, Li J, Chan MF, et al. Machine learning for patient-specific quality assurance of VMAT: prediction and classification accuracy. Int J Radiat Oncol Biol Phys. (2019) 105:893–902. doi: 10.1016/j.ijrobp.2019.07.049
17. Hirashima H, Ono T, Nakamura M, Miyabe Y, Mukumoto N, Iramina H, et al. Improvement of prediction and classification performance for gamma passing rate by using plan complexity and dosiomics features. Radiother Oncol. (2020) 153:250–7. doi: 10.1016/j.radonc.2020.07.031
18. Ono T, Hirashima H, Iramina H, Mukumoto N, Miyabe Y, Nakamura M, et al. Prediction of dosimetric accuracy for VMAT plans using plan complexity parameters via machine learning. Med Phys. (2019) 46:3823–32. doi: 10.1002/mp.13669
19. Hasse K, Scholey J, Ziemer BP, Natsuaki Y, Morin O, Solberg TD, et al. Use of receiver operating curve analysis and machine learning with an independent dose calculation system reduces the number of physical dose measurements required for patient-specific quality assurance. Int J Radiat Oncol Biol Phys. (2021) 109:1086–95. doi: 10.1016/j.ijrobp.2020.10.035
20. Kusunoki T, Hatanaka S, Hariu M, Kusano Y, Yoshida D, Katoh H, et al. Evaluation of prediction and classification performances in different machine learning models for patient-specific quality assurance of head-and-neck VMAT plans. Med Phys. (2022) 49:727–41. doi: 10.1002/mp.15393
21. Quintero P, Benoit D, Cheng Y, Moore C. Machine learning-based predictions of gamma passing rates for virtual specific-plan verification based on modulation maps, monitor unit profiles, and composite dose images. Physics in Medicine & Biology (2022) 67(24):245001. doi: 10.1088/1361-6560/aca38a
22. Sahiner B, Pezeshk A, Hadjiiski LM, Wang X, Drukker K, Cha KH, et al. Deep learning in medical imaging and radiation therapy. Med Phys. (2019) 46:e1–e36. doi: 10.1002/mp.13264
23. Arabi H, Zaidi H. Applications of artificial intelligence and deep learning in molecular imaging and radiotherapy. European Journal of Hybrid Imaging (2020) 4:17. doi: 10.1186/s41824-020-00086-8
24. Shen C, Nguyen D, Zhou Z, Jiang SB, Dong B, Jia X. An introduction to deep learning in medical physics: advantages, potential, and challenges. Phys Med Biol. (2020) 65:05tr01. doi: 10.1088/1361-6560/ab6f51
25. Hao Y, Zhang X, Wang J, Zhao T, Sun B. Improvement of IMRT QA prediction using imaging-based neural architecture search. Med Phys. (2022) 49:5236–43. doi: 10.1002/mp.15694
26. Interian Y, Rideout V, Kearney VP, Gennatas E, Morin O, Cheung J, et al. Deep nets vs expert designed features in medical physics: An IMRT QA case study. Med Phys. (2018) 45:2672–80. doi: 10.1002/mp.12890
27. Tomori S, Kadoya N, Takayama Y, Kajikawa T, Shima K, Narazaki K, et al. A deep learning-based prediction model for gamma evaluation in patient-specific quality assurance. Med Phys. (2018) 45(9):4055–4065. doi: 10.1002/mp.13112
28. Huang Y, Pi Y, Ma K, Miao X, Fu S, Chen H, et al. Virtual patient-specific quality assurance of IMRT using UNet++: classification, gamma passing rates prediction, and dose difference prediction. Front Oncol. (2021) 11:700343. doi: 10.3389/fonc.2021.700343
29. Matsuura T, Kawahara D, Saito A, Yamada K, Ozawa S, Nagata Y. A synthesized gamma distribution-based patient-specific VMAT QA using a generative adversarial network. Med Phys. (2023) 50:2488–98. doi: 10.1002/mp.16210
30. Matsuura T, Kawahara D. Predictive gamma passing rate of 3D detector array-based volumetric modulated arc therapy quality assurance for prostate cancer via deep learning. Physical and Engineering Sciences in Medicine. (2022) 45:1073–81. doi: 10.1007/s13246-022-01172-w
31. Yoganathan SA, Ahmed S, Paloor S, Torfeh T, Aouadi S, Al-Hammadi N, et al. Virtual pretreatment patient-specific quality assurance of volumetric modulated arc therapy using deep learning. Med Phys. (2023) 50:7891–903. doi: 10.1002/mp.16567
32. Zhu TC, Stathakis S, Clark JR, Feng W, Georg D, Holmes SM, et al. Report of AAPM Task Group 219 on independent calculation-based dose/MU verification for IMRT. Med Phys. (2021) 48:e808–29. doi: 10.1002/mp.15069
33. Siochi RA, Molineu A, Orton CG. Point/Counterpoint. Patient-specific QA for IMRT should be performed using software rather than hardware methods. Med Phys. (2013) 40:070601. doi: 10.1118/1.4794929
34. Smith JC, Dieterich S, Orton CG. Point/counterpoint. It is still necessary to validate each individual IMRT treatment plan with dosimetric measurements before delivery. Med Phys. (2011) 38:553–5. doi: 10.1118/1.3512801
35. Ma CM, Pawlicki T, Jiang SB, Li JS, Deng J, Mok E, et al. Monte Carlo verification of IMRT dose distributions from a commercial treatment planning optimization system. Phys Med Biol. (2000) 45:2483–95. doi: 10.1088/0031-9155/45/9/303
36. Pisaturo O, Moeckli R, Mirimanoff RO, Bochud FO. A Monte Carlo-based procedure for independent monitor unit calculation in IMRT treatment plans. Phys Med Biol. (2009) 54:4299–310. doi: 10.1088/0031-9155/54/13/022
37. Hoffmann L, Alber M, Söhn M, Elstrøm UV. Validation of the Acuros XB dose calculation algorithm versus Monte Carlo for clinical treatment plans. Med Phys. (2018) 45(8):3909–3915. doi: 10.1002/mp.13053
38. Molinelli S, Mairani A, Mirandola A, Vilches Freixas G, Tessonnier T, Giordanengo S, et al. Dosimetric accuracy assessment of a treatment plan verification system for scanned proton beam radiotherapy: one-year experimental results and Monte Carlo analysis of the involved uncertainties. Phys Med Biol. (2013) 58:3837–47. doi: 10.1088/0031-9155/58/11/3837
39. Xu Y, Zhang K, Liu Z, Liang B, Ma X, Ren W, et al. Treatment plan prescreening for patient-specific quality assurance measurements using independent Monte Carlo dose calculations. Front Oncol. (2022) 12:1051110. doi: 10.3389/fonc.2022.1051110
40. Xu Y, Xia W, Ren W, Ma M, Men K, Dai J. Is it necessary to perform measurement-based patient-specific quality assurance for online adaptive radiotherapy with Elekta Unity MR-Linac? J Appl Clin Med Phys. (2023) 25(2):e14175. doi: 10.1002/acm2.14175
41. Miften M, Olch A, Mihailidis D, Moran J, Pawlicki T, Molineu A, et al. Tolerance limits and methodologies for IMRT measurement-based verification QA: Recommendations of AAPM Task Group No. 218. Med Phys. (2018) 45:e53–83. doi: 10.1002/mp.12810
42. Han C, Zhang J, Yu B, Zheng H, Wu Y, Lin Z, et al. Integrating plan complexity and dosiomics features with deep learning in patient-specific quality assurance for volumetric modulated arc therapy. Radiat Oncol. (2023) 18:116. doi: 10.1186/s13014-023-02311-7
43. Xu XG, Liu T, Su L, Du X, Riblett M, Ji W, et al. ARCHER, a new Monte Carlo software tool for emerging heterogeneous computing environments. Ann Nucl Energy. (2015) 82:2–9. doi: 10.1016/j.anucene.2014.08.062
44. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. (2014) 15:1929–58.
Keywords: gamma passing rate, deep learning, Monte Carlo, quality assurance, ArcCHECK
Citation: Miao J, Xu Y, Men K and Dai J (2025) A feasibility study of deep learning prediction model for VMAT patient-specific QA. Front. Oncol. 15:1509449. doi: 10.3389/fonc.2025.1509449
Received: 11 October 2024; Accepted: 04 March 2025;
Published: 26 March 2025.
Edited by:
Gene A. Cardarelli, Brown University, United StatesReviewed by:
Jia-Ming Wu, Wuwei Cancer Hospital of Gansu Province, ChinaCopyright © 2025 Miao, Xu, Men and Dai. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jianrong Dai, ZGFpX2ppYW5yb25nQGNpY2Ftcy5hYy5jbg==; Kuo Men, bWVua3VvMTI2QDEyNi5jb20=
†These authors have contributed equally to this work and share first authorship
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.