Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 23 September 2022
Sec. Genitourinary Oncology
This article is part of the Research Topic Adaptive Radiotherapy for Localized Prostate Cancer View all 5 articles

Synthetic CT generation for MRI-guided adaptive radiotherapy in prostate cancer

Shu-Hui Hsu*Shu-Hui Hsu*Zhaohui HanZhaohui HanJonathan E. LeemanJonathan E. LeemanYue-Houng HuYue-Houng HuRaymond H. MakRaymond H. MakAtchar SudhyadhomAtchar Sudhyadhom
  • Department of Radiation Oncology, Brigham and Women’s Hospital and Dana-Farber Cancer Institute, Boston, MA, United States

Current MRI-guided adaptive radiotherapy (MRgART) workflows require fraction-specific electron and/or mass density maps, which are created by deformable image registration (DIR) between the simulation CT images and daily MR images. Manual density overrides may also be needed where DIR-produced results are inaccurate. This approach slows the adaptive radiotherapy workflow and introduces additional dosimetric uncertainties, especially in the presence of the magnetic field. This study investigated a method based on a conditional generative adversarial network (cGAN) with a multi-planar method to generate synthetic CT images from low-field MR images to improve efficiency in MRgART workflows for prostate cancer. Fifty-seven male patients, who received MRI-guided radiation therapy to the pelvis using the ViewRay MRIdian Linac, were selected. Forty-five cases were randomly assigned to the training cohort with the remaining twelve cases assigned to the validation/testing cohort. All patient datasets had a semi-paired DIR-deformed CT-sim image and 0.35T MR image acquired using a true fast imaging with steady-state precession (TrueFISP) sequence. Synthetic CT images were compared with deformed CT images to evaluate image quality and dosimetric accuracy. To evaluate the dosimetric accuracy of this method, clinical plans were recalculated on synthetic CT images in the MRIdian treatment planning system. Dose volume histograms for planning target volumes (PTVs) and organs-at-risk (OARs) and dose distributions using gamma analyses were evaluated. The mean-absolute-errors (MAEs) in CT numbers were 30.1 ± 4.2 HU, 19.6 ± 2.3 HU and 158.5 ± 26.0 HU for the whole pelvis, soft tissue, and bone, respectively. The peak signal-to-noise ratio was 35.2 ± 1.7 and the structural index similarity measure was 0.9758 ± 0.0035. The dosimetric difference was on average less than 1% for all PTV and OAR metrics. Plans showed good agreement with gamma pass rates of 99% and 99.9% for 1%/1 mm and 2%/2 mm, respectively. Our study demonstrates the potential of using synthetic CT images created with a multi-planar cGAN method from 0.35T MRI TrueFISP images for the MRgART treatment of prostate radiotherapy. Future work will validate the method in a large cohort of patients and investigate the limitations of the method in the adaptive workflow.

1 Introduction

Online prostate adaptive radiotherapy allows for the generation of an optimal treatment plan based on daily changes to anatomy and has the potential to improve target coverage and reduce the toxicity to the surrounding organs such as rectum, bladder and urethra. Recently, the MR-linac treatment modality offers on-board MR images with high soft tissue contrast for improved tumor and normal tissue delineation compared to standard cone-beam CT based radiotherapy (1). Additionally, the system can deliver gated treatment by monitoring the prostate or critical organ positions to ensure high-precision treatment. Therefore, MRI-guided adaptive radiotherapy (MRgART) can potentially improve treatment outcomes by reducing inter-fraction and intra-fraction uncertainties in prostate treatments (24).

Current clinical procedures in MRgART require CT image acquisitions for dose calculations in treatment planning because of the lack of electron/mass density information on MR images. For each fraction, daily MR images are acquired and deformably registered to the original simulation MR image, calculating a daily deformation vector field (DVF). The deformable image registration (DIR) DVF is then applied to original simulation CT images to propagate the electron/mass density information in the coordinate frame of daily MR images. Any remaining differences in the body contour and air pockets between the two image sets are compensated for by means of user-driven override structures with densities equal to soft tissue and air. The process of DIR and density overrides must be repeated for each fraction in the adaptive radiotherapy workflow. This approach is labor intensive, subject to user bias and interpretation, slows the adaptive radiotherapy workflow and introduces additional dosimetric uncertainties, especially in the presence of the magnetic field (5, 6). Thus, generating synthetic CT from daily MRI images would streamline the MRgART workflow and reduce uncertainties caused by image registration and manual density overrides.

Challenges in generating synthetic CT from MRI include the lack of correlation between the MRI intensity and tissue electron density, the dependence of the MRI contrast on scanning parameters, and the difficulty in distinguishing air from bone on MR images as both air and dense bone have little to no signal on MRI. Several methods have been proposed for generating synthetic CT images from MR images in various body sites and have been reviewed comprehensively in prior studies (711). These methods included bulk density assignment, atlas-based methods, voxel-based methods and machine learning methods. Most studies focused on MR simulators (1.5T and 3T). Additionally, these methods may require special scanning sequences or multiple pulse sequences, the most common being Dixon and zero echo time (ZTE)/ultrashort TE (UTE). Few studies have reported synthetic CT methods for low-field MRgART, such as ViewRay MRIdian (0.35T) (1214). Generating synthetic CT images from low-field MR images can be more challenging due to the lower signal-to-noise ratio compared to higher field strength and limited selection of scanning protocols (usually using a true fast imaging with steady-state precession (TrueFISP) sequence for fast and low-noise imaging at low-field MRI). Therefore, tissue classification via tissue-contrast optimized scanning protocols may not be feasible for low-field MRgART because of the need to compromise with fast imaging, particularly for breath-hold imaging and treatments.

Deep learning methods have shown promising results in medical imaging processing, such as image classification and reconstruction. Using deep learning methods in synthetic CT generation may be more feasible than voxel-based methods for low-field MRgART to obtain better image quality and may produce more accurate electron/mass density maps for dose calculations. Cusumano et al. (12) demonstrated that using a conditional generative adversarial network (cGAN) was feasible to generate synthetic CT from low-field MR images for abdominal and pelvic cases. Fu et al. (14) compared synthetic CT images generated from 0.35T MR images using the cGAN and the cycle-consistent generative adversarial network (cycleGAN) for liver radiotherapy and reported that both cGAN and cycleGAN achieved accurate dose calculations. Olberg et al. (13) proposed a deep spatial pyramid convolution framework to generate synthetic CT from 0.35T MRI for breast radiotherapy and demonstrated improved performance compared to a conventional GAN framework.

Based on previous studies, GAN-based deep learning methods have the potential to generate high quality synthetic CT images from low-field MR images. However, most of these studies trained networks using a single two-dimensional (2D) plane approach, typically using axial slices. Therefore, discontinuities across slices were often observed with this single 2D plane method. Using three-dimensional (3D) volumes as an input may generate a more accurate synthetic CT volume; however, this approach may require larger datasets for training to obtain high quality results. The aim of the study was to evaluate the quality of synthetic CT images generated by a cGAN using a multi-planar method as inputs in the application of low-field MRgART for prostate radiotherapy.

2 Materials and methods

2.1 Patient selection and image data

Fifty-seven male patients, who received MRI-guided pelvic radiation therapy on a ViewRay 0.35T MRIdian Linac (ViewRay Inc., Oakwood, OH, USA) between 2019 and 2022, were retrospectively enrolled in this study. All patients were scanned and treated with a free breathing technique. The selection criteria included absence of hip implants and less imaging artifacts, determined by visual interpretation. Each patient had both CT-sim and 0.35T MR-sim images acquired on the same day and in the treatment position. The CT images were acquired on a Siemens SOMATOM Confidence CT simulator (Siemens Healthcare, Malvern, PA, USA) or GE LightSpeed RT16 (GE Healthcare, Chicago, Illinois, USA), with a tube voltage of 120 kVp, slice thicknesses of 2 to 3 mm and in-plane pixel sizes of 0.98 × 0.98 and 1.27 × 1.27 mm2 (Table 1). MR images were acquired using a TrueFISP sequence on the 0.35T MRIdian, with two different fields of view (Table 1). All patients were simulated on both the CT simulator and the MR-linac with a modest bladder filling and enema rectum preparation (3).

TABLE 1
www.frontiersin.org

Table 1 CT imaging parameters (simulator model and voxel size), MR imaging parameters (field of view, scan time, voxel size, flip angle, repetition time and echo time), treatment sites and clinical plan information for training and validation/testing datasets.

A subset of forty-five cases were used for model training. The treatment sites included pelvic nodes, prostate, perirectal area, sacrum and ureter (Table 1). The task for the model was to learn male pelvic anatomy and CT-to-MRI mapping for a large patient cohort. Thus, the training cases were not limited to only prostate cases. The remaining twelve cases were used for model validation/testing. These validation/testing datasets were limited to prostate stereotactic body radiotherapy (SBRT) patients. The treatment plan parameters of these prostate cases are shown in Table 1.

2.2 Synthetic CT generation

2.2.1 Image preprocessing

In our current clinical workflow, the CT images were deformably registered to 0.35T MR images using MIM (MIM Software Inc., Cleveland, OH, USA) to get electron/mass density information for dose calculations. The quality of deformed CT images was carefully reviewed in the planning process. For our study, the deformed CT and MR images were exported, and additional image preprocessing was performed.

All MR images were corrected for residual intensity non-uniformity using a commonly applied post-processing bias-field correction algorithm (N4itk) (15) and implemented in a publicly available image analysis software environment (SLICER, surgical processing laboratory, Brigham and Women’s Hospital, Boston, MA, USA). The bias field was estimated within the volume defined by the skin surface. The N4itk optimization parameters included: BSpline order of 3, BSpline grid resolutions of (1, 1, 1), a shrink factor of 4, maximum numbers of 50, 40 and 30 iterations at each of the 3 resolution levels, and a convergence threshold of 0.0001. All CT and MR images were resampled to 256×256×256 with a resolution of 1.5×1.5×1 mm3. The pixel value of the volume outside the skin surface was set to -1000 for CT images and 0 for MR images.

The pixel value of CT images ranged from -1000 to 3095 while the pixel value for MR images ranged from 0 to a varying maximum value. The CT image intensity was normalized by dividing 4095 and scaled such that the CT number of -1000 HU was mapped to -1 and the CT number of 3095 HU was mapped to 1. The MR intensity was also normalized and mapped from -1 to 1 using a z-score method (16). The z-score method standardizes the image by centering the intensity histogram at a mean of 0 and dividing the intensity by the standard deviation. When calculating the mean and standard deviation for each MRI dataset, the areas with an intensity smaller than 40 were excluded from the calculations. This threshold value is often used in our clinical workflow to define the air/gas region on TrueFISP MR images. After normalizing the MR intensities, any intensity value less than or greater than three standard deviations from the mean were truncated to be -1 and 1, respectively. For example, after normalization, the pixel value in the air region was -1 and the value in the brightest area on MR images was 1.

2.2.2 Deep learning network and loss function

A cGAN based on the pix2pix architecture with some modifications was employed to generate synthetic CT images from MR images (17). The GAN has two neural networks, a generator and a discriminator, contesting with each other (Figure 1A). The generator (G) learns to generate synthetic CT images while the discriminator (D) learns to distinguish the generator’s synthetic CT images from real CT images. The generator was based on U-Net architecture with skip-connections while the discriminator was based on 16×16 PatchGAN. The objective (LT) was composed of an adversarial loss (LcGAN), pixel reconstruction loss (LL1) with a weighting factor (λ1) and mutual information loss (LMI) with a weighting factor (λ2).

FIGURE 1
www.frontiersin.org

Figure 1 (A) A synthetic CT training process using a multi-planar method. Three orthogonal planes from paired MR-CT image sets are used to train generator and discriminator networks with loss functions (LG and LD). (B) A synthetic CT generation process for validation/testing. A 3D MRI volume is sampled in three orthogonal directions, generating three MRI sets as inputs to the generator. Three corresponding synthetic CT sets are generated (sCTax, sCTcor, sCTsag) and combined to get the final synthetic CT (sCTave).

LT=argminGmaxDLcGAN(G,D)+λ1LL1(G)+λ2LMI(G)
LcGAN(G,D)=Ex,y[D(x,y)2]+Ex[(D(x,G(x))1)2]
LL1(G)=Ex,y[||yG(x)||1]=1ni=1n|yiG(xi)|
LMI(G)=Ex,y[1nMI(y,G(x))]

where λ1 and λ2 were 100 and 5, respectively, and x, y and G(x) represent MR images, real CT images and generated synthetic CT images, respectively. For LcGAN, a least-squares objective was used rather than the log likelihood objective because former exhibited better stability during training (18). The LL1 objective was used to calculate the mean-absolute-error between real CT and generated synthetic CT images for all voxels (n). The LMI objective was not in the original pix2pix implementation but was added in the loss function to minimize the effect of the uncertainty due to the misalignment between CT and MR images. This objective was to calculate image similarity between real CT and synthetic CT images, and the mutual information was normalized (nMI) so it ranged from 0 to 1.

To reduce the model oscillation, the discriminator was updated using a history of generated images rather than the ones produced by the latest generator (18). The optimization was performed using an adaptive moment estimation (Adam) solver, with an initial learning rate of 0.0002 for the generator and 0.00005 for the discriminator and momentum parameters of 0.5 and 0.999 (17). The learning rate was decreased by 1% per increasing epoch. A total of 100 epochs were trained.

The network training was performed in MATLAB (MathWorks, Inc., Natick, MA, USA) using a RTX A5000 GPU (NVIDIA, Sunnyvale, CA, USA). The GAN was trained on three orthogonal directions. Each image batch in the iteration of the training process included 2D axial, sagittal and coronal planes of the paired MR and CT images (Figure 1A). In each iteration, three orthogonal planes were randomly extracted from 3D volumes (unused slices in prior iterations) in both MR and CT image sets. If no anatomy information existed in any extracted planes, that image batch was disregarded. Maximum iterations in each epoch were 256 (slices)×45(number of training cases).

2.2.3 Synthetic CT generation for validation/testing

The generator network was trained with the multi-planar method (Figure 1A). Thus, an MRI volume sampled in any orthogonal direction can be used as an input to generate a synthetic CT volume. To improve the robustness of synthetic CT generation and reduce discontinuities across slices, three image sets sampled in three orthogonal directions from each MRI volume were input into the generator and the corresponding synthetic CT image sets in 2D axial (sCTax), coronal (sCTcor) and sagittal (sCTsag) planes were created, as illustrated in Figure 1B. As a result, three predictions were acquired for each voxel. Predicted values from the outputs in all three directions were then averaged to obtain the final synthetic CT volume (sCTave). The synthetic CT generation was performed using an in-house program, written in MATLAB.

2.3 Evaluation of synthetic CT quality

2.3.1 Image quality of synthetic CT volumes

To evaluate the quality of synthetic CT images, their similarity to a ground truth CT, on a voxel-wise basis, was calculated. The CT images closest to the ground truth in this study were deformed CT images. Therefore, mean-absolute-errors (MAE) and mean errors (ME) in CT numbers, peak signal-to-noise ratios (PSNRs) and structural similarity index measures (SSIMs) between synthetic CT and deformed CT were calculated following the equations shown below.

MAE=1n|sCTiCTi|n
ME=1nsCTiCTin
PSNR=10log(MaxCT2MSE)
SSIM=(2μsCTμCT+c1)(2σsCT,CT+c2)(μsCT2+μCT2+c1)(σsCT2+σCT2+c2)

where n is the voxel number in the region of interest (ROI); MAXCT is the maximum pixel value of deformed CT images; MSE is the mean-squared-error; μsCT and μCT are the means of synthetic CT and deformed CT images;  σsCT and σCT are the standard deviations;  σsCT, CT is cross-covariance; c1 and c2 are regularization constants for the luminance and contrast.

The ROI in calculating the SSIM was the whole 3D volume (256×256×256), while the ROI in calculating the PSNR, MAE and ME was the 3D volume but only included the voxels within the body contour. In particular, the MAE and ME were also calculated in three different tissue segments, including air, soft tissue and bone. The three tissue segments were identified by applying image intensity thresholds on the deformed CT. The voxels with CT numbers smaller than -200 HU was included in the air segment, and the voxels with CT numbers larger than 200 HU was included in the bone segment. The residual voxels were included in the soft tissue segment, i.e. CT numbers from -200 HU to 200 HU. No morphological post-processing on these segments was performed.

2.3.2 Dosimetric accuracy of synthetic CT volumes

Clinical intensity-modulated radiotherapy plans were calculated on deformed CT images. As the deformation was not likely perfect, differences in body contours and gas pockets between MRI and deformed CT datasets were compensated by overriding the mass density (relative to water) in these structures with soft tissue (1.02) and air (0.0012). Dose distributions were calculated using a Monte Carlo algorithm implemented in the MRIdian treatment planning system, including the effect of the magnetic field, 0.2 cm calculation grid and 0.5% calculation uncertainty. The clinical plans were then recalculated on the synthetic CT images with the same fluence and calculation parameters for each validation case. The CT calibration curve used in the synthetic CT recalculation was the same as the one in the initial plan, depending on which simulator was used for CT image acquisition.

Dose volume histograms (DVHs) were compared between synthetic CT and user-corrected deformed CT (dCTcorr) calculations of each clinical plan. Two DVH metrics (D95% and D2%) were evaluated for planning target volumes (PTVs) and D2% was evaluated for organs-at-risk (OARs), including rectum, bladder and urethra, for prostate SBRT.

Three-dimensional dose distributions were exported to SLICER and 3D gamma analyses were performed in a ROI where the dose was larger than 10% of the maximum dose in each clinical plan. Two different criteria in gamma analyses were compared, 1%/1 mm and 2%/2 mm.

3 Results

3.1 Image quality of synthetic CT volumes

The image quality of all synthetic CT image sets generated along different orientations was similar to the quality of deformed CT image sets (Figure 2). However, for sCTax image sets, the axial view had smoother edges than the sagittal and coronal views. Similar behaviors were observed for sCTsag and sCTcor image sets. The sCTave image set was created by averaging sCTax, sCTsag and sCTcor image sets, so the discontinuities across slices, observed in sCTax, sCTsag and sCTcor image sets, were reduced.

FIGURE 2
www.frontiersin.org

Figure 2 MR (top row), deformed CT (2nd row), sCTax (3rd row), sCTsag (4th row), sCTcor (5th row) and sCTave (bottom row) for one of the validation cases.

Regarding quantitative analyses, the MAEs ranged from 30.1 HU to 33.6 HU in the whole pelvis area for all four synthetic CT image sets; the MEs ranged from -4.3 HU to -7.9 HU; PSNRs ranged from 33.8 to 35.2; SSIMs ranged from 0.9694 to 0.9758 (Table 2). In terms of CT numbers in various tissue segments, the accuracy was the best in the soft tissue segment and worst in the air segment, due to changes in bowel gas positioning between scans. Comparing sCTave with the other three synthetic CT sets, MAEs were ~10% lower, PSNRs were 3% higher, and SSIMs were 0.5% higher in the whole pelvis.

TABLE 2
www.frontiersin.org

Table 2 MAEs and MEs for whole pelvis and individual segments (air, soft tissue and bone) and PSNRs and SSIMs for whole pelvis.

3.2 Dosimetric accuracy of synthetic CT volumes

Clinical plans were recalculated on sCTave image sets. Figure 3 plots the case that exhibited the best agreement in CT numbers (23.7 HU in MAE for the whole pelvis) and DVH metrics (< 0.2 Gy), while Figure 4 shows the case that had the largest differences in CT numbers (36.1 HU in MAE for the whole pelvis) and DVH metrics (< 0.9 Gy). The DVHs for the best and worst cases are shown in Figure 5. Large dose differences were observed in the PTV, urethra and rectum for the worst case.

FIGURE 3
www.frontiersin.org

Figure 3 Isodose color wash (124%, 100%, 69%, 50% and 30% of 36.25 Gy in the prescribed dose) displayed on MR and sCTave images in axial, sagittal and coronal views for the best case (SIB case).

FIGURE 4
www.frontiersin.org

Figure 4 Isodose color wash (100%, 69%, 50% and 30% of 36.25 Gy in the prescribed dose) displayed on MR and sCTave images in axial, sagittal and coronal views for the worst case.

FIGURE 5
www.frontiersin.org

Figure 5 (A) DVHs for the case shown in Figure 3 and (B) DVHs for the case shown in Figure 4. Sold lines: dCTcorr; dotted lines: sCTave.

In general, the DVH metrics calculated on sCTave image sets were higher than those calculated on dCTcorr image sets (Figure 6 and Table 3). The mean dose differences were within 0.25 Gy for the PTV metrics and 0.35 Gy for the OAR metrics, representing 0.6% and 0.9%, respectively (Table 3). All validation cases had dose differences less than 0.7 Gy in both PTV and OAR structures except one case (the worst case shown in Figure 4) that had a deviation in rectal D2% of 0.9 Gy (2.3%) (Figure 6).

FIGURE 6
www.frontiersin.org

Figure 6 Box-and-whisker plots of (A) dose difference (Gy) and (B) relative dose difference (%) between sCTave and dCTcorr for PTV and OAR metrics. The bottom and top of the box represent the 1st and 3rd quartiles; the band inside the box is the median; the ends of the whiskers represent 95% range; the crosses represent outliers.

TABLE 3
www.frontiersin.org

Table 3 DVH metrics of dCTcorr and sCTave, the comparison between sCTave and dCTcorr in DVH metrics and 3D gamma analyses with 1%/1 mm and 2%/2 mm in a ROI where the dose is larger than 10% of the maximum dose in each clinical plan.

Comparing dose distributions using 3D gamma analyses, good agreement was found for all validation cases, with 99% and 99.9% pass rates for 1%/1 mm and 2%/2 mm criteria, respectively (Table 3).

4 Discussion

The multi-planar cGAN deep learning method used in this study to generate synthetic CT images reduced discontinuities across slices that were often observed in the 2D planar method. The MAE in CT numbers was reduced by ~10% with the multi-planar method compared to the 2D planar method. Regarding its dosimetric accuracy compared to clinical plans on dCTcorr, the mean relative difference was less than 1% for PTV and OAR metrics, and 3D gamma analyses showed good agreement, achieving a 99% pass rate for 1%/1 mm criteria. This study demonstrated that the presented method generated synthetic CT images comparable to the deformed CT method that is currently used in the low-field MRgART workflow for treating prostate cancer.

Some outliers were observed when evaluating the dosimetric accuracy, particularly in the rectum (Figures 4, 5B, 6). Several factors may contribute to the large deviations. First, the MRIdian TPS uses a Monte Carlo method, which has an inherent calculation uncertainty (0.5% in this study). Second, deformed CT images with air and soft tissue overrides were used as ground truth to validate the accuracy of synthetic CT images. However, anatomical changes may be observed between simulation CT and MR scans, despite their acquisition on the same day, e.g., changes in bladder filling (as seen on the MR and CT images in Figure 2) and variation of gas pockets in the bowel and rectum (resulting in large deviations for air segments in Table 2). To compensate for the change in gas morphology, manual density overrides with pure air were used when planning. This process can introduce additional uncertainties because gas pockets may not be correctly identified nor necessarily be composed of pure air. As shown in Figure 4, a large gas pocket was found in the rectum adjacent to the prostate. The accuracy of its size and density was crucial for this case because of the electron return effect in the presence of the magnetic field. It is therefore difficult to assert that the deformed CT with density overrides represents the ground truth. Ultimately this resulted in higher measured error.

Compared with previous synthetic CT generating studies in the prostate or pelvis using deep learning methods, our results were superior or comparable to those performed with either low-field (12, 19) or high-field (14, 2025) MRI. For example, Cusumano et al. (12) generated synthetic CT from 0.35T MRI using a 2D cGAN and reported that MAEs in CT numbers were 54.3 ± 11.9 HU, 40.4 ± 9.2 HU, 224.4 ± 35.5 HU for whole pelvis, soft tissue and bone, respectively, as compared to our results, 30.1 ± 4.2 HU, 19.6 ± 2.3 HU and 158.5 ± 26.0 HU for whole pelvis, soft tissue and bone, respectively. Fu et al. (14) used T1-weighted images from 1.5T MRI to synthesize CT using a 3D convolutional neural network (CNN) and reported 37.6 ± 5.1 HU, 26.2 ± 4.5 HU and 154.3 ± 22.3 HU for whole pelvis, soft tissue and bone, respectively. Chen et al. (22) used T2-weighted images from 3T MRI to synthesize CT using a 2D U-net model and reported 30.0 ± 4.9 HU, 19.6 ± 2.5 HU and 122.5 ± 10.5 HU for whole pelvis, soft tissue and bone, respectively.

Regarding the calculation time, the time required by the neural network to generate synthetic CT was ~12 seconds with the GPU system used in this study. The presented method would be more efficient and provide more accurate electron/mass density information as the density override and review processes take ~4 minutes in our current MRgART workflow.

The image resolution and size combination used in this study may cut off part of the anatomy for fields of view larger than 38.4 cm×38.4 cm and limit potential clinical implementation in larger patients. A follow-up study will consider increasing the image size to ensure the method can be applied to large-size patients.

In addition, future work will investigate the accuracy, robustness, and limitations of the presented method in a large patient cohort, apply the method to different body sites for MRgART and explore a method to independently verify the accuracy of the synthetic CT generation for a quality assurance (QA) process in the clinical workflow, such as catching outliers to trigger further review in MRgART. A QA process will ensure that the deep learning method can be safely and reliably implemented in clinical workflows to generate synthetic CT volumes from MRI volumes and to improve the workflow efficiency in MRI-guided radiotherapy.

5 Conclusion

The quality of synthetic CT images generated from 0.35T MRI using a multi-planar cGAN method was evaluated for prostate radiotherapy. The MAE was 30.1 ± 4.2 HU in the whole pelvis compared to the deformed CT. Calculated doses on synthetic CT images agreed well with the doses in the clinical plans with a gamma pass rate of 99% for 1%/1 mm criteria, and the difference was less than 1% on average for all PTV and OAR metrics. Our study demonstrates the potential of using synthetic CT created from 0.35T MRI TrueFISP images in the adaptive workflow of prostate radiotherapy. Future work will validate the method in a large cohort of patients and investigate the limitations of the method in the adaptive workflow.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author/s.

Ethics statement

The studies involving human participants were reviewed and approved by IRB - 2020P000292: Retrospective Data Analysis of Patients Treated on the MR-Linac. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author contributions

SH designed the study, developed the method, analyzed data and drafted the manuscript. All authors contributed to the article and approved the submitted version.

Acknowledgments

AS’ effort on this project was supported by the NIBIB of the National Institutes of Health under award number R21EB026086. The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.

Conflict of interest

JL has received research funding from ViewRay and NH TherAguix.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Hall WA, Paulson ES, van der Heide UA, Fuller CD, Raaymakers BW, Lagendijk JJW, et al. The transformation of radiation oncology using real-time magnetic resonance guidance: A review. Eur J Cancer. (2019) 122:42–52. doi: 10.1016/j.ejca.2019.07.021

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Bruynzeel AME, Tetar SU, Oei SS, Senan S, Haasbeek CJA, Spoelstra FOB, et al. A prospective single-arm phase 2 study of stereotactic magnetic resonance guided adaptive radiation therapy for prostate cancer: Early toxicity results. Int J Radiat Oncol Biol Phys (2019) 105(5):1086–94. doi: 10.1016/j.ijrobp.2019.08.007

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Leeman JE, Cagney DN, Mak RH, Huynh MA, Tanguturi SK, Singer L, et al. Magnetic resonance–guided prostate stereotactic body radiation therapy with daily online plan adaptation: Results of a prospective phase 1 trial and supplemental cohort. Adv Radiat Oncol (2022) 7(5):100934. doi: 10.1016/j.adro.2022.100934

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Christiansen RL, Dysager L, Hansen CR, Jensen HR, Schytte T, Nyborg CJ, et al. Online adaptive radiotherapy potentially reduces toxicity for high-risk prostate cancer treatment. Radiother Oncol (2022) 167:165–71. doi: 10.1016/j.radonc.2021.12.013

PubMed Abstract | Google Scholar

5. Cusumano D, Teodoli S, Greco F, Fidanzio A, Boldrini L, Massaccesi M, et al. Experimental evaluation of the impact of low tesla transverse magnetic field on dose distribution in presence of tissue interfaces. Phys Med (2018) 53:80–5. doi: 10.1016/j.ejmp.2018.08.007

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Raaijmakers AJ, Raaymakers BW, Lagendijk JJ. Experimental verification of magnetic field dose effects for the MRI-accelerator. Phys Med Biol (2007) 52(14):4283–91. doi: 10.1088/0031-9155/52/14/017

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Spadea MF, Maspero M, Zaffino P, Seco J. Deep learning based synthetic-CT generation in radiotherapy and PET: A review. Med Phys (2021) 48(11):6537–66. doi: 10.1002/mp.15150

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Cusumano D, Boldrini L, Dhont J, Fiorino C, Green O, Gungor G, et al. Artificial intelligence in magnetic resonance guided radiotherapy: Medical and physical considerations on state of art and future perspectives. Phys Med (2021) 85:175–91. doi: 10.1016/j.ejmp.2021.05.010

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Owrangi AM, Greer PB, Glide-Hurst CK. MRI-Only treatment planning: benefits and challenges. Phys Med Biol (2018) 63(5):05TR1. doi: 10.1088/1361-6560/aaaca4

CrossRef Full Text | Google Scholar

10. Johnstone E, Wyatt JJ, Henry AM, Short SC, Sebag-Montefiore D, Murray L, et al. Systematic review of synthetic computed tomography generation methodologies for use in magnetic resonance imaging-only radiation therapy. Int J Radiat Oncol Biol Phys (2018) 100(1):199–217. doi: 10.1016/j.ijrobp.2017.08.043

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Edmund JM, Nyholm T. A review of substitute CT generation for MRI-only radiation therapy. Radiat Oncol (2017) 12(1):28. doi: 10.1186/s13014-016-0747-y

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Cusumano D, Lenkowicz J, Votta C, Boldrini L, Placidi L, Catucci F, et al. A deep learning approach to generate synthetic CT in low field MR-guided adaptive radiotherapy for abdominal and pelvic cases. Radiother Oncol (2020) 153:205–12. doi: 10.1016/j.radonc.2020.10.018

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Olberg S, Zhang H, Kennedy WR, Chun J, Rodriguez V, Zoberi I, et al. Synthetic CT reconstruction using a deep spatial pyramid convolutional framework for MR-only breast radiotherapy. Med Phys (2019) 46(9):4135–47. doi: 10.1002/mp.13716

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Fu J, Yang Y, Singhrao K, Ruan D, Chu FI, Low DA, et al. Deep learning approaches using 2D and 3D convolutional neural networks for generating male pelvic synthetic computed tomography from magnetic resonance imaging. Med Phys (2019) 46(9):3788–98. doi: 10.1002/mp.13672

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Tustison NJ, Avants BB, Cook PA, Zheng Y, Egan A, Yushkevich PA, et al. N4ITK: improved N3 bias correction. IEEE Trans Med Imaging. (2010) 29(6):1310–20. doi: 10.1109/TMI.2010.2046908

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Reinhold JC, Dewey BE, Carass A, Prince JL. Evaluating the impact of intensity normalization on MR image synthesis. Proc SPIE Int Soc Opt Eng (2019) 10949:109493H. doi: 10.1117/12.2513089

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Isola P, Zhu J-Y, Zhou T, Efros AA. Image-to-Image translation with conditional adversarial networks (2016). Available at: https://ui.adsabs.harvard.edu/abs/2016arXiv161107004I.

Google Scholar

18. Zhu J-Y, Park T, Isola P, Efros AA. Unpaired image-to-Image translation using cycle-consistent adversarial networks (2017). Available at: https://ui.adsabs.harvard.edu/abs/2017arXiv170310593Z.

Google Scholar

19. Fetty L, Lofstedt T, Heilemann G, Furtado H, Nesvacil N, Nyholm T, et al. Investigating conditional GAN performance with different generator architectures, an ensemble model, and different MR scanners for MR-sCT conversion. Phys Med Biol (2020) 65(10):105004. doi: 10.1088/1361-6560/ab857b

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Bahrami A, Karimian A, Fatemizadeh E, Arabi H, Zaidi H. A new deep convolutional neural network design with efficient learning capability: Application to CT image synthesis from MRI. Med Phys (2020) 47(10):5158–71. doi: 10.1002/mp.14418

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Bird D, Nix MG, McCallum H, Teo M, Gilbert A, Casanova N, et al. Multicentre, deep learning, synthetic-CT generation for ano-rectal MR-only radiotherapy treatment planning. Radiother Oncol (2021) 156:23–8. doi: 10.1016/j.radonc.2020.11.027

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Chen S, Qin A, Zhou D, Yan D. Technical note: U-net-generated synthetic CT images for magnetic resonance imaging-only prostate intensity-modulated radiation therapy treatment planning. Med Phys (2018) 45(12):5659–65. doi: 10.1002/mp.13247

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Largent A, Barateau A, Nunes JC, Mylona E, Castelli J, Lafond C, et al. Comparison of deep learning-based and patch-based methods for pseudo-CT generation in MRI-based prostate dose planning. Int J Radiat Oncol Biol Phys (2019) 105(5):1137–50. doi: 10.1016/j.ijrobp.2019.08.049

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Lei Y, Harms J, Wang T, Liu Y, Shu HK, Jani AB, et al. MRI-Only based synthetic CT generation using dense cycle consistent generative adversarial networks. Med Phys (2019) 46(8):3565–81. doi: 10.1002/mp.13617

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Maspero M, Savenije MHF, Dinkla AM, Seevinck PR, Intven MPW, Jurgenliemk-Schulz IM, et al. Dose evaluation of fast synthetic-CT generation using a generative adversarial network for general pelvis MR-only radiotherapy. Phys Med Biol (2018) 63(18):185001. doi: 10.1088/1361-6560/aada6d

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: MRI-guided therapy, adaptive radiotherapy, synthetic CT, deep learning, prostate radiotherapy

Citation: Hsu S-H, Han Z, Leeman JE, Hu Y-H, Mak RH and Sudhyadhom A (2022) Synthetic CT generation for MRI-guided adaptive radiotherapy in prostate cancer. Front. Oncol. 12:969463. doi: 10.3389/fonc.2022.969463

Received: 15 June 2022; Accepted: 26 August 2022;
Published: 23 September 2022.

Edited by:

Neil B. Desai, University of Texas Southwestern Medical Center, United States

Reviewed by:

Chenyang Shen, University of Texas Southwestern Medical Center, United States
Jeffrey Tuan, National Cancer Centre Singapore, Singapore
Jing Yuan, Hong Kong Sanatorium and Hospital, Hong Kong SAR, China

Copyright © 2022 Hsu, Han, Leeman, Hu, Mak and Sudhyadhom. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shu-Hui Hsu, shsu0@bwh.harvard.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.