Skip to main content

ORIGINAL RESEARCH article

Front. Bioeng. Biotechnol., 15 June 2023
Sec. Biomechanics

A benchmark study of convolutional neural networks in fully automatic segmentation of aortic root

Tingting YangTingting Yang1Guangyu Zhu
Guangyu Zhu1*Li CaiLi Cai2Joon Hock YeoJoon Hock Yeo3Yu MaoYu Mao4Jian Yang
Jian Yang4*
  • 1School of Energy and Power Engineering, Xi’an Jiaotong University, Xi’an, China
  • 2School of Mathematics and Statistics, Northwestern Polytechnical University, Xi’an, China
  • 3School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore, Singapore
  • 4Department of Cardiac Surgery, Xijing Hospital, The Fourth Military Medical University, Xi’an, China

Recent clinical studies have suggested that introducing 3D patient-specific aortic root models into the pre-operative assessment procedure of transcatheter aortic valve replacement (TAVR) would reduce the incident rate of peri-operative complications. Tradition manual segmentation is labor-intensive and low-efficient, which cannot meet the clinical demands of processing large data volumes. Recent developments in machine learning provided a viable way for accurate and efficient medical image segmentation for 3D patient-specific models automatically. This study quantitively evaluated the auto segmentation quality and efficiency of the four popular segmentation-dedicated three-dimensional (3D) convolutional neural network (CNN) architectures, including 3D UNet, VNet, 3D Res-UNet and SegResNet. All the CNNs were implemented in PyTorch platform, and low-dose CTA image sets of 98 anonymized patients were retrospectively selected from the database for training and testing of the CNNs. The results showed that despite all four 3D CNNs having similar recall, Dice similarity coefficient (DSC), and Jaccard index on the segmentation of the aortic root, the Hausdorff distance (HD) of the segmentation results from 3D Res-UNet is 8.56 ± 2.28, which is only 9.8% higher than that of VNet, but 25.5% and 86.4% lower than that of 3D UNet and SegResNet, respectively. In addition, 3D Res-UNet and VNet also performed better in the 3D deviation location of interest analysis focusing on the aortic valve and the bottom of the aortic root. Although 3D Res-UNet and VNet are evenly matched in the aspect of classical segmentation quality evaluation metrics and 3D deviation location of interest analysis, 3D Res-UNet is the most efficient CNN architecture with an average segmentation time of 0.10 ± 0.04 s, which is 91.2%, 95.3% and 64.3% faster than 3D UNet, VNet and SegResNet, respectively. The results from this study suggested that 3D Res-UNet is a suitable candidate for accurate and fast automatic aortic root segmentation for pre-operative assessment of TAVR.

1 Introduction

Transcatheter aortic valve replacement (TAVR) is a revolutionary technology for treating aortic stenosis (AS), with the advantages of no thoracotomy, less trauma, and faster recovery than surgical valve replacement (Elattar et al., 2014; Eleid and Holmes, 2017). With the emerging trend of expanding TAVR indications to younger and low-surgical-risk AS patients, the safety of TAVR has become a universal concern which requires careful patient-specific pre-operative planning and assessment (Walther et al., 2015; Leon et al., 2016; Puri et al., 2017; Youssefi et al., 2017; Mack et al., 2019). In an attempt to gap the limitation of the medical imaging assessment, recent developments provide powerful tools to support clinicians for precise patient-specific pre-operative morphological and functional assessment, such as image-based patient-specific 3D reconstruction, 3D printing, virtual reality (VR), augmented reality (AR) and numerical simulation techniques represented by finite element analysis (FEA) and computational fluid dynamics (CFD) (Levin et al., 2020; Jian et al., 2021; Otto et al., 2021; Wang et al., 2021; Chessa et al., 2022; Li et al., 2022).

As the basis of these applications, the rapid and accurate segmentation and 3D reconstruction of medical images have gradually become the key issues for personalized medicine. Traditional segmentation procedures mainly rely on manual operations requiring specialist knowledge, which are still time-consuming and labor-intensive (Kasel et al., 2013; Maragiannis et al., 2014; De Jaegere et al., 2016; Qian et al., 2017; Haghiashtiani et al., 2020). Especially when it comes to pre-operative TAVR assessment, manual segmentation generally cost tens of minutes due to the complex anatomy of the aortic root, including the ascending aorta, aortic valve, and coronary arteries. Thus, manual segmentation cannot provide a time-sensitive clinical recommendation with the increasingly large scale of patients receiving TAVR and the urgency of TAVR when patients with severe AS occasionally present with an acute decompensated state (Elbadawi et al., 2020). Moreover, the quality and reproducibility of manual segmentation are difficult to guarantee, and human errors induced in the manual segmentation stage could result in inaccurate or even wrong analysis results (Bertolini et al., 2022). These drawbacks prohibit its daily application in clinical centers.

To address the difficulties mentioned above, studies on the automatic segmentation of medical images started in the middle of the 1990s. The early stage of automatic image segmentation is featured by the methods based on supervised techniques such as active shaped models, which still require human intervention in extracting discriminant features from the images (Cootes et al., 1995; Cribier, 2002; Bengio et al., 2013; Kenny and Monaghan, 2015). With the growing clinical demand and the development of artificial intelligence techniques, fully automatic medical image segmentation methods based on deep learning (DL) overcame the drawbacks mentioned above and have become the technique of choice over the past decade (Smith et al., 2011; Litjens et al., 2017; Shen et al., 2017; Minaee et al., 2020). Among the segmentation-dedicated DL methods, 3D CNNs based architectures that were capable of analyzing volumetric medical image data provided powerful tools for the 3D reconstruction of lesions and organs (Thalji et al., 2014; Ranschaert et al., 2019; Panayides et al., 2020; Tsakanikas et al., 2020; Banerjee et al., 2021; Romaszko et al., 2021; Harrison et al., 2022). Among these, 3D UNet was the first medical image segmentation-dedicated 3D CNN, which was proposed in 2016 (Çiçek et al., 2016). It was first used to segment kidneys and then quickly extended to other tissues or organs, such as cardiac, brain and lung (Lundervold and Lundervold, 2019; Chen et al., 2020; Liu et al., 2021a). Since then, several novel 3D CNNs have been proposed for medical image segmentation, including VNet, variants of 3D UNet (3D Res-UNet, 3D Dense-UNet, 3D Attention-UNet, etc.), SegResNet, etc (Milletari et al., 2016; Fang et al., 2019; Myronenko, 2019; Liu et al., 2021b).

In recent years, these 3D CNNs have been gradually used to segment the aortic root automatically (Fassa et al., 2013; Fan et al., 2019; Ravichandran et al., 2019; Macruz et al., 2022; Sieren et al., 2022). Sieren et al. compared the performance between manual and 3D UNet segmentation based on CTA exams of the aorta of 191 patients and demonstrated that automated aorta segmentation by using 3D UNet is feasible (Sieren et al., 2022). Macruz et al. presented a 3D UNet-based framework for automated segmentation of the thoracic aorta in thoracic CT studies, which provides the basis for determining aortic diameter measurements and accurately predicting thoracic aortic aneurysms (Macruz et al., 2022). Ravichandran et al. evaluated the performance of 3D UNet and its variants in 3D segmentation of the aortic root under small samples by using a single evaluation metric Dice similarity coefficient (DSC), which indicated that the segmentation quality of the variants of 3D UNet is better than that of 3D UNet (Ravichandran et al., 2019). Although the current state-of-the-art literature focusing on CNN-based segmentation of aortic root is promising, there is a lack of a comprehensive evaluation of both segmentation quality and efficiency of the current popular 3D CNN architectures, especially for TAVR procedure which requires rapid and accurate pre-operative assessment.

Therefore, in this study, we compared the comprehensive segmentation performance of the four popular segmentation-dedicated 3D CNNs (3D UNet, VNet, 3D Res-UNet and SegResNet) under small sample low-dose CT datasets and realized a fully automated, accurate segmentation and reconstruction framework, which provides a reliable guarantee for pre-operative morphological and functional assessment of TAVR.

2 Materials and methods

2.1 Data preparation

2.1.1 Imaging data

In this study, 98 sets of anonymized chest CTA images acquired from low-dose multidetector 128-slice CT scanners (uCT 760, United Imaging Healthcare, Shanghai, China) were retrospectively collected from the pre-operative TAVR examination database of patients without calcification. All the scans included the aorta with 192–339 slices in the Z-axis. The slice size, thickness, and tube voltage are 512 × 512 pixels, 0.75 mm, and 120 kV, respectively. The pixel spacing of the scans varied between 0.25 and 0.60 mm. The axial spacing between slices is 0.5 mm. The image sets in Digital Imaging and Communications in Medicine (DICOM) format were manually inspected, and three image sets were excluded due to low signal-noise ratio or severe artifacts.

2.1.2 Label annotation

To reduce human error, two clinical engineers experienced in image segmentation and a senior radiologist participated in the labeling process. Firstly, the patient CT data in DICOM format were imported into Materialise Mimics version 21.0 software (Materialise, Leuven, Belgium) and three orthogonal sections were manually established based on the aortic annulus plane using the multiplane reconstruction function of the software (Ya’qoub et al., 2022). After that, the engineers individually labeled the mask of the aortic root with the help of the threshold segmentation algorithm (Underwood et al., 2000) (Figure 1). The mask covered the entire aortic root, an approximately 40 mm long segment starting from the aortic annulus plane. What’s more, the position of the coronary ostium is crucial for pre-operative assessment of TAVR to avoid coronary obstruction. Therefore, the coronary arteries about 10 mm distal from the coronary ostium were included in the mask as well. The average manual mask time for each patient is about 26 min. After the label annotation, the senior radiologist decides the label to be used in the experiments and the ground truth based on the quality assessment.

FIGURE 1
www.frontiersin.org

FIGURE 1. Schematic diagram of 3D label of aortic root. (A) Sagittal, coronal, horizontal view of label. (B) 3D visualization of label of aortic root.

2.1.3 Image pre-processing

Due to the original CTA images being too large and noisy, a series of data pre-processing is conducted before training and validation. Firstly, noise reduction and normalization of the images were performed to enhance the signal-to-noise ratio of the images. Next, the pixel and axial spacing were adjusted to 1 and 0.75 using bilinear interpolation to compress the data. In addition, to increase data diversity and improve training robustness, data augmentation was conducted on the pre-processed images following the methods described by Chlap et al. (2021) and He et al. (2011). All the volume data in the training set were augmented by randomly cropping into four sub-volumes, with the center being a foreground or background voxel based on the Positive-to-Negative Ratio of 1:1. Especially the size of the cropped sub-volumes used in this study was selected by the pre-tests which compared the segmentation performance of the three different sub-volume sizes (64 × 64 × 64 pixels, 96 × 96 × 96 pixels, 128 × 128 × 128 pixels) on the validation set. After the augmentation, the original training set size was multiplied four times.

2.2 CNN architectures for evaluation

Four 3D CNN architectures were evaluated in this study, including 3D UNet, 3D Res-UNet, VNet, and SegResNet. The specific differences between the evaluated 3D CNN architectures are listed in Table 1.

TABLE 1
www.frontiersin.org

TABLE 1. Model architecture comparison of different 3D CNNs.

2.2.1 3D UNet model

The 3D UNet model is a popular architecture based on 3D CNN, which consists of an encoder path and a decoder path (Çiçek et al., 2016). The 3D UNet used in this study has five layers with 16, 32, 64, 128 and 256 feature channels, respectively. In the encoder path, each layer contains two 3 × 3 × 3 convolutions, and then a 2 × 2 × 2 max pooling with strides of two in each dimension. In the decoder path, each layer consists of an up-convolution of 2 × 2 × 2 by strides of two in each dimension, followed by two 3 × 3 × 3 convolutions. The shortcut connections of equal-resolution layers in the encoder path provide high-resolution features for the decoder path. Instance normalization is used to prevent contrast shifting, ensuring input image contrast is not skewed by being batched with images with significantly different contrast ranges.

2.2.2 VNet model

The VNet model is another widely used architecture for 3D medical image segmentation (Milletari et al., 2016). In this architecture, a residual block was added to each layer, and the pooling operations in 3D UNet were replaced by 2 × 2 × 2 strided convolutions. In addition, the channel concatenation in the skip connections of 3D UNet was replaced with the element-wise summation. Benefiting from replacing pooling with convolution, VNet not only reduces memory demand during training but also makes the training process better understood and analyzed. In addition, the use of residual block alleviates the gradient vanishing problem in the deep networks. In this study, the VNet has 5 layers. The residual block in the first, second, and rest layers contain one, two, and three 5 × 5 × 5 convolutions, respectively.

2.2.3 3D Res-UNet

The 3D Res-UNet model is one of the most famous derivations of the 3D UNet model. Compared with 3D UNet, 3D Res-UNet improved the accuracy and efficiency in image segmentation by introducing residual block into the down-sampling path and up-sample unit in the up-sampling path (Yang et al., 2019). It also applied the 2 × 2 × 2 convolutions to replace the pooling operation. Furthermore, parametric rectifying linear units allow the network to learn a better activation, improving segmentation performance.

2.2.4 SegResNet

The SegResNet model was recently proposed and performed well in 3D MRI brain tumor segmentation as an encoder-decoder-based asymmetrical 3D CNN architecture with a larger encoder to extract image features and a smaller decoder to segment the image (Myronenko, 2019). Each encoder layer has a different number of ResNet-like blocks. Different from other networks, each decoder layer first begins with upsizing to reduce the number of features by using 1 × 1 × 1 convolution and double the spatial dimension by using 3D bilinear up-sampling, followed by the addition encoder output of the equivalent spatial level.

2.3 Training and testing

2.3.1 Implementation details

All the 3D CNNs were trained for 500 epochs using the Adam optimizer with the following parameters: learning rate = 0.0001, β1 = 0.9, β2 = 0.999, ε = 1e-8, batch size = 1. The Dice coefficient loss function (Diceloss) is used as loss function to evaluate the convergency of the training in all experiments. The definition of Diceloss is given in Eq. 1, where ygt and ypred are the ground truth and binary predictions from CNNs, ε was a constant set to 1e-5.

Diceloss=12ygtypred+εygt+ypred+ε(1)

The entire dataset was randomly divided into a training set (76 images) and a validation set (19 images). To explore the impact of the size of the training set on the segmentation performance, a sensitivity analysis was first conducted by setting up the three different sizes of the training set, 25%, 50% and 100% of the entire training set, respectively. After that, a five-fold cross-validation scheme was adopted for each 3D CNN under the entire training set. Due to the large original image, a 160 × 160 × 160 pixels sliding window is used to divide the validation set images to improve computing efficiency.

All experiments were implemented on a workstation powered by an NVIDIA GeForce RTX 3090 GPU with 24 GB of RAM. The code was implemented with PyTorch 1.9.1 in Windows 10. The overall experimental design of the workflow diagram is shown in Figure 2.

FIGURE 2
www.frontiersin.org

FIGURE 2. The overall experimental design of the workflow diagram.

2.3.2 Evaluation metrics

To quantitively compare the segmentation performances of the 3D CNNs, five metrics widely used in segmentation quality assessment were evaluated in this study, including recall, Dice similarity coefficient (DSC), Jaccard index, Hausdorff distance (HD) and 3D deviation. The specific definitions of the metrics are listed below.

Recall, also called Sensitivity or True Positive Rate (TPR), measures the portion of positive voxels in the ground truth identified as positive by the predictions of CNNs (Eq. 2).

Recall=ygtypredygt(2)

DSC is the pair-wise overlap ratio between the predictions and the ground truth, ranging from 0 to 1 (Eq. 3).

DSC=2ygtypredygt+ypred(3)

Jaccard index also reflects the overlap ratio between the intersection and union of predictions and the ground truth with the range between 0 and 1 (Eq. 4).

Jaccard=ygtypredygtypred(4)

HD measures the distance between ygt and ypred by indicating the greatest distances from a point in ypred to the closest point in ygt (Eq. 5).

HD=maxdygtypred,dypredygt=maxmaxxygtminyypreddx,y,maxyypredminxygtdx,y(5)

The CNN-based segmentation results and ground truth were 3D reconstructed using the Python platform and VTK library. The 3D deviation metrics between CNN-based segmentation and ground truth were analyzed using Geomagic software (Geomagic Inc., Research Triangle Park, NC). In addition, the segmentation efficiency of the 3D CNNs was compared by using the average segmentation time spent on the validation set.

3 Results

3.1 Impacts of sub-volume size on model performance

To explore the suitable sub-volume size of the training set for our study, the pre-tests of the impacts of sub-volume size on 3D CNNs performance were conducted. The three different sub-volume sizes (64 × 64 × 64 pixels, 96 × 96 × 96 pixels, 128 × 128 × 128 pixels) of the training set were selected to train and validate 3D Res-UNet for just one fold under the entire training set. Subject to hardware limitation, the sub-volume size of 128 × 128 × 128 pixels was the largest size implemented to train 3D Res-UNet in this study. The results showed that for these three sub-volume sizes, the convergence rate was negatively related to the sub-volume size, while segmentation quality performance is positively related to that (Figure 3) (Figure 4). This might indicate that the larger sub-volume size would produce better segmentation quality. In the subsequent studies, 304 (76 × 4) augmented 128 × 128 × 128 pixels sub-volumes were employed to train different 3D CNNs.

FIGURE 3
www.frontiersin.org

FIGURE 3. The influence of different sizes of sub-volumes on the convergence of the loss function during training.

FIGURE 4
www.frontiersin.org

FIGURE 4. The influence of different sizes of sub-volumes on the segmentation performance of validation set samples.

3.2 Impacts of training dataset sizes on model performance

3.2.1 Training performance

Figure 5 shows the Diceloss profiles of training and validation under different sizes of the training set. It demonstrated that the convergence rate of all the 3D CNNs was positively related to the size of the training set. When trained by 25% of the entire training set, only VNet performed a notable overfitting phenomenon which might be caused by its higher model complexity and the limited dataset. When trained by 100% of the entire training set, the training set error and the validation set error of 3D CNNs had the same downward trend, indicating that all 3D CNNs found a good fit between under-fitting and over-fitting in the training process. Therefore, the subsequent five-fold validation was conducted to compare the segmentation performance of the four 3D CNNs using the entire training set and the validation set (Section 3.3). In addition, compared with VNet and SegResNet, 3D UNet and 3D Res-UNet performed a faster convergence rate by reaching a training Diceloss of nearly 0.15 in less than 50 epochs with 100% of the entire training set.

FIGURE 5
www.frontiersin.org

FIGURE 5. The performance of the training and validation process. The changes of the Diceloss of (A) 3D UNet, (B) VNet, (C) 3D Res-UNet and (D) SegResNet during training and validation process under different training set sizes.

3.2.2 Segmentation performance

Figure 6 shows the segmentation performance of the 3D CNNs under the different sizes of the training sets. The segmentation performance of all the 3D CNNs was affected by the size of the training dataset. Among these 3D CNNs, the segmentation quality of VNet is more sensitive to the size of the training set. Table 2 showed that the evaluation metrics notably deteriorated with the decrease of training set size in VNet, while the performance deterioration in other CNNs is mild.

FIGURE 6
www.frontiersin.org

FIGURE 6. The segmentation performance of the four 3D CNNs trained by different size of the training set, including (A) Recall, (B) DSC, (C) Jaccard, and (D) HD metrics.

TABLE 2
www.frontiersin.org

TABLE 2. The segmentation quality of the four 3D CNNs under the different size of the training dataset.

In detail, when the size of the training set was reduced to 50% of the entire training set, the segmentation quality of the 3D CNNs was comparable on the evaluation metrics. However, when the size of the training set was reduced to 25% of the entire training set, the segmentation quality of VNet showed the most notable deterioration on recall, DSC, and Jaccard by 3.83%, 6.97%, and 11.54%, while that of the other three 3D CNNs was decreased by less than 4%. Even though the recall, DSC, and Jaccard of 3D UNet, 3D Res-UNet and SegResNet did not show a great decrease, HD of them was sharply deteriorated by 13.96%, 56.41% and 34.67%, respectively, indicating that the inadequate training dataset could severely affect the CNN-based segmentation performance.

3.3 Segmentation performance comparison

3.3.1 Segmentation quality evaluation

Table 3 shows the segmentation performance metrics of the 3D CNNs. All the 3D CNNs achieved a mean DSC higher than 0.95 over the five-fold cross-validation, and the differences in mean DSC between groups were less than 1%. Similar to DSC, the differences in other metrics between groups were less than 1%, including recall and Jaccard index. In terms of HD, VNet (7.72 ± 1.89) and 3D Res-UNet (8.56 ± 2.28) performed better than SegResNet (10.74 ± 5.41) and 3D UNet (15.96 ± 8.71). The difference between VNet and 3D Res-UNet is within 10%, while the difference between VNet and 3D UNet as well as SegResNet is beyond 30%.

TABLE 3
www.frontiersin.org

TABLE 3. Cross validation results of different models.

To provide the details of 3D deviations between CNN-based segmentation and ground truth, 3D surface reconstruction using the marching cubes algorithm based on Python platform and VTK library was conducted to obtain the 3D reconstruction models from CNN-based segmentation and ground truth. Specifically, due to the bilinear interpolation conducted to compress the data in image pre-processing, it is essential to re-sample the size of CNN-based automatic segmentation to the original image size before 3D reconstruction. After automatic 3D reconstruction, we processed the reconstruction models from CNN-based segmentation and ground truth by surface smoothing (Figure 7). To evaluate the 3D deviations between the 3D reconstruction models from CNN-based segmentation and ground truth, four metrics widely used in 3D model deviation analysis were evaluated in this study, including maximum distance, average distance, standard deviation (STD), and root mean square (RMS) value (Table 4). It was found that the segmentation results of 3D UNet have the largest maximum and averaged deviation from the ground truth compared with other evaluated 3D CNNs, which might be attributed to the lack of residual block.

FIGURE 7
www.frontiersin.org

FIGURE 7. From left to right, automatic 3D reconstruction visualization results based on 3D UNet, VNet, 3D Res-UNet and SegResNet, respectively.

TABLE 4
www.frontiersin.org

TABLE 4. 3D deviation analysis based on automatic 3D reconstruction results.

Furthermore, the results from this study also suggested that even though the CNN-based segmentation results have the same DSC and HD, there is no necessary connection between the 3D deviation locations and classical evaluation metrics. In some cases of high DSC (>0.94) and low HD (<7), a large deviation was observed at the location of interest, including the aortic valve and the bottom of the aortic root, which is crucial for pre-operative morphological and the subsequent functional assessment of TAVR (Figure 8A). Therefore, it is incomplete and limited to evaluate the segmentation performance of 3D CNNs on the aortic root solely relying on the classical metrics. In response to this problem, a statistical analysis of the maximum deviation location was performed based on the 3D deviation cloud between the CNN-based segmentation and ground truth, including ascending aorta, coronary arteries, aortic valve, and the bottom of the aortic sinus (Figure 9). It indicated that different 3D CNNs exhibited different tendencies in the four prone deviation local positions. Despite 3D Res-UNet and VNet being more prone to conduct a large deviation at ascending aorta and coronary arteries, they performed better at the bottom of the aortic root compared with 3D UNet and SegResNet, which would be more accurate for pre-operative morphological and functional assessment of TAVR.

FIGURE 8
www.frontiersin.org

FIGURE 8. 3D deviation graph of 3D automatic reconstruction models of manual segmentation and 3D automatic reconstruction models of automatic segmentation for two cases with higher DSC (A) and lower DSC (B) in the validation set. Four each case, from left to right, we show the 3D deviation map of 3D reconstruction models of manual segmentation and 3D reconstruction models of automatic segmentation based on 3D UNet, VNet, 3D Res-UNet, SegResNet, respectively. The above is a side perspective, the below is a bottom view.

FIGURE 9
www.frontiersin.org

FIGURE 9. Statistics of the large deviation local position of automatic 3D reconstruction based on the four CNNs in the validation set. A, B, C, and D represent the ascending aorta, coronary arteries, aortic valve and the bottom of the aortic root, respectively.

3.3.2 Segmentation efficiency evaluation

Regarding segmentation efficiency, compared with manual segmentation, all the evaluated CNN-based average segmentation time consumed was reduced by more than 500 times from 26 min to several seconds (Table 3). 3D Res-UNet (0.1s) and SegResNet (0.28s) performed the faster segmentation time compared with that of 3D UNet (1.13s) and VNet (2.14s). Among these, the average automatic segmentation time of 3D Res-UNet was significantly shorter than that of other 3D CNNs (p < 0.001), which are 91.2%, 95.3%, and 64.3% faster than 3D UNet, VNet and SegResNet, respectively. Therefore, 3D Res-UNet shows greater potential for real-time pre-operative morphological assessment of TAVR among the evaluated 3D CNNs.

4 Discussion

In this study, the comprehensive performances of the four popular segmentation-dedicated 3D CNNs for the aortic root segmentation were quantitively evaluated. To the best of our knowledge, this is the first study focusing on the comprehensive comparison of CNN-based aortic root segmentation performance for the pre-operative assessment of TAVR. The automatic segmentation quality and efficiency results suggest that 3D Res-UNet is the most accurate and efficient 3D CNN architecture, which could provide a rapid and reliable guarantee for pre-operative morphological assessment and subsequent functional assessment of TAVR.

4.1 Image pre-processing

This study introduced a series of image pre-processing before training and validating 3D CNNs. Firstly, due to statistical uncertainty in all physical measurements, the inevitable noise is introduced in CT images (Diwakar and Kumar, 2018). Therefore, the median filter was conducted in the image pre-processing to improve the image quality in this study which has been demonstrated to help improve the lymph segmentation performance effectively (Zhang et al., 2021). In addition, due to the wide range of Hounsfield Unit (HU) in CT images from −1024 to 3071, it is necessary to enhance the contrast between the background and the target area by image intensity normalization, which has been proven to be beneficial to improve the segmentation performance based on CNNs (Jacobsen et al., 2019). What’s more, subject to hardware limitations, this study conducted bilinear interpolation as the image resampling method which is widely used in data compression to reduce the image resolution and further reduce the data size. Furthermore, data augmentation has become an effective solution to the limited data problem by increasing the sufficiency and diversity of the training set, especially for acquiring limited medical images (Shorten and Khoshgoftaar, 2019). Therefore, this study conducted one of the basic image augmentation methods by randomly cropping into four sub-volumes which has been proven to help reduce overfitting and the error rate of CNN models (Krizhevsky et al., 2017).

4.2 The selection of sub-volume size

Several studies have suggested that choosing an adequate sub-volume size in the training of 3D CNN is one of the essential factors in achieving a good segmentation result (Sabottke and Spieler, 2020). However, sub-volume size selection is a tradeoff between computation resource, time, and accuracy. In the research field of CNN-based aorta segmentation, Ravichandran et al. deployed 64 × 64 × 64 pixels sub-volume training on three NVIDIA TITAN V GPUs with 36 GB RAM (Ravichandran et al., 2019; Rukundo, 2023). Fan et al. cropped a sub-volume size of 48 × 128 × 128 pixels with a GTX 1080Ti GPU (Fan et al., 2019). To explore the suitable sub-volume size for our study, the pre-tests of the impacts of sub-volume size on 3D CNN performance were conducted. Although the pre-tests showing a larger sub-volume size would result in better segmentation quality, it also demands a faster GPU and larger RAM. This result agrees well with the previous studies (Sabottke and Spieler, 2020). Thus, subject to hardware limitation, the sub-volume size of 128 × 128 × 128 pixels was selected in this study.

4.3 The selection of training set size

In general, DL models represented by 3D CNNs require a large amount of training data to effectively learn the target task (Ghadimi et al., 2021). However, due to the scarcity of annotated medical images, it is hard to build a large dataset at the initial stage of developing such 3D CNNs in the field of medical image segmentation. Therefore, it is significant to investigate the CNN-based segmentation performance of the aortic root under the small sample size. Due to that the selecting of an appropriate dataset size for 3D CNNs is still a challenging open problem (Cheung et al., 2021; Nu˜nez-Garcia et al., 2019), this study conducted a sensitivity analysis to explore the impact of the training set size on the CNN-based segmentation performance of the aortic root.

The results show that the training performance and segmentation performance of VNet is most sensitive to the size of the training dataset, which might be due to its more network parameters caused by the larger convolutional kernel size (Table 1), further leading to the need for more training data to guarantee the segmentation performance. Compared with VNet, even though the recall, DSC and Jaccard of 3D UNet, 3D Res-UNet and SegResNet did not show a notable decrease with the reduction of the training dataset, HD of them was sharply deteriorated, which might indicate that less training data could severely worsen the capacity of 3D CNNs to capture complex boundary details in the aortic root segmentation. Therefore, when selecting a suitable 3D CNN under small samples, it should be considered that CNN models with higher complexity could bring the problem of reliance on large datasets, despite their excellent learning ability.

4.4 Segmentation performance

4.4.1 Segmentation quality

4.4.1.1 Classical metrics analysis

Segmentation quality is the most important aspect in evaluating CNN-based automatic segmentation. Several classical segmentation quality assessment metrics have been proposed, including recall, DSC, Jaccard index, and HD.

Among the metrics, similarity coefficients (SCs), such as DSC and Jaccard index, are the essential spatial-based metrics representing the overlap ratio between segmentation results and ground truth. As one of the classical SCs, DSC was widely used to evaluate image segmentation quality, which is traditionally considered positively correlated with segmentation quality. However, its threshold of distinguishing good and poor segmentation quality varied by several factors, including segmentation objects, CNN architectures, image quality, and dataset size. In the field of cardiovascular segmentation, a DSC of 0.7 has been well-accepted as the segmentation quality threshold (Robinson et al., 2018; Robinson et al., 2019; Hann et al., 2021). In this study, all the evaluated 3D CNNs achieved a mean DSC higher than 0.95 over the five-fold cross-validation, and the differences in mean DSC between groups were less than 1%. This result is comparable with the previous studies focused on aorta segmentation based on private CT datasets (Zou et al., 2004; Fan et al., 2019; Ravichandran et al., 2019; Macruz et al., 2022; Sieren et al., 2022), which has an average DSC of 0.917 ± 0.054 (Table 5). In addition, the Jaccard index derived from this study also agrees well with the related study (Fan et al., 2019). The comparison results demonstrated that the CNN-based automatic segmentation framework established in this study was essentially reliable.

TABLE 5
www.frontiersin.org

TABLE 5. Comparison of the segmentation performance of this study with the previous studies.

However, a single metric (DSC) cannot comprehensively reflect the segmentation quality. Therefore, HD is conducted to evaluate the segmentation quality based on the spatial distance that can quantitively indicate the largest error between the boundary of CNN-based segmentation and ground truth. HD is generally sensitive to boundary morphology, which makes it suitable for evaluating cases where the complex boundary is of interest (Taha and Hanbury, 2015). Our results showed that the average HD of 3D UNet (15.96 ± 8.71) is notably higher than that of VNet (7.72 ± 1.89), 3D Res-UNet (8.56 ± 2.28), and SegResNet (10.74 ± 5.41), which suggested that the residual block in the latter three 3D CNNs contributed to the better capture of complex boundary details in the aortic root segmentation. This result is consistent with the theoretical analysis of the function of the residual block (He et al., 2016).

4.4.1.2 3D deviation analysis

Although the above classical metrics are widely used for segmentation quality assessment, they are based on the overall morphological characteristics, which cannot reflect the important geometric details for pre-operative morphological and the subsequent functional assessment of TAVR. Therefore, the 3D deviation analysis was carried out to demonstrate the detailed local errors between CNN-based segmentation and ground truth intuitively. As shown in Figure 9, all the evaluated 3D CNNs were prone to large deviations in the ascending aorta and coronary arteries due to the absence of anatomical boundaries (Figure 8B). In addition, the thin structures and ambiguous boundaries of the aortic valve are difficult to distinguish in CTA images, making it also difficult for CNNs to accurately segment (Figure 8A). Therefore, all the CNN-based segmentation are prone to large deviations in the ascending aorta, coronary arteries, and aortic valve.

Even though the deviation at the distal ends of the aortic root, such as the ascending aorta and coronary arteries, does not affect the pre-operative morphological assessment of TAVR, the segmentation quality of the aortic valve and the bottom of the aortic root is critical for successful TAVR. Therefore, compared with 3D UNet and SegResNet, 3D Res-UNet and VNet performed better at the aortic valve and the bottom of the aortic root, which would be more accurate for the pre-operative morphological and functional assessment of TAVR.

4.4.2 Segmentation efficiency

In the cardiovascular field, manual segmentation is labor-intensive and time-consuming. The segmentation time varies from minutes to hours for different segmented objects, which limits the clinical applications of patient-specific analysis that rely on image segmentation, especially for TAVR which requires rapid preoperative assessment (Byrne et al., 2016).

In our study, benefitting from the improvement of the hardware platform, all 3D CNNs shorten the segmentation time by more than 500 times compared with the manual segmentation time, which agrees well with the related study in order of magnitude (Bratt et al., 2019). Among these 3D CNNs, the average segmentation time of 3D Res-UNet and SegResNet is less than 1 s. VNet uses the largest convolution kernel size which tends to be disproportionally expensive in terms of computational cost, resulting in the longest segmentation time (Szegedy et al., 2016). 3D UNet uses max pooling layers as the traditional up/down sampling function enlarging the model size and memory occupation which may lead to a longer segmentation time (Ayachi et al., 2020). In comparison, both 3D Res-UNet and SegResNet use the smaller convolutional kernel size and convolution operation instead of max pooling layers as the up/down sampling function which greatly reduces the segmentation time across orders of magnitude.

4.5 Limitation

Although the results of this study are promising, there are some limitations. Firstly, although this study provides a relatively comprehensive assessment of the segmentation performance of the 3D CNNs under the small dataset, it should be noted that the dataset of this study was collected from a single clinical center and the same scanner vendor, which limits the generalizability across multiple centers and scanner vendors (Campello et al., 2021). In addition, this study only used the annotation results of one annotator as a benchmark and did not consider the differences between different annotators. The next step is to compare the results of the same sample annotated by different annotators. Furthermore, FEA and CFD simulation will be conducted to evaluate the influence of CNN-based segmentation results on the mechanical and hemodynamic environment for accurate patient-specific preoperative functional assessment of TAVR.

5 Conclusion

The segmentation performances of four segmentation-dedicated 3D CNNs on the aortic root segmentation were evaluated in this study. Although 3D Res-UNet and VNet are evenly matched in the aspect of classical segmentation quality metrics and 3D deviation location of interest analysis, 3D Res-UNet is the most efficient CNN architecture which provides a rapid and reliable guarantee for pre-operative morphological assessment and subsequent functional assessment of TAVR.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

GZ, TY contributed to the conception, design and manuscript draft of this study. TY contributed to the data analysis and visualization. LC, JY and JHY contributed to the data interpretation and commented on the manuscript. JY and YM contributed to the data collection and preparation. All authors contributed to the article and approved the submitted version.

Funding

This research was funded by the National Natural Science Foundation of China (NSFC) (12272289, 12271440).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ayachi, R., Afif, M., Said, Y., and Atri, M. (2020). “Strided convolution instead of max pooling for memory efficiency of convolutional neural networks BT,” in Proceedings of the 8th international conference on sciences of electronics, technologies of information and telecommunications (SETIT’18). Editors M. S. Bouhlel, and S. Rovetta (Berlin, Germany: Springer International Publishing), 234–243.

CrossRef Full Text | Google Scholar

Banerjee, A., Camps, J., Zacur, E., Andrews, C. M., Rudy, Y., Choudhury, R. P., et al. (2021). A completely automated pipeline for 3D reconstruction of human heart from 2D cine magnetic resonance slices. Philos. Trans. R. Soc. A Math. Phys. Eng. Sci. 379, 20200257. doi:10.1098/rsta.2020.0257

CrossRef Full Text | Google Scholar

Bengio, Y., Courville, A., and Vincent, P. (2013). Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 35, 1798–1828. doi:10.1109/TPAMI.2013.50

PubMed Abstract | CrossRef Full Text | Google Scholar

Bertolini, M., Luraghi, G., Belicchi, I., Migliavacca, F., and Colombo, G. (2022). Evaluation of segmentation accuracy and its impact on patient-specific CFD analysis. Int. J. Interact. Des. Manuf. 16, 545–556. doi:10.1007/s12008-021-00802-z

CrossRef Full Text | Google Scholar

Bratt, A., Kim, J., Pollie, M., Beecy, A. N., Tehrani, N. H., Codella, N., et al. (2019). Machine learning derived segmentation of phase velocity encoded cardiovascular magnetic resonance for fully automated aortic flow quantification. J. Cardiovasc. Magn. Reson. 21, 1. doi:10.1186/s12968-018-0509-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Byrne, N., Velasco Forte, M., Tandon, A., Valverde, I., and Hussain, T. (2016). A systematic review of image segmentation methodology, used in the additive manufacture of patient-specific 3D printed models of the cardiovascular system. JRSM Cardiovasc. Dis. 5, 204800401664546. doi:10.1177/2048004016645467

PubMed Abstract | CrossRef Full Text | Google Scholar

Campello, V. M., Gkontra, P., Izquierdo, C., Martin-Isla, C., Sojoudi, A., Full, P. M., et al. (2021). Multi-centre, multi-vendor and multi-disease cardiac segmentation: The M&amp;Ms challenge. IEEE Trans. Med. Imaging 40, 3543–3554. doi:10.1109/TMI.2021.3090082

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, C., Qin, C., Qiu, H., Tarroni, G., Duan, J., Bai, W., et al. (2020). Deep learning for cardiac image segmentation: A review. Front. Cardiovasc. Med. 7, 25. doi:10.3389/fcvm.2020.00025

PubMed Abstract | CrossRef Full Text | Google Scholar

Chessa, M., Van De Bruaene, A., Farooqi, K., Valverde, I., Jung, C., Votta, E., et al. (2022). Three-dimensional printing, holograms, computational modelling, and artificial intelligence for adult congenital heart disease care: An exciting future. Eur. Hear. J. 43, 2672–2684. doi:10.1093/eurheartj/ehac266

CrossRef Full Text | Google Scholar

Cheung, W. K., Bell, R., Nair, A., Menezes, L. J., Patel, R., Wan, S., et al. (2021). A computationally efficient approach to segmentation of the aorta and coronary arteries using deep learning. IEEE Access 9, 108873–108888. doi:10.1109/ACCESS.2021.3099030

PubMed Abstract | CrossRef Full Text | Google Scholar

Chlap, P., Min, H., Vandenberg, N., Dowling, J., Holloway, L., and Haworth, A. (2021). A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 65, 545–563. doi:10.1111/1754-9485.13261

PubMed Abstract | CrossRef Full Text | Google Scholar

Çiçek, Ö., Abdulkadir, A., Lienkamp, S. S., Brox, T., and Ronneberger, O. (2016). 3D U-net: Learning dense volumetric segmentation from sparse annotation. Lect. Notes Comput. Sci. 2016, 424–432. doi:10.1007/978-3-319-46723-8_49

CrossRef Full Text | Google Scholar

Cootes, T. F., Taylor, C. J., Cooper, D. H., and Graham, J. (1995). Active shape models-their training and application. Comput. Vis. Image Underst. 61, 38–59. doi:10.1006/cviu.1995.1004

CrossRef Full Text | Google Scholar

Cribier, A., Eltchaninoff, H., Bash, A., Borenstein, N., Tron, C., Bauer, F., et al. (2002). Percutaneous transcatheter implantation of an aortic valve prosthesis for calcific aortic stenosis: First human case description. Circulation 106, 3006–3008. doi:10.1161/01.CIR.0000047200.36165.B8

PubMed Abstract | CrossRef Full Text | Google Scholar

De Jaegere, P., De Santis, G., Rodriguez-Olivares, R., Bosmans, J., Bruining, N., Dezutter, T., et al. (2016). Patient-specific computer modeling to predict aortic regurgitation after transcatheter aortic valve replacement. JACC Cardiovasc. Interv. 9, 508–512. doi:10.1016/j.jcin.2016.01.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Diwakar, M., and Kumar, M. (2018). A review on CT image noise and its denoising. Biomed. Signal Process. Control 42, 73–88. doi:10.1016/j.bspc.2018.01.010

CrossRef Full Text | Google Scholar

Elattar, M. A., Wiegerinck, E. M., Planken, R. N., Vanbavel, E., Van Assen, H. C., Baan, J., et al. (2014). Automatic segmentation of the aortic root in CT angiography of candidate patients for transcatheter aortic valve implantation. Med. Biol. Eng. Comput. 52, 611–618. doi:10.1007/s11517-014-1165-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Elbadawi, A., Elgendy, I. Y., Mentias, A., Saad, M., Mohamed, A. H., Choudhry, M. W., et al. (2020). Outcomes of urgent versus nonurgent transcatheter aortic valve replacement. Catheter. Cardiovasc. Interv. 96, 189–195. doi:10.1002/ccd.28563

PubMed Abstract | CrossRef Full Text | Google Scholar

Eleid, M. F., and Holmes, D. R. (2017). Transcatheter aortic valve replacement: State of the art and future directions. Annu. Rev. Med. 68, 15–28. doi:10.1146/annurev-med-101615-020427

PubMed Abstract | CrossRef Full Text | Google Scholar

Fan, B., Tomii, N., Tsukihara, H., Maeda, E., Yamauchi, H., Nawata, K., et al. (2019). Attention-guided decoder in dilated residual network for accurate aortic valve segmentation in 3D CT scans. Berlin, Germany: Springer International Publishing. doi:10.1007/978-3-030-33327-0_15

CrossRef Full Text | Google Scholar

Fang, Y., Wang, J., Chen, S., Shen, S., Zhang, Z., and Hu, W. (2019). Impact of dataset size on deep learning-based auto segmentation for head and neck cancer. Int. J. Radiat. Oncol. 105, E129–E130. doi:10.1016/j.ijrobp.2019.06.2258

CrossRef Full Text | Google Scholar

Fassa, A.-A., Himbert, D., and Vahanian, A. (2013). Mechanisms and management of TAVR-related complications. Nat. Rev. Cardiol. 10, 685–695. doi:10.1038/nrcardio.2013.156

PubMed Abstract | CrossRef Full Text | Google Scholar

Ghadimi, S., Auger, D. A., Feng, X., Sun, C., Meyer, C. H., Bilchick, K. C., et al. (2021). Fully-automated global and segmental strain analysis of DENSE cardiovascular magnetic resonance using deep learning for segmentation and phase unwrapping. J. Cardiovasc. Magn. Reson. 23, 20. doi:10.1186/s12968-021-00712-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Haghiashtiani, G., Qiu, K., Sanchez, J. D. Z., Fuenning, Z. J., Nair, P., Ahlberg, S. E., et al. (2020). 3D printed patient-specific aortic root models with internal sensors for minimally invasive applications. Sci. Adv. 6, 1–12. doi:10.1126/sciadv.abb4641

PubMed Abstract | CrossRef Full Text | Google Scholar

Hann, E., Popescu, I. A., Zhang, Q., Gonzales, R. A., Barutçu, A., Neubauer, S., et al. (2021). Deep neural network ensemble for on-the-fly quality control-driven segmentation of cardiac MRI T1 mapping. Med. Image Anal. 71, 102029. doi:10.1016/j.media.2021.102029

PubMed Abstract | CrossRef Full Text | Google Scholar

Harrison, K., Pullen, H., Welsh, C., Oktay, O., Alvarez-Valle, J., and Jena, R. (2022). Machine learning for auto-segmentation in radiotherapy planning. Clin. Oncol. 34, 74–88. doi:10.1016/j.clon.2021.12.003

PubMed Abstract | CrossRef Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27-30 June 2016 (IEEE) 770–778. doi:10.1109/CVPR.2016.90

CrossRef Full Text | Google Scholar

He, L., Chao, Y., and Suzuki, K. (2011). Two efficient label-equivalence-based connected-component labeling algorithms for 3-D binary images. IEEE Trans. Image Process. 20, 2122–2134. doi:10.1109/TIP.2011.2114352

PubMed Abstract | CrossRef Full Text | Google Scholar

Jacobsen, N., Deistung, A., Timmann, D., Goericke, S. L., Reichenbach, J. R., and Güllmar, D. (2019). Analysis of intensity normalization for optimal segmentation performance of a fully convolutional neural network. Z. Med. Phys. 29, 128–138. doi:10.1016/j.zemedi.2018.11.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Jian, Y., Lee, A., and Vida, L. V. (2021). in Cardiovascular 3D printing. Editors J. Yang, A. P.-W. Lee, and V. L. Vida (Singapore: Springer Singapore). doi:10.1007/978-981-15-6957-9

CrossRef Full Text | Google Scholar

Kasel, A. M., Cassese, S., Bleiziffer, S., Amaki, M., Hahn, R. T., Kastrati, A., et al. (2013). Standardized imaging for aortic annular sizing. JACC Cardiovasc. Imaging 6, 249–262. doi:10.1016/j.jcmg.2012.12.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Kenny, C., and Monaghan, M. (2015). How to assess aortic annular size before transcatheter aortic valve implantation (TAVI): The role of echocardiography compared with other imaging modalities. Heart 101, 727–736. doi:10.1136/heartjnl-2013-304689

PubMed Abstract | CrossRef Full Text | Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2017). ImageNet classification with deep convolutional neural networks. Commun. ACM 60, 84–90. doi:10.1145/3065386

CrossRef Full Text | Google Scholar

Leon, M. B., Smith, C. R., Mack, M. J., Makkar, R. R., Svensson, L. G., Kodali, S. K., et al. (2016). Transcatheter or surgical aortic-valve replacement in intermediate-risk patients. N. Engl. J. Med. 374, 1609–1620. doi:10.1056/NEJMoa1514616

PubMed Abstract | CrossRef Full Text | Google Scholar

Levin, D., Mackensen, G. B., Reisman, M., McCabe, J. M., Dvir, D., and Ripley, B. (2020). 3D printing applications for transcatheter aortic valve replacement. Curr. Cardiol. Rep. 22, 23. doi:10.1007/s11886-020-1276-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, J., Yan, W., Wang, W., Wang, S., and Wei, L. (2022). Comparison of balloon-expandable valve and self-expandable valve in transcatheter aortic valve replacement: A patient-specific numerical study. J. Biomech. Eng. 144, 104501. doi:10.1115/1.4054332

PubMed Abstract | CrossRef Full Text | Google Scholar

Litjens, G., Kooi, T., Bejnordi, B. E., Setio, A. A. A., Ciompi, F., Ghafoorian, M., et al. (2017). A survey on deep learning in medical image analysis. Med. Image Anal. 42, 60–88. doi:10.1016/j.media.2017.07.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, X., Li, K. W., Yang, R., and Geng, L. S. (2021a). Review of deep learning based automatic segmentation for lung cancer radiotherapy. Front. Oncol. 11, 717039–717116. doi:10.3389/fonc.2021.717039

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, X., Song, L., Liu, S., and Zhang, Y. (2021b). A review of deep-learning-based medical image segmentation methods. Sustain 13, 1224–1229. doi:10.3390/su13031224

CrossRef Full Text | Google Scholar

Lundervold, A. S., and Lundervold, A. (2019). An overview of deep learning in medical imaging focusing on MRI. Z. Med. Phys. 29, 102–127. doi:10.1016/j.zemedi.2018.11.002

PubMed Abstract | CrossRef Full Text | Google Scholar

Mack, M. J., Leon, M. B., Thourani, V. H., Makkar, R., Kodali, S. K., Russo, M., et al. (2019). Transcatheter aortic-valve replacement with a balloon-expandable valve in low-risk patients. N. Engl. J. Med. 380, 1695–1705. doi:10.1056/NEJMoa1814052

PubMed Abstract | CrossRef Full Text | Google Scholar

Macruz, F. B. C., Lu, C., Strout, J., Takigami, A., Brooks, R., Doyle, S., et al. (2022). Quantification of the thoracic aorta and detection of aneurysm at CT: Development and validation of a fully automatic methodology. Radiol. Artif. Intell. 4, e210076. doi:10.1148/ryai.210076

PubMed Abstract | CrossRef Full Text | Google Scholar

Maragiannis, D., Jackson, M. S., Igo, S. R., Chang, S. M., Zoghbi, W. A., and Little, S. H. (2014). Functional 3D printed patient-specific modeling of severe aortic stenosis. J. Am. Coll. Cardiol. 64, 1066–1068. doi:10.1016/j.jacc.2014.05.058

PubMed Abstract | CrossRef Full Text | Google Scholar

Milletari, F., Navab, N., and Ahmadi, S. A. (2016). V-Net: Fully convolutional neural networks for volumetric medical image segmentation. Proc. - 2016 4th Int. Conf. 3D Vis. 2016, 565–571. doi:10.1109/3DV.2016.79

CrossRef Full Text | Google Scholar

Minaee, S., Boykov, Y., Porikli, F., Plaza, A., Kehtarnavaz, N., and Terzopoulos, D. (2020). Image segmentation using deep learning: A survey. J. Phys. Conf. Ser. 1712, 3523–3542. doi:10.1109/tpami.2021.3059968

CrossRef Full Text | Google Scholar

Myronenko, A. (2019). 3D MRI brain tumor segmentation using autoencoder regularization. Lect. Notes Comput. Sci. 2019, 311–320. doi:10.1007/978-3-030-11726-9_28

CrossRef Full Text | Google Scholar

Nu˜nez-Garcia, M., Zhuang, X., Sanroma, G., Li, L., Xu, L., and Butakoff, C. (2019). Statistical atlases and computational models of the heart. Atrial segmentation and LV quantification challenges. Berlin, Germany: Springer International Publishing. doi:10.1007/978-3-030-12029-0

CrossRef Full Text | Google Scholar

Otto, C. M., Nishimura, R. A., Bonow, R. O., Carabello, B. A., Erwin, J. P., Gentile, F., et al. (2021). 2020 ACC/AHA guideline for the management of patients with valvular heart disease: A report of the American college of cardiology/American heart association joint committee on clinical practice guidelines. Circulation 143, e72–e227. doi:10.1161/CIR.0000000000000923

PubMed Abstract | CrossRef Full Text | Google Scholar

Panayides, A. S., Amini, A., Filipovic, N. D., Sharma, A., Tsaftaris, S. A., Young, A., et al. (2020). AI in medical imaging informatics: Current challenges and future directions. IEEE J. Biomed. Heal. Inf. 24, 1837–1857. doi:10.1109/JBHI.2020.2991043

PubMed Abstract | CrossRef Full Text | Google Scholar

Puri, R., Chamandi, C., Rodriguez-Gabella, T., and Rodés-Cabau, J. (2017). Future of transcatheter aortic valve implantation — evolving clinical indications. Nat. Rev. Cardiol. 15, 57–65. doi:10.1038/nrcardio.2017.116

PubMed Abstract | CrossRef Full Text | Google Scholar

Qian, Z., Wang, K., Liu, S., Zhou, X., Rajagopal, V., Meduri, C., et al. (2017). Quantitative prediction of paravalvular leak in transcatheter aortic valve replacement based on tissue-mimicking 3D printing. JACC Cardiovasc. Imaging 10, 719–731. doi:10.1016/j.jcmg.2017.04.005

PubMed Abstract | CrossRef Full Text | Google Scholar

Ranschaert, E. R., Morozov, S., and Algra, P. R. (2019). in Artificial intelligence in medical imaging: Opportunities, applications and risks. Editors E. R. Ranschaert, S. Morozov, and P. R. Algra (Berlin, Germany: Springer International Publishing). doi:10.1007/978-3-319-94878-2

CrossRef Full Text | Google Scholar

Ravichandran, S. R., Nataraj, B., Huang, S., Qin, Z., Lu, Z., Katsuki, A., et al. (2019). “3D inception u-net for aorta segmentation using computed tomography cardiac angiography,” in 2019 IEEE EMBS Int. Conf. Biomed. Heal. Informatics, BHI 2019 - Proc, Chicago, IL, USA, 19-22 May 2019 (IEEE) 6–9. doi:10.1109/BHI.2019.8834582

CrossRef Full Text | Google Scholar

Robinson, R., Oktay, O., Bai, W., Valindria, V. V., Sanghvi, M. M., Aung, N., et al. (2018). “Real-time prediction of segmentation quality,” in 21st International Conference on Medical Image Computing and Computer Assisted Intervention, MICCAI 2018, Granada, Spain, 15-20 September 2020 (Germany: Springer) 578–585. doi:10.1007/978-3-030-00937-3_66

CrossRef Full Text | Google Scholar

Robinson, R., Valindria, V. V., Bai, W., Oktay, O., Kainz, B., Suzuki, H., et al. (2019). Automated quality control in image segmentation: Application to the UK biobank cardiovascular magnetic resonance imaging study. J. Cardiovasc. Magn. Reson. 21, 18. doi:10.1186/s12968-019-0523-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Romaszko, L., Borowska, A., Lazarus, A., Dalton, D., Berry, C., Luo, X., et al. (2021). Neural network-based left ventricle geometry prediction from CMR images with application in biomechanics. Artif. Intell. Med. 119, 102140. doi:10.1016/j.artmed.2021.102140

PubMed Abstract | CrossRef Full Text | Google Scholar

Rukundo, O. (2023). Effects of image size on deep learning. Electronics 12, 985. doi:10.3390/electronics12040985

CrossRef Full Text | Google Scholar

Sabottke, C. F., and Spieler, B. M. (2020). The effect of image resolution on deep learning in radiography. Radiol. Artif. Intell. 2, e190015. doi:10.1148/ryai.2019190015

PubMed Abstract | CrossRef Full Text | Google Scholar

Shen, D., Wu, G., and Suk, H.-I. (2017). Deep learning in medical image analysis. Annu. Rev. Biomed. Eng. 19, 221–248. doi:10.1146/annurev-bioeng-071516-044442

PubMed Abstract | CrossRef Full Text | Google Scholar

Shorten, C., and Khoshgoftaar, T. M. (2019). A survey on image data augmentation for deep learning. J. Big Data 6, 60. doi:10.1186/s40537-019-0197-0

CrossRef Full Text | Google Scholar

Sieren, M. M., Widmann, C., Weiss, N., Moltz, J. H., Link, F., Wegner, F., et al. (2022). Automated segmentation and quantification of the healthy and diseased aorta in CT angiographies using a dedicated deep learning approach. Eur. Radiol. 32, 690–701. doi:10.1007/s00330-021-08130-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Smith, C. R., Leon, M. B., Mack, M. J., Miller, D. C., Moses, J. W., Svensson, L. G., et al. (2011). Transcatheter versus surgical aortic-valve replacement in high-risk patients. N. Engl. J. Med. 364, 2187–2198. doi:10.1056/NEJMoa1103510

PubMed Abstract | CrossRef Full Text | Google Scholar

Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., and Wojna, Z. (2016). “Rethinking the inception architecture for computer vision,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit, Las Vegas, NV, USA, 27-30 June 2016 (IEEE) 2818–2826. doi:10.1109/CVPR.2016.308

CrossRef Full Text | Google Scholar

Taha, A. A., and Hanbury, A. (2015). Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med. Imaging 15, 29. doi:10.1186/s12880-015-0068-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Thalji, N. M., Suri, R. M., Greason, K. L., and Schaff, H. V. (2014). Risk assessment methods for cardiac surgery and intervention. Nat. Rev. Cardiol. 11, 704–714. doi:10.1038/nrcardio.2014.136

PubMed Abstract | CrossRef Full Text | Google Scholar

Tsakanikas, V. D., Siogkas, P. K., Mantzaris, M. D., Potsika, V. T., Kigka, V. I., Exarchos, T. P., et al. (2020). “A deep learning oriented method for automated 3D reconstruction of carotid arterial trees from MR imaging,” in 2020 42nd Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Montreal, QC, Canada, 20-24 July 2020 (IEEE) 2408–2411. doi:10.1109/EMBC44109.2020.9176532

PubMed Abstract | CrossRef Full Text | Google Scholar

Underwood, M. J., El Khoury, G., Deronck, D., Glineur, D., and Dion, R. (2000). The aortic root: Structure, function, and surgical reconstruction. Heart 83, 376–380. doi:10.1136/heart.83.4.376

PubMed Abstract | CrossRef Full Text | Google Scholar

Walther, T., Hamm, C. W., Schuler, G., Berkowitsch, A., Kötting, J., Mangner, N., et al. (2015). Perioperative results and complications in 15,964 transcatheter aortic valve replacements: Prospective data from the GARY registry. J. Am. Coll. Cardiol. 65, 2173–2180. doi:10.1016/j.jacc.2015.03.034

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, D. D., Qian, Z., Vukicevic, M., Engelhardt, S., Kheradvar, A., Zhang, C., et al. (2021). 3D printing, computational modeling, and artificial intelligence for structural heart disease. JACC Cardiovasc. Imaging 14, 41–60. doi:10.1016/j.jcmg.2019.12.022

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang, S., Kweon, J., Roh, J.-H., Lee, J.-H., Kang, H., Park, L.-J., et al. (2019). Deep learning segmentation of major vessels in X-ray coronary angiography. Sci. Rep. 9, 16897. doi:10.1038/s41598-019-53254-7

PubMed Abstract | CrossRef Full Text | Google Scholar

Ya’qoub, L., Faza, N., Khalique, O., Parwani, P., Cader, A., Swaminathan, M., et al. (2022). Intraprocedural cardiac complications of transcatheter aortic and mitral valve interventions: The eyes do not see what the mind does not know. Cardiovasc. Revascularization Med. 36, 144–152. doi:10.1016/j.carrev.2021.05.001

CrossRef Full Text | Google Scholar

Youssefi, P., Gomez, A., He, T., Anderson, L., Bunce, N., Sharma, R., et al. (2017). Patient-specific computational fluid dynamics—Assessment of aortic hemodynamics in a spectrum of aortic valve pathologies. J. Thorac. Cardiovasc. Surg. 153, 8–20.e3. doi:10.1016/j.jtcvs.2016.09.040

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, X., Liao, W. T., and Xu, G. (2021). Median filter helps lymph node segmentation in deep learning via PET/CT. ACM Int. Conf. Proceeding Ser. 2021, 67–71. doi:10.1145/3506651.3506662

CrossRef Full Text | Google Scholar

Zou, K. H., Warfield, S. K., Bharatha, A., Tempany, C. M. C., Kaus, M. R., Haker, S. J., et al. (2004). Statistical validation of image segmentation quality based on a spatial overlap index. Acad. Radiol. 11, 178–189. doi:10.1016/S1076-6332(03)00671-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: TAVR, aortic root, automatic segmentation, 3D CNN, deep learning

Citation: Yang T, Zhu G, Cai L, Yeo JH, Mao Y and Yang J (2023) A benchmark study of convolutional neural networks in fully automatic segmentation of aortic root. Front. Bioeng. Biotechnol. 11:1171868. doi: 10.3389/fbioe.2023.1171868

Received: 22 February 2023; Accepted: 06 June 2023;
Published: 15 June 2023.

Edited by:

Fuyou Liang, Shanghai Jiao Tong University, China

Reviewed by:

Aike Qiao, Beijing University of Technology, China
Lijian Xu, The Chinese University of Hong Kong, China

Copyright © 2023 Yang, Zhu, Cai, Yeo, Mao and Yang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Guangyu Zhu, emh1Z3Vhbmd5dUB4anR1LmVkdS5jbg==; Jian Yang, eWFuZ2ppYW4xMjEyQGhvdG1haWwuY29t

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.