Skip to main content

ORIGINAL RESEARCH article

Front. Cardiovasc. Med., 09 March 2022
Sec. Heart Valve Disease
This article is part of the Research Topic New Approaches in Heart Valve Research View all 7 articles

Aortic Valve Leaflet Shape Synthesis With Geometric Prior From Surrounding Tissue

  • 1Institute for Robotics and Cognitive Systems, University of Lübeck, Lübeck, Germany
  • 2Department of Engineering Science, University of Oxford, Oxford, United Kingdom
  • 3Department of Cardiac Surgery, University-Hospital Schleswig-Holstein, Lübeck, Germany

Even though the field of medical imaging advances, there are structures in the human body that are barely assessible with classical image acquisition modalities. One example are the three leaflets of the aortic valve due to their thin structure and high movement. However, with an increasing accuracy of biomechanical simulation, for example of the heart function, and extense computing capabilities available, concise knowledge of the individual morphology of these structures could have a high impact on personalized therapy and intervention planning as well as on clinical research. Thus, there is a high demand to estimate the individual shape of inassessible structures given only information on the geometry of the surrounding tissue. This leads to a domain adaptation problem, where the domain gap could be very large while typically only small datasets are available. Hence, classical approaches for domain adaptation are not capable of providing sufficient predictions. In this work, we present a new framework for bridging this domain gap in the scope of estimating anatomical shapes based on the surrounding tissue's morphology. Thus, we propose deep representation learning to not map from one image to another but to predict a latent shape representation. We formalize this framework and present two different approaches to solve the given problem. Furthermore, we perform a proof-of-concept study for estimating the individual shape of the aortic valve leaflets based on a volumetric ultrasound image of the aortic root. Therefore, we collect an ex-vivo porcine data set consisting of both, ultrasound volume images as well as high-resolution leaflet images, evaluate both approaches on it and perform an analysis of the model's hyperparameters. Our results show that using deep representation learning and domain mapping between the identified latent spaces, a robust prediction of the unknown leaflet shape only based on surrounding tissue information is possible, even in limited data scenarios. The concept can be applied to a wide range of modeling tasks, not only in the scope of heart modeling but also for all kinds of inassessible structures within the human body.

1. Introduction

Despite ongoing advancements of medical imaging techniques, there are structures in the human body that are difficult to visualize using typical medical imaging modalities like Computed Tomography (CT), Magnetic Resonance Imaging (MRI), or Ultrasound Imaging (US). However, the knowledge of these structures' shape is highly relevant for clinical decision making and intervention, e.g., for biomechanical simulations or for personalized prosthesis shaping. One example of such a structure is the aortic valve. This valve consists of three thin leaflets embedded in the aortic root that are pressed together during diastole to prevent the blood in the aorta from flowing back into the left ventricle (1). The geometry of the aortic valve, i.e., the shape of its three leaflets, shows an enormous inter-patient variability (2). As the aortic valve and root form a complex biomechanical system with a lot of interaction between different anatomical structures (3), this individual geometry is crucial for the correct functionality of the valve. The valve's leaflets are extremely thin and flutter in the blood stream, hence, imaging them using typical medical imaging modalities is challenging. However, the knowledge about the individual geometry of the aortic valve is necessary for many applications, ranging from heart modeling and simulation to the development of personalized prostheses. For both applications, the leaflet shape should be assessed in an unpressurized state to avoid stress-related deformations. Due to the high impact of the individual valve geometry on a procedure's outcome, a personalization of this prosthesis would be beneficial but is out of scope as the desired prosthesis shape is not assessible using typical in-vivo imaging modalities. Hence, the synthesis of personalized leaflet shapes presents a promising approach for solving this problem.

In an ex-vivo setting, it is possible to extract the aortic valve leaflets and collect high-resolution images of them in a planar state (4). Such a dataset could serve as a training set for a generative model aiming on leaflet shape synthesis. An additional advantage of this approach is that the synthesized leaflet shapes are already in a planar shape, allowing for directly manufacturing a prosthesis out of typical material, e.g., pericardium. However, the shape synthesis should be tailored to the individual patient. Thus, the generative model should receive a prior in the form of information of relevant surrounding tissue to estimate the patient's individual leaflet shapes. In this case, the three-dimensional (3D) shape of the aortic root might be a sufficient prior as its geometry should be closely related to the shape of the individual leaflet. Additionally, the aortic root is clearly visible in ultrasound images acquired using transesophageal echocardiography (TEE).

However, the domain gap between a 3D ultrasound image of the aortic root where the leaflets are barely visible and an RGB image of the leaflet shape in its planar state is huge. The image's appearance highly differs between both modalities and even though the acquired organ is the same, both modalities assess very different parts of the organ with only little overlap across the modalities. Typical methods for bridging domain gaps utilize adversarial training, i.e., generative adversarial networks (GAN), to transfer an image from one domain to the other one (5). However, GANs need vast amounts of training data to converge that is typically not available in medical imaging. Especially for structure shape synthesis based on surrounding tissue, data collection is very time-consuming and requires extra effort due to ex-vivo experiments. Additionally, GANs are prone to sometimes synthesizing unrealistic images (6), which should be avoided in the scope of medical decision support systems or prosthesis manufacturing.

In this work, we present a robust approach for synthesizing aortic valve leaflet shapes with the individual aortic root shape as geometric prior based on shape encoding with autoencoders. First, we collect and present a dataset containing 3D US images of porcine aortic roots as well as 2D planar images of the corresponding leaflets. Then, we formalize the given problem and present different ways to solve it. We evaluate all these approaches on the collected dataset, including a hyperparameter analysis and a comparison of the different approaches.

1.1. Contribution of This Work

The contribution of this work is three-fold. First, to the best of our knowledge, we present the first sufficiently large dataset containing 3D aortic root shapes as well as high resolution images of the corresponding valve leaflets. Second, we describe a novel methodology for bridging big domain gaps that works robustly even on small data. As this method is not limited to aortic valve leaflet synthesis, our methodological contribution might be of great interest in the medical image analysis community as well as for general computer vision researchers. Third, our proposed method can be directly applied in the scope of personalized aortic valve modeling, for example for prosthesis development or cardiac simulation, highlighting not only the methodological contribution but also the clinical applicability and translational value of this study.

2. Related Work

In general, dealing with data from different domains is referred to as domain adaptation (7). Lots of studies aimed at generalizing across different domains (810). However, the typical focus is dealing with input data from different domains, not to transfer an image to another domain (11).

The problem of estimating an image based on another image is called image-to-image translation and was introduced by (5). Typically, generative adversarial networks (GAN) (12) are utilized to learn a distribution over the target images, conditioned by the input image (5). Such approaches have been applied to a wide range of topics, for example semantic image synthesis (13), image segmentation (14), style transfer (15), or image inpainting (16). However, most of these approaches are used in settings with a moderate domain gap while the general objects in the input image remain unchanged in their shape.

Several studies focused on image-to-image-translation for shape deformation, aiming on bridging a wider domain gap (17). However, all these approaches were evaluated on large datasets that are typically not available in clinical applications. Even though, Liu et al. (18) proposed a few-shot approach for image-to-image translation, the model has to be pre-trained on a large dataset that is relatively similar to the small one. Lin et al. (19) present a framework based on a cascaded GAN structure to learn image-to-image translation from one image pair only. However, such a model is not capable of generating the full variance of anatomies present across patients as it is trained to fit one anatomy as closely as possible. In contrast, our approach directly models this variance by encoding all observed anatomies into latent space descriptions.

Several studies aimed at personalized modeling of the aortic valve leaflet shape. Typically, this is either done by segmenting the leaflets from CT data (20) or by deforming a template leaflet shape according to the surrounding tissue anatomy extracted from a CT (21). Both approaches incorporate a substantial amount of expert knowledge as well as manual interaction. Additionally, due to the deformation constraints, they are biased toward predicting an average leaflet shape.

In a previous study, it could be shown that the aortic root shape carries enough information to estimate specific features of the valve leaflets, i.e., the commissure contour line (22). Thus, a support vector regression model was trained based on manually identified geometric features. Even though the commissure line was predictable, the model was not capable of estimating the full leaflet shape, i.e., synthesizing images of the leaflets. Additionally, the evaluation was performed on a small dataset and the ultrasound volumes were stitched together and interpolated from several rotated 2D imaging planes. The approach to use autoencoders to derive a sufficient latent space containing the shape of aortic valve leaflets was presented in Hagenah et al. (23). However, this study focused on shape typification in latent space and no connection between the leaflet shape and the aortic root geometry was identified.

3. Materials and Methods

In this chapter, our methodology is presented. First, we present the data collection procedure and the preprocessing applied to the raw data (section 3.1). Then, we formalize the problem of domain mapping for leaflet shape synthesis and present general approaches to solve it (section 3.2). In sections 3.3 and 3.4, our proposed methods for solving the problem are presented. As there is a wide variety of interpretations of the term shape, it is important to note that we interpret the shape of a leaflet as its segmentation in a 2D image that shows the leaflet in a spreaded state.

3.1. Data Collection

One big challenge of collecting a sufficiently large dataset is that acquiring high-resolution images of the aortic valve leaflets is barely possible in-vivo. One typical method is to extract the single leaflets and photograph them in a planar shape (4, 23). Hence, we followed an ex-vivo approach examining porcine hearts. The pig heart's anatomy and physiology are quite similar to the human one making it a common animal model (24). This also holds for the aortic valve apparatus as porcine valves are even used as xenograft prostheses (25). As the pig hearts were bought at a slaughterhouse, there are no ethical concerns raised by this study. All in all, we collected data of 29 porcine hearts. From each of these hearts, we extracted and acquired a volumetric ultrasound image of the aortic root, followed by an extraction of the leaflets and their image acquisition. Therefore, the collected data consists of 29 ultrasound volumes and photographs of the right-coronary, left-coronary, and non-coronary leaflet, respectively, so 87 leaflet images in total. Details on the setups, the workflow and data preprocessing are described in the following paragraphs.

3.1.1. Aortic Root Imaging

Imaging of the aortic root mainly followed the method described in Hagenah et al. (26). From the fresh pig heart, the aortic root was extracted by exposing the aorta and then cutting the root out of the left ventricle around the ventriculoarterial junction. The coronary arteries were clamped, the root was attached to a vertical tube within a water basin and, using a water column within this tube, a constant, physiologically realistic diastolic pressure was applied. Hence, the valve appeared in a closed state. Figure 1 shows photographs of the extraction steps. Within the water basin, an ultrasound probe for transesophageal echocardiography (TEE) was installed such that the viewing angle as well as the distance to the aortic root mimics a TEE examination. Then, a volumetric ultrasound image of the root was acquired. We used a GE Vivid E95 ultrasound system with the 6VT-D probe. The size of the image was 84 × 202 × 84 with a voxel size of 0.71 × 0.49 × 0.71mm.

FIGURE 1
www.frontiersin.org

Figure 1. Experimental workflow. At first, the aortic root is extracted from the porcine heart (A). After clamping the coronary arteries (B), an ultrasound image of the aortic root can be acquired. Then, the root is cut vertically and the leaflets can be extracted (C).

From this ultrasound volume, we cropped out a subvolume that shows the aortic root. Therefore, we manually identified the commissure plane, i.e., the horizontal slice through the aortic root that shows all three commissure points, i.e., the highest point of the commissures. We defined this plane as the uppermost slice of the subvolume and added the 31 layers below to it. Hence, the subvolume covered 22.72mm of the aortic root's height. Keeping this dimension fixed, we cropped the other two dimensions, i.e., horizontal slice images through the aortic root, to a size of 128 × 128 such that the aortic root was roughly in the image center of the slice images. We performed zero padding around the cropped root, i.e., areas of the new volume where no image information was available were filled with a gray value of zero. After rearrangement of the dimensions for convenience, the final image size was 32 × 128 × 128.

To get rid of imaging artifacts and background noise, we applied thresholding, setting all gray values smaller or equal to t = 80 to 0. Finally, the ultrasound images were scaled with the factor 1255 to lie in the range [0, 1].

3.1.2. Leaflet Imaging

After the acquisition of ultrasound images of the aortic root, the three leaflets were extracted and assessed with high resolution, following the method of Hagenah et al. (23). Thus, the root was opened by a vertical cut in between the right-coronary and the non-coronary sinus. Then, the single leaflets were cut off the root wall and spread on a diffusive glass plate. Special care was taken to preserve the leaflet's original shape during extraction and spreading. From below the plate, blue illumination (470nm) was applied as this wavelength is absorbed in the collagen structures of the leaflets, leading to a high contrast. Then, a photograph was taken from above using a Canon DS 6041 SLR camera with a fixed exposure time of 16s. The resolution of this image was 0.037mmpixel.

The preprocessing of the leaflets also followed (23). The individual leaflets were cropped out and the images were transformed to grayscale (range: [0, 255]) and inverted. All background pixels were set to 0 using thresholding with manually selected, individual thresholds for each leaflet to avoid holes in the segmentation. All thresholds were in the range from 158 to 168. The leaflets were centered in the image by translating the center of mass into the image's mid point. Then, the leaflet was aligned by rotating it around the image center so that the commissure points were vertically aligned. The resulting images were downsampled to a size of 128 × 64 pixels, leading to a resolution of 0.34mmpixel, and the grayscale values were scaled by the factor 1255 to lie in the range [0, 1]. Figure 2 shows an example raw image from the dataset as well as one preprocessed leaflet. In the scope of this study, we also made use of the leaflet image dataset presented in Hagenah et al. (23), containing 168 images of leaflets from 56 valves. Note that for these images, no corresponding ultrasound volume is known and they are used as an auxiliary dataset to reliably cover the full variety of leaflet shapes.

FIGURE 2
www.frontiersin.org

Figure 2. Leaflet image acquisition. (A) Raw photograph of illuminated plate with right-coronary, left-coronary, and non-coronary leaflet of one valve. (B) Right-coronary leaflet after preprocessing.

Furthermore, we assessed the distribution of leaflet shapes in the collected dataset. Therefore, we identified the length, width and area of each leaflet automatically. The length was measured as the maximal leaflet spreading among all image columns and the width as the maximum leaflet spreading among all image rows. The area was assessed as the number of pixels that show a part of the leaflet. The necessary segmentations were performed using thresholding (t = 0.45, corresponding to a grayscale value of 115). The resulting distribution of the dataset regarding the length, width and area for each leaflet type is visualized in Figure 3. Note that the distribution is visualized across all available leaflets, i.e., the data collected in this study as well as the auxiliary dataset. The high variance of leaflet shapes highlights the need for personalized modeling.

FIGURE 3
www.frontiersin.org

Figure 3. Distribution of leaflet length (A), width (B), and area (C) with regard to the leaflet type. The white dot represents the median value.

3.2. Problem Formulation

As described in section 3.1, the data lies in two different spaces: The volumetric ultrasound data space DUS32×128×128 and the leaflet data spaces DLfl128×64,l{rc,lc,nc} for the right-coronary, left-coronary and non-coronary leaflets, respectively. In addition, we define a latent representation Ƶ for each data space, leading to the latent spaces ƵUSml and ƵLfnl. Note that we assume a shared latent space representation for all three leaflet types, i.e., right-coronary, left-coronary and non-coronary. The latent space descriptions can be derived from the data space using representation learning on the corresponding dataset, respectively. Figure 4 shows the different coordinate spaces and their connection.

FIGURE 4
www.frontiersin.org

Figure 4. Sketch of our proposed approaches. As a direct mapping between the two data domains, also called image to image translation, is barely possible in the given scenario, we propose to encode the data to latent space descriptions using representation learning. Then, the latent leaflet description can be predicted based on the 3D ultrasound image, called shape estimation, or based on the latent description of the ultrasound volume, referred to as domain mapping.

The general goal is to synthesize the unknown personalized leaflet images IrcDLfrc,IlcDLflc and IncDLfnc of one patient based on the information contained in the patient's individual aortic root geometry given by VDUS. The direct mapping from V to Irc, Ilc, and Inc is known as image-to-image translation and typically requires a vast amount of training data as it is usually solved using GANs (see section 2). First encoding V to its latent representation zUS and then mapping to Irc, Ilc, and Inc is still interpreted as image-to-image translation as the only difference is an unsupervised pretraining of a subnetwork, which is a well-known technique in deep learning on small data (27). By encoding Irc, Ilc, and Inc to their latent representations zLfrc,zLflc,zLfncƵLf using the encoder qLf, the image to image translation problem can be reformulated as a regression problem where zLfrc,zLflc, and zLfnc should be estimated based on the ultrasound image data. If a decoder pLf is known, the corresponding shape leaflet images can be synthesized. Due to intense preprocessing of the leaflet image data, ƵLf mainly contains shape information. Hence, predicting this latent information based on the volumetric ultrasound data V is referred to as shape estimation. Another option is to not only encode the leaflet images but also the ultrasound data. We assume that there is a connection between both latent spaces as they represent different parts and states of the same organ. However, the difference between both latent spaces is the domain gap between the data sets. One way to overcome this domain gap is to train a machine learning model to map from ƵUS to ƵLf. Adding the encoder qUS and the decoder pLf, a full path from an ultrasound volume to synthesized leaflet images is given. This approach is called domain mapping.

As described above, training an image-to-image translation model on very limited data until convergence is a challenging task. In addition, it is likely that an adversarial model learns the full distribution of leaflets without taking the prior information V into account, making the model fail in personalized shaping. Hence, we propose to follow the other two approaches, namely shape estimation and domain mapping. Details on our implementation of these approaches are given in the following sections.

3.3. Shape Estimation Approach

The goal of the shape estimation approach is to encode the leaflet shape, given as an image, into a latent space description and predict the individually optimal latent shape based on a 3D US image of the aortic root. Thus, the approach consists of two models: the leaflet autoencoder for shape encoding and the regression network for mapping from the US image to the latent representation.

For the leaflet autoencoder, we used a parameterized architecture (28). The encoder qLf consists of nc convolutional layers with nf filters, ReLU activation (29) and followed by a 2 × 2 average pooling each. Following this convolutional part, a flattening operation and fully connected layer (ReLU activation) with as many neurons as outputs of the last pooling layer connect to the bottleneck layer with nl neurons, featuring linear activation. The decoder pLf follows the mirrored encoder architecture utilizing upconvolutisonal and upsampling. The autoencoder was trained using the adam optimizer, mean squared error loss and a batch size of 32 for 100 epochs (29).

It is assumed that there is a shared latent space description over all leaflet types, i.e., right-coronary, left-coronary and non-coronary leaflets. Hence, the autoencoder is trained on all three kinds of leaflets, leading to the dataset DLf=DLfrcDLflcDLfnc. To ensure a reliable representation, the data collected in Hagenah et al. (23) was added to the training data as an auxiliary dataset DLfaux, containing 168 images of leaflets from 56 valves.

For the regression network, we propose to use a 3D convolutional neural network (CNN) with a VGG-like architecture (30). Thus, the network consists of kb convolutional blocks. Each of these blocks consists of kc 3D convolutional layers with kf filters and ReLU activation. Each block is followed by a 2 × 2 average pooling layer. Behind the last convolutional block and a flattening operation, kd fully connected layers with ReLU activation and kn neurons each are attached, followed by the output layer with kl neurons and linear activation.

We assume that the prediction of the individual shapes of the right-coronary, left-coronary, and non-coronary leaflets are independent. Hence, three regression models are created and trained, predicting one leaflet type each. The three models share the same architecture but are trained independently.

3.3.1. Hyperparameter Analysis

To identify an optimal autoencoder architecture, i.e., an optimal combination of the hyperparameters nc, nf, and nl, we assessed multiple of these combinations regarding the model's ability to reconstruct the input image after propagating it through the model in a grid-search approach. Thus, we performed 10-fold Monte-Carlo crossvalidation (80% train, 20% test) on DLf (28). After training the autoencoder on the training data, we propagated the test data through the model and compared the resulting reconstruction to the original image using root mean square error (RMSE). As mentioned before, the auxiliary dataset DLfaux was added to the training data within each fold. Table 1 shows all combinations of hyperparameters tested.

TABLE 1
www.frontiersin.org

Table 1. Combinations of hyperparameters of the leaflet autoencoder assessed during hyperparameter analysis.

After identifying an optimal autoencoder architecture, the hyperparameters of the regression network were optimized in a similar way. For multiple combinations of the hyperprameters kb, kc, kf, kd, and kn, the model's performance on predicting the latent shape representation of the corresponding leaflets for unseen ultrasound images is assessed using a 10-fold Monte Carlo crossvalidation (80% train, 20% test) on DUS and DLf. Therefore, the autoencoder was trained on the training leaflet images using the optimal hyperparameters. Then, all training leaflet images were encoded to the latent space and the three regression networks were trained to predict the latent space representation of the respective leaflet based on the corresponding 3D ultrasound image. The accuracy was assessed by predicting the shape representation for the test US images, reconstruct images of the predicted leaflet shapes using the decoder and comparing these predicted images to the ground truth leaflet images, once again using RMSE. Table 2 shows all hyperparameter combinations assessed in this study.

TABLE 2
www.frontiersin.org

Table 2. Combinations of hyperparameters of the regression networks assessed during hyperparameter analysis.

3.3.2. Performance Analysis

After identifying optimal sets of hyperparameters for the autoencoder and the regression networks, we analyzed the performance of the shape estimation approach. Thus, we trained the autoencoder and the three regression networks on training data, using the respective optimal hyperparameters, and predicted the leaflet shapes for unseen test data. We did this using a 10-fold Monte Carlo crossvalidation (80% train, 20% test) on DUS and DLf, while once again DLfaux was added to the training data for training the autoencoder. We compared the predicted leaflet images to their corresponding ground truth using four metrics: Jaccard similarity, Hausdorff distance, average symmetric contour distance (ASCD) and RMSE. The ASCD of two sets of contour points X and Y is given as

ASCD(X,Y)=ACD(X,Y)+ACD(Y,X)2    (1)

with the average contour distance (ACD) defined as

ACD(X,Y)=xXminyYd(x,y)X,    (2)

where d(x, y) is the euclidian distance between the points x and y (31).

To compute the Jaccard similarity, the Hausdorff distance and the ASCD, the leaflets were segmented in the predicted images utilizing thresholding (t = 0.45, corresponding to a grayscale value of 115).

3.4. Domain Mapping Approach

The key idea of the domain mapping approach is to encode both, the leaflet images as well as the volumetric US data, into a latent space description, respectively, and train a model to map from one latent space to the other. One important advantage is that the dimensionality of the mapping is much smaller than in the shape estimation approach. Hence, it is possible to learn this mapping using classical machine learning methods like random forests (RF) or multi-layer perceptrons (MLP) (28).

The autoencoder for the leaflet images is the same a presented in section 3.3. To find a representation of the US data, we propose a similar architecture but featuring 3D convolution and pooling. Hence, the encoder qUS consists of mc 3D convolutional layers with mf filters, ReLU activation and followed by a 2 × 2 × 2 average pooling each. Following this convolutional part, a flattening operation and fully connected layer (ReLU activation) with as many neurons as outputs of the last pooling layer connect to the bottleneck layer with ml neurons, featuring linear activation. Once again, the decoder pUS follows the mirrored encoder architecture.

To find a mapping between the two latent spaces ƵUS and ƵLf, we evaluated a random forest with t decision trees and an MLP with lh hidden layers and ln neurons in each hidden layer. For the MLP, all hidden layers used ReLU activation, while the output layer utilized linear activation.

3.4.1. Hyperparameter Analysis

The optimal hyperparameters of the US autoencoder were identified in a similar way as for the leaflet autoencoder. Thus, we performed 10-fold Monte Carlo crossvalidation (80% train, 20% test) on DUS and assessed the reconstruction accuracy on the test data using RMSE. Table 3 shows all evaluated combinations of hyperparameters.

TABLE 3
www.frontiersin.org

Table 3. Combinations of hyperparameters of the ultrasound autoencoder assessed during hyperparameter analysis.

For finding the mapping between the latent spaces, we assessed different values of the hyperparameters for both learning methods. For the random forest approach, we analyzed the values 50, 100, 150, 200, 250 for t. For the MLP approach, we assessed the values 1, 2, 3, 4, 5 for lh and 50, 100, 150, 200 for ln.

3.4.2. Performance Analysis

After identifying optimal sets of hyperparameters, we evaluated the performance of the domain mapping approach on predicting the individual leaflet shapes based on a 3D US image on unseen data. Once again, we performed 10-fold Monte Carlo crossvalidation (80% train, 20% test) on DUS and DLf. Thus, we trained both autoencoders on the training data (including DLfaux for the leaflet autoencoder) and encoded the training samples to their latent space representation. Then we trained models to predict zLf for a given zUS. As described for the shape estimation approach, we propose to use three different models, one for each leaflet type, i.e., right-coronary, left-coronary and non-coronary. We evaluated RFs as well as MLPs for this regression, each of them using the optimal hyperparameters. After training the models, the test US images were encoded using qUS, the latent leaflet representations were predicted and corresponding images were synthesized using the decoder pLf. These predicted leaflet images were compared to the ground truth using Jaccard similarity, Hausdorff distance, ASCD and RMSE, once again with a thresholding if necessary (t = 0.45, corresponding to a grayscale value of 115).

4. Results and Discussion

At first, the results of the hyperparameter analysis are presented, followed by the performance analysis of both approaches. Afterwards, the results are discussed in detail and their impact on future research is given in an outlook paragraph.

4.1. Hyperparameter Analysis

As described above, all hyperparameters were optimized regarding a minimal RMSE. For both autoencoders, an architecture featuring three convolutional blocks with 16 filters each and a latent dimension size of 20 was identified as optimal, with an RMSE of 0.0617±0.0106 for the leaflet autoencoder and 0.0678±0.0200 for the ultrasound autoencoder.

For assessing the hyperparameters of the shape estimation approach, we used the optimal leaflet autoencoder hyperparameters and evaluated the parameter combinations of the CNN mapping from the ultrasound volume space DUS to the latent space of the leaflet ƵLf. The optimal architecture features four convolutional blocks with three convolutional layers using 16 filters each, followed by a single fully connected layer with 100 neurons. This indicates that feature extraction needs a certain degree of abstraction, but the identified features are meaningful and can be processed with a simple classification model, i.e., a single fully connected layer. The RMSE between the true leaflet samples in the latent space description and the predicted ones reached with this architecture was 0.1331±0.0392.

For the domain mapping approach, we used the previously identified, optimal hyperparameters for both autoencoders and evaluated the hyperparameter influence on the model mapping from ƵUS to Ƶleaflet. In the case of a Random Forest regression, 200 trees were found to perform best with an RMSE of 0.1365±0.0278, once again measured in the leaflet's latent space between the predicted and the true latent representation of the leaflet. Using an MLP for domain mapping, the optimal architecture consisted of four hidden layers with 100 neurons each, reaching an RMSE of 0.1386±0.0409. All hyperparameters that were identified to be optimal are marked in bold within Tables 13.

4.2. Latent Space Exploration

The proposed method heavily relies on the identification of a sufficiently accurate representation of the leaflet shape. Therefore, we performed an explorative analysis of the consistency and smoothness of the leaflet autoencoder's latent space. Figure 5 visualizes the distribution of leaflet shapes in the latent space as a t-SNE embedding (32). There is a substantial overlap between the shape representation of right-coronary, left-coronary and non-coronary leaflets. This finding validates our approach of identifying a unified representation for all leaflets DLf instead of distinct representations for each leaflet type.

FIGURE 5
www.frontiersin.org

Figure 5. t-SNE embedding of the leaflet's latent space with the encoded dataset, divided into right-coronary (rc), left-coronary (lc), and non-coronary (nc) leaflets. There is a substantial overlap between the shape representations of the three leaflet types.

To assess the identified leaflet representation regarding its consistency and smoothness, we performed an exploration study in the latent space of the leaflet autoencoder. Therefore, after training the model, we encoded all leaflets from the dataset to the latent space and computed the median shape over all samples. Then, we manipulated the individual dimensions of this median shape by adding an offset o ∈ {−5, −4, −3, −2, −1, 1, 2, 3, 4, 5} to its value in the current dimension. Then, an image of the resulting manipulated shape was synthesized using the decoder network. Thus, the influence of the individual dimensions of the latent space, i.e., the identified abstract features, can be visualized and assessed. We found that most of the synthesized shapes were realistic with a small number of outliers for large offsets. Furthermore, the influence of the dimensions on the leaflet shape were consistent and complementary while the shape changes were smooth. Figure 6 exemplarily shows the synthesized images for the first five dimensions of the latent space. Based on these results, we assume that the leaflet representation is sufficiently consistent and smooth to use it for personalized shape synthesis.

FIGURE 6
www.frontiersin.org

Figure 6. Exploration of the leaflet autoencoder's latent space. The median shape was manipulated by adding an offset to a specific dimension. Then, an image of this manipulated shape was synthesized. The resulting images are shown exemplarily for the first five dimensions of the latent space. Besides few outliers for extreme offsets, all shapes are realistic. The influences of the individual dimensions on the leaflet shape are consistent and complementary and the shape changes are smooth.

4.3. Performance Analysis

To compare the different approaches for leaflet shape synthesis, we evaluated all of them in a 10-fold crossvalidation. Figure 7 shows qualitative synthesis results exemplarily for four valves drawn from a test set within one fold of the crossvalidation. In general, all three synthesis methods are capable of providing realistic leaflet shapes in all cases. The shape estimation approach shows vanishing inter-patient variance and tends to predict the same leaflet for each individual. This is supported by the shape variance of the predicted leaflets regarding their length, width, and area compared to the ground truth distribution in the full dataset (see Figure 8). While the domain mapping approach with RF provides a fair variance, domain mapping with MLP shows the highest variance and coverage of the true shape distribution, highlighting its individualization performance. The latter approach is capable of following individual shapes relatively accurately in some cases, e.g., the right-coronary leaflet of valve 1, while it tends to overestimate the leaflet length. All presented methods struggle in predicting atypical leaflets, e.g., the right-coronary leaflet of valve 2.

FIGURE 7
www.frontiersin.org

Figure 7. Qualitative results for leaflet shape synthesis with a geometric prior, exemplarily given for four valves (columns) with their right-coronary (rc), left-coronary (lc), and non-coronary (nc) leaflet, respectively. The rows show the ground truth, the reconstruction of the ground truth after propagating it through the whole autoencoder, as well as the synthetic leaflets produced by the different approaches, i.e., shape estimation, domain mapping with random forests, and domain mapping with multilayer perceptrons. All four valves were drawn from the test set of one fold of the crossvalidation.

FIGURE 8
www.frontiersin.org

Figure 8. Distribution of the predicted leaflet shapes in comparison to the true distribution of the dataset DLf. The columns show the distribution of the length (left), width (middle), and area (right) in dependency of the leaflet type for the true distribution given in the dataset (blue) and the distribution in the predicted test set (orange). The rows correspond to the utilized prediction model, i.e., shape estimation (top), domain mapping with RF (middle), and domain mapping with MLP (bottom). While shape estimation fails in predicting a variety of shapes, the MLP-based domain mapping approach outperforms the RF-based one regarding the coverage of the full distribution of observed leaflet shapes.

Figure 9 shows the correlation between the length, width, and area of the predicted leaflets and the corresponding ground truth for each prediction method. It is clearly visible that the shape estimation approach predicts a very limited number of possible shapes and therefor does not provide satisfying results. In contrast, both domain mapping approaches are capable of predicting the leaflet length well, with some severe outliers in the MLP-based approach. This highlights the necessity of encoding the ultrasound images to a latent representation and visualizes the higher robustness of the RF-based approach compared to utilizing MLPs. Both domain-mapping approaches tend to overestimate the size of smaller leaflets.

FIGURE 9
www.frontiersin.org

Figure 9. Correlation between geometric features of the predicted leaflets and the corresponding ground truth. The columns show the predicted length (left), width (middle), and area (right) in dependency of the true one, respectively. The assessed data points from the test set of one fold (blue) are scattered around the line sketching optimal correlation (orange). The rows correspond to the utilized prediction model, i.e., shape estimation (top), domain mapping with RF (middle), and domain mapping with MLP (bottom). While shape estimation only predicts a limited number of possible outputs, both domain mapping approaches approximate the leaflet length sufficiently, with some severe outliers in the MLP-based approach. Both methods tend to overestimate the leaflet size. Please not the different scaling of the y-axes.

These findings are supported by a qualitative assessment of the leaflet contour line prediction across all different models. Figure 10 exemplarily shows the comparison of predicted contour lines, together with the ground truth, for three typical cases. If the leaflet's shape is close to the average one, all models provide a satisfying contour line prediction. For leaflets that differ from the average shape, the domain mapping approach with MLP typically shows the best prediction regarding the contour line. However, for small leaflets, all prediction models tend to overestimate the size of the leaflet.

FIGURE 10
www.frontiersin.org

Figure 10. Comparison of predicted leaflet contour lines for the different prediction models as well as the ground truth, exemplarily shown for three typical cases. Typically, the domain mapping approach with MLP shows the best approximation (A). For leaflets that are close to the average shape, all models provide a satisfying prediction (B). If the leaflet is small, all prediction models tend to overestimate the size of the leaflet (C). The contour lines are given in pixel coordinates.

Table 4 shows the results of this study regarding Jaccard similarity, Hausdorff distance, ASCD and RMSE, averaged over all folds and all leaflet types. All three methods are capable of achieving a Jaccard similarity of more than 80%, while the shape estimation approach reaches the maximal value of around 83% coverage. This is most likely due to the high overlap between the test leaflets and an average one. However, regarding the contour line accuracy, the domain mapping approach provides lower ASCD values, indicating that it outperforms the shape estimation approach regarding finer details of the leaflet shape. As the image resolution was 0.34mmpixel, the domain mapping with an MLP achieves an ASCD of 0.97mm and hence is shown to predict the leaflet contour with sub-millimeter accuracy, with a maximum distance, i.e., Hausdorff distance, of 2.39mm.

TABLE 4
www.frontiersin.org

Table 4. Accuracy of leaflet shape prediction for all methods, regarding RMSE, Jaccard similarity, Hausdorff distance, and ASCD.

In Figure 11, the ASCD is given for each method regarding the accuracy in predicting the different leaflet types, i.e., right-coronary, left-coronary, and non-coronary. The results show that all methods are by far better in predicting the non-coronary leaflet shape than the other two leaflet shapes. This might be due to the fact that the inter-patient shape variance is higher for the right- and left-coronary leaflets than for the non-coronary one (see Figure 3). Additionally, the shape estimation performs worse in predicting the left-coronary leaflet than the right-coronary one, while both accuracy values are comparable in the case of domain mapping, regardless of the regression method. This indicates a higher robustness of the latter approach by utilizing meaningful features that describe the aortic root geometry that were identified during representation learning.

FIGURE 11
www.frontiersin.org

Figure 11. Accuracy of all methods regarding the predicted leaflet type, given as the ASCD (mean ± standard deviation) in pixels.

4.4. Discussion and Outlook

Our results indicate that bridging the high domain gap between volumetric ultrasound images of the aortic root and the planar shape of its three leaflets is possible even on a small amount of data. As the domain mapping with MLP provided personalized synthetic shapes, the ultrasound image carries enough information for this mapping when it is combined with a proper leaflet representation. However, the relevant information seems to be hidden within the ultrasound image and hard to extract as the shape estimation approach was not capable of providing adequately personalized synthetic images. By encoding the images to a latent space representation, it is possible to extract the relevant information, but still a relatively high level of abstraction is needed to make use of it as the optimal MLP featured four hidden layers. Accordingly, even though the RF grasps some parts of this information, its performance lags behind the MLP-based approach. Hence, in this scenario, the domain mapping approach with MLP clearly outperformed the other methods. However, this might be highly task-dependant and it is conceivable that the shape estimation method might perform better in tasks with lower-dimensional surrounding tissue data, i.e., 2D images. A broader analysis of the performance of the proposed methods for different applications and scenarios could provide important insights.

There are only few studies aiming at personalized aortic valve prosthesis shaping (33, 34). As the influence of personalized leaflet shaping on cardiac simulations or prosthesis performance has barely been studied, it is not clear what quantitative value for the accuracy of leaflet synthesis can be considered as clinically acceptable. The methods proposed in this work present the first model to personalize leaflet shapes in a data-driven way and hence open up possibilities for the asssessment of the influence of individual leaflet shaping, for example in a whole-heart simulation or, after fabricating a personalized prosthesis, in a left-heart simulator setup.

In this study, we only aimed on predicting the leaflets' shapes. However, regarding personalized biomechanical modeling or prosthesis development, a prediction of the leaflet's inner structure, i.e., the distribution of the prominent collagen fibers, would also be of high interest. The autoencoder used in this study smooths the inner structure and focuses on representing the general shape, which is desired in the scope of this work. However, if the leaflet representation could be extended to also contain information about the inner structure, the approaches described here would be capable of predicting the full leaflet structure, i.e., the shape and the inner structure. This once again highlights the flexibility of the presented framework.

It is important to note that in this study, all examined aortic roots and valves can be assumed to be healthy. In several clinical scenarios, this is not a limitation, e.g., for personalized aortic valve prosthesis shaping for patients suffering from aortic valve stenosis. In this case, the aortic root shape should not be altered significantly and hence, a prediction should still be possible. In contrast, there are clinical applications of the proposed framework that demand for predicting healthy leaflet shapes from pathological priors as well. This arises for patients suffering from morphological changes of the aortic root, e.g., due to an aneurysm. Therefore, the proposed framework should be evaluated in a setting where the geometric prior corresponds to a pathological state in the scope of a future study. As an alternative, it is possible to estimate the individual healthy aortic root shape based on the pathological one (35). The integration of this estimation step into the clinical pipeline might be a promising approach to apply the proposed framework also in scenarios where only a pathological prior can be assessed.

The output of the proposed method is a planar shape of the leaflet. Regarding the fabrication of personalized aortic valve prostheses, this is actually desirable since the planar shape can serve as a stencil to cut the leaflet out of the fabrication material. If the material mimics the leaflet's biomechanical properties sufficiently, the in-vivo shape under realistic prestrains will mimic the original healthy leaflets shape. However, it should be notes that in some application, the 3D curved shape of the leaflets under prestrains is desired, e.g., for when the fabrication material does not sufficiently reproduce the complex biomechanics or in digital twin scenarios. In these cases, a prediction of the planar, ex-vivo state without external strain is not sufficient. In contrast, Xu et al. (33) or Hsu et al. (34) aimed at directly modeling the 3D curved shape of the leaflet within the aortic root. But it should be noted that the presented framework is very flexible and can easily be adapted to other imaging modalities. Hence, if it is possible to collect image data of the leaflets in-vivo with a sufficient resolution, the framework could be used to predict this curved in-vivo shape as well. The only difference to the presented approach would be a different input for the leaflet autoencoder, the mapping approaches stay the same. This highlights the flexibility of the proposed framework and opens up new possibilities beyond personalized heart valve prosthesis shaping.

5. Conclusion

In this work, we presented a new framework to synthesize the shape of unknown anatomical structures based on the geometry of surrounding tissue by solving a domain mapping problem. We formalized the problem and proposed two general approaches to solve it. In an evaluation of this framework for the application of synthesizing aortic valve leaflet shapes based on volumetric ultrasound images of the aortic root, we could show that our method is capable of reliably synthesizing realistic leaflet shapes and that the geometric prior carries enough information to synthesize personalized leaflet shapes when both domains are encoded into a latent space and an MLP is used to learn a mapping between both latent spaces. This proof-of-concept study does not only open up plenty of applications of personalized aortic valve modeling but also presents a transferable approach for anatomical shape synthesis with geometric prior in general.

Data Availability Statement

The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://gitlab.rob.uni-luebeck.de/robPublic/avlid.

Author Contributions

JH developed, implemented, and evaluated the presented methods. Furthermore, he contributed to the data collection and wrote most of the manuscript. MS contributed to the data collection, the experimental setups and workflows, and revised the manuscript. FE contributed to the general methodological framework as well as to the manuscript. Furthermore, he provided scientific advise during data collection and method implementation. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the KI-LAB Lübeck [funded by the German Federal Ministry of Education and Research (BMBF), grant number 01IS19069] by providing access to an NVIDIA DGX A100, where all computations presented in this work were executed. Furthermore, this study was supported by the German Federal Ministry of Education and Research (grant number 13GW0228) by providing the ultrasound machine and probe. We acknowledge financial support by Land Schleswig-Holstein within the funding programme Open Access Publikationsfons.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

The authors would like to thank Kristian Küssner for his help collecting the data.

References

1. Dominik J, Zacek P. Heart Valve Surgery. Springer Berlin Heidelberg (2010). doi: 10.1007/978-3-642-12206-4

CrossRef Full Text | Google Scholar

2. Labrosse MR, Beller CJ, Boodhwani M, Hudson C, Sohmer B. Subject-specific finite-element modeling of normal aortic valve biomechanics from 3D+t TEE images. Med Image Anal. (2015) 20:162–72. doi: 10.1016/j.media.2014.11.003

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Yankah CA, Weng YG, Hetzer R editors. Aortic root surgery - the biological solution. In: Aortic Valve Repair and Valve Sparing Root Procedures. Heidelberg: Steinkopff (2010). p. 87–191. doi: 10.1007/978-3-7985-1869-8

CrossRef Full Text | Google Scholar

4. Sacks MS, Smith DB, Hiester ED. The aortic valve microstructure: effects of transvalvular pressure. J Biomed Mater Res. (1998) 41:131–41. doi: 10.1002/(SICI)1097-4636(199807)41:1<131::AID-JBM16=3.0.CO;2-Q

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Isola P, Zhu JY, Zhou T, Efros AA. Image-to-image translation with conditional adversarial networks. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Honululu, HI). (2017). p. 5967–76. doi: 10.1109/CVPR.2017.632

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Li J, Adry AM, Peebles J, Schmidt L. On the limitations of first-order approximation in GAN dynamics. In: Dy J, Krause A, editors. Proceedings of the 35th International Conference on Machine Learning (Stockholm). Vol. 80 (2018). p. 3005–13.

Google Scholar

7. Kouw WM, Loog M. A review of domain adaptation without target labels. IEEE Trans Pattern Anal Mach Intell. (2021) 43:766–85. doi: 10.1109/TPAMI.2019.2945942

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Gholami A, Subramanian S, Shenoy V, Himthani N, Yue X, Zhao S, et al. A novel domain adaptation framework for medical image segmentation. In: Crimi A, Bakas S, Kuijf H, Keyvan F, Reyes M, van Walsum T, editors. Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries. Lecture Notes in Computer Science, Vol. 11384. Cham: Springer (2019). p. 289–98. doi: 10.1007/978-3-030-11726-9_26

CrossRef Full Text | Google Scholar

9. Ilse M, Tomczak JM, Louizos C, Welling M, Nl MW. DIVA: domain invariant variational autoencoders. In: Arbel T, Ben Ayed I, de Bruijne M, Descoteaux M, Lombaert H, editors. Proceedings of the Third Conference on Medical Imaging with Deep Learning, Proceedings of Machine Learning Research. Vol. 121 (2020). p. 322–48. Available online at: https://github.com/AMLab-Amsterdam/DIVA

Google Scholar

10. Khandelwal P, Yushkevich P. Domain generalizer: a few-shot meta learning framework for domain generalization in medical imaging. In: Albarqouni S, editor. Domain Adaptation and Representation Transfer, and Distributed and Collaborative Learning. Vol. 12444. Cham: Springer (2020). p. 73–84. doi: 10.1007/978-3-030-60548-3_8

CrossRef Full Text | Google Scholar

11. Sun S, Shi H, Wu Y. A survey of multi-source domain adaptation. Inform Fus. (2015) 24:84–92. doi: 10.1016/j.inffus.2014.12.003

CrossRef Full Text | Google Scholar

12. Goodfellow I, Pouget-Abadie J, Mirza M, Xu B, Warde-Farley D, Ozair S, et al. Generative adversarial nets. In: Ghahramani Z, Welling M, Cortes C, Lawrence N, Weinberger K Q, editors. Advances in Neural Information Processing Systems. Vol. 27. Montreal, QC: Curran Associates, Inc. (2014). p. 2672–80.

Google Scholar

13. Park T, Liu MY, Wang TC, Zhu JY. Semantic image synthesis with spatially-adaptive normalization. In: 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) (Long Beach, CA). (2019). p. 2332–41. doi: 10.1109/CVPR.2019.00244

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Li R, Cao W, Jiao Q, Wu S, Wong HS. Simplified unsupervised image translation for semantic segmentation adaptation. Pattern Recogn. (2020) 105:107343. doi: 10.1016/j.patcog.2020.107343

CrossRef Full Text | Google Scholar

15. Zhu JY, Park T, Isola P, Efros AA. Unpaired image-to-image translation using cycle-consistent adversarial networks. In: 2017 IEEE International Conference on Computer Vision (ICCV) (Venice). (2017). p. 2242–51. doi: 10.1109/ICCV.2017.244

CrossRef Full Text | Google Scholar

16. Pathak D, Krahenbuhl P, Donahue J, Darrell T, Efros AA. Context encoders: feature learning by inpainting. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Las Vegas, NV). (2016). p. 2536–44. doi: 10.1109/CVPR.2016.278

CrossRef Full Text | Google Scholar

17. Pang Y, Lin J, Qin T, Chen Z. Image-to-image translation: methods and applications. IEEE Trans Multimedia. (2021). doi: 10.1109/TMM.2021.3109419

CrossRef Full Text | Google Scholar

18. Liu MY, Huang X, Mallya A, Karras T, Aila T, Lehtinen J, et al. Few-shot unsupervised image-to-image translation. In: 2019 IEEE/CVF International Conference on Computer Vision (ICCV) (Seoul). (2019). p. 10550–9. doi: 10.1109/ICCV.2019.01065

CrossRef Full Text | Google Scholar

19. Lin J, Pang Y, Xia Y, Chen Z, Luo J. TuiGAN: learning versatile image-to-image translation with two unpaired images. In: Vedaldi A, Bischof H, Brox T, Frahm JM, editors. Computer Vision - ECCV 2020. ECCV 2020. Lecture Notes in Computer Science. Vol. 12349 (2020). p. 18–35. doi: 10.1007/978-3-030-58548-8_2

CrossRef Full Text | Google Scholar

20. Pouch AM, Wang H, Takabe M, Jackson BM, Sehgal CM, Gorman JH, et al. Automated segmentation and geometrical modeling of the tricuspid aortic valve in 3D echocardiographic images. Med Image Comput Comput Assist Interv. (2013) 16(Pt 1):485–92. doi: 10.1007/978-3-642-40811-3_61

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Liang L, Kong F, Martin C, Pham T, Wang Q, Duncan J, et al. Machine learning-based 3-D geometry reconstruction and modeling of aortic valve deformation using 3-D computed tomography images. Int J Numer Methods Biomed Eng. (2017) 33:e2827. doi: 10.1002/cnm.2827

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Hagenah J, Evers T, Scharfschwerdt M, Schweikard A, Ernst F. A support vector regression-based data-driven leaflet modeling approach for personalized aortic valve prosthesis development. In: 2018 Computing in Cardiology Conference (CinC) (Maastricht). (2018). p. 1–4. doi: 10.22489/CinC.2018.096

CrossRef Full Text | Google Scholar

23. Hagenah J, Kuehl K, Scharfschwerdt M, Ernst F. Cluster analysis in latent space: identifying personalized aortic valve prosthesis shapes using deep representations. In: Cardoso MJ, Feragen A, Glocker B, Konukoglu E, Oguz I, Unal G, et al., editors. Proceedings of the 2nd International Conference on Medical Imaging with Deep Learning, Proceedings of Machine Learning Research (London). Vol. 102. (2019). p. 236–49.

Google Scholar

24. Crick SJ, Sheppard MN, Ho SY, Gebstein L, Anderson RH. Anatomy of the pig heart: comparisons with normal human cardiac structure. J Anat. (1998) 193:105–19. doi: 10.1046/j.1469-7580.1998.19310105.x

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Oyer PE, Stinson EB, Reitz BA, Craig Miller D, Rossiter SJ, Shumway NE. Long-term evaluation of the porcine xenograft bioprosthesis. J Thorac Cardiovasc Surg. (1979) 78:343–50. doi: 10.1016/S0022-5223(19)38100-0

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Hagenah J, Werrmann E, Scharfschwerdt M, Ernst F, Metzner C. Prediction of individual prosthesis size for valve-sparing aortic root reconstruction based on geometric features. In: 2016 38th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC) (Orlando, FL). (2016). p. 3273–6. doi: 10.1109/EMBC.2016.7591427

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Bengio Y, Lamblin P, Popovici D, Larochelle H. Greedy layer-wise training of deep networks. In: Schölkopf B, Platt J, Hoffman T, editors. Advances in Neural Information Processing Systems, Vol. 19. Vancouver, BC: MIT Press (2007).

28. Russell S, Norvig P. Artificial Intelligence: A Modern Approach. 2nd Edn. Boston, MA: Pearson Education Inc. (2003).

Google Scholar

29. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. (2015) 521:436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. In: 3rd International Conference on Learning Representations, ICLR 2015. San Diego, CA (2015). Available online at: http://arxiv.org/abs/1409.1556

Google Scholar

31. Ito KL, Kim H, Liew S. A comparison of automated lesion segmentation approaches for chronic stroke T1-weighted MRI data. Hum Brain Mapp. (2019) 40:4669–85. doi: 10.1002/hbm.24729

PubMed Abstract | CrossRef Full Text | Google Scholar

32. van der Maaten L, Hinton G. Visualizing data using t-SNE. J Mach Learn Res. (2008) 9:2579–605. Available online at: http://jmlr.org/papers/v9/vandermaaten08a.html

Google Scholar

33. Xu F, Morganti S, Zakerzadeh R, Kamensky D, Auricchio F, Reali A, et al. A framework for designing patient-specific bioprosthetic heart valves using immersogeometric fluid-structure interaction analysis. Int J Numer Methods Biomed Eng. (2018) 34:e2938. doi: 10.1002/cnm.2938

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Hsu MC, Kamensky D, Bazilevs Y, Sacks MS, Hughes TJR. Fluid-structure interaction analysis of bioprosthetic heart valves: significance of arterial wall deformation. Comput Mech. (2014) 54:1055–71. doi: 10.1007/s00466-014-1059-4

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Hagenah J, Mehdi M, Ernst F. Generating healthy aortic root geometries from ultrasound images of the individual pathological morphology using deep convolutional autoencoders. In: 2019 Computing in Cardiology (CinC) (Singapore). (2019). p. 1–4. doi: 10.22489/CinC.2019.091

CrossRef Full Text | Google Scholar

Keywords: deep learning, shape synthesis, representation learning, domain gap, aortic valve

Citation: Hagenah J, Scharfschwerdt M and Ernst F (2022) Aortic Valve Leaflet Shape Synthesis With Geometric Prior From Surrounding Tissue. Front. Cardiovasc. Med. 9:772222. doi: 10.3389/fcvm.2022.772222

Received: 07 September 2021; Accepted: 31 January 2022;
Published: 09 March 2022.

Edited by:

Sasha A. Singh, Brigham and Women's Hospital and Harvard Medical School, United States

Reviewed by:

Ankush Aggarwal, University of Glasgow, United Kingdom
Takaharu Asano, Brigham and Women's Hospital and Harvard Medical School, United States

Copyright © 2022 Hagenah, Scharfschwerdt and Ernst. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Floris Ernst, ZXJuc3QmI3gwMDA0MDtyb2IudW5pLWx1ZWJlY2suZGU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.