- 1Division of Cardiology, Department of Paediatrics, The Hospital for Sick Children, University of Toronto, Toronto, ON, Canada
- 2Department of Pediatric Cardiology, Baylor College of Medicine, Houston, TX, United States
- 3Centre for Innovative Ultrasound Solutions and Department of Circulation and Medical Imaging, Norwegian University of Science and Technology, Trondheim, Norway
Artificial intelligence (AI) is frequently used in non-medical fields to assist with automation and decision-making. The potential for AI in pediatric cardiology, especially in the echocardiography laboratory, is very high. There are multiple tasks AI is designed to do that could improve the quality, interpretation, and clinical application of echocardiographic data at the level of the sonographer, echocardiographer, and clinician. In this state-of-the-art review, we highlight the pertinent literature on machine learning in echocardiography and discuss its applications in the pediatric echocardiography lab with a focus on automation of the pediatric echocardiogram and the use of echo data to better understand physiology and outcomes in pediatric cardiology. We also discuss next steps in utilizing AI in pediatric echocardiography.
Introduction
Artificial intelligence (AI), the study and development of computational algorithms that mimic human cognitive functions such as learning and thinking, is used in non-medical fields to assist with automation and decision-making (1). There are multiple tasks AI is designed for that could improve the quality, interpretation, and clinical application of echocardiographic data at the level of the sonographer, echocardiographer, and clinician. Machine learning (ML), algorithms whose goal is to learn patterns from data to improve at a given task, is a field within AI that has been applied to many tasks in medical research. ML algorithms often require some engineering of the input features (i.e., the process of creating new variables from the data such as extracting pixel density, peak velocity from a spectrogram, or a measurement from an image) prior to developing the model which can be both time consuming and challenging especially for large high dimensional datasets like images and videos. Deep learning (DL) is a subset of ML algorithms that allow more flexibility in approximating the underlying structure of the data which leads to less feature engineering requirements to obtain accurate predictions. While this approach is especially appealing for high dimensional data such as echocardiograms, this flexibility is at the cost of increased complexity which has its own shortcomings (e.g., ‘black-box' decision-making and computational cost) (2).
Advances in pediatric cardiac imaging have proven challenging due to the complexity of pediatric heart disease and the impact of growth. The potential for ML in pediatric cardiology, especially in the pediatric echocardiography laboratory, is very high. While there are several excellent recent reviews that have highlighted the clinical applications of AI in medicine and cardiology, reviews specific to AI applied in the pediatric echocardiography laboratory are lacking (1, 3–5). In this state-of-the-art review, we identify needs unique to the pediatric echocardiography laboratory that could be addressed by AI, the recent applications of ML in pediatric echocardiography, and perspectives on future directions of AI in the pediatric echocardiography laboratory.
Part I: Optimizing the pediatric echocardiogram
Echocardiography is one of the fundamental technologies that helps guide the diagnosis and treatment of children with congenital heart disease (CHD). AI, especially DL, has been implemented in other fields where it has excelled in tasks using unstructured data (e.g., raw images and video clips) such as facial recognition and automated driving (2). Here, we describe the recent literature on using deep learning to automate and optimize echocardiographic acquisition, image optimization, measurements, and diagnosis in pediatric cardiology (Tables 1, 2).
Facilitating image acquisition and automating optimization
In congenital and pediatric cardiology, the quality of ultrasound image acquisitions is guided by the proper implementation of published practice guidelines (16). However, despite best efforts to adhere to consensus standards, misdiagnoses may still occur as echocardiography is a highly operator-dependent technique. This may be related to errors at the acquisition level such as sonographic planes being incorrectly obtained or the defect visualized on screen but not recognized by the operator (17) (Figure 1). AI has been shown to achieve human-level performance in some medical imaging analysis tasks (18). This raises the potential for automating aspects of the pediatric ultrasound scan, including automated image identification and measurements in real-time. In fetal CHD screening, proof of concept studies using AI tools during the acquisition phase have been shown to positively impact the efficiency and quality of the fetal examination compared to a standard manual scan. Recently, Matthew et al. showed that using AI-embedded tools to reduce repetitive tasks (e.g., measuring fetal biometry and manually acquiring video clips of standard fetal views) may allow more attention to be directed to obtaining accurate morphological diagnoses (19). They used an ensemble of convolutional neural networks (CNN), a type of DL algorithm that is versatile in performing tasks related to images and videos, that were trained for anatomic measurements and image classification (20). Komatsu et al. (6) also found that CNNs could be trained to achieve an automatic detection of each cardiac substructure in fetal ultrasound videos, and showed this could be applied to assist in detecting cardiac structural abnormalities.
Figure 1. Overview of tasks in the pediatric echocardiography laboratory that could be assisted with artificial intelligence workflows.
Transthoracic echocardiographic measurements and interpretation are heavily reliant on optimal probe positioning and insonation angle (16). Østvik et al. (21) trained a CNN to automatically identify if the operator has positioned the probe correctly to obtain optimal angles for seven different cardiac views (e.g., parasternal long axis, apical 2/3/4 chamber). They demonstrated the CNN's ability to do this in real time such that it could facilitate optimal acquisition at the bedside. Issues of image optimization for automation can be more challenging in fetal and pediatric echocardiography, despite having generally better image quality. Indeed, contrary to adult cardiology, multiple ultrasound probe types can be used at different frequencies and different frame rates. This issue of acquisition heterogeneity is especially troublesome in AI as the impact on image quality and the ensuing lack of homogeneity is detrimental in AI-model training if not accounted for (22). For this purpose, AI-assisted feedback systems have been proposed to facilitate optimization of parameters including depth, gain and frequency (23). Additionally, AI-based denoising and artifact removal tools have been recently developed for transthoracic echocardiographic imaging in congenital heart disease that can further standardize image quality (7). This allows a new generation of standardized high-quality images that can be used for AI-based tools for measurement, segmentation, and classification.
While these proof-of-concept studies support that AI can address a strong clinical need within echocardiography, each study has developed their own proprietary means to integrate their AI algorithms and operator interface with the ultrasound scanner. Extensive collaboration and development with industry is needed before these novel innovations can be incorporated into readily available commercial packages.
Automating echocardiography laboratory tasks
After image acquisition and prior to the comprehensive interpretation of an echocardiogram, several intermediate steps need to be manually performed including view classification, segmentation of cardiac structures (e.g., “tracing” a clinically relevant structure within an image), and other quantitative measurements that rely on view classification and segmentation [e.g., ejection fraction (EF)].
While identifying an echocardiographic image to an experienced operator is a simple task, it is because training a sonographer or cardiologist in finding the correct view is often the first step in understanding echocardiography. Subsequently, training an algorithm to classify echocardiographic views is an important first step in creating an AI workflow, especially for automation tasks (2). This has been demonstrated to be feasible in adult studies (21, 24, 25). More recently, Arnaout et al. used 1,326 fetal echocardiograms to train a CNN to identify five standard fetal views (e.g., three-vessel view, left ventricular outflow tract view) with an AUC range of 0.72–0.88 (8). Furthermore, Gearhart et al. used a similar approach to perform automatic image view classification on 12,067 individual transthoracic pediatric echocardiographic images (9). The authors showed this model identified 28 preselected views with 90% accuracy.
In echocardiography, we perform segmentation of cardiac structures to assess abnormalities in cardiac morphology (e.g., left ventricular end diastolic diameter to assess for dilation) and to be used in other quantitative measurements (e.g., end systolic/diastolic left ventricular area to derive ejection fraction). When an AI algorithm is tasked to use an image to segment a cardiac structure, it needs to be trained on a pre-labeled set of training images, supervised learning. The accurate labeling (i.e., establishing the ground truth) of this dataset, which is often manually performed, is a key aspect to creating a strong model. However, in clinical medicine, such labels are not always clear-cut. For example, if a researcher was interested in developing a model to facilitate the echocardiographic screening of pediatric hypertrophic cardiomyopathy (HCM), the current guidelines on the diagnosis of HCM recommend a ventricular wall thickness z-score of >2.5 as a potential cutoff to screen asymptomatic children (26). Yet, it is unclear which z-score criteria to use (e.g., Detroit, Boston, PHN) and there may be inconsistency in the measurement itself (e.g., determining how to exclude right ventricular muscle bundles for interventricular septal diameter) all of which lead to imperfect labeling and thus a less accurate model. With that said, segmentation tasks with clearly defined labels have been proven to be comparable to manual segmentation in adult echocardiographic studies (25). However, in pediatrics sample size and variability in cardiac morphology are common challenges to the development of pediatric-specific AI models. Guo et al. (10) was able to develop a pediatric-specific DL model that is based on a fully convolutional network (FCN), a type of CNN that is designed for segmentation tasks common in medical imaging, to segment and perform measurements on the left atrium and left ventricle. They developed novel methods to accommodate for variability in size and heart rate prevalent in children. Other methods are being developed to augment the abilities of deep learning on smaller datasets such as the use of generative adversarial networks (GAN) (27, 28). GANs generate simulated data based on the original training images to improve the model's ability to perform a certain task; Arafati et al. developed an FCN to perform segmentation of atrial and ventricular chambers in the 4-chamber view and used a GAN to augment the 450 adult echocardiograms used resulting in a dice metric of 86%−92%, a measure of the degree of overlap between the model's segmentation and the manual segmentation.
It is possible for a deep learning algorithm to replicate the steps a human would take to perform a quantification task. For example, Zhang et al. (25) developed a CNN that automatically identified the apical four chamber view, selected the frames that best represented end-diastole and end-systole, traced the LV endocardial border, and then derived an ejection fraction from those steps. In contrast, Ouyang et al. used a CNN architecture that considers both spatial and temporal features (i.e., spatiotemporal convolution) in predicting EF (Simpson's) and did not restrict the model predictions solely based on end-diastolic volumes derived by endocardial area segmentation. This permitted the model to have more freedom to derive its own spatiotemporal features leading to a more accurate and consistent prediction (29). The downside to this approach is that is less transparent than the former model whose pipeline makes each part of the segmentation process explicit. This model was recently adapted for children where it was retrained on a pediatric dataset to not only use apical four chamber views but also parasternal short views to estimate EF by 5/6 area length method with an R2 of 0.78 (11). This so-called ‘transfer learning' technique of fine-tuning a model previously trained on a different dataset to be optimized for a different task is a methodology unique to DL. In practice it allows the practitioner to develop a model with significantly less data, a problem that frequently occurs in pediatric cardiology (30, 31).
Apart from traditional methods to assess ventricular function, other functional measurements including strain imaging can be fully automated. However, compared to adult echocardiography labs the uptake of the method has been slower in pediatric cardiology (32). This can be explained by the anatomical variability present in congenital cardiology, the absence of dedicated post-processing software adjusting for different probe frequencies, and the limitations in frame rates relative to the higher heart rates present in younger children, especially in infants (33). Some of these limitations could be addressed using an AI-based approach to strain imaging. An adult study has recently demonstrated that, using B-mode image acquisitions, an AI algorithm could automatically derive global longitudinal strain measurements with minimal measurement variability (34). Applying AI-based automated strain measurements in children would potentially result in improving its applicability in pediatric heart disease and better understanding of the factors that influence strain imaging in children (35, 36).
Several of the studies in this section have obtained large multicenter retrospective datasets to train and test their algorithms. However, this does not preclude these algorithms from prospective real-world trials to test their efficacy and generalizability. Indeed, a number of these studies rely on open-source datasets which have its own set of limitations (e.g., variable quality and number of images, poor labeling, etc.) which could induce bias if not properly accounted for (22). To address this, imaging biobanks are being developed whose goal is to provide standardized medical image collections for the development of higher quality models (37). Furthermore, ML is the norm within the field of radiomics, the study of deriving and analyzing imaging biomarkers from conventional medical images to aid in clinical decision support systems, and there is a strong push toward the standardization of not only image acquisition but of the entire radiomics pipeline from the development of features and ML algorithms, to its implementation in the clinical workspace (e.g., the image biomarker standardization initiative) (38).
Diagnosis of heart disease
The past 5 years have seen research efforts in developing AI-based tools to assist with congenital heart disease diagnosis. This would be useful as democratizing the diagnosis of significant congenital heart disease to non-expert echocardiography users would make the screening for CHD more efficient in particular in fetal echocardiography. First-line screening for CHD is typically provided by obstetricians during routine anatomic scans during the second trimester. Despite newer guidelines recommending a more comprehensive approach to screening including additional fetal views, CHD continues to be missed (17, 39). Having AI-based methods for screening would likely increase the detection rates for significant CHD. Chotzoglou et al. (12) were interested in developing an AI algorithm to screen for abnormal hearts on fetal echocardiography using a method called one-class anomaly detection. This method trains a deep learning model on normal echocardiographic data using generative adversarial networks and an autoencoder (α-GAN). Here, the GAN involved two separate networks: the first network (generator) creates simulated echocardiographic images based on the training data given to it, and a second model (discriminator) tries to discriminate if a given image is simulated or real. With this method, the α-GAN model was able to distinguish normal hearts from HLHS with an AUC of 0.81 when exposed to a fetal ultrasound dataset from a single center. To understand if the model was making clinically intelligible decisions, they applied a gradient-weighted class activation map to the model which visualizes which pixels the model deemed most important in detecting an abnormal image (Figure 2). This approach demonstrated that an AI-based model could help in screening for specific types of critical CHD. This will need to be tested for other types of CHD since HLHS was deliberately chosen as the initial lesion as it is grossly abnormal in the four-chamber view and thus identifiable from a single imaging plane. Arnaout et al. (8) further assessed the clinical applicability of deep learning as a screening tool by testing an ensemble of neural networks on a larger more heterogenous dataset of 107,823 images derived from multiple sources. They utilized a novel CNN algorithm for classification of fetal images and found an AUC of 0.95–0.99 discriminating normal from abnormal from a test set of many complex CHD.
Figure 2. Gradient-weighted class activation mapping (Grad-CAM) helps us understand what aspects of an image are important in a deep learning model's decision-making. Chotzoglou et al. developed a DL model that classified fetal images as normal or abnormal cardiac anatomy. They used Grad-CAM mapping to demonstrate that the model identifies cardiac structures as important in understanding if the fetal ultrasound image represents a normal or abnormal heart. Figure adapted from Chotzoglou et al. (12).
As a decision support tool, AI could further help transthoracic echocardiography operators in diagnosing different types of CHD. For instance, two studies demonstrated the ability for DL models to diagnose septal defects including ASDs, VSDs, and AVSDs (13, 40). Additionally, Diller et al. (41) trained a CNN to classify apical 4-chamber and parasternal short-axis images as congenitally corrected transposition of the great arteries (ccTGA), d-TGA, and normal with an accuracy of 98%. Furthermore, they used transfer learning to adapt a CNN that was previously developed for biomedical image segmentation tasks to segment the endocardial border of the systemic ventricle (Dice score 0.79 for ccTGA). Of note, current error rates in CHD diagnosis in experienced pediatric echocardiography laboratories are extremely low relative to the number of studies performed, often with limited therapeutic impact. Instead, DL could prove more useful in developing models for specific interventions that diagnostic imaging is used to help with. For example, a CNN could be trained on echocardiographic data to accurately predict the correct atrial septal defect occlusion device, patent ductus arteriosus closure device, or pulmonary valve replacement device size and type. Using ML to assist in specific interventions in CHD is further elaborated below (Part II, subsection: Congenital Heart Disease).
Part II: Machine learning to understand pediatric heart disease
While AI applications in echocardiography have largely focused on automation of different tasks, it can also be used to obtain a deeper statistical understanding of data by identifying cardiac phenotypes or associated factors to clinical outcomes. In the pediatric echocardiography lab, this could be used to understand how our non-invasive assessments reflect cardiac physiology.
Explainability in machine learning
Supervised ML is where an algorithm's goal is to best predict the label (e.g., outcome) for a given set of pre-labeled observations. Some supervised algorithms, especially DL models, are often considered “black box” techniques in that it is difficult to understand how predictions are derived for a given model. The field of interpretable machine learning has been developed in order to explain model predictions which could help improve clinical acceptance of the ML model (42). For instance, for a given set of features (i.e., variables), we can compute how important an individual feature is by evaluating the impact on prediction accuracy in a dataset when those feature's values are not used in its prediction (e.g., permutation feature importance). To understand how features are used in a model, we can develop a global surrogate model to approximate the predictions of the original comprehensive model. Surrogate models provide a more transparent way of assessing feature usage (e.g., regression coefficient table in a logistic regression or feature value cutoffs in a decision tree) and by assessing the performance of the approximated model to the original comprehensive model, we can understand how accurate the surrogate model is (43, 44).
In unsupervised machine learning, there are no labels or outcomes annotated with each observation, and instead the algorithm's task is to understand the structure of the data with tasks typically involving reducing the amount of redundant variables/features (e.g., dimensionality reduction) or quantifying the similarity between patients/observations (e.g., cluster analysis; Figure 3). Thus, the end-goal in unsupervised learning is finding relationships in the data itself. The potential of this technique to identify imaging phenotypes in echocardiography is strong due to the high dimensional and complex data generated from an echocardiogram when representing cardiac form and function (5). Unsupervised techniques are often used in combination with other techniques in a ML pipeline. For example, dimensionality reduction (e.g., principal component analysis, multiple kernel learning, autoencoders) can be used initially to identify the most important features in order to facilitate the efficiency and interpretation of a subsequent cluster analysis (15). It can also be combined with a supervised learner such as in generative adversarial networks (7). Finally, it can be used on the output of the supervised learning model to improve our understanding of how it chose to group patients (24).
Figure 3. Unsupervised learning techniques such as dimensionality reduction and cluster analysis have improved our ability to identify imaging phenotypes [Garcia-Canadilla et al. (15)]. (A) Non-parametric dimensionality reduction techniques (multiple kernel learning) quantify the similarity between patients based on their echocardiographic inputs (Doppler velocity and ventricular strain tracings). This plot is a representation of how patients are positioned by their similarities based on dimensionality reduction. K-means clustering then identified five separate groups of patients based on these similarities (different colors represent different clusters). (B) Each cluster has clinically distinct characteristics. Clusters 1 and 2 were healthy volunteers. Clusters 3–5 were DCM patients. In particular, cluster 5 had the oldest patients and had a relatively increased usage of oral and IV medications. (C,D) Representative mitral, aortic and pulmonary vein Doppler velocity patterns normalized for one full cardiac cycle for each group are seen here. Patients with DCM (Clusters 3–5) have relatively abnormal Doppler tracings compared to healthy volunteers (clusters 1–2). Figure adapted from Garcia-Canadilla et al. (15).
Deep learning has been very successful in performing tasks on high dimensional data and can be given both supervised and unsupervised tasks. If one considers each pixel to be an individual input feature, then an individual echocardiogram could potentially have millions of features (45). Therefore, the high dimensionality of imaging data like echocardiograms proves to be an excellent substrate for DL. The advantage of DL over traditional ML algorithms is that it can derive very complex abstract relationships for a given input with relatively less engineering of those input features. This capability is, in large part, due to several layers that analyze features in highly non-linear ways to establish mapping between the features and labels/outcomes. A major trade-off of this complexity is that it its decision-making process is increasingly abstract. Methods have been developed to specifically address explainability for this unique aspect of DL. For example, Chotzoglou et al. (12) used gradient-weighted class activation maps (Grad-CAM) to visualize which parts of an image are most important in detecting if the fetal echocardiogram was normal or abnormal (Figure 2). This provides a way for the clinician to both verify if the model decided correctly and to assist the provider in deciding if additional views are needed. As another example, Madani et al. used t-distributed stochastic neighbor embedding (t-SNE) to visualize how their deep learning model classified images into different views (Figure 4). t-SNE is a dimensionality reduction technique that has been designed to organize observations based on how similar their learned features are (i.e., activation maps) according to the DL model. Using this technique, error analysis of the mis-classified observations can facilitate understanding the strengths and inherent biases of the model. Other ways of understanding model bias include occlusion experiments and test data with artificially created image artifacts (7, 24). Finally, exploring the causal relationship between a set of features and an output is one of the goals of clinical inference (46). eXplainable AI is a field of research that attempts to reduce the issues of black-box methods including developing AI models within a framework of causality. One of its goals is to develop a “Human-AI interface” which allows the user to interrogate the model (e.g., counterfactual “what-if” questions) to gain insight into the model's decision-making process (47). While the terms interpretable and explainable are sometimes used interchangeably, some consider eXplainable AI the approach to develop models that explain why it came to its decisions while interpretable ML seeks to describe how a model came to its decision (48).
Figure 4. T-distributed stochastic neighbor embedding (t-sne) is a deep learning-specific dimensionality reduction visualization tool that plots how similar each image is according to the deep learning model. Madani et al. developed a convolutional neural network (CNN) to classify echocardiographic images into 18 different views, and they used t-sne to understand how the CNN analyzed each image with respect to each other. Visually distinct views tended to group farther away from each other such as continuous wave Doppler (black; cw) and apical 4 chamber view (blue; a4c) while pulse wave (gray; pw) was very similar to continuous wave and tended to overlap. Figure adapted from Madani et al. (24).
Congenital heart disease
Since the introduction of B-mode echocardiography into clinical practice over 50 years ago (49), congenital heart specialists have honed and exceled at deriving anatomic relationships non-invasively to guide surgical decision making. Nevertheless, there are many surgical management options that rely on echocardiography where there is still clinical equipoise such as timing of neonatal Tetralogy of Fallot repair or surgical management of the patient with borderline left ventricle (LV). Most echocardiographic clinical research on borderline LV involves a reductionist approach whereby clinically accepted measures on an echocardiogram that are thought to potentially relate to prognostic information are studied to assess their strength in relating to outcomes. This approach is necessary in clinical echocardiographic research due to the abundance of information that each echo provides without a methodically robust way of identifying patterns within them. The risk scores that have been developed over the past 20 years with this approach have largely focused on aortic valve, left ventricle length, and mitral valve size (50). Yet, these heuristic-based algorithms account for some, but not all, of the complex patterns in an echocardiogram that could be useful in diagnosis (51).
Meza et al. (14) helped demonstrate that unsupervised machine learning could help identify patterns in echocardiographic data that could be clinically relevant to diagnosis and prognosis of patients with borderline left ventricle. They collected 194 functional and morphologic variables in each echocardiogram of neonates with ductal-dependent hypoplastic left-sided structures or aortic stenosis/atresia. They performed hierarchical clustering using this echo data alone without any additional clinical or outcomes data to reduce bias in understanding similarity between patients. It identified three distinct groups of patients which corresponded to multi-level LV hypoplasia, hypoplastic left heart syndrome, and critical aortic stenosis. Accordingly, surgical decision and mortality were distinguishable between groups with mortality and single ventricle palliation being the highest in the hypoplastic left heart group. Within these groups, they found that aortic valve atresia and LV end-diastolic volume best distinguished between the groups as determined by multinomial regression and linear discriminant analysis. Mitral valve characteristics and pulmonary vein anomalies, parameters often used in clinical practice to help guide clinical management, were not found to be significant in distinguishing between the three groups. This study is important in that it used echocardiographic data on congenital heart disease patients to define statistically driven variables of importance with a technique that was free of any a priori assumptions about the relationships between the variables.
Cardiomyopathy
The manual extraction of measurements for a given set of data, like peak E wave, deceleration time, E/A ratio from a mitral inflow Doppler spectrogram, summarizes the data source into a set of expert-crafted features (e.g., Meza et al. extracted a set of 194 features). Performing manual extraction of a set of inputs can be time consuming, and subtle patterns not previously identified may be lost with this approach. To overcome this limitation, Garcia-Canadilla et al. explored the use of unsupervised learning directly on left ventricular longitudinal strain, aortic outflow Doppler, pulmonary vein Doppler, and mitral inflow Doppler velocity tracings as inputs to assess whether echocardiographic imaging phenotypes could be associated with clinical characteristics and outcomes in dilated cardiomyopathy. They used multiple kernel learning to perform dimensionality reduction and organize patients in accordance with their similarity in feature values, performed k-means cluster analysis to identify groups with similar phenotypes, and most importantly they were able to explore how each of the strain/Doppler tracings are represented in the output space. In other words, they were able to visualize how strain and Doppler velocity tracings would look like for a given group of patients. With this approach, we can potentially understand and identify how systolic and diastolic dysfunction in DCM can be represented by subtle patterns in strain and Doppler velocity data; and further, how they relate to clinical course and risk for adverse outcome. The analysis of strain and Doppler patterns rather than absolute values provides useful information. Moreover, this study used Doppler and strain data over the entire cardiac cycle, thereby providing more comprehensive data than current approaches which typically measure data at single point (e.g., peak systolic strain).
Both cluster analysis and dimensionality reduction techniques described are excellent at using underlying statistical patterns in echocardiographic data to relate observations and features to each other. However, just like in supervised learning methods, unsupervised algorithms do not necessarily make explicit how they came to their decisions. For example, cluster methods do not quantify which features are similar within a group of observations, and dimensionality reduction techniques often render feature values into abstract estimations. While the output of these unsupervised learners may be adequate for a given situation, there are methods being developed to help elucidate underlying statistical patterns including model-agnostic interpretability methods to describe feature importance in cluster analysis (52). For dimensionality reduction, separate algorithms have been developed that retain feature values after the dataset has been reduced into the low output space (CUR) (53). Finally, while the unsupervised learning techniques described in Part II are excellent at identifying key patient groups and features, they are not validated as a predictive model. Thus, in isolation, they are hypothesis generating techniques for the purposes of understanding how physiology is reflected in patterns within echocardiographic data. Unsupervised learning is often paired as a data preparation step with supervised learners and these algorithms could be used in an ML pipeline to accurately predict surgical technique (e.g., determining surgery for LV obstruction in CHD) or risk of diastolic heart failure (relating echocardiographic phenotypes of diastolic function in DCM).
Perspectives and future steps
The non-invasive assessment of the heart through echocardiography provides us representations of the interplay between cardiac anatomy and physiology. We often perform a reductionist approach to identify the key features of an echo that are most associated with the underlying pathophysiology of the heart as it relates to clinical signs and symptoms. This approach, though easy to perform, disregards complexity and nuances in an echocardiogram that could potentially improve detection of physiologic/anatomic changes in cardiac health and disease and their association with clinical outcomes. Machine learning has been used successfully over the past few years in identifying more subtle and complex patterns in echocardiographic data that can strengthen our understanding of how cardiac (patho)physiology is represented in an echocardiogram (5). This ability to learn and categorize patterns, in conjunction with modern day computing power, provides us the ability to not only improve our understanding of cardiac anatomy and physiology, but optimize and automate logistic tasks in the clinical echocardiography laboratory.
Challenges in pediatric cardiology
Validation of translational diagnostic tools is necessary to promote their use in clinical practice with the goal of becoming standard of care. This is especially true in pediatric cardiology where a range of anatomic variability and loading conditions, which may change over time, pose a challenge (54). Although the standard of practice in AI workflows includes a step where the model is tested on unseen data (e.g., a test or holdout set), unidentified biases and poor generalizability are still issues if the training/test set is limited in size and heterogeneity (e.g., data exclusively from developed countries, a single center, one type of imaging machine/vendor, etc.). Thus, prospective multicenter validation is still required to assess for generalizability and help identify previously unrecognized biases in the model (2, 55). Indeed, these considerations, in addition to the small sample sizes prevalent in pediatric cardiology and the inherent difficulty in translating predictive models into clinical practice, are all hurdles that need to be overcome for widespread use to occur (45, 56, 57). Ways to facilitate bias reduction and prospective validation of a model to promote widespread use includes decentralizing the AI algorithm, so-called federated learning. This developing approach is an alternative to sending de-identified data to a central storage system where the algorithm is then trained. Instead, the model is brought to each collaborating center where it is trained on the data locally which reduces the effort of data de-identification/transfer and can potentially expedite the adoption of the AI model (22). The potential downside to this approach is that quality control may be more challenging without all images being assessed and processed in a central core laboratory. Finally, while AI innovations continue to advance at a rapid pace, it is critical that the proper governance and regulatory oversight is in place to establish a secure and ethical standard (58). For example, equitable inclusion of patients during model development or maintaining high security standards against data breaches will promote acceptance of AI as a whole in the clinical community.
AI and the clinician
Other reasons why AI is not as readily accepted in clinical practice is due to skepticism and unfamiliarity of AI with the clinician. On top of the standard rigor needed for a tool to be clinically validated, it is important that the user (i.e., the sonographer and practicing physician) is familiar with the strengths and weaknesses of the tool, is facile in its use, and can ultimately trust in its abilities. While there are conventional therapies routinely used in medicine whose mechanism of action is not fully-elucidated, understanding the process by which a diagnostic tool achieves its decision is very important in the acceptance of the model in clinical practice. Indeed, certain AI modalities (e.g., DL) are notorious for their black box nature, but the fields of interpretable machine learning and eXplainable AI have been developed to address this issue (Please see Explainability in machine learning). Further, it should be noted that fully automated AI have gained notoriety for not only being successful at mimicking human intelligence, but also for the critical errors that inevitably occur (e.g., the self-driving car that runs a red light). It is unlikely that fully-automated AI will exist in clinical practice without physician oversight and instead it will fulfill a much-needed role as a reliable partially automated tool (59). In addition, with the adoption of any form of automation, there is an increased risk of automation bias, an error where the user trusts the automated calculation despite overt clinical evidence suggesting it is incorrect (1). Thus, the active training of physicians to critically appraise AI models as well as training in how to use it in the clinical workspace is needed.
This review has discussed literature on optimizing and improving the clinical pediatric echocardiogram workflow. Though there is little research on this currently in pediatric echocardiography, using AI to understand clinical narratives in the echocardiogram report can both improve report consistency and enrich its diagnostic utility. Natural Language Processing (NLP) is a branch of AI and linguistics devoted to performing tasks related to speech and text, and there is a growing body of work on applying it to radiology report data (60). Tasks NLP models were designed to address include improving quality compliance by identifying if patient indications for scans adhered to study guidelines and institutional protocols (61). Many studies focused on disease surveillance including extracting relevant information for a particular disease and tracking key features longitudinally over several diagnostic reports (60).
AI, echocardiography, and the patient
While AI has great potential to improve the standard of care in patients who are diagnosed with echocardiography, equity is an issue that needs to be accounted for early in the development of AI algorithms. Centers who perform AI research will naturally include more of their patients relative to other centers, and consequently the models will tend to be trained and tested on these populations. If not accounted for, the fitted model could make incorrect decisions if applied to a new setting due socioeconomic factors as well as variations in local practice (e.g., indications for echocardiogram, differences in image acquisition protocol) leading to significant sampling bias (22). For example, AI can improve prenatal detection of CHD especially in practices that serve disadvantaged communities where pediatric cardiology expertise may not be readily available. However, under/overrepresented medical conditions endemic to that community could impact the accuracy of the DL model if not adequately addressed (62). In addition, poor insurance coverage may limit the access to deep learning tools (62).
AI and translational medicine
AI provides a means for rapid innovation in medicine as it is designed to perform tasks efficiently on data structures commonly used in medical research (e.g., images, video clips, and tabular datasets). However, despite all the promising results in the studies featured in this review, AI is still a translational technology which carries with it unique problems both old and new. Namely, the process of translating a novel idea into a tool used in standard of care is a complex one which involves more than just the rigorous academic stages from in vitro experimentation to multicenter clinical validation. Indeed, widespread clinical use of a novel technology requires regulatory approval from regional government agencies as well as industry partnership to help facilitate the accessibility of the technical innovation. It is impossible for one person to be expert in all of these facets of translational medicine, and thus a multidisciplinary team of collaborators who partner well between clinicians, academia, industry, and government is needed to shepherd these novel AI tools to clinical implementation (63).
Conclusion
The echocardiogram is the first-line imaging tool for the cardiologist due to its ability to allow the clinician to quickly identify cardiac anatomy and physiology. With the advent of AI in medical imaging, we can extend the utility of a cardiac ultrasound beyond what is immediately apparent to explore patterns previously unseen and make diagnoses more accurately and efficiently. Indeed, based on current trends, we expect that the next era of pediatric echocardiography will be data-centric where AI will augment and integrate the role of the sonographer, echocardiographer, and clinician to improve patient care. The development of the AI-based pediatric echocardiography laboratory of the future will however be a long path with many expected obstacles, given the complexity of pediatric heart disease.
Author contributions
MN and LM contributed to the design of the study. MN and OV wrote the first draft of the manuscript. MF, LL, CR, and LM wrote sections of the manuscript. MN and LM performed final revisions of the manuscript. All authors contributed to the article and approved the submitted version.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Coppola F, Faggioni L, Gabelloni M, De Vietro F, Mendola V, Cattabriga A, et al. Human, all too human? An all-around appraisal of the “artificial intelligence revolution” in medical imaging. Front Psychol. (2021) 12:1–15. doi: 10.3389/fpsyg.2021.710982
2. Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. (2019) 25:44–56. doi: 10.1038/s41591-018-0300-7
3. Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. (2019) 380:1347–58. doi: 10.1056/NEJMra1814259
4. Quer G, Arnaout R, Henne M, Arnaout R. Machine learning and the future of cardiovascular care: JACC State-of-the-Art Review. J Am Coll Cardiol. (2021) 77:300–13. doi: 10.1016/j.jacc.2020.11.030
5. Seetharam K, Raina S, Sengupta PP. The role of artificial intelligence in echocardiography. Curr Cardiol Rep. (2020) 22:99. doi: 10.1007/s11886-020-01329-7
6. Komatsu M, Sakai A, Komatsu R, Matsuoka R, Yasutomi S, Shozu K, et al. Detection of cardiac structural abnormalities in fetal ultrasound videos using deep learning. Appl Sci. (2021) 11:1–12. doi: 10.3390/app11010371
7. Diller GP, Lammers AE, Babu-Narayan S, Li W, Radke RM, Baumgartner H, et al. Denoising and artefact removal for transthoracic echocardiographic imaging in congenital heart disease: utility of diagnosis specific deep learning algorithms. Int J Cardiovasc Imaging. (2019) 35:2189–96. doi: 10.1007/s10554-019-01671-0
8. Arnaout R, Curran L, Zhao Y, Levine JC, Chinn E, Moon-Grady AJ. An ensemble of neural networks provides expert-level prenatal detection of complex congenital heart disease. Nat Med. (2021) 27:882–91. doi: 10.1038/s41591-021-01342-5
9. Gearhart A, Goto S, Powell AJ, Deo RC. Abstract 10614: an automated view identification model for pediatric echocardiography using artificial intelligence. Circulation. (2021) 144(Suppl\_1):A10614. doi: 10.1161/circ.144.suppl_1.10614
10. Guo L, Lei B, Chen W, Du J, Frangi AF, Qin J, et al. Dual attention enhancement feature fusion network for segmentation and quantitative analysis of paediatric echocardiography. Med Image Anal. (2021) 71:102042. doi: 10.1016/j.media.2021.102042
11. He B, Ouyang D, Lopez L, Zou J, Reddy CD. Abstract 10345: video-based deep learning model for automated assessment of ejection fraction in pediatric patients. Circulation. (2021) 144(Suppl\_1):A10345. doi: 10.1161/circ.144.suppl_1.10345
12. Chotzoglou E, Day T, Tan J, Matthew J, Lloyd D, Razavi R, et al. Learning normal appearance for fetal anomaly screening: application to the unsupervised detection of Hypoplastic Left Heart Syndrome. arXiv. (2020) 1–25. http://arxiv.org/abs/2012.03679 (accessed February 20, 2022).
13. Wang J, Liu X, Wang F, Zheng L, Gao F, Zhang H, et al. Automated interpretation of congenital heart disease from multi-view echocardiograms. Med Image Anal. (2021) 69:101942. doi: 10.1016/j.media.2020.101942
14. Meza JM, Slieker M, Blackstone EH, Mertens L, DeCampli WM, Kirklin JK, et al. A novel, data-driven conceptualization for critical left heart obstruction. Comput Methods Programs Biomed. (2018) 165:107–16. doi: 10.1016/j.cmpb.2018.08.014
15. Garcia-canadilla P, Sanchez-martinez S, Mart PM, Slorach C, Hui W, Piella G, et al. Machine-learning based exploration to identify remodelling patterns associated with death or heart-transplant in paediatric dilated cardiomyopathy. J Hear Lung Transplant. (2021) 41:516–26. doi: 10.1016/j.healun.2021.11.020
16. Lopez L, Colan SD, Frommelt PC, Ensing GJ, Kendall K, Younoszai AK, et al. Recommendations for quantification methods during the performance of a pediatric echocardiogram: a report from the pediatric measurements writing group of the american society of echocardiography pediatric and congenital heart disease council. J Am Soc Echocardiogr. (2010) 23:465–95. doi: 10.1016/j.echo.2010.03.019
17. van Nisselrooij AEL, Teunissen AKK, Clur SA, Rozendaal L, Pajkrt E, Linskens IH, et al. Why are congenital heart defects being missed? Ultrasound Obstet Gynecol. (2020) 55:747–57. doi: 10.1002/uog.20358
18. Liu X, Faes L, Kale AU, Wagner SK, Fu DJ, Bruynseels A, et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: a systematic review and meta-analysis. Lancet Digit Heal. (2019) 1:e271–97. doi: 10.1016/S2589-7500(19)30123-2
19. Matthew J, Skelton E, Day TG, Zimmer VA, Gomez A, Wheeler G, et al. Exploring a new paradigm for the fetal anomaly ultrasound scan: artificial intelligence in real time. Prenat Diagn. (2022) 42:49–59. doi: 10.2139/ssrn.3795326
20. Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, et al. Review of deep learning: concepts, CNN architectures, challenges, applications, future directions. J Big Data. (2021) 8:53. doi: 10.1186/s40537-021-00444-8
21. Østvik A, Smistad E, Aase SA, Haugen BO, Lovstakken L. Real-time standard view classification in transthoracic echocardiography using convolutional neural networks. Ultrasound Med Biol. (2019) 45:374–84. doi: 10.1016/j.ultrasmedbio.2018.07.024
22. Willemink MJ, Koszek WA, Hardell C, Wu J, Fleischmann D, Harvey H, et al. Preparing medical imaging data for machine learning martin. Radiology. (2020) 295:4–15. doi: 10.1148/radiol.2020192224
23. Annangi P, Ravishankar H, Patil R, Tore B, Aase SA, Steen E. AI assisted feedback system for transmit parameter optimization in cardiac ultrasound. IEEE Int Ultrason Symp IUS. (2020) 2020:1–4. doi: 10.1109/IUS46767.2020.9251501
24. Madani A, Arnaout R, Mofrad M, Arnaout R. Fast and accurate view classification of echocardiograms using deep learning. NPJ Digit Med. (2018) 1:1–8. doi: 10.1038/s41746-017-0013-1
25. Zhang J, Gajjala S, Agrawal P, Tison GH, Hallock LA, Beussink-Nelson L, et al. Fully automated echocardiogram interpretation in clinical practice: feasibility and diagnostic accuracy. Circulation. (2018) 138:1623–35. doi: 10.1161/CIRCULATIONAHA.118.034338
26. Ommen SR, Mital S, Burke MA, Day SM, Deswal A, Elliott P, et al. 2020 AHA/ACC guideline for the diagnosis and treatment of patients with hypertrophic cardiomyopathy. Circulation. (2020) 142:e558–631. doi: 10.1161/CIR.0000000000000938
27. Arafati A, Morisawa D, Avendi MR, Amini MR, Assadi RA, Jafarkhani H, et al. Generalizable fully automated multi-label segmentation of four-chamber view echocardiograms based on deep convolutional adversarial networks. J R Soc Interface. (2020) 17:20200267. doi: 10.1098/rsif.2020.0267
28. Yi X, Walia E, Babyn P. Generative adversarial network in medical imaging: a review. Med Image Anal. (2019) 58:101552. doi: 10.1016/j.media.2019.101552
29. Ouyang D, He B, Ghorbani A, Yuan N, Ebinger J, Langlotz CP, et al. Video-based AI for beat-to-beat assessment of cardiac function. Nature. (2020) 580:252–6. doi: 10.1038/s41586-020-2145-8
30. Ishizu T. Deep learning brings new era in echocardiography. Circ J. (2021) 86:96–7. doi: 10.1253/circj.CJ-21-0663
31. Raghu M, Zhang C, Kleinberg J, Bengio S. Transfusion: understanding transfer learning for medical imaging. Adv Neural Inf Process Syst. (2019) 32(NeurIPS):1–22. doi: 10.48550/arXiv.1902.07208
32. Colquitt JL, Pignatelli RH. Strain imaging: the emergence of speckle tracking echocardiography into clinical pediatric cardiology. Congenit Heart Dis. (2016) 11:199–207. doi: 10.1111/chd.12334
33. Ziebell D, Bettermann E, Lipinski J, Border WL, Sachdeva R. Current practice and barriers to implementation of strain imaging in pediatric echocardiography labs: a national survey. J Am Soc Echocardiogr. (2021) 34:316–8. doi: 10.1016/j.echo.2020.11.011
34. Salte IM, Østvik A, Smistad E, Melichova D, Nguyen TM, Karlsen S, et al. Artificial intelligence for automatic measurement of left ventricular strain in echocardiography. JACC Cardiovasc Imaging. (2021) 14:1918–28. doi: 10.1016/j.jcmg.2021.04.018
35. Dey D, Slomka PJ, Leeson P, Comaniciu D, Shrestha S, Sengupta PP, et al. Artificial intelligence in cardiovascular imaging: JACC State-of-the-Art Review. J Am Coll Cardiol. (2019) 73:1317–35. doi: 10.1016/j.jacc.2018.12.054
36. Gearhart A, Gaffar S, Chang AC. A primer on artificial intelligence for the paediatric cardiologist. Cardiol Young. (2020) 30:934–45. doi: 10.1017/S1047951120001493
37. Scapicchio C, Gabelloni M, Barucci A, Cioni D, Saba L, Neri E, et al. Deep look into radiomics. Radiol Medica. (2021) 126:1296–311. doi: 10.1007/s11547-021-01389-x
38. Zwanenburg A, Vallières M, Abdalah MA, Aerts HJWL, Andrearczyk V, Apte A, et al. The image biomarker standardization initiative: standardized quantitative radiomics for high-throughput image-based phenotyping. Radiology. (2020) 295:328–38. doi: 10.1148/radiol.2020191145
39. AIUM practice parameter for the performance of fetal echocardiography. J Ultrasound Med. (2020) 39:E5–16. doi: 10.1002/jum.15263
40. Nova R, Nurmaini S, Partan RU, Putra ST. Automated image segmentation for cardiac septal defects based on contour region with convolutional neural networks: a preliminary study. Informatics Med Unlocked. (2021) 24:100601. doi: 10.1016/j.imu.2021.100601
41. Diller GP, Babu-Narayan S, Li W, Radojevic J, Kempny A, Uebing A, et al. Utility of machine learning algorithms in assessing patients with a systemic right ventricle. Eur Heart J Cardiovasc Imaging. (2019) 20:925–31. doi: 10.1093/ehjci/jey211
42. Molnar C. Interpretable machine learning. A Guide for Making Black Box Models Explainable” Lulu. (2019).
43. Harrell FE. Describing, resampling, validating, and simplyfing the model. In: Regression Modeling Strategies: With Applications to Linear Models, Logistic and Ordinal Regression, and Survival Analysis, 2nd ed. Switzerland: Springer (2015), p. 103–21. doi: 10.1007/978-3-319-19425-7_5
44. Nguyen MB, Dragulescu A, Chaturvedi R, Fan CS, Villemain O, Friedberg MK, et al. Understanding complex interactions in pediatric diastolic function assessment. J Am Soc Echocardiogr. (2022) 35:868–77.e5. doi: 10.1016/j.echo.2022.03.017
45. Kutty S. The 21st annual feigenbaum lecture: beyond artificial: echocardiography from elegant images to analytic intelligence. J Am Soc Echocardiogr. (2020) 33:1163–71. doi: 10.1016/j.echo.2020.07.016
46. Pearl J. Linear models: a useful “microscope” for causal analysis. J Causal Inference. (2013) 1:155–70. doi: 10.1515/jci-2013-0003
47. Holzinger A. Explainable AI and multi-modal causability in medicine. I-Com. (2021) 19:171–9. doi: 10.1515/icom-2020-0024
48. Linardatos P, Papastefanopoulos V, Kotsiantis S. Explainable ai: a review of machine learning interpretability methods. Entropy. (2021) 23:1–45. doi: 10.3390/e23010018
49. Newman PG, Rozycki GS. The history of ultrasound. Surg Clin North Am. (1998) 78:179–95. doi: 10.1016/S0039-6109(05)70308-X
50. Ma XJ, Huang GY. Prediction of biventricular repair by echocardiography in borderline ventricle. Chin Med J. (2019) 132:2105–8. doi: 10.1097/CM9.0000000000000375
51. Sengupta Partho P., Shrestha S. Machine learning for data-driven discovery: the rise and relevance. JACC Cardiovasc Imaging. (2019) 12:690–2. doi: 10.1016/j.jcmg.2018.06.030
52. Ellis CA, Sendi MSE, Geenjaar EPT, Plis SM, Miller RL, Calhoun VD. Algorithm-agnostic explainability for unsupervised clustering. arxiv. (2021). http://arxiv.org/abs/2105.08053 (accessed February 20, 2022).
53. Hendryx EP, Rivière BM, Sorensen DC, Rusin CG. Finding representative electrocardiogram beat morphologies with CUR. J Biomed Inform. (2018) 77:97–110. doi: 10.1016/j.jbi.2017.12.003
54. Kempny A, Fraisse A, Dimopoulos K. Risk stratification in congenital heart disease - a call for protocolised assessment and multicentre collaboration. Int J Cardiol. (2019) 276:114–5. doi: 10.1016/j.ijcard.2018.11.101
55. Bernardino G, Benkarim O., Sanz-de la Garza M, Prat-Gonzàlez S, Sepulveda-Martinez A, Crispi F, et al. Handling confounding variables in statistical shape analysis - application to cardiac remodeling. Med Image Anal. (2020) 65:101792. doi: 10.1016/j.media.2020.101792
56. Wessler BS, Yh LL, Kramer W, Cangelosi M, Raman G, Lutz JS, et al. Clinical prediction models for cardiovascular disease: tufts predictive analytics and comparative effectiveness clinical prediction model database. Circ Cardiovasc Qual Outcomes. (2015) 8:368–75. doi: 10.1161/CIRCOUTCOMES.115.001693
57. Diller GP, Arvanitaki A, Opotowsky AR, Jenkins K, Moons P, Kempny A, et al. Lifespan perspective on congenital heart disease research: jacc state-of-the-art review. J Am Coll Cardiol. (2021) 77:2219–35. doi: 10.1016/j.jacc.2021.03.012
58. Dzobo K, Adotey S, Thomford NE, Dzobo W. Integrating artificial and human intelligence: a partnership for responsible innovation in biomedical engineering and medicine. Omi A J Integr Biol. (2020) 24:247–63. doi: 10.1089/omi.2019.0038
59. Tokodi M, Shrestha S, Bianco C, Kagiyama N, Casaclang-Verzosa G, Narula J, et al. Interpatient similarities in cardiac function: a platform for personalized cardiovascular medicine. JACC Cardiovasc Imaging. (2020) 13:1119–32. doi: 10.1016/j.jcmg.2019.12.018
60. Casey A, Davidson E, Poon M, Dong H, Duma D, Grivas A, et al. A systematic review of natural language processing applied to radiology reports. BMC Med Inform Decis Mak. (2021) 21:1–18. doi: 10.1186/s12911-021-01533-7
61. Mabotuwana T, Hombal V, Dalal S, Hall CS, Gunn M. Determining adherence to follow-up imaging recommendations. J Am Coll Radiol. (2018) 15:422–8. doi: 10.1016/j.jacr.2017.11.022
62. Morris S, Lopez K. Deep learning for detecting congenital heart disease in the fetus. Nat Med. (2021) 27:759–61. doi: 10.1038/s41591-021-01354-1
Keywords: pediatric cardiology, pediatric echocardiography, echocardiography, pediatrics, artificial intelligence
Citation: Nguyen MB, Villemain O, Friedberg MK, Lovstakken L, Rusin CG and Mertens L (2022) Artificial intelligence in the pediatric echocardiography laboratory: Automation, physiology, and outcomes. Front. Radiol. 2:881777. doi: 10.3389/fradi.2022.881777
Received: 23 February 2022; Accepted: 01 August 2022;
Published: 09 September 2022.
Edited by:
Fàtima Crispi, Hospital Clinic Barcelona, SpainReviewed by:
Lorenzo Faggioni, University of Pisa, ItalyAshraf Harahsheh, Children's National Hospital, United States
Copyright © 2022 Nguyen, Villemain, Friedberg, Lovstakken, Rusin and Mertens. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Luc Mertens, bHVjLm1lcnRlbnMmI3gwMDA0MDtzaWNra2lkcy5jYQ==