Skip to main content

ORIGINAL RESEARCH article

Front. Neurol., 19 September 2018
Sec. Neuromuscular Disorders and Peripheral Neuropathies

Segmentation of Peripheral Nerves From Magnetic Resonance Neurography: A Fully-Automatic, Deep Learning-Based Approach

\r\nFabian BalsigerFabian Balsiger1Carolin SteindelCarolin Steindel2Mirjam ArnMirjam Arn2Benedikt WagnerBenedikt Wagner2Lorenz GrunderLorenz Grunder2Marwan El-KoussyMarwan El-Koussy2Waldo Valenzuela,Waldo Valenzuela1,2Mauricio ReyesMauricio Reyes1Olivier Scheidegger,*Olivier Scheidegger2,3*
  • 1Institute for Surgical Technology and Biomechanics, University of Bern, Bern, Switzerland
  • 2Support Center for Advanced Neuroimaging, Institute for Diagnostic and Interventional Neuroradiology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland
  • 3Department of Neurology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

Diagnosis of peripheral neuropathies relies on neurological examinations, electrodiagnostic studies, and since recently magnetic resonance neurography (MRN). The aim of this study was to develop and evaluate a fully-automatic segmentation method of peripheral nerves of the thigh. T2-weighted sequences without fat suppression acquired on a 3 T MR scanner were retrospectively analyzed in 10 healthy volunteers and 42 patients suffering from clinically and electrophysiologically diagnosed sciatic neuropathy. A fully-convolutional neural network was developed to segment the MRN images into peripheral nerve and background tissues. The performance of the method was compared to manual inter-rater segmentation variability. The proposed method yielded Dice coefficients of 0.859 ± 0.061 and 0.719 ± 0.128, Hausdorff distances of 13.9 ± 26.6 and 12.4 ± 12.1 mm, and volumetric similarities of 0.930 ± 0.054 and 0.897 ± 0.109, for the healthy volunteer and patient cohorts, respectively. The complete segmentation process requires less than one second, which is a significant decrease to manual segmentation with an average duration of 19 ± 8 min. Considering cross-sectional area or signal intensity of the segmented nerves, focal and extended lesions might be detected. Such analyses could be used as biomarker for lesion burden, or serve as volume of interest for further quantitative MRN techniques. We demonstrated that fully-automatic segmentation of healthy and neuropathic sciatic nerves can be performed from standard MRN images with good accuracy and in a clinically feasible time.

1. Introduction

Current state-of-the-art to diagnose and monitor the effects of potentially available treatments for peripheral neuropathies relies on clinical examination and electrodiagnostic studies (EDX). Certain regions of the body are less amenable to EDX, or may show ambiguous symptoms and signs regarding localization of lesions when affected: e.g., brachial and lumbosacral plexus, nerves situated deeply in the extremities such as the sciatic nerve, and nerves close to the trunk. Magnetic resonance neurography (MRN) (1, 2) has emerged as a complementary diagnostic instrument for peripheral neuropathy, especially where neurological examinations are difficult or inconclusive. While MRN has mostly been used as a qualitative diagnostic method, image-derived morphometric parameters (e.g., cross-sectional areas, CSA) and quantitative MR measures based on relaxometry, magnetization transfer, or diffusion-weighted imaging of peripheral nerves have increasingly been reported recently (311), and could potentially serve as outcome measures (12).

Quantitative assessment of peripheral nerves from MRN typically proceeds either by assessment of CSA or by the identification of regions of interest in which abnormal signal behavior or quantitative MR parameters are further analyzed. In both cases, a segmentation of the peripheral nerve at interest is typically performed manually. It has been shown that for clinically relevant image-based quantification, fully- and semi-automatic computer-assisted segmentation is favorable to manual segmentation regarding reproducibility, time-efficiency, and cost-efficiency (13, 14). Computer-assisted segmentation of peripheral nerves from MRN images has been addressed by Felisaz et al. (15). They proposed a semi-automatic method to compartmentalize the tibial nerve in micro-neurography images and showed the potential of computer-assisted segmentation by associating peripheral neuropathy with decreased fascicular-to-nerve volume ratio, increased nerve volume, and increased CSA (16). Unfortunately, potential MR-based outcome measures of peripheral nerves will be of little clinical use if the assessment remains a manual, user-dependent, and tedious task for physicians or trained personnel, or is limited to a small field of view (FOV) as in Felisaz et al. (16). Fully- or semi-automatic computer-assisted segmentation of peripheral nerves with a large extremities coverage may be an important future step to obtain quantitative imaging outcome measures at a larger scale for disease-specific diagnosis and treatment monitoring of peripheral neuropathies.

We present, to the best of our knowledge, the first attempt of a fully-automatic, deep learning-based segmentation of peripheral nerves obtained from a larger coverage MRN image set of the thigh acquired in a clinical setting, containing healthy volunteers and patients suffering from peripheral nervous system (PNS) disorders.

2. Materials and Methods

2.1. Healthy Volunteer and Patient Data

All healthy volunteers and patients receiving a MRN examination at our institution between 2013 and 2017 have been enrolled in a registry. We developed and evaluated the proposed method on retrospectively chosen images of the human thigh from healthy volunteers, and all patients with clinically and electrophysiologically diagnosed sciatic neuropathy enrolled in our registry. The healthy cohort consisted of 10 volunteers (4 female, 6 male; age = 25.0 ± 2.6). The patient cohort consisted of 42 patients (21 female, 21 male; age = 55.7 ± 15.7), with sciatic neuropathy as confirmed by senior physicians (over 10 years of experience) in our neuromuscular disease unit based on clinical examination and EDX. The study was approved by the local ethical committee, and informed consent was obtained from all participants.

2.2. MR Acquisition

For this retrospective study, a turbo spin-echo T2-weighted sequence without fat suppression (T2) was chosen, which is part of the routine MRN examination protocol at our institution. The sequence was acquired using either a circular 15-channel knee coil or an anterior body surface and posterior built-in spine coil in a clinical scanner running at 3 Tesla (Siemens MAGNETOM Verio, Siemens Healthcare GmbH, Erlangen, Germany) with following sequence parameters: repetition time (TR) of 4690 ms, echo time (TE) of 82 ms, FOV of 384 × 330 mm2, flip angle (FA) of 134 °, turbo factor of 12, voxel size of 0.52 × 0.52 × 4.0 mm3, and 60 axial-oriented slices with an inter-slice gap of 0.4 mm. The acquisition time was 4 min 43 s. Anatomical coverage was one image stack for each volunteer and patient, with image stacks positioned at the distal thigh in all volunteers, and variably between the distal thigh up to the head of the femur in patients, respectively.

2.3. Manual Ground Truth Segmentation

Three physicians (experienced in clinical, electrophysiological, and imaging-based assessment of neuromuscular diseases: author OS, senior neurologist and neuroradiologist > 14 years, author BW, neuroradiologist > 4 years, and author LG, neuroradiologist > 2 years) manually segmented the sciatic nerve including its branches the tibial and peroneal nerve. Each physician individually segmented the available data to study the inter-rater agreement for manual sciatic nerve segmentation. A consensus segmentation (referred hereafter as consensus ground truth) for the evaluation of our computer-assisted segmentation approach was obtained using majority voting of the three rater segmentations, i.e., a voxel belongs to peripheral nerve if at least two raters segmented it. All segmentations were performed with the ITK-SNAP software1 (17) using the polygon or paintbrush tool.

2.4. Computer-Assisted Segmentation

We developed a fully-automatic, deep learning-based approach to segment peripheral nerves from MRN images. The input of our method was the T2 image, and the output of the method was a binary segmentation of the T2 image into peripheral nerve and background. Our method bases on fully-convolutional neural networks (fCNNs), often referred to as deep learning algorithms or simply as neural network, which have shown excellent performances in various medical image segmentation tasks (18). The main working principle behind fCNNs is supervised learning, a machine learning paradigm in which a computer learns to distinguish peripheral nerve from background tissue using training data (Figure 1A). In our case, the training data consisted of T2 images as well as the corresponding consensus ground truth images. Our fCNN processed the images slice-wise, i.e., the input was a T2 image slice, and the output was a segmented binary mask of the T2 image slice. This segmented image slice was then compared to the image slice of the consensus ground truth to improve the segmentation during the training of the fCNN. Once a fCNN is trained, new and previously unseen images can be segmented without the need of any ground truth image (Figure 1B). During this so-called testing of the fCNN, the neural network segments a T2 image slice into peripheral nerve and background without the need for a manually segmented ground truth image slice. Segmenting all T2 image slices resulted in an entirely segmented peripheral nerve in the T2 image. The following sections describe the pre-processing of our data, the neural network, and our training strategy.

FIGURE 1
www.frontiersin.org

Figure 1. Overview of the proposed peripheral nerve segmentation method. (A) Training of the neural network with T2 and ground truth image slices, and (B) testing of the trained neural network yields a segmentation of the peripheral nerve without the need of a ground truth.

2.4.1. Pre-processing

We pre-processed the T2 images by intensity normalization, and cropping. First, we normalized the intensities of each T2 image to zero mean and unit variance. Second, we cropped each T2 image to 320 × 320 pixels in the axial plane to be able to reduce the image size evenly, which is a requirement of our neural network architecture.

2.4.2. Neural Network

Our fCNN architecture adopts the fully-convolutional DenseNet (19). It consists of four transition down (TD) in the downsampling path and four transition up (TU) in the upsampling path. We defined a layer in our neural network to be the following sequence of operations: batch normalization (20), rectified linear unit (ReLU) activation function, 3 × 3 convolution with stride one and same padding, and dropout (21) with probability p = 0.2. Our dense blocks consist of four consecutive layers each with 12 filters (growth rate 12). A TD applied batch normalization, ReLU, 1 × 1 convolution with stride one and same padding, dropout with p = 0.2, and 2 × 2 max pooling with stride two. Therefore, the input resolution was 20 × 20 pixels in the bottleneck dense block. A TU applied a transposed 3 × 3 convolution with stride two and same padding. Before the first dense block in the downsampling path, a 3 × 3 convolution with stride one, same padding, and 48 filters was applied to the input. Similarly, we applied a 1 × 1 convolution with stride one, same padding, and two filters, the desired number of classes, after the last dense block in the upsampling path. Finally, a softmax non-linearity was applied to calculate the pixel-wise probability distribution for background and peripheral nerve. The fCNN was implemented using PyTorch 0.4.0 (Facebook, Inc., Menlo Park, CA, U.S.) with Python 3.6 (Python Software Foundation, Wilmington, DA, U.S.).

2.4.3. Training

We trained the neural network on the 10 healthy volunteers and 42 patient images using a randomly generated four-fold cross-validation. That is, the network was trained on 39 images, and its performance was tested on 13 images. This procedure was repeated for all four folds, such that the network's performance on each image could be assessed. During the fold randomization, we balanced healthy volunteer and patient images (i.e., we ensured that each fold contained at least two volunteer images among the 13 test images). For training, we used a cross entropy loss and the Adam optimizer (22) with a learning rate of 1 × 10-3, β1 = 0.9, and β2 = 0.999. We decreased the learning rate to 1 × 10-4 after 30 epochs, and to 1 × 10-5 after 80 epochs. The training was run with a batch size of eight image slices for 100 epochs, which we empirically found to be sufficient. Additionally, we used data augmentation during the training to prevent memorization of training data and to introduce artificial variety: random flipping, random translation, and random elastic deformation. Note that we tuned the neural network on one randomly chosen cross-validation split and did not use the other splits to develop and tune the method.

2.5. Evaluation and Statistical Analysis

We compared the performance of our method to the consensus ground truth (Auto-GT), and additionally to the inter-rater variability (R-R). The inter-rater variability quantifies the difficulty of peripheral nerve segmentation and served as baseline for our method. It was obtained by comparing the manual segmentations of each possible rater-rater pair (i.e., OS-BW, OS-LG, and BW-LG) and aggregating these comparisons per cohort [cf. Section III-E in (23)]. Therefore, the inter-rater variability consisted of n = 30 (healthy volunteer cohort) and n = 126 (patient cohort) results for every evaluation metric.

We used the following three evaluation metrics to evaluate the performances: (1) The Dice coefficient (24), which measures the spatial overlap between the segmentation and the ground truth. (2) The Hausdorff distance (HD) (25), which measures the 95th percentile distance between the segmentation and the ground truth boundaries. (3) The volumetric similarity (VS) (26), which measures the absolute difference between the segmentation and ground truth volume divided by the sum of the two volumes. We used the open-source evaluation metrics implementation presented in Taha and Hanbury (27) (version 2017.04.25) and refer the reader to the reference for mathematical details. Furthermore, we measured the execution time of our method and the raters' segmentation time .

We hypothesized that there is no statistically significant difference between our method (Auto-GT) and the inter-rater baseline (R-R). The hypothesis was tested for both cohorts and the three evaluation metrics independently. We used an unpaired Mann-Whitney U-test to confirm the hypothesis due to non-normal distributed data and different sample size between Auto-GT and the aggregated R-R results. All tests were performed with a significance level of 0.05 (95 % confidence interval) using R (R Core Team, Vienna, Austria) version 3.5.0.

3. Results

The proposed approach yielded Dice coefficients of 0.859 ± 0.061 and 0.719 ± 0.128 for the healthy and patient cohorts, respectively. Regarding the HD metric, the approach yielded 13.9 ± 26.6 and 12.4 ± 12.1 mm, respectively. Finally, the VS metric resulted in 0.930 ± 0.054 and 0.897 ± 0.109, respectively. The inter-rater performance was 0.869 ± 0.031 and 0.786 ± 0.093 for the Dice coefficient, 0.70 ± 0.67 and 11.2 ± 19.0 mm for the HD, and 0.937 ± 0.043 and 0.897 ± 0.087 for the VS. Overall, for each metric, the segmentation performance was better for the healthy volunteer cohort than the patient cohort (Figures 2A–C). No statistical significant differences (p > 0.05) between Auto-GT and R-R were found for the volunteer Dice coefficients (p=0.6), the volunteer VS (p=0.8), and the patient VS (p = 0.3). Statistical significant differences (p < 0.05) were found for patient Dice coefficients (p = 0.002), volunteer HD (p = 0.02), and patient HD (p ≤ 0.001). The detailed results of the segmentation analysis between individual raters, the consensus ground truth, and our proposed approach for both cohorts are summarized in Table 1.

FIGURE 2
www.frontiersin.org

Figure 2. Segmentation evaluation metrics of our method (Auto-GT) compared to inter-rater variability (R-R). Boxplot of the (A) Dice coefficient, (B) Hausdorff distance, (C) volume similarity, and (D) segmentation time separated by healthy volunteer and patient cohort. *The segmentation time of our method is less than 1 s and therefore barely visible in the boxplot.

TABLE 1
www.frontiersin.org

Table 1. Detailed segmentation evaluation metrics for both healthy volunteer and patient cohorts.

A post-hoc analysis revealed that the distribution of image stack location along the thigh was more variable in patients than in healthy volunteers. For healthy volunteers, slice stacks were predominantly taken from the mid-thigh, while in patients the stacks were variously taken from the distal, medial, and proximal portions of the thigh. To analyze the location dependence of the segmentation, we subdivided the patient cohort into locations distal, medial, and proximal thigh, with medial being approximately at the same location as the healthy volunteer cohort. Dice coefficients were 0.687 ± 0.144 for distal thigh (n = 14), 0.720 ± 0.109 for medial thigh (n = 7), and 0.741 ± 0.123 for proximal thigh (n = 21) in the patient cohort.

A color-coded three-dimensional rendering of a segmented sciatic nerve for a representative patient result, showing the similarity between the segmentation and the ground truth, is depicted in Figure 3. A higher surface-to-surface distance is present at over-segmented locations as can be seen on the T2 image sections with the segmentation (green) and ground truth boundaries (blue).

FIGURE 3
www.frontiersin.org

Figure 3. Segmentation of the sciatic nerve of a patient. (Left) 3-dimensional rendering of the segmentation. The color map encodes the surface-to-surface distance of the segmentation to the ground truth. (Right) The segmentation boundaries (green) and ground truth boundaries (blue) on the T2 image are shown for three slices along the nerve course.

Segmenting the image stack of one subject with our method on a standard desktop computer equipped with a graphics processing unit (GPU) required less than 1 s for the volunteer and the patient cohort (Ubuntu 16.04 LTS, 3.2 GHz Intel Core i7-3930K, 64 GB memory, NVIDIA TITAN Xp with 12 GB memory). The time required to segment the ground truth of one subject manually was 21.1 ± 7.68 and 18.2 ± 7.42 min for the volunteer and patient cohort, respectively (Figure 2D). Note that the one-time training (without any user interaction) of the fCNN before segmenting new subjects required approximately 5 h.

4. Discussion

A fully-automatic, deep learning-based segmentation of peripheral nerves for T2-weighted MRN images was evaluated on thigh scans acquired in a clinical setting. The proposed method was successful in segmenting peripheral nerves with and without lesions with good accuracy both in healthy volunteers and patients suffering from sciatic neuropathy. Our fully-automatic method results in a significant time gain for sciatic nerve segmentation, compared to manual segmentation.

The peripheral nerves in our images can be considered as small structures with a volume fraction (i.e., nerve volume divided by background volume) of 0.143 ± 0.049 %. Unfortunately, evaluating the segmentation performance of methods applied to small structures is difficult because commonly used metrics such as the Dice coefficient and the VS are sensitive to small structures (27). While distance metrics are more suitable for evaluating small structures, they might not always be of interest to clinicians as they can be sensitive to outliers (27). Our inter-rater study revealed that our method reaches human level performance for volunteer Dice coefficients, and VS in both healthy volunteers and patients. Inter-rater disagreement for small cutaneous branches of the tibial and peroneal nerve was present in patients mostly because of more distally located image stacks in this cohort, whereas in our healthy volunteers the cutaneous branches were mostly not branching from the tibial or peroneal nerves yet. Missing such a branch on one or several image slices results in a higher mean and standard deviation of the HD metric compared to the volunteer cohort (see Figure 4A for an example). Furthermore, large HD metrics arise in the case of false positive segmentations of our method. A volunteer example with a Dice coefficient of 0.772 but a HD of 64.1 mm is shown in Figure 4B, where a falsely segmented vein contributes to the large HD. Overall, interpreting the segmentation performance for peripheral nerves is challenging, especially in the case of the HD metric, and may be misleading when comparing metrics to other segmentation tasks such as brain tumors.

FIGURE 4
www.frontiersin.org

Figure 4. Interpretability of the Hausdorff distance (HD) metric for peripheral nerves. (A) The same nerves segmented by the three raters (left to right) are depicted in red. One rater does not segment all branches (arrows), which results in a large HD. (B) The consensus ground truth (left) compared to the segmentation results by our method (right). A falsely segmented vein (arrow) by our method results in a large HD.

Our approach has several limitations and difficulties, which resulted in a slightly decreased segmentation accuracy for the patient cohort. The difference in performance between the two cohorts is mainly assumed to be caused by the variability in PNS appearance, differences in the image quality and location of the image stacks. While the volunteers were imaged with a standardized positioning protocol for MRN sequence quality assurance, the patient images were obtained retrospectively from a clinical setting for which no standardized protocol was used. Consequently, different peripheral nerve lesion types and muscle pathologies were present in the images, and the position of the MR acquisition differed considerably, which also might cause the image stack to no longer be perpendicular to the main nerve direction (e.g., in proximal parts of the thigh near the hip or at the knee region). Further, the contrast between peripheral nerve and its surrounding tissues was sometimes very limited on distal image slices. These aspects ultimately increase the complexity of the segmentation task, which is reflected in the lower Dice coefficients found for distally located regions on patient images, as compared to the regions located proximally or at mid-thigh in volunteer and patient images. Besides increasing the amount of training data, a potential workaround could be the use of isotropic 3-D T2-weighted sequences for MRN despite their lower resolution [e.g., SPACE, Sampling Perfection with Application optimized Contrasts using different flip angle Evolution (28)]. In addition, the quality of the patient images was lower than the quality of the volunteer images. Movement artifacts of varying strength were noticeable mostly in the patient cohort due to uncomfortable scanning positioning, while the volunteers could be imaged without noticeable movement artifacts. More frequently present in the patient cohort were also signal distortion artifacts, stemming from an off-center positioning of the extremities during the MR acquisition. Due to the retrospective type of study performed, healthy volunteers and patient cohorts differed regarding mean age, and detailed clinical data on the patient cohort was not available. Hence, we can neither rule out an effect of age, nor of severity of the neuropathy on segmentation performance.

We see our contribution as a proof of concept for segmentation of peripheral nerves for a wider clinical usage of quantitative MRN as an adjunct to clinical assessment and EDX, especially in more proximal body regions not well accessible using aforementioned standard work-up. Despite the discussed limitations, the partially achieved human-level performance in this retrospective analysis of MRN images demonstrates the promising usefulness and strength of fully-automatic segmentation. This is additionally emphasized by the fact that achieving a high agreement of a manual segmentation to a consensus ground truth requires experienced personnel with a long training history (see better results of rater OS in Table 1), while partially achieved human-level performance of the fCNN required just 5 h of unassisted training. We think that such peripheral nerve segmentation is essential toward obtaining quantitative imaging outcome measures, which could be an integral part of MRN-based examinations of PNS and PNS disorders, both for diagnostic and monitoring purposes. Our method could assist clinicians by providing a quantification of nerve volume, CSA, and lesion burden through a combination of morphometry and signal intensity analysis as shown in Figure 5. Such analyses might allow distinguishing between healthy nerve and diseased nerve, and peripheral neuropathy type as it has been shown by others (5, 7, 911). Additionally, our segmentation could serve as a volume of interest for other quantitative MR methods like diffusion-weighted imaging, magnetization transfer imaging, and relaxometry (38, 10), which might further be applied to assess total lesion burden, distribution pattern of lesions, and multi-model characterization of lesions. However, such potential outcome measures will only find acceptance in a clinical setting, if they can be obtained in an accurate and time-efficient manner without tedious and labor-intensive manual segmentation.

FIGURE 5
www.frontiersin.org

Figure 5. Potential of computer-assisted segmentation of peripheral nerves for imaging biomarkers. (Left column) 3-dimensional renderings of the sciatic nerve with lesions colored red: (Top row) a healthy volunteer, (Middle row) a patient with a focal lesion, (Bottom row) and a patient with an extended lesion. (Middle column) Cross-sectional area evolution and (Right column) lesion burden evolution obtained from the segmentation could be used as biomarkers to assess disease severity and progression, or to categorize the lesion type. Note that not all peripheral nerve lesion types show morphometric abnormalities, hence a combination with signal intensity (or other quantifiable MR parameters) is necessary to assess the lesion burden. The quantified signal intensity evolution was assessed by segmenting hyperintense nerve fascicle bundles on a co-registered T2-weighted sequence with fat suppression using inversion recovery.

Further refinements of the method could include algorithmic as well as MRN related changes. Algorithmic changes could include three-dimensional convolutions to enrich the contextual information during the segmentation. Other than that, a dedicated post-processing of the segmentation could reduce false positive segmentation by reconstructing the peripheral nerves under physiological constraints similar to Rempfler et al. (29). Regarding changes in MRN, additional MRN images could be used to enrich the information about the peripheral nerves, especially for regions showing problems with fully-automated segmentation. Also, combining the complementary information from MRN images with diffusion-weighted images (e.g., from tractography of peripheral nerves) may add additional imaging information for better discrimination of peripheral nerves from surrounding tissue [e.g., (4, 30)]. In general, the segmentation could be extended to peripheral nerves of other body regions, to peripheral nerve lesions, and to muscles (31) with the aim of a holistic computer-assisted quantification of neuromuscular diseases.

In conclusion, we proposed, to the best of our knowledge, the first fully-automatic, deep learning-based segmentation of peripheral nerves from the thigh in a clinical setting. Our method segments healthy and diseased peripheral nerves from MRN images with good accuracy and in clinically feasible time, and is a promising approach toward quantitative outcome measures for the diagnosis of peripheral neuropathies.

Ethics Statement

This study was carried out in accordance with the recommendations of the Swiss legislation for human research. All subjects gave written informed consent. The protocol was approved by the local ethics committee of the Canton of Bern, Switzerland.

Author Contributions

FB, WV, MR, and OS conceived the ideas and designed the study. FB developed the method and analyzed the results. CS, MA, BW, and OS contributed to data collection and registry enrollment. BW, LG, and OS manually segmented the ground truth. ME-K supervised the work. FB, WV, MR, and OS contributed to manuscript writing. All authors revised the manuscript. OS is the scientific guarantor of the study.

Funding

This research was supported by the Swiss Foundation for Research on Muscle Diseases (ssem), grant attributed to OS.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors thank the NVIDIA Corporation for their GPU donation and Dr. R. McKinley for proofreading the manuscript.

Footnotes

1. ^Available from www.itksnap.org

References

1. Howe FA, Filler AG, Bell BA, Griffiths JR. Magnetic resonance neurography. Magn Reson Med. (1992) 28:328–38.

PubMed Abstract | Google Scholar

2. Filler AG, Howe FA, Hayes CE, Kliot M, Winn HR, Bell BA, et al. Magnetic resonance neurography. Lancet (1993) 341:659–661.

PubMed Abstract | Google Scholar

3. Gambarota G, Mekle R, Mlynárik V, Krueger G. NMR properties of human median nerve at 3 T: proton density, T1, T2, and magnetization transfer. J Magn Reson Imaging (2009) 29:982–6. doi: 10.1002/jmri.21738

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Simon NG, Lagopoulos J, Gallagher T, Kliot M, Kiernan MC. Peripheral nerve diffusion tensor imaging is reliable and reproducible. J Magn Reson Imaging (2016) 43:962–9. doi: 10.1002/jmri.25056

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Kronlage M, Schwehr V, Schwarz D, Godel T, Heiland S, Bendszus M, et al. Magnetic resonance neurography. Clin Neuroradiol. (2017) doi: 10.1007/s00062-017-0633-5. [Epub ahead of print].

CrossRef Full Text | Google Scholar

6. Kronlage M, Schwehr V, Schwarz D, Godel T, Uhlmann L, Heiland S, et al. Peripheral nerve diffusion tensor imaging (DTI): normal values and demographic determinants in a cohort of 60 healthy individuals. Eur Radiol. (2018) 28:1801–8. doi: 10.1007/s00330-017-5134-z

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Kronlage M, Bäumer P, Pitarokoili K, Schwarz D, Schwehr V, Godel T, et al. Large coverage MR neurography in CIDP: diagnostic accuracy and electrophysiological correlation. J Neurol. (2017) 264:1434–43. doi: 10.1007/s00415-017-8543-7

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Lichtenstein T, Sprenger A, Weiss K, Slebocki K, Cervantes B, Karampinos D, et al. MRI biomarkers of proximal nerve injury in CIDP. Ann Clin Transl Neurol. (2018) 5:19–28. doi: 10.1002/acn3.502

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Pitarokoili K, Kronlage M, Bäumer P, Schwarz D, Gold R, Bendszus M, et al. High-resolution nerve ultrasound and magnetic resonance neurography as complementary neuroimaging tools for chronic inflammatory demyelinating polyneuropathy. Ther Adv Neurol Disord. (2018) 11:175628641875997. doi: 10.1177/1756286418759974

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Ratner S, Khwaja R, Zhang L, Xi Y, Dessouky R, Rubin C, et al. Sciatic neurosteatosis: relationship with age, gender, obesity and height. Eur Radiol. (2018) 28:1673–80. doi: 10.1007/s00330-017-5087-2

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Jende JME, Groener JB, Oikonomou D, Heiland S, Kopf S, Pham M, et al. Diabetic neuropathy differs between type 1 and type 2 diabetes: insights from magnetic resonance neurography. Ann. Neurol. (2018) 83:588–98. doi: 10.1002/ana.25182

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Noguerol TM, Barousse R, Socolovsky M, Luna A. Quantitative magnetic resonance (MR) neurography for evaluation of peripheral nerves and plexus injuries. Quantit Imaging Med Surg. (2017) 7:398–421. doi: 10.21037/qims.2017.08.01

CrossRef Full Text | Google Scholar

13. Porz N, Bauer S, Pica A, Schucht P, Beck J, Verma RK, et al. Multi-modal glioblastoma segmentation: man versus machine. PLoS ONE (2014) 9:e96873. doi: 10.1371/journal.pone.0096873

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology (2016) 278:563–77. doi: 10.1148/radiol.2015151169

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Felisaz PF, Balducci F, Gitto S, Carne I, Montagna S, De Icco R, et al. Nerve fascicles and epineurium volume segmentation of peripheral nerve using magnetic resonance micro-neurography. Acad Radiol. (2016) 23:1000–7. doi: 10.1016/j.acra.2016.03.013

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Felisaz PF, Maugeri G, Busi V, Vitale R, Balducci F, Gitto S, et al. MR Micro-neurography and a segmentation protocol applied to diabetic neuropathy. Radiol Res Pract. (2017) 2017:2761818. doi: 10.1155/2017/2761818

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Yushkevich PA, Piven J, Hazlett HC, Smith RG, Ho S, Gee JC, et al. User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability. Neuroimage (2006) 31:1116–28. doi: 10.1016/j.neuroimage.2006.01.015

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med Image Anal. (2017) 42:60–88. doi: 10.1016/j.media.2017.07.005

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Jégou S, Drozdzal M, Vazquez D, Romero A, Bengio Y. The one hundred layers tiramisu: fully convolutional DenseNets for semantic segmentation. In: 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). Los Alamitos, CA: IEEE Computer Society (2016). p. 1175–83.

20. Ioffe S, Szegedy C. Batch normalization: accelerating deep network training by reducing internal covariate shift. In: Bach F, Blei D, editors. Proceedings of the 32nd International Conference on International Conference on Machine Learning. ICML'15. Lille: PMLR (2015). p. 448–56.

Google Scholar

21. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res. (2014) 15:1929–58. doi: 10.1214/12-AOS1000

CrossRef Full Text | Google Scholar

22. Kingma DP, Ba JL. Adam: a method for stochastic optimization. In: International Conference on Learning Representations 2015. Ithaca, NY (2015). p. 1–15.

Google Scholar

23. Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, et al. The multimodal Brain Tumor Image Segmentation Benchmark (BRATS). IEEE Trans Med Imaging (2015) 34:1993–2024. doi: 10.1109/TMI.2014.2377694

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Dice LR. Measures of the amount of ecologic association between species. Ecology (1945) 26:297–302.

Google Scholar

25. Huttenlocher DP, Klanderman GA, Rucklidge WJ. Comparing images using the Hausdorff distance. IEEE Trans Pattern Anal Mach Intell. (1993) 15:850–63.

Google Scholar

26. Cárdenes R, de Luis-García R, Bach-Cuadra M. A multidimensional segmentation evaluation for medical image data. Comput Methods Programs Biomed. (2009) 96:108–24. doi: 10.1016/j.cmpb.2009.04.009

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Taha AA, Hanbury A. Metrics for evaluating 3D medical image segmentation: analysis, selection, and tool. BMC Med Imaging (2015) 15:29. doi: 10.1186/s12880-015-0068-x

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Madhuranthakam A, Lenkinski R. Technical advancements in MR neurography. Semin Musculoskelet Radiol. (2015) 19:86–93. doi: 10.1055/s-0035-1547370.

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Rempfler M, Schneider M, Ielacqua GD, Xiao X, Stock SR, Klohs J, et al. Reconstructing cerebrovascular networks under local physiological constraints by integer programming. Med Image Anal. (2016) 25:86–94. doi: 10.1016/j.media.2015.03.008

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Zhou Y, Narayana PA, Kumaravel M, Athar P, Patel VS, Sheikh KA. High resolution diffusion tensor imaging of human nerves in forearm. J Magn Reson Imaging (2014) 39:1374–83. doi: 10.1002/jmri.24300

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Gadermayr M, Disch C, Müller M, Merhof D, Gess B. A comprehensive study on automated muscle segmentation for assessing fat infiltration in neuromuscular diseases. Magn Reson Imaging (2017) 48:20–6. doi: 10.1016/j.mri.2017.12.014

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: health, sciatic nerve, peripheral nervous system diseases, magnetic resonance imaging, magnetic resonance neurography, machine learning, segmentation

Citation: Balsiger F, Steindel C, Arn M, Wagner B, Grunder L, El-Koussy M, Valenzuela W, Reyes M and Scheidegger O (2018) Segmentation of Peripheral Nerves From Magnetic Resonance Neurography: A Fully-Automatic, Deep Learning-Based Approach. Front. Neurol. 9:777. doi: 10.3389/fneur.2018.00777

Received: 15 July 2018; Accepted: 27 August 2018;
Published: 19 September 2018.

Edited by:

Massimiliano Filosto, Asst degli Spedali Civili di Brescia, Italy

Reviewed by:

Anna Pichiecchio, Fondazione Istituto Neurologico Nazionale Casimiro Mondino (IRCCS), Italy
Paola Sandroni, Mayo Clinic, United States

Copyright © 2018 Balsiger, Steindel, Arn, Wagner, Grunder, El-Koussy, Valenzuela, Reyes and Scheidegger. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Olivier Scheidegger, b2xpdmllci5zY2hlaWRlZ2dlckBpbnNlbC5jaA==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.