Skip to main content

METHODS article

Front. Neurosci., 24 December 2021
Sec. Perception Science
This article is part of the Research Topic Imaging in the Visual System Disorders View all 11 articles

DW-Net: Dynamic Multi-Hierarchical Weighting Segmentation Network for Joint Segmentation of Retina Layers With Choroid Neovascularization

\r\nLianyu WangLianyu Wang1Meng WangMeng Wang1Tingting WangTingting Wang1Qingquan MengQingquan Meng1Yi ZhouYi Zhou1Yuanyuan PengYuanyuan Peng1Weifang ZhuWeifang Zhu1Zhongyue ChenZhongyue Chen1Xinjian Chen,*\r\nXinjian Chen1,2*
  • 1School of Electronics and Information Engineering, Soochow University, Suzhou, China
  • 2State Key Laboratory of Radiation Medicine and Protection, Soochow University, Suzhou, China

Choroid neovascularization (CNV) is one of the blinding factors. The early detection and quantitative measurement of CNV are crucial for the establishment of subsequent treatment. Recently, many deep learning-based methods have been proposed for CNV segmentation. However, CNV is difficult to be segmented due to the complex structure of the surrounding retina. In this paper, we propose a novel dynamic multi-hierarchical weighting segmentation network (DW-Net) for the simultaneous segmentation of retinal layers and CNV. Specifically, the proposed network is composed of a residual aggregation encoder path for the selection of informative feature, a multi-hierarchical weighting connection for the fusion of detailed information and abstract information, and a dynamic decoder path. Comprehensive experimental results show that our proposed DW-Net achieves better performance than other state-of-the-art methods.

Introduction

The choroid is an important tissue of the human eye. It is a soft and smooth brown film located between the retina and the sclera (Hageman et al., 1995; Bressler, 2002). Optical coherence tomography (OCT) is a noninvasive, high-resolution biological imaging technology that can be used for in vivo measurement of fundus structures such as the retina, retinal nerve fiber layer, macula, and optic disc (Huang et al., 1991; Fercher et al., 1993). In OCT image, the normal retinal structure presents multiple interconnected retinal layers (Srinivasan et al., 2014; Zanet et al., 2019); from the inside to the outside are: the nerve fiber layer (NFL), Ganglion cell layer (GCL), inner plexiform layer (IPL), inner nuclear layer (INL), outer plexiform layer (OPL), outer nuclear layer (ONL), outer photoreceptor segment layer (OPSL), and retinal pigment epithelium (RPE). Figure 1 shows the OCT image with normal retinal layers.

FIGURE 1
www.frontiersin.org

Figure 1. Optical coherence tomography (OCT) image of the normal retinal layer. (A) Original image. (B) Label. NFL, nerve fiber layer; GCL, ganglion cell layer; IPL, inner plexiform layer; INL, inner nuclear layer; OPL, outer plexiform layer; ONL, outer nuclear layer; OPSL, outer photoreceptor segment layer; RPE, retinal pigment epithelium.

Choroid neovascularization (CNV), also known as sub-retinal neovascularization, refers to the pathologically proliferating blood vessels that extend from the choroid to the sub-retinal pigment epithelium, the sub-retinal space, or a combination of the above (Lopez et al., 1991; Laud et al., 2006). Figure 2 shows the OCT image of the retina with CNV. Due to the high permeability of the vascular wall of neovascularization, it may lead to sub-retinal hemorrhage, lipid exudation, detachment of the retinal pigment epithelium and choroid, and the formation of fibrotic scars (Zhang et al., 2017). The main symptoms are visual loss, distortion of vision, and central or para-central dark spots, which eventually lead to blindness (Saxe et al., 1993; Grossniklaus and Green, 2004). Therefore, early detection and quantitative measurement of CNV are crucial for the establishment of subsequent treatment plans.

FIGURE 2
www.frontiersin.org

Figure 2. Optical coherence tomography (OCT) image of the normal retinal layer containing choroid neovascularization (CNV). (A) Original image. (B) Label.

Medical-aided diagnosis segmentation algorithm based on computer vision can quickly obtain the shape, size, location, and optical density value, which can provide reliable and accurate quantitative information for the diagnosis and treatment of CNV (Chen et al., 2012, 2016; Gao et al., 2015). Therefore, the development of a reliable and automatic OCT-based CNV segmentation method requires further attention.

However, accurate segmentation of CNV still faces great challenges. Firstly, the structure of the retina is complex due to the multiple retinal layers it contains (Garvin et al., 2009; Roy et al., 2017). Secondly, with the existence of CNV or fluid, the adjacent retinal layers will deform greatly, resulting in a decrease in contrast (Shi et al., 2015). Thirdly, some CNVs are small objects that are hard to discriminate, resulting in performance degradation.

Therefore, focusing on these problems, we propose a new dynamic multi-hierarchical weighting segmentation network (DW-Net) for the joint segmentation of CNV and retinal layers in retinal OCT images. To alleviate the increase in the difficulty of CNV segmentation due to the complexity of the retinal layer structure, we developed a joint framework for the simultaneous segmentation of the retinal layers and CNV. To reduce the impact of partial deformation of the retinal layers and improve the segmentation performance on small CNVs, multiple multi-hierarchical connections are introduced in our proposed network, thus making full use of contextual information. Comprehensive experimental results suggest that our proposed DW-Net achieves superior performance in OCT-based segmentation of retinal layers with CNV compared with several state-of-the-art methods.

The major contributions of this paper can be summarized as follows. Firstly, we create an end-to-end deep learning framework for the simultaneous segmentation of the retinal layers and CNV. Secondly, we develop multiple multi-hierarchical connections to extract and fuse the features in a contextual-driven manner. Thirdly, we evaluate the proposed methods on OCT images of the retina, with experimental results suggesting the effectiveness of our methods.

The rest of the paper is organized as follows. We first briefly review related work in section “Related Work.” Then, we introduce the proposed dynamic multi-hierarchical weighting segmentation network (DW-Net) in section “Methods”. In section “Experiments and Results,” we present the experimental settings, experimental results, ablation study, and the materials used in this study. The ablation study and the limitations of our current work are shown in section “Discussion,” as well as possible future directions. Finally, we conclude this paper in section “Conclusion.”

Related Work

In recent years, several automatic CNV segmentation methods of the retinal layers and CNV have been proposed. Lu et al. (2010) segmented the retinal blood vessel into multiple vascular and non-vascular slices, smoothed and filtered to refine the layer boundary. Song et al. (2013) further used arc-based graph representation, combined extensive prior information through paired energy terms, and calculated the maximum flow in low-order polynomial time. In the same year, Dufour et al. (2013) proposed a graph-based automatic multi-surface segmentation algorithm to add prior information from the learning model and further improved the accuracy of segmentation. Xu et al. (2013) used the Iowa reference algorithm to segment 10 retinal layers, followed by a combined graph search/graph cut method to segment pairs of adjacent retinal layers and any present fluid-associated abnormality detection region in 3D. Xi et al. (2017, 2018) developed a structure prior method based on sparse representation classification and local latent function to capture the global spatial structure and local similarity structure prior, which improved the segmentation robustness of CNVs of different sizes.

At present, deep neural networks have been widely used for the segmentation of retinal images and CNV. Su et al. (2020) proposed a differential amplification block to extract the contrast information of the foreground and background, which is integrated into the U-shaped convolutional neural network for CNV segmentation. Based on density cell-like P systems, Xue et al. (2018) proposed an automatic quantification method of the CNV total lesion area on outer retinal OCT angiograms to improve the accuracy of the segmentation boundaries. To simultaneously segment layers and neovascularization, Xiang et al. (2018) extracted well-designed features to find the coarse surfaces of different OCTs and introduced a constrained graph search algorithm to accurately detect retinal surfaces. Wang et al. (2020) trained two independent convolutional neural networks to classify the input scans according to the presence or absence of CNVs in a complementary manner, forming a powerful CNV description system.

Methods

Overview

The encoder–decoder structure (Ronneberger et al., 2015; Zhao et al., 2017; Feng et al., 2020) has been proven to be an efficient architecture for pixel-wise semantic segmentation among many deep learning-based methods; therefore, we propose a novel joint segmentation framework to solve the challenges in retinal CNV segmentation based on this. As shown in Figure 3A, the proposed DW-Net consists of three parts: residual aggregation encoder path, dynamic multi-hierarchical weighting connection, and dynamic decoder path.

FIGURE 3
www.frontiersin.org

Figure 3. (A) Architecture of the proposed dynamic multi-hierarchical weighting segmentation network (DW-Net). The dark yellow part in (B,C) indicate the residual aggregation encoder path and the dynamic multi-hierarchical weighting connection, respectively.

Residual Aggregation Encoder Path

In the conventional encoder path, encoders are composed of stacked convolutional layers and pooling layers, which are used to extract rich semantic information and global features layer by layer. However, continuous convolution and pooling will reduce the resolution of semantic features, which may lead to the loss of some small objects (such as small CNVs). To reduce the loss of resolution and enhance the selectivity of the feature encoder, we utilized the residual module as our encoder in this paper. By fusing the current feature maps with previous feature maps, the residual module can obtain informative feature maps that are more conducive to subsequent segmentation.

As shown in Figure 3B, the input data X ∈ ℝH×W×C is encoded by a convolutional layer and four encoders, as follows:

{ X i , 0 = Conv ( X )        i = 0       X i , 0 = R e s N e t ( X i - 1 , 0 )          1 i 4 (1)

where H, W, and C denote the height, width, and channels of the input data, respectively, and X0,0 represents the output of the first convolutional layer. Xi,0(1 ≤ i ≤ 4) denotes the output feature maps of four encoders, with channel numbers of 64, 128, 256, and 512, respectively. To improve the feature extraction ability and save computing resources, we used the pre-trained model of layers 1–4 in ResNet18 (He et al., 2016) to initialize the parameters of the encoders.

Dynamic Multi-Hierarchical Weighting Connection

Encoders of different hierarchies can extract features of different levels. The local features extracted by the low-level encoder are relatively simple and are more inclined to the basic components of images such as points, lines, and contours, while the high-level encoder is able to extract more complex features, such as abstract globe information. As for the semantic segmentation tasks, abstract global features can improve the overall positioning ability of the object, while fine local features can refine the edges of the segmented object.

To make full use of the feature maps in multilevel encoders, as in Figure 4 (Zhou et al., 2018) developed UNet++. They concatenated the features of encoders in order layer by layer directly (gray dotted line in Figure 4), thus improving the performance of the segmentation network. However, the output feature map of the encoder usually contains interference information such as background and noise, which need to be selected and filtered. Also, the output features of each level have different contributions to the segmentation task; therefore, direct concatenation cannot highlight the importance of each part. In addition, concatenation in each hierarchy will greatly increase the parameters of the network, which may reduce the training and increase the risk of overfitting.

FIGURE 4
www.frontiersin.org

Figure 4. Architecture of the UNet++ by Zhou et al. (2018).

In response to the above problems, we proposed a dynamic multi-hierarchical weighting connection, which aims to take full advantage of the multi-scale extracted features that are conducive to segmentation in a contextual-driven manner and to filter irrelevant information. Figure 3C shows the structure of our proposed dynamic multi-hierarchical weighting connection, and its calculation process is as follows:

X i , j = Conv ( X i , 0 + D B ( X x + 1 , j - 1 ) )    0 i 3 , j = 1 (2)
X i , j = Conv ( k = 0 j - 2 α i , 2 j + k * X + i , k X i , j - 1 + D B ( X x + 1 , j - 1 ) )        0 i 4 - j , 2 j 4 (3)

where i and j denote the layer index and column index of the feature map Xi,j, respectively, and DB represents a decode module composed of a 3 × 3 convolutional layer and an upsampling layer. αi,2j+k is a learnable parameter, which is optimized through multiple iterations. To make full advantage of the known detailed information and abstract information at all hierarchies, we performed pixel addition on all higher-level feature maps and current-level feature maps according to their weight, thereby dynamically enhancing the segmentation ability of the current-level decoder.

Dynamic Decoder Path

The dynamic decoder path contains four decoders, and the channels of the output feature map X4−i,i (0 ≤ i ≤ 3) are 256, 128, 64, and 32, respectively. The decoder path is composed of stacked convolutional layers and upsampling layers, which aims to upsample the feature maps with strong semantic information from a high level and restore the spatial resolution layer by layer. Zhou et al. (2018), conducted pixel-wise averaging on the output feature map of the decoder path and output feature maps at the same hierarchy (the black straight line in the upper part of Figure 4), as follows:

Y = soft  max ( 1 4 j = 1 4 Conv ( X 0 , j ) ) (4)

where Conv is a simple 1 × 1 convolutional layer for compressing the output feature channel. This strategy directly merges different feature maps without considering their depths. However, in convolutional neural networks, segmentation tasks are sensitive to the depth of the network; thus, a reasonable design of its depth will improve the performance (Simonyan and Zisserman, 2014). For this consideration, we modified the decoder path of UNet++ (Zhou et al., 2018) to extract more informative prediction results.

Y = soft  max ( 1 4 j = 1 4 D N ( X 0 , j ) ) (5)

where DN is a dynamic fusion module consisting of a bilinear upsampling layer used to restore the input spatial resolution and two 1 × 1 convolutional layers followed by a normalization layer and a Relu nonlinear activation layer. Then, the 1 × 1 convolutional layer is applied for channel compression. Finally, we performed pixel-wise averaging on the output of DN, followed by a softmax layer. Y′ represents the predicted probability map.

Loss Function

In the task of semantic segmentation of medical images, the pixel-by-pixel cross-entropy loss, ℒCE, is a commonly used loss function that compares the predicted probability map with the gold standard (GT) in order according to the spatial position.

Y = soft  max ( 1 4 j = 1 4 D N ( X 0 , j ) ) (6)

where k denotes the number of objects and Y represents the gold standard.

Dice loss, ℒDice, is another widely used loss function (Milletari et al., 2016) that aims to measure the overlap ratio of two samples, and its value ranges from 0 to 1.

D i c e = 1 - 1 k c = 0 k - 1 2 Y Y + ξ Y + Y + ξ (7)

where ξ is set to a very small constant to ensure that the divisor is not equal to 0. The final loss function we used is as follows:

= D i c e + C E (8)

Experiments and Results

Dataset and Implementation

Dataset

To evaluate the effectiveness of the proposed method, we conducted comprehensive experiments. The dataset we used in the experiment was collected by the Joint Shantou International Eye Center of Shantou University and The Chinese University of Hong Kong. The acquisition process lasted 13 months, and 6,016 retinal OCT images from 47 three-dimensional retinal OCT volumes with CNV were completely acquired through the Zeiss canner. The size of the actual scanning area is 6 mm × 2 mm × 6 mm (X × Y × Z), and the number of voxels is 512 × 1,024 × 128. Pixel-level annotations of NFL, GCL, IPL, INL, OPL, ONL, OPSL+RPE, and CNV were provided by senior ophthalmologists.

Implementation Details

The implementation of our proposed DW-Net is based on the public platform Pytorch 1.8.0 with CUDA 11.0 parallel computing library and GeForce RTX 3090 GPU with 24-GB memory.

To save computing resources and increase network receptivity, each slice was resized to 512 × 512 by bilinear interpolation. We divided the 6,016 retinal OCT images into four groups, with the slice number as balanced as possible. Fourfold cross-validation was conducted on the divided dataset (1,664, 1,792, 1,280, and 1,280). The Adam optimizer with a learning rate of 1e-4 was adopted as our optimizer. The batch size and epoch were set to 4 and 100, respectively. For fair comparison, we used the same training strategy in all experiments.

Evaluation Metrics

Five metrics including dice similarity coefficients (DSCs), intersection-over-union (IoU), accuracy (Acc), sensitivity (Sen), and precision (Pre) (Garcia-Garcia et al., 2017) were used to fully and fairly evaluate the performance, where TN, TP, FN, and FP represent true negative, true positive, false negative, and false positive, respectively.

DSC = 2 T P 2 T P + FP + FN (9)
IoU = TP TP + FP + FN (10)
Acc = TP + TN TP + FP + TN + FN (11)
Sen = TP TP + FN (12)
Pr e = TP TP + FP (13)

Results

We first compared our proposed DW-Net with other excellent convolutional neural network (CNN)-based methods, including UNet (Ronneberger et al., 2015), AttUNet (Oktay et al., 2018), CE-Net (Gu et al., 2019), Multi-ResUNet (Ibtehaz and Rahman, 2020), R2UNet (Alom et al., 2018), and DeepLab v3 (Chen et al., 2017). In addition, UNet++ (Zhou et al., 2018) was applied as our backbone. Tables 1, 2 show the mean joint segmentation results of the 7 retinal layers containing CNV and the joint segmentation results of CNV, respectively.

TABLE 1
www.frontiersin.org

Table 1. Mean segmentation results (in percent) of the contrast experiments and ablation studies (mean ± SD).

TABLE 2
www.frontiersin.org

Table 2. Choroid neovascularization (CNV) segmentation results (in percent) of the contrast experiments and ablation studies (mean ± SD).

From Table 1, it is worth noting that the proposed DW-Net achieves better performance than all of the above methods, with DSC, IoU, Acc, Sen, and Pre of 95.38, 91.26, 99.41, 95.44, and 95.36%, respectively. As for CNV, the performance of our proposed joint segmentation realized 2.52, 4.42, 0.09, 2.65, and 1.96% improvements in terms of DSC, IoU, Acc, Sen, and Pre, respectively, over the backbone, as shown in Table 2.

The performance of CE-Net (Gu et al., 2019) was comparable to that of the proposed DW-Net for CNV Sen, while being slightly lower in other metrics. In Figure 5, we plotted the visualization results of the different methods, where the red, green, dark blue, yellow, light blue, purple, white, and navy blue areas represent NFL, GCL, IPL, INL, OPL, ONL, OPSL+RPE, and CNV, respectively. It can be seen that our proposed DW-Net can accurately segment each retinal layer and CNV, which is closer to the GT compared with the other methods.

FIGURE 5
www.frontiersin.org

Figure 5. Visualization results of the different methods.

Furthermore, we carried out a quantitative analysis of the experimental results. Figure 6 shows a histogram of the comparison between the size of the actual CNV and the segmented CNV using DW-Net, which are represented by blue and orange bars, respectively. It can be seen from the qualitative and quantitative results in the figure that the volume difference between the prediction of DW-Net and GT is generally small, which further proves the effectiveness and stability of the joint segmentation network and suggest promising clinical value and application prospects.

FIGURE 6
www.frontiersin.org

Figure 6. Histogram of choroid neovascularization (CNV) volume comparison.

Discussion

In this section, we first conduct a series of ablation experiments. Then, we study the contribution of the information on the retinal layers to the CNV segmentation task. Finally, we introduce the limitations of this work and possible future research directions.

Ablation Experiments for Residual Aggregation Encoder Path

To evaluate the effectiveness of the residual aggregation encoder path, we further compared the backbone with its counterparts (called Res18UNet++). Specifically, Res18UNet++ directly applies residual aggregation encoder path based on UNet++ (Zhou et al., 2018) and replaces concatenation by pixel addition, as shown in Figure 7A, where α is a constant that is fixed to 1. Table 3 reports the segmentation results.

FIGURE 7
www.frontiersin.org

Figure 7. Architecture of Res18UNet++ (A) and AdaptiveUNet++ (B).

TABLE 3
www.frontiersin.org

Table 3. Ablation experiments (mean ± SD).

It can be seen that our proposed Res18UNet++ achieves better performance over the backbone on all metrics, which suggests that the residual aggregation encoder path can retain more effective features as possible to alleviate the resolution loss caused by network deepening.

Ablation Experiments for Dynamic Multi-Hierarchical Weighting Connection

We also compared the backbone with another counterpart (called AdaptiveUNet++), as shown in Figure 7B. Here, αi,2j+k is a learnable parameter that is multiplied with the output feature map of the corresponding encoder. Its value during the training process is shown in Figure 8.

FIGURE 8
www.frontiersin.org

Figure 8. Value of the learnable parameter αi,2j+k during training.

We can conclude from Table 3 and Figure 8 that our proposed AdaptiveUNet++ enables the encoders to utilize multi-scale context information and filter irrelevant information. In addition, residual aggregation encoder path and dynamic multi-hierarchical weighting connection can influence and promote each other, thereby further improving the overall joint segmentation performance of the network, as shown in the results of DW-Net in Table 3.

Ablation Experiments for Retinal Layer Information

All the experiments above were based on the assumption that the introduction of complex retinal layer information is conducive to improving the performance of CNV segmentation. Therefore, we performed joint segmentation of the retinal layers and CNV. In this section, we set out to verify the assumption.

Pre-processing

We considered CNV as the foreground, and the corresponding spatial label was set to 1, then the remaining area including the retinal layers was regarded as background, with the label of 0. Here, the joint segmentation was transformed into a foreground–background segmentation. A variant of DW-Net, named DW-Net-2, was applied for a single CNV segmentation, where the last layer of the network was modified to sigmoid function, and the number of output channels was set to 1. Table 4 shows the segmentation results of DW-Net-2 and DW-Net.

TABLE 4
www.frontiersin.org

Table 4. Choroid neovascularization (CNV) segmentation experiments without retinal layers (mean ± SD).

It can be clearly seen that the performance of DW-Net is superior, which proves that the introduction of retinal layer information is conducive to distinguishing the features of background, retinal layers, and CNV, thereby improving the segmentation performance of CNV.

Limitations and Future Work

The current work still has many limitations. Our proposed DW-Net contains many learnable parameters, which will increase the computational burden; therefore, further compression is needed in practical applications. The dataset used in our experiment needs further expansion, which is also one of our future works. We will conduct experiments on more datasets to verify the effectiveness and generalization of the proposed DW-Net.

Conclusion

CNV segmentation is a fundamental task in medical image analysis. In this paper, we proposed a novel end-to-end dynamic multi-hierarchical weighting segmentation network (DW-Net) for the simultaneous segmentation of the retinal layers and CNV. Specifically, the proposed network is composed of a residual aggregation encoder path for the selection of informative feature, a multi-hierarchical weighting connection for the fusion of detailed information and abstract information, and a dynamic decoder path. Comprehensive experimental results show the effectiveness and stability of our proposed DW-Net and suggest promising clinical value and application prospects.

Data Availability Statement

The datasets presented in this article are not readily available because constrained by ethics and patient privacy. Requests to access the datasets should be directed to corresponding author and LW, lywang12@126.com.

Ethics Statement

The studies involving human participants were reviewed and approved by the Soochow University. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author Contributions

LW conceptualized and designed the study, wrote the first draft of the manuscript, and performed data analysis. MW, TW, QM, YZ, YP, ZC, and XC performed the experiments, collected and analyzed the data, and revised the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This study was supported by the National Key R&D Program of China (2018YFA0701700).

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Alom, M. Z., Hasan, M., Yakopcic, C., Taha, T. M., and Asari, V. K. (2018). Recurrent residual convolutional neural network based on U-Net (R2U-Net) for medical image segmentation. arXiv [Preprint]. arXiv:1802.06955.

Google Scholar

Bressler, N. M. (2002). Verteporfin therapy of subfoveal choroidal neovascularization in age-related macular degeneration-Reply. Am. J. Ophthalmol. 133, 857–859. doi: 10.1016/s0002-9394(02)01423-x

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, H., Xia, H., Qiu, Z., Chen, W., and Chen, X. (2016). Correlation of optical intensity on optical coherence tomography and visual outcome in central retinal artery occlusion. Retina 36, 1964–1970. doi: 10.1097/IAE.0000000000001017

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, L.-C., Papandreou, G., Schroff, F., and Adam, H. (2017). Rethinking atrous convolution for semantic image segmentation. arXiv [Preprint]. arXiv:1706.05587.

Google Scholar

Chen, X., Zhang, L., Sohn, E. H., Lee, K., Niemeijer, M., Chen, J., et al. (2012). Quantification of external limiting membrane disruption caused by diabetic macular edema from SD-OCT. Invest. Ophthalmol. Vis. Sci. 53, 8042–8048. doi: 10.1167/iovs.12-10083

PubMed Abstract | CrossRef Full Text | Google Scholar

Dufour, P. A., Ceklic, L., Abdillahi, H., Schröder, S., De Dzanet, S., Wolf-Schnurrbusch, U., et al. (2013). Graph-based multi-surface segmentation of OCT data using trained hard and soft constraints”. IEEE Trans. Med. Imag. 32, 531–543. doi: 10.1109/TMI.2012.2225152

PubMed Abstract | CrossRef Full Text | Google Scholar

Feng, S., Zhao, H., Shi, F., Cheng, X., Wang, M., Ma, Y., et al. (2020). CPFNet: context pyramid fusion network for medical image segmentation. IEEE Trans. Med. Imaging 39, 3008–3018. doi: 10.1109/TMI.2020.2983721

PubMed Abstract | CrossRef Full Text | Google Scholar

Fercher, A. F., Hitzenberger, C. K., Drexler, W., Kamp, G., and Sattmann, H. (1993). In vivo optical coherence tomography in ophthalmology. Am. J. Ophthalmol. 116, 113–115.

Google Scholar

Gao, E., Chen, B., Yang, J., Shi, F., Zhu, W., Xiang, D., et al. (2015). Comparison of retinal thickness measurements between the topcon algorithm and a graph-based algorithm in normal and glaucoma eyes. PLoS One 10:e0128925. doi: 10.1371/journal.pone.0128925

PubMed Abstract | CrossRef Full Text | Google Scholar

Garcia-Garcia, A., Orts-Escolano, S., Oprea, S., Villena-Martinez, V., and Garcia-Rodriguez, J. (2017). A review on deep learning techniques applied to semantic segmentation. arXiv [Preprint]. arXiv:1704.06857.

Google Scholar

Garvin, M. K., Abramoff, M. D., Wu, X., Russell, S. R., Burns, T. L., and Sonka, M. (2009). Automated 3-D intraretinal layer segmentation of macular spectral-domain optical coherence tomography images. IEEE Trans. Med. Imag. 28, 1436–1447. doi: 10.1109/TMI.2009.2016958

PubMed Abstract | CrossRef Full Text | Google Scholar

Grossniklaus, H. E., and Green, W. R. (2004). Choroidal neovascularization. Am. J. Ophthalmol. 137, 0–503.

Google Scholar

Gu, Z., Cheng, J., Fu, H., Zhou, K., Hao, H., Zhao, Y., et al. (2019). CE-Net: context encoder network for 2D medical image segmentation. IEEE Trans. Med. Imaging 38, 2281–2292. doi: 10.1109/TMI.2019.2903562

PubMed Abstract | CrossRef Full Text | Google Scholar

Hageman, G. S., Gehrs, K., Johnson, L. V., and Anderson, D. (1995). “Age-related macular degeneration (AMD),” in The Orianization of the Retina and Visual System (Salt Lake City, UT: University of Utah Health Sciences Center).

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep residual learning for image recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Las Vegas, NV), 770–778.

Google Scholar

Huang, D., Swanson, E. A., Lin, C. P., Schuman, J. S., Stinson, W. G., Chang, W., et al. (1991). Optical coherence tomography. Science 254, 1178–1181.

Google Scholar

Ibtehaz, N., and Rahman, M. S. (2020). MultiResUNet: rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 121, 74–87. doi: 10.1016/j.neunet.2019.08.025

PubMed Abstract | CrossRef Full Text | Google Scholar

Laud, K., Spaide, P. F., and Freund, K. B. (2006). Treatment of choroidal neovascularization in pathologic myopia with intravitreal bevacizumab. Ratina 26, 960. doi: 10.1097/01.iae.0000240121.28034.c3

CrossRef Full Text | Google Scholar

Lopez, P. F., Grossniklaus, H. E., Lambert, H. M., Aaberg, T. M., Capone, A., Sternberg, P., et al. (1991). Pathologic features of surgically-excised subretinal neovascular membranes in age-related macular degeneration. Am. J. Ophthalmol. 112, 647–656. doi: 10.1016/s0002-9394(14)77270-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Lu, S., Cheung, C. Y.-L., Liu, J., Lim, J. H., Leung, C. K.-S., and Wong, T. Y. (2010). Automated layer segmentation of optical coherence tomography images. IEEE Trans. Biomed. Eng. 57, 2605–2608.

Google Scholar

Milletari, F., Navab, N., and Ahmadi, S. (2016). “V-net: fully convolutional neural networks for volumetric medical image segmentation,” in Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA.

Google Scholar

Oktay, O., Schlemper, J., Le Folgoc, L., Lee, M., Heinrich, M., Misawa, K., et al. (2018). Attention U-Net: learning where to look for the pancreas. arXiv [Preprint]. arXiv:1804.03999.

Google Scholar

Ronneberger, O., Fischer, P., and Brox, T. (2015). “U-net: convolutional networks for biomedical image segmentation,” in International Conference on Medical Image Computing and Computer-Assisted Intervention, eds N. Navab, J. Hornegger, W. Wells, and A. Frangi (Cham: Springer), 234–241.

Google Scholar

Roy, A. G., Conjeti, S., Karri, S. P. K., Sheet, D., Katouzian, A., Wachinger, C., et al. (2017). ReLayNet: retinal layer and fluid segmentation of macular optical coherence tomography using fully convolutional networks. Biomed. Optics Exp. 8, 3627–3642. doi: 10.1364/BOE.8.003627

PubMed Abstract | CrossRef Full Text | Google Scholar

Saxe, S. J., Grossniklaus, H. E., Lopez, P. F., and Sternberg, P.L’Hernault, N. (1993). Ultrastructural features of surgically-excised subretinal neovascular membranes in the ocular histoplasmosis syndrome. Arch. Ophthalmol. 111, 88–95. doi: 10.1001/archopht.1993.01090010092033

PubMed Abstract | CrossRef Full Text | Google Scholar

Shi, F., Chen, X., Zhao, H., Zhu, W., Xiang, D., Gao, E., et al. (2015). Automated 3-D retinal layer segmentation of macular optical coherence tomography images with serous pigment epithelial detachments. IEEE Trans. Med. Imag. 34, 441–452. doi: 10.1109/TMI.2014.2359980

PubMed Abstract | CrossRef Full Text | Google Scholar

Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. arXiv [Preprint]. arXiv: 1409.1556.

Google Scholar

Song, Q., Bai, J., Garvin, M. K., Sonka, M., Buatti, J. M., and Wu, X. (2013). Optimal multiple surface segmentation with shape and context priors. IEEE Trans. Med. Imag. 32, 376–386. doi: 10.1109/TMI.2012.2227120

PubMed Abstract | CrossRef Full Text | Google Scholar

Srinivasan, P. P., Heflin, S. J., Izatt, J. A., Arshavsky, V. Y., and Farsiu, S. (2014). Automatic segmentation of up to ten layer boundaries in SD-OCT images of the mouse retina with and without missing layers due to pathology. Biomed. Optics Exp. 5, 348–365. doi: 10.1364/BOE.5.000348

PubMed Abstract | CrossRef Full Text | Google Scholar

Su, J., Chen, X., Ma, Y., Zhu, W., and Shi, F. (2020). Segmentation of choroid neovascularization in OCT images based on convolutional neural network with differential amplification blocks. SPIE Med. Imaging 2020 Image Process.

Google Scholar

Wang, J., Hormel, T. T., Gao, L., Zang, P., Guo, Y., Wang, X., et al. (2020). Automated diagnosis and segmentation of choroidal neovascularization in OCT angiography using deep learning. Biomed. Optics Exp. 11, 927–944. doi: 10.1364/BOE.379977

PubMed Abstract | CrossRef Full Text | Google Scholar

Xi, X., Meng, X., Yang, L., Nie, X., Yin, Y., and Chen, X. (2017). Learned local similarity prior embedded active contour model for choroidal neovascularization segmentation in optical coherence tomography images. Sci. China Inform. Sci.

Google Scholar

Xi, X., Meng, X., YangX, L., Nie, X., Yin, Y., and Chen, X. (2018). Automated segmentation of choroidal neovascularization in optical coherence tomography images using multi-scale convolutional neural networks with structure prior. Multimedia Syst. 25, 1–8. doi: 10.1117/1.jmi.6.2.024009

CrossRef Full Text | Google Scholar

Xiang, D., Tian, H., Yang, X., Shi, F., Zhu, W., Chen, H., et al. (2018). Automatic segmentation of retinal layer in OCT images with choroidal neovascularization. IEEE Trans. Image Process. 25, 5880–5891. doi: 10.1109/TIP.2018.2860255

PubMed Abstract | CrossRef Full Text | Google Scholar

Xu, X., Zhang, L., Lee, K., Wahle, A., Chen, X., Wu, X., et al. (2013). Automated choroidal neovascularization associated abnormality detection and quantitative analysis from clinical SD-OCT. Invest. Ophthalmol. Visual Sci. 54:5510.

Google Scholar

Xue, J., Camino, A., Bailey, S. T., Liu, X., Li, D., and Jia, Y. (2018). Automatic quantification of choroidal neovascularization lesion area on OCT angiography based on density cell-like P systems with active membranes. Biomed. Opt. Exp. 9, 3208–3219. doi: 10.1364/BOE.9.003208

PubMed Abstract | CrossRef Full Text | Google Scholar

Zanet, S. D., Ciller, C., Apostolopoulos, S., Wolf, S., and Sznitman, R. (2019). “Pathological OCT retinal layer segmentation using branch residual u-shape networks,” in Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (Cham: Springer) 2017: 294–301.

Google Scholar

Zhang, Q., Chen, C. L., Chu, Z., Zheng, F., Miller, A., Roisman, L., et al. (2017). Automated quantitation of choroidal neovascularization: a comparison study between spectral-domain and sweptsource OCT angiograms. Invest. Ophthalmol. Vis. Sci. 58, 1506–1513. doi: 10.1167/iovs.16-20977

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao, H., Shi, J., Qi, X., Wang, X., and Jia, J. (2017). “Pyramid scene parsing network,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (Piscataway, NJ: IEEE), 2881–2890.

Google Scholar

Zhou, Z., Siddiquee, M. M. R., Tajbakhsh, N., and Liang, J. (2018). “UNet++: a nested U-Net architecture for medical image segmentation,” in Deep Learningin Medical Image Analysis and Multimodal Learning for Clinical DecisionSupport, ed. D. Stoyanov (Cham: Springer), 3–11. doi: 10.1007/978-3-030-00889-5

CrossRef Full Text | Google Scholar

Keywords: multi-target segmentation, choroid neovascularization, convolutional neural network, optical coherence tomography, medical image processing, attention mechanism

Citation: Wang L, Wang M, Wang T, Meng Q, Zhou Y, Peng Y, Zhu W, Chen Z and Chen X (2021) DW-Net: Dynamic Multi-Hierarchical Weighting Segmentation Network for Joint Segmentation of Retina Layers With Choroid Neovascularization. Front. Neurosci. 15:797166. doi: 10.3389/fnins.2021.797166

Received: 18 October 2021; Accepted: 22 November 2021;
Published: 24 December 2021.

Edited by:

Jian Zheng, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences (CAS), China

Reviewed by:

Shouping Zhu, Xidian University, China
Xiaopeng Ma, Shandong University, China

Copyright © 2021 Wang, Wang, Wang, Meng, Zhou, Peng, Zhu, Chen and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xinjian Chen, xjchen@suda.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.