- 1School of Informatics, Hunan University of Chinese Medicine, Changsha, Hunan, China
- 2Big Data Analysis Laboratory of Traditional Chinese Medicine, Hunan University of Chinese Medicine, Changsha, Hunan, China
Background: Segmentation and evaluation of infarcts on medical images are essential for diagnosis and prognosis of acute ischemic stroke (AIS). Computed tomography (CT) is the first-choice examination for patients with AIS.
Methods: To accurately segment infarcts from the CT images of patients with AIS, we proposed an automated segmentation method combining the convolutional attention mechanism and residual Deformable Transformer in this article. The method used the encoder-decoder structure, where the encoders were employed for downsampling to obtain the feature of the images and the decoder was used for upsampling and segmentation. In addition, we further applied the convolutional attention mechanism and residual network structure to improve the effectiveness of feature extraction. Our code is available at: https://github.com/XZhiXiang/AIS-segmentation/tree/master.
Results: The proposed method was assessed on a public dataset containing 397 non-contrast CT (NCCT) images of AIS patients (AISD dataset). The symptom onset to CT time was less than 24 h. The experimental results illustrate that this work had a Dice coefficient (DC) of 58.66% for AIS infarct segmentation, which outperforms several existing methods. Furthermore, volumetric analysis of infarcts indicated a strong correlation (Pearson correlation coefficient = 0.948) between the AIS infarct volume obtained by the proposed method and manual segmentation.
Conclusion: The strong correlation between the infarct segmentation obtained via our method and the ground truth allows us to conclude that our method could accurately segment infarcts from NCCT images.
1. Introduction
Stroke refers to sudden brain dysfunction caused by cerebral blood circulation disorder and is one of the most prevalent fatal illnesses. Stroke can be grouped into two types: ischemic and hemorrhagic stroke. These are caused by blockage or rupture of cerebral blood vessels, respectively. Most patients with stroke suffer from acute ischemic stroke (AIS) (1). Due to the fast speed and low expense of CT, it has become the first-choice imaging technique for diagnosis and prognosis of stroke (2). In addition, the comparative analysis of the infarct volume in non-contrast CT (NCCT) images before and after treatment can facilitate judgment of the effectiveness of treatment. NCCT imaging is one of the essential methods for AIS diagnosis (3), and rapid segmentation of infarcts is crucial for AIS diagnosis. Manual segmentation is mainly utilized in clinical practice to ensure segmentation accuracy. However, problems exist with this method, such as excessive reliance on prior knowledge in the medical field and human evaluation errors (4). Therefore, the segmentation of medical images can be challenging in medical image analysis (5). Many methods have treated the segmentation of AIS infarcts as an anomaly detection task, determining the differences between the infarct and surrounding tissues (6). Nevertheless, detecting the infarct can only provide a rough assessment of AIS, which cannot effectively guide diagnosis or the development of corresponding treatment plans. To segment AIS infarcts from medical images, several machine learning (ML)-based (7) segmentation methods, such as SVM and random forest (8), have been proposed. For example, Kuang et al. (9) proposed a convex optimization method based on random forest classification.
In order to enhance the segmentation performance, segmentation methods based on deep learning (DL) (10) have been proposed. The CNN-based AIS infarct segmentation method achieved impressive performance (11). A full convolutional neural network (FCN) removed the fully connected layers in CNN and elevated image segmentation from image-level to pixel-level. Zhang et al. (12) combined 3D FCN with dense connections to automatically segment AIS infarcts. Compared to ordinary images, medical images have a wide grayscale range and unclear boundaries. To address these issues, U-Net (13) based on FCN was proposed. U-Net employs an encoder-decoder architecture with skip connections between downsampled and upsampled information to improve segmentation accuracy. U-Net has been refined through advancements such as U-Net++ (14) and Res-UNet (15), and these improvements have yielded noteworthy outcomes in numerous image processing domains. Ni et al. (16) proposed a novel asymmetry disentanglement network (ADN), where asymmetric disentanglement of the input NCCTs is first conducted to produce various 3D asymmetric maps. Subsequently, a synthesized intrinsic asymmetry-compensated pathologically enhanced NCCT volume is created and utilized as the input for the segmentation network to segment the AIS infarct, achieving good performance. The backbone of ADN is 3D ResidualUnet, which leverages convolutions as the fundamental operations to perform feature extraction and encoding on the NCCT. However, the limited receptive field of convolutional operations constrains their ability to capture global dependencies (17). Therefore, many studies have explored the use of Transformers in image analysis, which are known for their strong global modeling capabilities. A typical example is the Vision Transformer (ViT) (18), which utilizes a pure Transformer structure to process image patches and demonstrated remarkable achievements in image recognition tasks. SETR (19) employs a Transformer as the encoder and a CNN architecture as the decoder to create a superior-performing segmentation model. The encoder in TransUNet (20) connects CNN with the Transformer and employs the latter to process the final layer features produced by the CNN, resulting in remarkable performance and effectiveness. However, the complexity of training a Transformer for image tasks is considerable and requires high-performance computers.
Therefore, to solve the above problems, we aimed to develop an automatic segmentation method in this work to accurately segment infarcts from NCCT images of AIS patients. Our method is distinguished by the following four characteristics: (1) by combining CNN and a Transformer, we optimize their performance by mitigating their limitations and enhancing their advantages; (2) to improve the efficiency of local feature extraction, the convolutional block attention module (CBAM) is used to direct attention to the key areas of segmentation; (3) to simplify the Transformer for image tasks and to improve the efficiency of model training, deformable multi-heads self-attention (DMSA) is introduced to distribute attention to a few key points around the sampling points instead of to all points in the feature map; (4) given the information loss during transmission via the Transformer layer, we integrate a residual connection before and after the Transformer encoder to enhance the information and achieve better segmentation.
2. Materials and methods
We conducted the experiment on the AIS dataset (AISD) (21) in this study. It consists of 397 NCCT scans of acute ischemic strokes acquired within 24 h of the patient's symptom onset. In addition, patients had a diffusion-weighted MRI (DWI) within a day of receiving the CT scan. The NCCT scans had a slice thickness of 5 mm. Labels were manually annotated by a doctor and carefully checked by another senior doctor. According to the data division in the original article, 345 scans of patients were employed for model training and parameter tuning, and the remaining 52 NCCT scans were employed for evaluating the proposed method.
We employed the Z-score method to normalize the contextual feature information of the original NCCT dataset. Image normalization can optimize the efficiency of DL in segmentation tasks (22). To avoid overfitting problems caused by the limited training data, we used data augmentation methods to diversify it. The methods used include random rotation, Flip, zoom, adding Gaussian white noise, Gaussian blur, adjusting accuracy and contrast, gamma transform, and simulating low resolution.
The segmentation method proposed in this work consisted of two encoders and a decoder, including a CNN Encoder, Transformer Encoder, and Decoder. The encoders obtained the features, and the feature maps were passed to the Decoder for upsampling. They were restored to the same resolution as the source image and segmentation was finally achieved. The CNN Encoder extracted the local features of the image and then modeled global dependency on them via the Transformer Encoder. In addition, the CNN Encoder used CBAM to direct attention to the feature maps. In the Transformer Encoder, DMSA was used to make it simpler, and a residual connection was employed before and after the Transformer Encoder to enhance the information. The structure is shown in Figure 1.
Figure 1. Pipeline of the proposed infarct segmentation method. The structure of the CNN Encoder is shown in the left part of the figure, the middle part shows the Transformer Encoder, and the right part shows the structure of the Decoder. The gray rectangle in the figure is the CNN block, the yellow circle is the convolutional block attention module, and the green rectangle is the Deformable Transformer layer. The CNN Encoder extracts the image's local features, connects the last two layers of features, and passes them into the Transformer Encoder for global relationship modeling. Then, the local and global features are sent to the decoder for processing and segmentation.
2.1. CNN Encoder
The CNN Encoder encoded the input image with multiple convolutional layers, similar to convolutional pyramids. It extracted the local features of the image, in which the CBAM (23) was used to compute effective local attention maps, strengthen the influence of the infarct area, and reduce feature redundancy for segmentation.
The CNN Encoder contained a Conv-IN-ReLU block and three residual convolutional blocks (ResConvBlock). The Conv-IN-ReLU block comprised a convolutional layer with a large kernel and followed processes with instance normalization (IN) (22) and ReLU activation. The intermediate feature map was obtained after the Conv-IN-ReLU block. The given input image was x ∈ RC×D×H×W. D, H, W, and C individually represented depth, height, width, and channels. After each ResConvBlock, D, H, and W were halved. The ResConvBlock downsampled the upper-layer feature maps to high-level and coarse-resolution. The CBAM was employed to integrate channel and spatial attention at local levels, allowing the CNN encoder to effectively capture significant features and their locations, leading to improved performance.
The detailed layout of the CNN Encoder is shown in Figure 2A and contains a Conv-IN-ReLU block, three CBAM blocks, and three ResConvBlock, which consisted of three, three, and two 3D residual convolution operations, respectively. The specific structure of ResConvBlock is shown in Figure 2D.
Figure 2. Detailed structure blueprint of the CNN Encoder (A), Decoder (B), and CBAM (C). The blue “3D Conv” in the CNN Encoder consists of a 3D convolutional layer and processes with IN and ReLu activation. The detailed structure of the yellow “CBAM” is shown in (C), consisting of the CAM and SAM. The specific structure of the green “3D ResConv” is shown in (D). The orange “3D Conv” is a 3D convolutional layer. The gray “TransposeConv” is a 3D transposed convolutional layer. The blue arrows represent the skip connections. The numbers in the convolutional blocks indicate the convolution kernels and stride, and those in the residual convolutional blocks denote the convolution kernels, stride, and convolutional layers.
2.2. CBAM
CBAM consisted of a channel attention module (CAM), followed by a spatial attention module (SAM; as illustrated in Figure 2C). Channels can be deemed to be feature detectors in feature maps so that the CAM can pay attention to features with greater importance. The intermediate feature map served as the input for the CAM, with feature information aggregated along the spatial axis using global average pooling and max pooling. Following that, the two pooled features were forwarded to a shared feed-forward network, resulting in a 1D channel attention map. It can be formulated as follows:
where M is the intermediate feature map, AP and MP represent the average and max pooling, and represent two pooled features on the spatial axis, respectively, and CAM(M) denotes the obtained channel attention map.
The channel attention map was fused with the intermediate feature map by element-wise multiplication, and the formulation is as follows:
where ⊗ represents element-wise product and M′ is the feature map with channel attention.
The SAM used M′ as the input. Similar to the operations of the CAM, M′ was aggregated in the direction of the channel axis using average-pooling and max-pooling operations, which can emphasize the location information of essential features (24). Then, the two-pooled information was concatenated and a spatial attention map in 2D was produced using a convolution operation. The formulation of this is as follows:
where and represent the two pooled features on the channel axis, and the generated spatial attention map is denoted by SAM(M′).
The SAM(M′) was fused with the obtained features of CAM to focus on essential features and locations. The formulation is as follows:
where M″ represents the final refined output.
2.3. Transformer Encoder
Since the receptive field of CNN was limited, it was difficult for convolution operations to capture the global dependency of the feature map (17); therefore, a Transformer Encoder was introduced. The Transformer achieved outstanding results in natural language processing. Its self-attention could obtain global dependency, making each word pay attention to other words at all positions in the sentence. However, in image tasks, slow convergence and high computational complexity would result if each point focused on other points in all positions. We used deformable self-attention (DSA) (25) to concentrate on a limited number of key points around each reference point to solve this problem and to improve model efficiency.
The results of the CNN Encoder served as the source information for the Transformer encoder. Since the Transformer is a sequential model, we need to convert the feature maps into a sequence. However, the process of feature serialization would cause the loss of position information; therefore, we used position embedding in the Transformer to complete the 3D position information. The following formula obtains the position embedding:
using the sine and cosine functions alternately to obtain the position embedding, where {D, H, W} represent different dimensions, EmD, EmH, and EmW form the 3D position embedding of the position P with k dimension, and . We added the position embedding to the serialized CNN encoding using the corresponding element summation method before being fed into the Transformer Encoder.
The Transformer Encoder used the multi-layer features of the CNN encoder, and the position embedding was different at the same position of each layer. Therefore, L represents the layers of the feature map, and fl denotes features in the l-th layer. Zq is the feature of the query q, and pq represents the 3D position of the reference point; thus, the DMSA can be formulated as:
where M is the number of heads of the self-attention, K represents the quantity of key sampled points, σ(•) denotes the Sigmoid function that adjusts pq to the feature of l-th level, Δpmlqk is the offset of the sampled point, and Amlqk represents the attention weight and is in the range of [0, 1].
The Transformer Encoder consisted of six stacked Deformable Transformer layers. One Deformable Transformer layer contained a DMSA layer and two normalization operations, followed by a feed-forward network. In addition, the input of the Transformer Encoder was residually added to the output to enhance the local information and to compensate for the loss caused by the transmission of information in the Transformer Encoder.
2.4. Decoder
The Decoder used the pure transposed convolution operations to restore the encoded feature maps and gradually upsampled them to D × H × W, the same as the input size. The output sequence was re-formed into the feature maps. After that, the transposed convolution operations were performed. The skip-connection between a certain level of the CNN-Encoder and the corresponding level of the Decoder added fine details to make the segmentation more accurate. The 3D residual convolutions were utilized to enhance the upsampled feature maps.
Figure 2B illustrates the Decoder's detailed structure, which contains four upsampling blocks. In the first three up-sampling blocks, the feature map was subjected to a transposed convolution operation and refined by a 3D residual convolution operation. The Decoder was skip-connected with the CNN Encoder. The last block contained an upsampling layer and a 3D convolutional with a kernel size of 1–1, which mapped the features to the number of categories for classification.
2.5. Loss function
Dice loss (26) was used to evaluate the correlation between two regions. It shows outstanding performance when the positive and negative samples are unequal in the data. CrossEntropy loss was used to find the overall average loss. We combined the Dice loss and CrossEntorpy loss (27) in our model, calculated as follows:
where Y, P, C, and N represent the infarct ground truth (GT), the infarcts segmented by the proposed module results, the number of categories, and the number of pixels, respectively.
2.6. Implementation details
During the model training phase, we randomly cropped patches with a size of 16×32×320 from NCCT scans as input images. The model underwent 350 epochs of training and 250 iterations per epoch. The learning rate was configured to 0.001, the stochastic gradient descent method was used to adjust it, and the momentum value was assigned to 0.99. According to the experiments, we set the number of 3D residual convolution layers contained in the three ResConvBlock stages to three, three, and two, the key points (K) to four, the heads (H) to six, and the number of layers in the Deformable Transformer (L) to six. The implementation and evaluation of our method were conducted on a server equipped with an NVIDIA A40 GPU.
The sliding window approach was employed for testing, and the window size was the same as the patch size in the training phase. To evaluate the effects of all compared infarct segmentation methods, we calculated the Dice score to assess the regional correlation between them; the range of possible values for the Dice score is [0, 1]. Furthermore, we calculated the F1-score, Recall, and Precision to evaluate the infarct level.
3. Results
3.1. Comparison with existed methods
We contrast the proposed method with several already existing methods, as follows: Unet (13) using a CNN architecture purely, an image-level method (Unet-IM) (28), a feature-level method (Unet-FT) (29), a method using a 3D convolutional block as the basic encoding block (HybridUnet) (30) and its implementation at the image-level, pixel-level (HybridUnet-IM, HybridUnet-FT), and a Symmetric Enhanced Attention Network (SEAN) (21). Table 1 presents the number of parameters, running time, and segmentation effectiveness of these methods. It can be observed that the Dice score and recall of our method are 58.66% and 0.6319, respectively, which are better than the existing methods. In addition, our method produces better results than the compared methods in terms of the F1-score (0.6298), although our method did not achieve the best result on precision. Since our method jointly considers the local features and global relationships of images, it could better process the detailed and overall information of the image so that the segmentation results are more similar to the GT. There was an increase in the number of parameters and running time for our method because of the Transformer encoder; however, the use of deformable self-attention rendered this increase within an acceptable range. Moreover, our method achieved the best segmentation performance. Two segmentation examples, the Dice score of the proposed method, and six compared methods for the AISD infarct segmentation task are shown in Figure 3.
Table 1. Quantitative comparison of our proposed method with seven baselines for AIS infarct segmentation, where M = 1,024 – 1,024, h represents hours, and s represents seconds.
Figure 3. Visual qualitative comparisons of the six compared methods with the proposed method for two AIS cases.
To statistically analyze the AIS infarct segmentation results, we conducted pair-wise comparisons (based on the Dice score) using the Wilcoxon rank-sum test between the proposed method and the six compared methods. When the p-value < 0.05, this suggests a statistically significant difference between the two methods being tested. Table 2 presents the results of the statistical analysis, which demonstrate that the proposed method outperformed the six compared methods in AIS infarct segmentation. The differences were statistically significant (all p-values < 0.05). Thus, the proposed method significantly improved the segmentation of AIS infarcts.
To evaluate the effectiveness and generalizability of our method on different datasets, we also compared our method with six other methods for infarct segmentation using the ISLES2018 (31, 32) training set. The ISLES2018 training dataset consists of multi-modal CT image data from 94 ischemic stroke patients, of which we only used CT and OT data. We trained and tested the method using five-fold cross-validation. As shown in Table 3, our method achieved a Dice score of 46.67%, an F1 score of 0.5242, a recall of 0.4724, and a precision of 0.5888, all of which outperformed the other methods and effectively segmented the infarcts, demonstrating the generality of our method on other datasets.
3.2. Volumetric analysis
We calculated the AIS infarct volume segmented based on the proposed method (VS) and the manually segmented infarct volume (VM). Pearson correlation calculations of VS with VM were performed and Bland-Altman plots were used to analyze the volumes. Furthermore, we calculated the volume difference between VS and VM: ΔVdiff = VS−VM and the absolute volume difference: |ΔVdiff| = |VS−VM|. Additionally, to show the clinical relevance of our results, we performed dichotomization, with 70 cc as the cut-off value in our analysis of the AIS infarct volume. We evaluated the dichotomization analysis using Accuracy, Kappa, and Specificity as metrics.
Table 4 displays the volumetric analysis results. The average volume difference between VS and VM is 2.82 cc and the absolute average volume difference is 19.86. The volume correlation between VS and VM is r=0.948 (95% confidence interval: 0.916–0.972, p < 0.001), reflecting the excellent correlation between them, see Figure 4A. The average volume difference between VS and VM is shown in the Bland-Altman diagram in Figure 4B, from which it could be seen that they have good consistency. The excellent correlation and consistency of VS and VM verify the utility of our method. Table 5 illustrates the results of the dichotomization analysis, with the cut-off set at 70 cc, for our method and six other compared methods. We can see from Table 3 that our method achieves an accuracy of 96.2% (95% confidence interval: 93.4–97.8), a Kappa of 0.88 (95% confidence interval: 0.79–0.93), and a Specificity of 0.95 (95% confidence interval: 0.92–0.97), surpassing the performance of other methods.
Figure 4. Results of volumetric analysis. (A) Pearson correlation plot of method segmentation versus manual segmentation of the infarct volume, where the abscissa and ordinate represent the infarct volume segmented manually and using the proposed method, respectively. (B) Blad-Altman plot for the two segmentation methods, where the x-axis is the average of VS and VM and the y-axis is the ΔVdiff.
Table 5. Classification performance when using 70 cc as the infarct volume cut-off (95% confidence interval is shown in parentheses).
3.3. Ablation study
In this study, CBAM was added to direct attention to the CNN Encoder and a residual connection was used on the Transformer encoder to enhance the detail features. To prove the effectiveness of these two methods, we conducted baseline experiments on the method with only CBAM added and the method with only residual structure added. As shown in Table 6, both structures enhance the segmentation efficiency of the baseline method, and the integration of these two operations can achieve better results. Two segmentation examples of each method and the Dice score in the AISD infarct segmentation task are shown in Figure 5.
4. Discussion
In the AIS infarct segmentation task of AISD, our proposed method achieved a Dice score of 58.66%, surpassing previous methods. The experimental results show that the fusion framework of CNN and Transformer effectively captured both local and global information from NCCT scans. Ablation studies highlight the contributions of CBAM and the residual structure of the Transformer Encoder improving segmentation efficiency. Volumetric analysis reveal a strong correlation between our method and manually segmented infarcts. Furthermore, our method achieves the best performance in a dichotomization analysis, with a cut-off value of 70 cc, demonstrating its accuracy in classifying infarct volumes. These results indicate the potential of our method to provide valuable infarct information for clinical diagnosis and practice.
5. Conclusion
To achieve automatic and accurate segmentation of AIS infarcts, we proposed a segmentation method in this study based on the convolutional attention mechanism and Deformable Transformer. Our method used a CNN Encoder to extract these features, augmented by the CBAM to enhance the importance of these features. Additionally, a Deformable Transformer Encoder is used to model the global dependencies and reduce complexity. We incorporated residual connections before and after the Transformer Encoder to enhance the local features. The segmentation results for AISD demonstrates the superior performance of our method, offering a novel solution for AIS infarct segmentation and improving segmentation accuracy.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found at: GitHub, https://github.com/GriffinLiang/AISD.
Ethics statement
Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent from the patients/participants or patients/participants' legal guardian/next of kin was not required to participate in this study in accordance with the national legislation and the institutional requirements.
Author contributions
ZX: methodology, formal analysis, visualization, and writing—review and editing. CD: conceptualization, review and editing, and funding acquisition. Both authors contributed to the article and approved the submitted version.
Funding
This work was supported in part by the Project of Changsha Science and Technology Bureau (No. kq2202265), the Key Scientific Research Project of Hunan Provincial Department of Education (No. 22A0255), and the Postgraduate Scientific Research Innovation Project of Hunan University of Chinese Medicine (No. 2022CX119).
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Zhang Q, Zhao YH. Therapeutic angiogenesis after ischemic stroke: Chinese medicines, bone marrow stromal cells (BMSCs) and their combinational treatment. Am J Chin Med. (2014) 42:61–77. doi: 10.1142/S0192415X14500049
2. Potter CA, Vagal AS, Goyal M, Nunez DB, Leslie-Mazwi TM, Lev MH. CT for treatment selection in acute ischemic stroke: a code stroke primer. Radiographics. (2019) 39:1717–38. doi: 10.1148/rg.2019190142
3. Kuang H, Menon BK, Sohn SI, Qiu W. EIS-Net: segmenting early infarct and scoring ASPECTS simultaneously on non-contrast CT of patients with acute ischemic stroke. Med Image Anal. (2021) 70:101984. doi: 10.1016/j.media.2021.101984
4. Kuang H, Najm M, Chakraborty D, Maraj N, Sohn S, Goyal M, et al. Automated ASPECTS on noncontrast CT scans in patients with acute ischemic stroke using machine learning. Am J Neuroradiol. (2019) 40:33–8. doi: 10.3174/ajnr.A5889
5. Xie Y, Zhang J, Shen C, Xia Y. COTR: efficiently bridging cnn and transformer for 3D medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part III 24. Berlin: Springer (2021). p. 171–180.
6. Karthik R, Menaka R, Johnson A, Anand S. Neuroimaging and deep learning for brain stroke detection—a review of recent advancements and future prospects. Comput Methods Programs Biomed. (2020) 197:105728. doi: 10.1016/j.cmpb.2020.105728
7. Yahiaoui AFZ, Bessaid A. Segmentation of ischemic stroke area from CT brain images. In: 2016 International Symposium on Signal, Image, Video and Communications (ISIVC). Tunis (2016). p. 13–7.
8. Subudhi A, Dash M, Sabut S. Automated segmentation and classification of brain stroke using expectation-maximization and random forest classifier. Biocybernet Biomed Eng. (2020) 40:277–89. doi: 10.1016/j.bbe.2019.04.004
9. Kuang H, Menon BK, Qiu W. Segmenting hemorrhagic and ischemic infarct simultaneously from follow-up non-contrast CT images in patients with acute ischemic stroke. IEEE Access. (2019) 7:39842–51. doi: 10.1109/ACCESS.2019.2906605
10. Liu L, Wang YP, Wang Y, Zhang P, Xiong S. An enhanced multi-modal brain graph network for classifying neuropsychiatric disorders. Med Image Anal. (2022) 81:102550. doi: 10.1016/j.media.2022.102550
11. Chen L, Bentley P, Rueckert D. Fully automatic acute ischemic lesion segmentation in DWI using convolutional neural networks. NeuroImage. (2017) 15:633–43. doi: 10.1016/j.nicl.2017.06.016
12. Zhang R, Zhao L, Lou W, Abrigo JM, Mok VC, Chu WC, et al. Automatic segmentation of acute ischemic stroke from DWI using 3-D fully convolutional DenseNets. IEEE Trans Med Imaging. (2018) 37:2149–60. doi: 10.1109/TMI.2018.2821244
13. Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. In: Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015, Proceedings, Part III 18. Berlin: Springer (2015). p. 234–41.
14. Zhou Z, Rahman Siddiquee MM, Tajbakhsh N, Liang J. U-net++: a nested u-net architecture for medical image segmentation. In: Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, September 20, 2018, Proceedings 4. Berlin: Springer (2018). p. 3–11.
15. Xiao X, Lian S, Luo Z, Li S. Weighted res-Unet for high-quality retina vessel segmentation. In: 2018 9th International Conference on Information Technology in Medicine and Education (ITME). Hangzhou: IEEE (2018). p. 327–31.
16. Ni H, Xue Y, Wong K, Volpi J, Wong ST, Wang JZ, et al. Asymmetry disentanglement network for interpretable acute ischemic stroke infarct segmentation in non-contrast CT scans. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2022: 25th International Conference, Singapore, September 18–22, 2022, Proceedings, Part VIII. Berlin: Springer (2022). p. 416–26.
17. Luo W, Li Y, Urtasun R, Zemel R. Understanding the effective receptive field in deep convolutional neural networks. Adv Neural Informat Process Syst. (2016) 29:4898–906. doi: 10.48550/arXiv.1701.04128
18. Dosovitskiy A, Beyer L, Kolesnikov A, Weissenborn D, Zhai X, Unterthiner T, et al. An image is worth 16x16 words: transformers for image recognition at scale. arXiv preprint arXiv:201011929. (2020). doi: 10.48550/arXiv.2010.11929
19. Zheng S, Lu J, Zhao H, Zhu X, Luo Z, Wang Y, et al. Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Nashville, TN (2021). p. 6881–90.
20. Chen J, Lu Y, Yu Q, Luo X, Adeli E, Wang Y, et al. TransUnet: transformers make strong encoders for medical image segmentation. arXiv preprint arXiv:210204306. (2021). doi: 10.48550/arXiv.2102.04306
21. Liang K, Han K, Li X, Cheng X, Li Y, Wang Y, et al. Symmetry-enhanced attention network for acute ischemic infarct segmentation with non-contrast CT images. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, September 27–October 1, 2021, Proceedings, Part VII 24. Berlin: Springer (2021). p. 432–41.
22. Ulyanov D, Vedaldi A, Lempitsky V. Instance normalization: the missing ingredient for fast stylization. arXiv preprint arXiv:160708022. (2016). doi: 10.48550/arXiv.1607.08022
23. Woo S, Park J, Lee JY, Kweon IS. CBAM: Convolutional block attention module. In: Proceedings of the European Conference on Computer Vision (ECCV). Munich (2018). p. 3–19.
24. Zagoruyko S, Komodakis N. Paying more attention to attention: improving the performance of convolutional neural networks via attention transfer. arXiv preprint arXiv:161203928. (2016). doi: 10.48550/arXiv.1612.03928
25. Zhu X, Su W, Lu L, Li B, Wang X, Dai J. Deformable detr: deformable transformers for end-to-end object detection. arXiv preprint arXiv:201004159. (2020). doi: 10.48550/arXiv.2010.04159
26. Milletari F, Navab N, Ahmadi SA. V-Net: fully convolutional neural networks for volumetric medical image segmentation. In: 2016 Fourth International Conference on 3D Vision (3DV). Stanford, CA: IEEE (2016). p. 565–71.
27. Zhang J, Xie Y, Zhang P, Chen H, Xia Y, Shen C. Light-weight hybrid convolutional network for liver tumor segmentation. IJCAI. (2019) 19:4271–7. doi: 10.24963/ijcai.2019/593
28. Qiu W, Kuang H, Teleg E, Ospel JM, Sohn SI, Almekhlafi M, et al. Machine learning for detecting early infarction in acute stroke with non–contrast-enhanced CT. Radiology. (2020) 294:638–44. doi: 10.1148/radiol.2020191193
29. Zhang H, Zhu X, Willke TL. Segmenting brain tumors with symmetry. arXiv preprint arXiv:171106636. (2017). doi: 10.48550/arXiv.1711.06636
30. Fang C, Li G, Pan C, Li Y, Yu Y. Globally guided progressive fusion network for 3D pancreas segmentation. In: Medical Image Computing and Computer Assisted Intervention–MICCAI 2019: 22nd International Conference, Shenzhen, China, October 13–17, 2019, Proceedings, Part II 22. Berlin: Springer (2019). p. 210–8.
31. Maier O, Menze BH, Von der Gablentz J, Häni L, Heinrich MP, Liebrand M, et al. ISLES 2015—a public evaluation benchmark for ischemic stroke lesion segmentation from multispectral MRI. Med Image Anal. (2017) 35:250–69. doi: 10.1016/j.media.2016.07.009
Keywords: Transformer, convolutional attention mechanism, infarct segmentation, volumetric analysis, acute ischemic stroke, non-contrast CT
Citation: Xu Z and Ding C (2023) Combining convolutional attention mechanism and residual deformable Transformer for infarct segmentation from CT scans of acute ischemic stroke patients. Front. Neurol. 14:1178637. doi: 10.3389/fneur.2023.1178637
Received: 10 March 2023; Accepted: 19 June 2023;
Published: 20 July 2023.
Edited by:
John R. Absher, Prisma Health Neuroscience Associates, United StatesReviewed by:
Hulin Kuang, Central South University, ChinaYuan Xue, Johns Hopkins University, United States
Copyright © 2023 Xu and Ding. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Changsong Ding, ZGluZ2NzMTk3NSYjeDAwMDQwO2hudWNtLmVkdS5jbg==