Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 14 May 2024
Sec. Visual Neuroscience

Application and visualization study of an intelligence-assisted classification model for common eye diseases using B-mode ultrasound images

  • 1School of Information Engineering, Huzhou University, Huzhou, China
  • 2Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, China
  • 3Department of Ophthalmology, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China

Aim: Conventional approaches to diagnosing common eye diseases using B-mode ultrasonography are labor-intensive and time-consuming, must requiring expert intervention for accuracy. This study aims to address these challenges by proposing an intelligence-assisted analysis five-classification model for diagnosing common eye diseases using B-mode ultrasound images.

Methods: This research utilizes 2064 B-mode ultrasound images of the eye to train a novel model integrating artificial intelligence technology.

Results: The ConvNeXt-L model achieved outstanding performance with an accuracy rate of 84.3% and a Kappa value of 80.3%. Across five classifications (no obvious abnormality, vitreous opacity, posterior vitreous detachment, retinal detachment, and choroidal detachment), the model demonstrated sensitivity values of 93.2%, 67.6%, 86.1%, 89.4%, and 81.4%, respectively, and specificity values ranging from 94.6% to 98.1%. F1 scores ranged from 71% to 92%, while AUC values ranged from 89.7% to 97.8%.

Conclusion: Among various models compared, the ConvNeXt-L model exhibited superior performance. It effectively categorizes and visualizes pathological changes, providing essential assisted information for ophthalmologists and enhancing diagnostic accuracy and efficiency.

1 Introduction

Eye B-mode ultrasonography is a prevalent medical classification approach employed to evaluate eye structures and identify eye diseases. This technique is particularly valuable for examining intraocular diseases, especially when obstructions like refractive media opacity or lesions situated within the orbit are present (Bates and Goett, 2022). Leveraging ultrasound technology, it produces real-time images of internal eye structures and holds extensive clinical utility. Diseases such as vitreous opacities, posterior vitreous detachment (PVD), retinal detachment, and choroidal detachment represent common eye ailments (Bellows et al., 1981; Ghazi and Green, 2002; Shinar et al., 2011; Gishti et al., 2019; Ryan, 2021), often discernible through eye B-mode ultrasound images.

Eye B-mode ultrasound image classification traditionally relies on the expertise of ophthalmologists. However, an uneven distribution of medical professionals across different regions in China presents a challenge (Jiao and Wang, 2022). The majority of ophthalmologists are concentrated in economically developed eastern coastal areas. Consequently, eye B-mode ultrasonography is often conducted by medical imaging specialists who may lack the specialized knowledge required for accurate interpretation. As a result, misdiagnoses occur due to the absence of proficiency in ocular ultrasonography interpretation. This underscores the current shortage of skilled professionals capable of effectively analyzing B-mode ultrasound images, particularly in numerous grassroots hospitals, leading to an inability to fulfill the demand for precise image analysis.

Recently, the integration of artificial intelligence (AI) into the medical domain has emerged as a prominent research area. Notably, machine learning and deep learning techniques have demonstrated substantial efficacy in medical diagnostics. Machine learning, characterized by the use of manually selected features and algorithms for disease identification, has yielded satisfactory outcomes (Rajan and Ramesh, 2015; Chowdhury and Banerjee, 2016; Hosoda et al., 2020; Xu et al., 2020; Kooner et al., 2022; Yang et al., 2023). On the other hand, deep learning harnesses convolutional neural networks to automatically extract features from images (Mirzania et al., 2021; Ho et al., 2022; Xie et al., 2022; Chen et al., 2023; Li et al., 2023). In a study by Nagasato et al., a comparison between machine learning and deep learning in branch retinal vein occlusion (RVO) detection using ultra-widefield retinal images revealed that deep learning exhibited superior sensitivity and specificity (Nagasato et al., 2019). This trend has consequently spurred numerous researchers to leverage deep learning methodologies for the diagnosis of eye diseases.

For instance, Zhu et al. harnessed machine learning to predict changes in spherical equivalent refraction (SER) and axial length (AL) in children, yielding impressive R2 values of 89.97% for SER and 75.46% for AL prediction (Zhu et al., 2023). Employing a deep learning model, Li et al. successfully discriminated between normal fundus images and eight prevalent eye diseases, attaining an exceptional area under the receiver operating characteristic curve (AUC) of 97.84% (Li et al., 2020). In their work, He et al. proposed a binary classification and segmentation model tailored for pterygium diagnosis, achieving a remarkable classification accuracy of 99% (Zhu et al., 2022). Meanwhile, Chen et al. devised a deep learning system for swift screening of rhegmatogenous retinal detachment (RRD), vitreous detachment (VD), and vitreous hemorrhage (VH), culminating in accuracies of 94, 90, 92, 94, and 0.91% for normal, VD, VH, RD, and other lesions, respectively (Chen et al., 2021). Notably, Zheng et al. crafted a common retinal disease classification model grounded in ResNet50 architecture using a dataset of 2000 fundus images, effectively diagnosing various eye diseases except macular degeneration (MD) (Zheng et al., 2021). Moreover, Google’s team, led by Gulshan et al., skillfully trained a deep learning model to diagnose diabetic retinopathy (DR) and assign a severity grade through fundus images. Their efforts culminated in compelling outcomes validated through clinical trials (Ahn et al., 2018).

To tackle the challenge of intelligent assistance in the analysis of eye disease B-mode ultrasound images, this study leverages the ConvNeXt-L model (Liu et al., 2022) to train a intelligence-assisted five-classification framework for common eye diseases. This is the first application of the ConvNeXt model in Eye B-mode ultrasound image classification for eye diseases. This framework aids non-specialist ophthalmologists in conducting preliminary patient diagnoses, facilitating the identification of no significant abnormalities and four common eye diseases (vitreous opacities, posterior vitreous detachment, retinal detachment, choroidal detachment), while also conducting visualization analysis. This innovative approach seeks to bridge the gap between the substantial patient volume encountered in grassroots hospitals and the limited accessibility of specialized ophthalmologists. By pursuing this approach, the objective is to deliver efficacious services to individuals with eye diseases. The subsequent findings are delineated as follows.

2 Materials and methods

2.1 Data source

This study encompassed experiments conducted on a dataset of common eye disease B-mode ultrasound images, generously contributed by collaborative hospitals. The research conformed to the ethical principles outlined in the Helsinki Declaration. To ensure data providers’privacy, the study did not impose any restrictions on patients’age or gender within the images. A rigorous anonymization process was undertaken, eliminating all personally identifiable information to uphold patient confidentiality. Consequently, no patient demographics are available. The research incorporated a total of 2064 images, with 1860 allocated for the training set and 204 for the test set, as detailed in Table 1.

Table 1
www.frontiersin.org

Table 1. Eye disease B-mode ultrasound image dataset.

2.2 ConvNeXt model

This investigation utilized the ConvNeXt-L model, which was pre-trained on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) dataset. Leveraging a dataset comprising 1860 images, encompassing “No Abnormalities” images alongside four common eye disease B-mode ultrasound images, a five-class intelligence-assisted diagnostic model for prevalent eye diseases was meticulously trained. This adeptly trained model exhibits proficiency in classifying both “No Abnormalities” images and the four prevalent eye diseases. Furthermore, it employs visualization techniques to unveil focal regions within the images, contributing to a comprehensive diagnostic process.

This study incorporated additional well-established deep learning classification models, namely ResNet50 (Targ et al., 2016) and EfficientNet-B4 (Tan and Le, 2019). These models principally comprise convolutional layers, pooling layers, and fully connected layers. The ConvNeXt model, a refined iteration of the ResNet-50 model, is available in distinct versions like ConvNeXt-T, ConvNeXt-S, ConvNeXt-B, ConvNeXt-L, and ConvNeXt-XL. Notably, these models differ primarily in their network structure’s depth and width.

The foundational structure of the ConvNeXt model encompasses convolutional layers, activation layers, and fully connected layers. The architecture of the ConvNeXt-L model integrates a 4×4 convolutional layer alongside 36 ConvNeXtBlocks of varying sizes. Each ConvNeXtBlock comprises three convolutional layers of different dimensions, along with LayerNorm, GELU activation, LayerScale, and DropPath components. For ConvNeXt-L, input images undergo preprocessing and resizing to 192×192 dimensions, derived from eye B-mode ultrasound images, culminating in the final five-class intelligent diagnostic model. Figure 1 visually outlines the architecture of the ConvNeXt-L model.

Figure 1
www.frontiersin.org

Figure 1. The structure of the ConvNeXt model.

To enhance the precision of eye disease classification in B-mode ultrasonography, this study conducted a comprehensive model comparison. The examined models encompassed ResNet50, EfficientNet-B4, DeiT3 (Touvron et al., 2022), and Swin Transformer V2 (Liu et al., 2022). Notably, ResNet50 and EfficientNet-B4 belong to the domain of deep convolutional neural networks, while DeiT3 and Swin Transformer V2 adopt attention-based neural network architectures. These four models were subjected to experimentation using the eye B-mode ultrasound image dataset. Subsequently, the most superior-performing model was meticulously identified for the accurate classification of eye B-mode ultrasonography diseases.

2.3 Statistical analysis

Statistical analysis was performed using SPSS 27.0 software. Firstly, count data was presented in the form of both image numbers and percentages. Subsequently, common metrics for diagnosing eye diseases, such as sensitivity, specificity, and F1 score, were calculated for both normal fundus images and images associated with four common diseases. Following this, the receiver operating characteristic (ROC) curve was plotted to evaluate the model’s performance. Additionally, we employed the kappa statistic to assess the agreement between the diagnoses made by expert clinicians and the model’s diagnostic results. In this context, the results from the expert clinician group were taken as the ground truth, and the kappa statistic was utilized to quantify the level of agreement. Typically, a kappa value between 0.61 and 0.80 indicates substantial agreement, while a kappa value exceeding 0.80 suggests high agreement. Through the aforementioned analyses, conclusions regarding the performance and consistency of the eye disease diagnostic model can be drawn.

3 Results

3.1 Expert classification results

The study employed 1860 eye B-mode ultrasound images for training and 204 images for testing. The image collection format remained consistent throughout. Among the entire set of eye B-mode ultrasound images, expert clinicians identified 451 as normal, 327 as vitreous opacity, 385 as posterior vitreous detachment, 474 as retinal detachment, and 427 as choroidal detachment.

3.2 Intelligent-assisted classification results

The confusion matrices for the diagnostic results of the ResNet50, EfficientNet-B4, DeiT3, Swin Transformer V2, and ConvNeXt-L models can be found in Tables 26.

Table 2
www.frontiersin.org

Table 2. ResNet50 model classification results.

Table 3
www.frontiersin.org

Table 3. EfficientNet-B4 model classification results.

Table 4
www.frontiersin.org

Table 4. DeiT3 model classification results.

Table 5
www.frontiersin.org

Table 5. Swin transformer V2 model classification results.

Table 6
www.frontiersin.org

Table 6. ConvNeXt-L model classification results.

3.3 Evaluation metric results

The evaluation metric results for the five models: ResNet50, EfficientNet-B4, DeiT3, Swin Transformer V2, and ConvNeXt-L, are presented in Table 7.

Table 7
www.frontiersin.org

Table 7. Evaluation metric results for the five models.

3.4 Visualization study

In this research, Grad-CAM (Selvaraju et al., 2017) was harnessed to produce visualizations for the ResNet50, EfficientNet-B4, DeiT3, Swin Transformer V2, and ConvNeXt-L models, as showcased in Figure 2. An observation from the images reveals that ConvNeXt-L’s visualization showcases the most accurate annotated regions. In comparison, the annotated regions in the visualizations for the remaining four models exhibit a slightly lower precision. It’s noteworthy that these visualizations for the five models are generated through the same algorithm. Consequently, as the overall performance of the models advances, the accuracy of the annotated regions in the visualizations proportionally improves.

Figure 2
www.frontiersin.org

Figure 2. B-mode ultrasound images and four common eye diseases, along with visualization results.

4 Discussion

Eye ultrasonography stands as a pivotal imaging technique for evaluating eye and adjacent tissue structures (Elabjer et al., 2017). In the realm of ophthalmology, it holds a prominent status as a diagnostic imaging approach, offering a secure, non-invasive avenue with instantaneous, real-time feedback. Its integration into ophthalmic evaluation traces back to the 1950s, spurred by the accessibility of eye anatomy (Bangal et al., 2016). Notably, the eye’s cystic nature renders it an ideal candidate for ultrasound examination facilitated by high-frequency sound waves (Uduma et al., 2019). The merits of eye ultrasound examination lie in its attributes of non-invasiveness, cost-effectiveness, real-time functionality, and the potential for repetition in diagnosing eye pathologies. As a consequence, eye ultrasound examination has evolved into an indispensable tool, effectively serving the diagnosis and treatment of a spectrum of ophthalmic diseases.

As scientific and technological progress continues, the landscape of artificial intelligence (AI) is undergoing rapid transformation, venturing into uncharted territories. An illustrative example is the landmark introduction of AlexNet by Krizhevsky et al. (2012) and colleagues, which remarkably outperformed competitors in the ImageNet competition. This breakthrough ignited widespread fascination with neural networks, laying the groundwork for tackling image-related challenges using deep neural networks. This catalyst subsequently led to a proliferation of exceptional accomplishments in the realm. In 2015, the introduction of ResNet (Simonyan and Zisserman, 2014), pioneered by Kaiming He and team, secured a triumph in the ImageNet competition, pushing image classification error rates to a remarkable low of 3.6%, even surpassing human recognition capacities.

Further fueling progress, Google’s application of the Transformer to image classification tasks in 2020 birthed the revolutionary Vision Transformer (ViT) model (Dosovitskiy et al., 2020). Another milestone occurred in 2022, with the unveiling of the ConvNeXt model by Liu et al. (2022). This innovation drew inspiration from the Swin Transformer architecture, achieving advancements by refining ResNet50. The ConvNeXt-L variant notably achieved heightened performance, boasting an accuracy of 87.8% on ImgNet 22 k. The pervasive influence of AI extends across diverse domains, with the medical sphere being no exception. Noteworthy instances include Zheng et al. (2021), who harnessed 2000 fundus images to devise a model employing ResNet50 for classifying common retinal diseases, and Zhu et al. (2022), who effectively utilized the EfficientNet-B4 model for diagnosing a range of retinal diseases, yielding promising results.

This study employed five models for experimentation: ResNet50, EfficientNet-B4, DeiT3, Swin Transformer V2, and ConvNeXt-L. Among them, the ConvNeXt-L model exhibited the best classification performance. Several factors could contribute to its superior performance over the other four models. Firstly, the ConvNeXt-L model ingeniously tweaks the ResNet50 architecture, infusing it with design principles inspired by Transformers while retaining convolutional layers instead of attention modules. This strategic fusion enhances the ConvNeXt-L model’s feature extraction capabilities, amplifying its strength in classification. Secondly, the ConvNeXt-L model underwent training on the ImgNet 22 k dataset, leveraging a pretrained model that was subsequently fine-tuned in this investigation. The model’s exposure to a diverse array of images during its pretraining phase ostensibly influenced its performance boost. The ConvNeXt model encompasses an assortment of five versions, namely ConvNeXt-T, ConvNeXt-S, ConvNeXt-B, ConvNeXt-L, and ConvNeXt-XL. The elemental divergence across these versions pertains to the depth and breadth of their network structures. Importantly, ConvNeXt-L distinguishes itself with its enhanced feature extraction prowess and a relatively streamlined parameter count, culminating in expedited execution speeds compared to the remaining four variants.

Using the Grad-CAM method, this study produced visualizations. From Figure 2, it can be observed that the ConvNeXt-T model exhibited the most optimal performance, with its annotated regions closely aligning with those identified by ophthalmologists. Among the remaining four models, apart from the DeiT3 model, which exhibited a substantial disparity between its visualized annotations and the ophthalmologists’ diagnostic regions, the performances were relatively comparable to the ConvNeXt-L model’s results. DeiT3’s classification performance was not poor; however, the poor visualized annotations might be attributed to a misalignment with the Grad-CAM methodology.

In Table 7, the models demonstrated relatively higher sensitivity and specificity scores for the diseases of Normal, Retinal Detachment, and Choroidal Detachment. However, for the diseases of Vitreous Opacity and Posterior Vitreous Detachment, the sensitivity and specificity scores were notably lower. This discrepancy arises from the models difficulty in distinguishing between Vitreous Opacity and Posterior Vitreous Detachment, likely due to the striking similarity of pathological regions in these two diseases on eye B-mode ultrasound images. The low sensitivity and specificity scores for these two diseases can be attributed to their confusion by the model, which could potentially be mitigated by an increased dataset size.

Tables 16 collectively demonstrate that these five models are capable of diagnosing Normal and the other four common eye diseases. However, errors are more prone to occur in the classification of the four common eye diseases, particularly Posterior Vitreous Detachment versus Vitreous Opacity, and Retinal Detachment versus Choroidal Detachment. Therefore, when the model diagnoses any of the four diseases apart from Normal, it is recommended for physicians to reconsider the diagnosis or for patients to seek further evaluation at a higher-level medical institution to prevent misdiagnosis or missed diagnosis.

Although, this research does have some limitations. Primarily, the model faces challenges in diagnosing detachment of the vitreous from the retina and vitreous opacities, as well as the immature application of transfer learning on medical imaging datasets. Additionally, there is still room for improvement in the model’s interpretability. The model frequently confuses Posterior Vitreous Detachment with Vitreous Opacity, and Choroidal Detachment with Retinal Detachment. Thus, the five-classification model based on ConvNeXt-L is only suitable for preliminary diagnosis of common eye diseases. This issue might stem from the fact that a single eye B-mode ultrasound image can have multiple labels, while the model provides only the most severe diagnosis, leading to incorrect results. The accuracy of the model in diagnosing Normal and the four common eye diseases ranges from 78 to 84%, indicating that the accuracy is not exceptionally high. The slightly lower accuracy is mainly due to the limited dataset and the presence of multiple labels per image. To elevate accuracy, specificity, and sensitivity, strategies like data augmentation, employing Generative Adversarial Networks (GANs) for image generation, and expanding the dataset’s scope could be harnessed. Due to the incomplete maturity of transfer learning in medical imaging datasets, we will further explore its potential on such datasets to enhance the classification accuracy of eye diseases. Meanwhile, considering the limitations of model interpretability, we plan to adopt ensemble methods to delve into the crucial features in the model decision-making process, aiming to further improve model interpretability.

5 Conclusion

In conclusion, this study proposes an intelligence-assisted five-classification model for common eye diseases B-mode ultrasound images, focusing on four common eye diseases. The model aims to intelligently analyze and visualize images of these four common eye diseases. This study employed the ConvNeXt-L model to design an intelligence-assisted five-classification model for common eye diseases in eye B-mode ultrasound images. This model contributes to the enhancement of classification accuracy and timeliness for common eye diseases in underdeveloped regions. In the future, this model will be refined to provide more comprehensive assisted classification results for patients with eye diseases, thus covering an even broader range of eye diseases and adaptable to various scenarios.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

SZ: Writing – original draft, Writing – review & editing. XL: Writing – review & editing, Writing – original draft. YL: Writing – review & editing. BZ: Writing – review & editing. MW: Writing – review & editing. XY: Writing – review & editing, Conceptualization, Data curation, Methodology. WY: Writing – review & editing, Conceptualization, Funding acquisition, Methodology. YG: Conceptualization, Funding acquisition, Methodology, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This study was supported by the National Natural Science Foundation of China (No. 61906066), Scientific Research Fund of Zhejiang Provincial Education Department (No. Y202250188 and No. Y202351126), the Key Science and Technology Project of Huzhou (No. 2023GG11), Shenzhen Fund for Guangdong Provincial High-level Clinical Key Specialties (SZGSP014) and Sanming Project of Medicine in Shenzhen (SZSM202311012).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Ahn, J. M., Kim, S., Ahn, K. S., Cho, S. H., Lee, K. B., and Kim, U. S. (2018). A deep learning model for the detection of both advanced and early glaucoma using fundus photography. PLoS One 13, 1–8. doi: 10.1371/journal.pone.0207982

Crossref Full Text | Google Scholar

Bangal, S. V., Bhandari, A. J., and Siddhiqui, F. (2016). Pattern of ocular pathologies diagnosed with B-scan ultrasonography in a hospital in rural India. Niger. J. Ophthalmol. 24, 71–75. doi: 10.4103/0189-9171.195199

Crossref Full Text | Google Scholar

Bates, A., and Goett, H. J. (2022). “Ocular ultrasound” in StatPearls. ed. H. J. Goett (Treasure Island (FL): StatPearls Publishing).

Google Scholar

Bellows, A. R., Chylack, L. T. Jr., and Hutchinson, B. T. (1981). Choroidal detachment: clinical manifestation, therapy and mechanism of formation. Ophthalmology 88, 1107–1115. doi: 10.1016/S0161-6420(81)34897-0

PubMed Abstract | Crossref Full Text | Google Scholar

Chen, D., Anran, E., Fang Tan, T., Ramachandran, R., Li, F., Cheung, C., et al. (2023). Applications of artificial intelligence and deep learning in glaucoma. Asia Pac. J. Ophthalmol. 12, 80–93. doi: 10.1097/APO.0000000000000596

Crossref Full Text | Google Scholar

Chen, D., Yu, Y., Zhou, Y., Peng, B., Wang, Y., Hu, S., et al. (2021). A deep learning model for screening multiple abnormal findings in ophthalmic ultrasonography (with video). Transl. Vis. Sci. Technol. 10:22. doi: 10.1167/tvst.10.4.22

PubMed Abstract | Crossref Full Text | Google Scholar

Chowdhury, A. R., and Banerjee, S. (2016). Detection of abnormalities of retina due to diabetic retinopathy and age related macular degeneration using SVM. Sci. J. Circ. Syst. Signal Process. 5, 1–7. doi: 10.11648/j.cssp.20160501.11

Crossref Full Text | Google Scholar

Dosovitskiy, A., Beyer, L., Kolesnikov, A., et al. (2020). An image is worth 16x16 words: transformers for image recognition at scale. arXiv 2:11929. doi: 10.48550/arXiv.2010.11929

Crossref Full Text | Google Scholar

Elabjer, B. K., Bušić, M., Tvrdi, A. B., et al. (2017). Ultrasound reliability in detection of retinal tear in acute symptomatic posterior vitreous detachment with vitreous hemorrhage. Int. J. Ophthalmol. 10, 1922–1924. doi: 10.18240/ijo.2017.12.21

Crossref Full Text | Google Scholar

Ghazi, N. G., and Green, W. R. (2002). Pathology and pathogenesis of retinal detachment. Eye 16, 411–421. doi: 10.1038/sj.eye.6700197

PubMed Abstract | Crossref Full Text | Google Scholar

Gishti, O., van den Nieuwenhof, R., Verhoekx, J., and van Overdam, K. (2019). Symptoms related to posterior vitreous detachment and the risk of developing retinal tears: a systematic review. Acta Ophthalmol. 97, 347–352. doi: 10.1111/aos.14012

PubMed Abstract | Crossref Full Text | Google Scholar

Ho, E., Wang, E., Youn, S., Sivajohan, A., Lane, K., Chun, J., et al. (2022). Deep ensemble learning for retinal image classification. Transl. Vis. Sci. Technol. 11:39. doi: 10.1167/tvst.11.10.39

PubMed Abstract | Crossref Full Text | Google Scholar

Hosoda, Y., Miyake, M., Yamashiro, K., Ooto, S., Takahashi, A., Oishi, A., et al. (2020). Deep phenotype unsupervised machine learning revealed the significance of pachychoroid features in etiology and visual prognosis of age-related macular degeneration. Sci. Rep. 10, 1–13. doi: 10.1038/s41598-020-75451-53

Crossref Full Text | Google Scholar

Jiao, S. Q., and Wang, C. G. (2022). Research on the equity of rural medical service supply under the background of medical and health community. J. Soc. Sci 2022, 156–243. doi: 10.13858/j.cnki.cn32-1312/c.20220325.018

Crossref Full Text | Google Scholar

Kooner, K. S., Angirekula, A., Treacher, A. H., Al-Humimat, G., Marzban, M. F., Chen, A., et al. (2022). Glaucoma diagnosis through the integration of optical coherence tomography/angiography and machine learning diagnostic models. Clin. Ophthalmol. 16, 2685–2697. doi: 10.2147/OPTH.S367722

Crossref Full Text | Google Scholar

Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Proces. Syst. 60, 84–90. doi: 10.1145/3065386

Crossref Full Text | Google Scholar

Li, Z., Guo, C., Nie, D., Lin, D., Zhu, Y., Chen, C., et al. (2020). Deep learning for detecting retinal detachment and discerning macular status using ultra-widefield fundus images. Commun. Biol. 3:15. doi: 10.1038/s42003-019-0730-x

Crossref Full Text | Google Scholar

Li, Z., Yang, J., Wang, X., and Zhou, S. (2023). Establishment and evaluation of intelligent diagnostic model for ophthalmic ultrasound images based on deep learning. Ultrasound Med. Biol. 49, 1760–1767. doi: 10.1016/j.ultrasmedbio.2023.03.022

PubMed Abstract | Crossref Full Text | Google Scholar

Liu, Z., Hu, H., Lin, Y., et al. Swin transformer v2: scaling up capacity and resolution. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. (2022). IEEE. New Orleans, LA.

Google Scholar

Liu, Z., Mao, H., Wu, C. Y., et al. (2022). A convnet for the 2020s. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. IEEE. New Orleans, LA.

Google Scholar

Mirzania, D., Thompson, A. C., and Muir, K. W. (2021). Applications of deep learning in detection of glaucoma: a systematic review. Eur. J. Ophthalmol. 31, 1618–1642. doi: 10.1177/1120672120977346

PubMed Abstract | Crossref Full Text | Google Scholar

Nagasato, D., Tabuchi, H., Ohsugi, H., Masumoto, H., Enno, H., Ishitobi, N., et al. (2019). Deep-learning classifier with ultrawide-field fundus ophthalmoscopy for detecting branch retinal vein occlusion. Int. J. Ophthalmol. 12, 94–99. doi: 10.18240/ijo.2019.01.15

PubMed Abstract | Crossref Full Text | Google Scholar

Rajan, A., and Ramesh, G. P. (2015). Automated early detection of glaucoma in waveletdomain using optical coherence tomography images. Biosci. Biotechnol. Res. Asia 12, 2821–2828. doi: 10.13005/bbra/1966

Crossref Full Text | Google Scholar

Ryan, E. H. (2021). Current treatment strategies for symptomatic vitreous opacities. Curr. Opin. Ophthalmol. 32, 198–202. doi: 10.1097/ICU.0000000000000752

PubMed Abstract | Crossref Full Text | Google Scholar

Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D.. Grad-cam:visual explanations from deep networks via gradient-based localization. In:Proceedings of the IEEE international conference on computer vision. Venice.(2017). 618–626. IEEE. Venice, Italy.

Google Scholar

Shinar, Z., Chan, L., and Orlinsky, M. (2011). Use of ocular ultrasound for the evaluation of retinal detachment. J. Emerg. Med. 40, 53–57. doi: 10.1016/j.jemermed.2009.06.001

PubMed Abstract | Crossref Full Text | Google Scholar

Simonyan, K., and Zisserman, A. (2014). Very deep convolutional networks for large-scale image recognition. Comput. Vis. Pattern Recog. 6:1556. doi: 10.48550/arXiv.1409.1556

Crossref Full Text | Google Scholar

Tan, M., and Le, Q.. (2019). EfficientNet: rethinking model scaling for convolutional neural networks. Proceedings of the 36th international conference on machine learning, ICML, Long Beach, CA.

Google Scholar

Targ, S., Almeida, D., and Lyman, K.. (2016). Resnet in resnet: generalizing residual architectures. Available at: https://arxiv.org/abs/1603.08029

Google Scholar

Touvron, H., Cord, M., and Jégou, H.. (2022). Deit iii: Revenge of the vit. In Computer Vision-ECCV. 17th European conference. Cham: Springer Nature Switzerland, 516–533.

Google Scholar

Uduma, F. U., Iquo-Abasi, A. S., and Rosemary, N. N. (2019). B-scan ophthalmic ultrasonography: a review corroborated with echograms. New Front. Ophthalmol. 5, 1–5. doi: 10.15761/NFO.1000233

Crossref Full Text | Google Scholar

Xie, X., Yang, L., Zhao, F., Wang, D., Zhang, H., He, X., et al. (2022). A deep learning model combining multimodal radiomics, clinical and imaging features for differentiating ocular adnexal lymphoma from idiopathic orbital inflammation. Eur. Radiol. 32, 6922–6932. doi: 10.1007/s00330-022-08857-6

PubMed Abstract | Crossref Full Text | Google Scholar

Xu, J., Yang, W., Wan, C., and Shen, J. (2020). Weakly supervised detection of central serous chorioretinopathy based on local binary patterns and discrete wavelet transform. Comput. Biol. Med. 127:104056. doi: 10.1016/j.compbiomed.2020.104056

PubMed Abstract | Crossref Full Text | Google Scholar

Yang, W. H., Shao, Y., Xu, Y. W., et al. (2023). Guidelines on clinical research evaluation of artificial intelligence in ophthalmology. Int. J. Ophthalmol. 16, 1361–1372. doi: 10.18240/ijo.2023.09.02

Crossref Full Text | Google Scholar

Zheng, B., Jiang, Q., Lu, B., He, K., Wu, M. N., Hao, X. L., et al. (2021). FiveCategory intelligent auxiliary diagnosis model of common fundus diseases based on fundus images. Transl. Vis. Sci. Technol. 10, 1–10. doi: 10.1167/tvst.10.7.20

Crossref Full Text | Google Scholar

Zhu, S., Fang, X., Qian, Y., He, K., Wu, M., Zheng, B., et al. (2022). Pterygium screening and lesion area segmentation based on deep learning. J. Healthc. Eng. 2022, 3942110–3942119. doi: 10.1155/2022/3942110

PubMed Abstract | Crossref Full Text | Google Scholar

Zhu, S., Lu, B., Wang, C., Wu, M., Zheng, B., Jiang, Q., et al. (2022). Screening of common retinal diseases using six-category models based on EfficientNet. Front. Med. 9:808402. doi: 10.3389/fmed.2022.808402

PubMed Abstract | Crossref Full Text | Google Scholar

Zhu, S., Zhan, H., Yan, Z., Wu, M., Zheng, B., Xu, S., et al. (2023). Prediction of spherical equivalent refraction and axial length in children based on machine learning. Indian J. Ophthalmol. 71, 2115–2131. doi: 10.4103/IJO.IJO_2989_22

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: B-mode ultrasonography, common eye diseases, deep learning, visualization, classification, image

Citation: Zhu S, Liu X, Lu Y, Zheng B, Wu M, Yao X, Yang W and Gong Y (2024) Application and visualization study of an intelligence-assisted classification model for common eye diseases using B-mode ultrasound images. Front. Neurosci. 18:1339075. doi: 10.3389/fnins.2024.1339075

Received: 27 November 2023; Accepted: 25 April 2024;
Published: 14 May 2024.

Edited by:

Kele Xu, National University of Defense Technology, China

Reviewed by:

Dingwei Wei, Second Affiliated Hospital of Chengdu Medical College, China
Bing Wang, Shandong Provincial Hospital, China
Shaolin Du, Dongguan Tungwah Hospital, China
Jiong Zhang, University of Southern California, United States

Copyright © 2024 Zhu, Liu, Lu, Zheng, Wu, Yao, Yang and Gong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Weihua Yang, benben0606@139.com; Xue Yao, 18925257121@163.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.