Skip to main content

ORIGINAL RESEARCH article

Front. Med.
Sec. Ophthalmology
Volume 11 - 2024 | doi: 10.3389/fmed.2024.1418048
This article is part of the Research Topic Predictive and Diagnostic Approaches for Systemic Disorders Using Ocular Assessment View all 6 articles

Enhancing the Ophthalmic AI Assessment with a Fundus Image Quality Classifier Using Local and Global Attention Mechanisms

Provisionally accepted
Shengzhan Wang Shengzhan Wang 1*Wenyue Shen Wenyue Shen 2Zhiyuan Gao Zhiyuan Gao 2*Xiaoyu Jiang Xiaoyu Jiang 3Yaqi Wang Yaqi Wang 4Yunxiang Li Yunxiang Li 4*Xiaoyu Ma Xiaoyu Ma 4*Wenhao Wang Wenhao Wang 1*Shuanghua Xin Shuanghua Xin 1*Weina Ren Weina Ren 1Kai Jin Kai Jin 2*Juan Ye Juan Ye 2*
  • 1 Affiliated Hospital, Ningbo University, Ningbo, Zhejiang Province, China
  • 2 Eye Center, Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, China
  • 3 College of Control Science and Engineering, Zhejiang University, Hangzhou, Zhejiang Province, China
  • 4 Communication University of Zhejiang, Hangzhou, Zhejiang Province, China

The final, formatted version of the article will be published soon.

    The assessment of image quality (IQA) plays a pivotal role in the realm of imagebased computer-aided diagnosis techniques, with fundus imaging standing as the primary method for the screening and diagnosis of ophthalmic diseases. Conventional studies on fundus IQA tend to rely on simplistic datasets for evaluation, predominantly focusing on either local or global information, rather than a synthesis of both. Moreover, the interpretability of these studies often lacks compelling evidence. In order to address these issues, this study introduces the Local and Global Attention Aggregated Deep Neural Network (LGAANet), an innovative approach that integrates both local and global information for enhanced analysis.The LGAANet was developed and validated using a Multi-Source Heterogeneous Fundus (MSHF) database, encompassing a diverse collection of images. This dataset includes 802 color fundus photography (CFP) images (302 from portable cameras), and 500 ultrawidefield (UWF) images from 904 patients with diabetic retinopathy (DR) and glaucoma, as well as healthy individuals.This dataset includes normal and pathological images from color fundus photography (CFP), images captured with a portable camera from healthy volunteers, and ultrawide-field (UWF) images from patients with diabetic retinopathy. The assessment of image quality was meticulously carried out by a trio of ophthalmologists, leveraging the human visual system as a benchmark. Furthermore, the model employs attention mechanisms and saliency maps to bolster its interpretability.In testing with the CFP dataset, LGAANet demonstrated remarkable accuracy in three critical dimensions of image quality (illumination, clarity and contrast), recording scores of 0.947, 0.924, and 0.947, respectively. Similarly, when applied to the UWF dataset, the model achieved accuracies of 0.889, 0.913, and 0.923, respectively. These results underscore the efficacy of LGAANet in distinguishing between varying degrees of image quality with high precision.To our knowledge, LGAANet represents the inaugural algorithm trained on an MSHF dataset specifically for fundus IQA, marking a significant milestone in the advancement of computer-aided diagnosis in ophthalmology. This research significantly contributes to the Abstract field, offering a novel methodology for the assessment and interpretation of fundus images in the detection and diagnosis of ocular diseases.

    Keywords: fundus photography, attention mechanism, image quality assessment, spatial information, Multiscale feature extraction

    Received: 15 Apr 2024; Accepted: 23 Jul 2024.

    Copyright: © 2024 Wang, Shen, Gao, Jiang, Wang, Li, Ma, Wang, Xin, Ren, Jin and Ye. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence:
    Shengzhan Wang, Affiliated Hospital, Ningbo University, Ningbo, Zhejiang Province, China
    Zhiyuan Gao, Eye Center, Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, China
    Yunxiang Li, Communication University of Zhejiang, Hangzhou, 310019, Zhejiang Province, China
    Xiaoyu Ma, Communication University of Zhejiang, Hangzhou, 310019, Zhejiang Province, China
    Wenhao Wang, Affiliated Hospital, Ningbo University, Ningbo, Zhejiang Province, China
    Shuanghua Xin, Affiliated Hospital, Ningbo University, Ningbo, Zhejiang Province, China
    Kai Jin, Eye Center, Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, China
    Juan Ye, Eye Center, Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, China

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.