Skip to main content

ORIGINAL RESEARCH article

Front. Med., 17 May 2022
Sec. Ophthalmology
This article is part of the Research Topic Big Data and Artificial Intelligence in Ophthalmology View all 14 articles

Artificial Intelligence for the Estimation of Visual Acuity Using Multi-Source Anterior Segment Optical Coherence Tomographic Images in Senile Cataract

\nHyunmin AhnHyunmin Ahn1Ikhyun Jun,Ikhyun Jun1,2Kyoung Yul SeoKyoung Yul Seo1Eung Kweon Kim,Eung Kweon Kim2,3Tae-im Kim,
Tae-im Kim1,2*
  • 1Department of Ophthalmology, Institute of Vision Research, Yonsei University College of Medicine, Seoul, South Korea
  • 2Corneal Dystrophy Research Institute, Yonsei University College of Medicine, Seoul, South Korea
  • 3Saevit Eye Hospital, Goyang, South Korea

Purpose: To investigate an artificial intelligence (AI) model performance using multi-source anterior segment optical coherence tomographic (OCT) images in estimating the preoperative best-corrected visual acuity (BCVA) in patients with senile cataract.

Design: Retrospective, cross-instrument validation study.

Subjects: A total of 2,332 anterior segment images obtained using swept-source OCT, optical biometry for intraocular lens calculation, and a femtosecond laser platform in patients with senile cataract and postoperative BCVA ≥ 0.0 logMAR were included in the training/validation dataset. A total of 1,002 images obtained using optical biometry and another femtosecond laser platform in patients who underwent cataract surgery in 2021 were used for the test dataset.

Methods: AI modeling was based on an ensemble model of Inception-v4 and ResNet. The BCVA training/validation dataset was used for model training. The model performance was evaluated using the test dataset. Analysis of absolute error (AE) was performed by comparing the difference between true preoperative BCVA and estimated preoperative BCVA, as ≥0.1 logMAR (AE≥0.1) or <0.1 logMAR (AE <0.1). AE≥0.1 was classified into underestimation and overestimation groups based on the logMAR scale.

Outcome Measurements: Mean absolute error (MAE), root mean square error (RMSE), mean percentage error (MPE), and correlation coefficient between true preoperative BCVA and estimated preoperative BCVA.

Results: The test dataset MAE, RMSE, and MPE were 0.050 ± 0.130 logMAR, 0.140 ± 0.134 logMAR, and 1.3 ± 13.9%, respectively. The correlation coefficient was 0.969 (p < 0.001). The percentage of cases with AE≥0.1 was 8.4%. The incidence of postoperative BCVA > 0.1 was 21.4% in the AE≥0.1 group, of which 88.9% were in the underestimation group. The incidence of vision-impairing disease in the underestimation group was 95.7%. Preoperative corneal astigmatism and lens thickness were higher, and nucleus cataract was more severe (p < 0.001, 0.007, and 0.024, respectively) in AE≥0.1 than that in AE <0.1. The longer the axial length and the more severe the cortical/posterior subcapsular opacity, the better the estimated BCVA than the true BCVA.

Conclusions: The AI model achieved high-level visual acuity estimation in patients with senile cataract. This quantification method encompassed both visual acuity and cataract severity of OCT image, which are the main indications for cataract surgery, showing the potential to objectively evaluate cataract severity.

Introduction

Cataract is the leading cause of blindness, with ~12.6 million cases of cataract worldwide (1). The visual impairment caused by cataract can be treated with advanced cataract surgery, which can ensure a progressively better quality of vision and fewer complications than in the past (25). The most important indications for cataract surgery are preoperative visual acuity and cataract grading, and advances in surgical technology have expanded the scope of the surgery to even include less severe cataracts (5).

Although the cataract grading system shows a good correlation with surgical difficulty, indicating that the surgery becomes more challenging as the cataract grading increases (6, 7), it shows limitations in reflecting the patient's visual symptoms, especially in cases with nuclear cataract and cortical opacity (8). Cataract grading depends on the subjective competence of the investigator (9). However, visual acuity reflects the patient's symptoms and influences surgical difficulty; therefore, the surgery becomes more challenging also as the visual acuity decreases (10, 11). Moreover, preoperative visual acuity can serve as an important predictor of postoperative vision in various diseases (1214).

Artificial intelligence (AI) is being increasingly used in medicine, and ophthalmology is one of the most active fields for its clinical application (15). Recent studies have attempted to use AI for cataract grading with various methods, including slit-lamp photography, fundus photography, and optical coherence tomography (OCT) (1622), and the results suggest that AI-based cataract grading shows acceptable performance with 70–90% accuracy. However, an AI-based approach for evaluation of visual acuity in patients with cataracts is still lacking. An approach linking objective image data with the subjective symptoms represented by visual acuity is particularly relevant, since the resultant method would encompass both visual acuity and cataract grade, which are the main indications for cataract surgery. Therefore, we attempted to implement an AI model that can evaluate cataract severity based on visual acuity by using multi-source OCT data and to assess the applicability of this AI model in actual clinical practice.

Methods

The study was conducted at the Department of Ophthalmology, Severance Hospital, Yonsei University College of Medicine in accordance with the ethical standards of the Declaration of Helsinki, and institutional review board approval was obtained for the study protocol (4-2021-1697). The institutional review boards waived the need for informed consent because of the retrospective and de-identified nature of the study.

Participants and Dataset

All medical records of patients who underwent cataract surgery between January 2019 and December 2021 were reviewed. The demographic and clinical information of the patients, including age, sex, and clinical history, was collected. We defined distinct inclusion criteria for the training/validation and test datasets. For the training/validation dataset, we collected 2,332 anterior segment OCT images of 2,332 eyes in 2,332 patients with senile cataract alone whose 1-month postoperative best-corrected visual acuity (BCVA) was 0.0 logMAR or better; the images were obtained between January 2019 and December 2020 by using swept-source OCT (ANTERION® swept-source OCT; Heidelberg Engineering, Heidelberg, Germany), optical biometry for intraocular lens calculation (IOLMaster® 700; Carl Zeiss Meditec AG, Jena, Germany), and a femtosecond laser platform (LenSx® laser system; Alcon Laboratories, Inc., Fort Worth, TX, USA). Through this process, we aimed to develop a pure cataract analyzer. In the test dataset, to evaluate the application of the analyzer in actual clinical practice, we collected 1,002 anterior segment OCT images of 1,002 eyes of 1,002 patients who underwent cataract surgery between January 2021 and December 2021; the images were obtained using optical biometry (IOLMaster® 700) and another femtosecond laser platform (CATALYS™ Precision Laser System; Johnson & Johnson Inc., New Brunswick, NJ, USA). When multi-axial images were obtained from one device, a vertical image was selected. When multi-source anterior segment OCT images of a patient were available, a study image was selected randomly using Python version 3.8. When data for both eyes of a patient were available, a study eye was selected randomly. In the training/validation and test datasets, the patients with no anterior segment OCT images of the crystalline lens, not obtained due to corneal opacity or other reasons, were excluded. All OCT images in the training/validation and test datasets were labeled with the preoperative BCVA (Figure 1).

FIGURE 1
www.frontiersin.org

Figure 1. Study flow diagram. For the training/validation dataset, we collected 2,332 anterior segment OCT images of 2,332 eyes in as many patients with isolated senile cataract, whose 1-month postoperative best-corrected visual acuity (BCVA) was 0.0 logMAR or better. The images were obtained between January 2019 and December 2020 using swept-source OCT (ANTERION®), optical biometry for intraocular lens calculation (IOLMaster® 700), and a femtosecond laser platform (LenSx® Laser System). With this process, we aimed to develop a pure cataract analyzer. For the test dataset, we collected 1,002 anterior segment OCT images of 1,002 eyes in as many patients who underwent cataract surgery between January 2021 and December 2021 to evaluate the application of the analyzer in actual clinical practice. The images were obtained using optical biometry (IOLMaster® 700) and a different femtosecond laser platform (CATALYS™ Precision Laser System). The multi-source OCT images varied in size, contour, and direction; an image augmentation method (ImageDataGenerator from the Keras library) was used for the training dataset. The AI modeling was based on an ensemble of the Inception-v4 convolutional neural network (CNN) and ResNet via the stacking technique. The mean absolute error, root mean square error, mean percentage error, and correlation coefficient between the actual and predicted preoperative BCVA were calculated to evaluate the model performance.

Artificial Intelligence Modeling

In this study, to improve the model performance and to analyze the technical error, we manipulated the AI modeling through reduction of input image size, use of image augmentation in the training dataset, and modification of the model architecture. The image size was reduced from 100 to 1% in 1% increments while maintaining the same height-to-width ratio. The multi-source OCT images varied in size, contour, and direction; the image augmentation method using ImageDataGenerator from Keras library was used for the training dataset. Finally, AI modeling was based on an ensemble of the Inception-v4 convolutional neural network (CNN) and ResNet via the stacking technique (23, 24). In the high-resolution and large-scale images, such as the OCT images in this study, a very deep CNN architecture can be expected to perform well until a certain level (25). However, exploding calculation and gradient vanishing are problems associated with very deep CNNs. To overcome these problems, architectures, such as the Inception network and ResNet, can be considered (23, 24). Ensemble modeling of the Inception network and ResNet was implemented using an aggregating method with a weighted average, and the hyperparameters of the model were modified to ensure that the model performed flexibly according to the proportion of the input image shape.

Clinical Assessments

Clinical assessments were used to evaluate AI performance and perform error analysis in the medical approach. All patients underwent detailed preoperative examinations, including slit-lamp biomicroscopy, non-contact tonometry, ophthalmoscopy, and manifest refraction for BCVA. The mean corneal power and corneal astigmatism were measured using autokeratometry (Topcon KR-800A; Topcon Corporation, Tokyo, Japan). Intraocular lens calculation was performed using the IOLMaster® 700, and axial length and lens thickness were measured. An A-scan ultrasound biometry was used for intraocular lens calculation when the IOLMaster® 700 was unavailable. Cataract grading was performed using the Lens Opacities Classification System (LOCS) III by an expert surgeon (T.K.) who evaluated the opacity of the cortex, nucleus, and posterior subcapsular portion of the crystalline lens (26). Postoperative examinations, BCVA, slit-lamp examination, and pupil-dilation were conducted 1 month after the cataract surgery. Vision-impairing disease was defined as a clearly diagnosed disease in the detailed pre- and postoperative examinations that satisfied all of the following criteria: (1) postoperative BCVA > 0.1 logMAR, (2) persistent disease (that leaves an irreversible visual sequelae), (3) existing before cataract surgery, and (4) not a complication of cataract surgery.

Model Performance

To evaluate the model performance, the mean absolute error (MAE), root mean square error (RMSE, which is influenced by large errors), and mean percentage error (MPE, which shows in percentage how much the forecasts of a model differ from the actual values) between the actual preoperative BCVA and predicted preoperative BCVA were calculated as follows:

MAE = | y  y^ |n
RMSE= (y  y^ )2n
MPE = (y  y^ )n ×100 (%)

y: true preoperative BCVA

ŷ: estimated preoperative BCVA

n: number of images in the test dataset

A correlation analysis was conducted between the true and estimated preoperative BCVAs with Pearson correlation coefficient.

Error Analysis

The participants were initially classified into two groups based on the absolute error (AE) between true preoperative BCVA and estimated preoperative BCVA, one with an AE of at least 0.1 logMAR (AE≥0.1), and the other with AE under 0.1 logMAR (AE <0.1). Next, the AE≥0.1 group was further divided into the underestimation group, in which the estimated preoperative BCVA was lower than the true preoperative BCVA, and the overestimation group, in which the estimated preoperative BCVA was higher than the true preoperative BCVA. Ordinal interference was considered if the clinical and statistical sequences were consistent.

Statistical Analysis

Comparative analyses of clinical assessments were conducted between the groups and between the subgroups with independent t-tests for continuous variables and Fisher's exact tests for categorical variables. The values from the A-scan ultrasound biometry were excluded in the comparisons of axial length and lens thickness. Statistical significance was set at p < 0.05.

Results

Table 1 shows the overall demographics and characteristics of the training/validation and test datasets. Although the OCT image source and underlying disease differed between the datasets, no remarkable differences were observed in age, sex, preoperative BCVA, axial length, lens thickness, and LOCS III cataract grade. The postoperative BCVA differed between the datasets (p < 0.001).

TABLE 1
www.frontiersin.org

Table 1. Demographics and characteristics of the training/validation dataset and test dataset.

In the test dataset, the MAE, RMSE, and MPE of the model performance were 0.050 ± 0.130 logMAR, 0.140 ± 0.134 logMAR, and 1.3 ± 13.9%, respectively (Table 2). The correlation coefficient (R) between the true preoperative BCVA and estimated preoperative BCVA was 0.969 (p < 0.001) (Figure 2). The percentage of cases in the AE ≥ 0.1 group was 8.4%, and 1.9% had AE ≥ 0.2.

TABLE 2
www.frontiersin.org

Table 2. Performance of the artificial intelligence model for prediction of preoperative best corrected visual acuity in patients with senile cataract.

FIGURE 2
www.frontiersin.org

Figure 2. Scatter plot between true preoperative BCVA and estimated preoperative BCVA. The solid black line shows that the true and estimated preoperative BCVAs were equal. The black dotted line shows the absolute error (AE) of the true and estimated preoperative BCVAs of 0.1 logMAR. The red dotted line describes the Pearson correlation coefficient. In the test dataset, 91.6% of cases had AE under 0.1 logMAR.

The preoperative BCVA values were significantly different between the AE≥0.1 and AE <0.1 groups (0.30 ± 0.08 vs. 0.12 ± 0.10, respectively; p < 0.001), and corneal astigmatism, lens thickness, and nucleus cataract in the LOCS III grading were significantly different (p < 0.001, < 0.001, and 0.024, respectively) (Table 3). Postoperative BCVA in the AE≥0.1 group was worse than that in the AE <0.1 group (p < 0.001). In the AE≥0.1 group, the incidence of postoperative BCVA > 0.1 was 21.4%, which was higher than that in the AE <0.1 group, and the proportion of vision-impairing disease was also higher than in the AE <0.1 group. The percentage of cases with vision-impairing diseases was significantly higher in the AE≥0.1 group than in the AE <0.1 group (54.8 vs. 12.9%; p < 0.001).

TABLE 3
www.frontiersin.org

Table 3. Comparison of the patients showing absolute error of BCVA of 0.1 and over (AE≥0.1) with those showing absolute error under 0.1 (AE<0.1) in the test dataset.

The AE≥0.1 group was further divided into underestimation and overestimation groups (Table 4). Pre- and postoperative BCVAs were significantly worse in the underestimation group (0.39 ± 0.10 vs. 0.15 ± 0.09 in overestimation, and 0.10 ± 0.15 vs. 0.01 ± 0.05; p < 0.001 and p < 0.001, respectively). The proportion of cases with postoperative BCVA > 0.1 was higher in the underestimation group (p < 0.001). The proportion of cases with vision-impairing disease was 95.7% in the underestimation group. Axial length was longer, and cortical opacity and posterior subcapsular opacity were more severe in the underestimation group (p < 0.001, 0.045, and 0.002, respectively).

TABLE 4
www.frontiersin.org

Table 4. Comparison of the overestimation and underestimation groups among patients showing an absolute error of 0.1 and over (AE≥0.1) in BCVA.

Error analysis for model performance was conducted by performing comparative analyses between the AE≥0.1 and AE<0.1 groups and between the underestimation and overestimation subgroups of the AE≥0.1 group (Table 5). Preoperative corneal astigmatism, lens thickness, and nuclear opacity in the LOCS III grading were higher in the AE≥0.1 group. Increased BCVA and vision-impairing disease were more significant in the underestimation subgroup. Axial length was longer, and the cortical/posterior subcapsular opacity was more severe in the overestimation subgroup than in the underestimation subgroup.

TABLE 5
www.frontiersin.org

Table 5. Error analysis of model performance in clinical practice.

In the AI modeling process, reducing the original image decreased the model performance, and the neural network could not estimate the BCVA when the area was reduced by ~≥90% (32% of width and height) (Figure 3A). Image augmentation increased the model performance in terms of the MAE (Figure 3B). The ensemble model had the lowest MAE, and Inception-v4 and ResNet were almost similar in terms of MAE (Figure 3C).

FIGURE 3
www.frontiersin.org

Figure 3. Mean absolute error (MAE) of the model performance in the AI training process. As the size of the OCT image reduced, MAE increased (A), and image augmentation using the image generator showed a lower MAE (B). The ensemble model of Inception-v4 and ResNet showed the lowest MAE compared to the individual architecture (C).

Discussion

With the test dataset of this study, over 90% of cases could be estimated in their BCVA under 0.1 logMAR of AE. Most of the underestimation errors were caused by vision-impairing disease, and about half of the decreased model performance could be explained by a clinical approach. The technical issues, image resolution, diversity of image forms in the training dataset, and model architecture also affected the model performance.

Objective clinical assessment for estimation of the subjective visual symptoms represented by visual acuity, which is considered the ultimate goal of ophthalmologic interventions, can be utilized in various ways from clinical to experimental; however, this is very difficult for clinicians due to confounding factors (27). AI is expected to help human evaluators perform difficult tasks (28). This study suggested that cataract severity can be quantified as visual acuity via AI using OCT images.

In this study, cross-instrument validation was performed using different combinations of anterior segment OCT image sources for the training/validation dataset and the test dataset. To enhance the model performance, the OCT images used in the training/validation dataset were obtained from three devices with different detection methods, image resolutions, and image directions. The test dataset was constructed to evaluate the practical application of the AI model in real-world scenarios by including all patients within a certain period and using commonly available OCT images in the process of cataract surgery and its preparation.

Model performance and error analysis were conducted from both clinical and engineering perspectives. Although medical issues have rarely been mentioned in previous AI research, for clinical application of medical AI, error analysis from the clinical perspective is indispensable (29, 30). The results suggest that the predisposing disease and the conditions, such as high values of corneal astigmatism, axial length, lens thickness, and severity of each cataract subtype, which are well known to cause visual impairment, were the main factors underlying errors in the medical approach (8, 3134). For engineering issues, this study suggested that low-resolution images led to the degradation of the model performance (see Figure 3A), and highlighted the importance of high-resolution images in analyzing precise medical observations (35). Multi-source images were used, and the transformation for direction and size through image augmentation in the training dataset was shown to yield better model performance (36). Although the AI model using a unified material may show a different performance from that using multi-sources materials, we focused on the versatility of the AI model by using OCT images acquired with multiple instruments.

The primary goal of cataract surgery is to restore the best possible vision as well as remove the natural crystalline lens (37). Thus, postoperative visual acuity is the target indicator of cataract surgery. Wei et al. reported the use of AI with fundus OCT for predicting postoperative visual acuity in patients with high myopia (38), and the result showed the lowest MAE of 0.16 logMAR and RMSE of 0.24 logMAR. However, this performance level was still insufficient for clinical application, and the findings implied that visual acuity was not determined by a single defined etiology. Clinicians frequently encounter patients with visual impairment without a specific pathologic lesion. This phenomenon may be caused by medical issues and unknown predisposing factors, including developmental problems, such as amblyopia, an extraocular disease, such as brain lesion, or a temporary problem, such as dry eye disease. Thus, approaches based on a single etiological factor showed limited ability to predict the postoperative BCVA. Since multiple factors, including the presence of both cataract and comorbid disease, can influence the accuracy of prediction of postoperative BCVA with AI based on the preoperative BCVA, our study could serve as the basis for future studies. Future studies should aim to predict postoperative visual acuity using a combination of preoperative visual acuity, anterior segment OCT, and other examinations.

Similar to previous medical AI research (16, 3941), this study did not completely overcome the limitations in interpreting the neural network. In this study, the attention areas of the AI model were analyzed using heatmap analysis (Figure 4), and we tried to analyze the relationships between the attention areas and the error analysis in the medical approach. Although model training was conducted using patients with only a clinical diagnosis of cataract, the AI also recognized the cornea that was laid on the visual axis and the angle in the anterior chamber. Thus, some factors affecting the corneal shape (e.g., corneal refractive surgery), anterior chamber angle (e.g., glaucoma), and lens thickness may be related.

FIGURE 4
www.frontiersin.org

Figure 4. Example of heatmap analysis of the Artificial Intelligence (AI) attention area. Original OCT image (A) and Heatmap image (B). The preoperative best-corrected visual acuity (BCVA) of this 67-year-old patient, who showed a senile cataract with grade 3 cortical opacity and grade 2 nuclear opacity in the Lens Opacities Classification System III, was 0.50 logMAR, and the AI-estimated BCVA was 0.52 logMAR.

The strengths of this study were not limited to the excellent performance of the AI model. Our attempt to prove the importance of clinical approach in model performance was successful and our findings suggest that the addition of various clinical information in AI modeling is crucial for improving model performance.

In conclusion, the AI developed using OCT images from multiple sources showed excellent performance in estimating visual acuity in patients with senile cataracts. This quantification method encompasses both visual acuity and cataract severity of the OCT images, which are the main indications for cataract surgery, and has the potential to allow objective evaluation of cataract severity. This AI model can be used when it is difficult to express or measure the subjective visual acuity due to various causes, such as an inability to communicate. Additionally, we would like to emphasize that this was a preliminary study to expand the prediction of visual acuity after cataract surgery in patients with other diseases, possibly accompanied by visual impairment.

Data Availability Statement

The datasets presented in this article are not readily available because of privacy and ethical concerns, neither the data nor the source of the data can be made available. Requests to access the datasets should be directed to HA, overhyun31@gmail.com.

Ethics Statement

The studies involving human participants were reviewed and approved by the Severance Hospital Clinical Research Ethics Committee. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author Contributions

HA and T-iK: conceptualization. HA: methodology, software, formal analysis, investigation, writing—original draft preparation, and visualization. IJ, KS, EK, and T-iK: validation. IJ, KS, and T-iK: resources. T-iK: writing—review and editing and supervision. All authors contributed to the article and approved the submitted version.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Abbreviations

AE, absolute error; AI, artificial intelligence; BCVA, best-corrected visual acuity; CNN, convolutional neural network; LOCS, Lens Opacities Classification System; MAE, mean absolute error; MPE, mean percentage error; OCT, optical coherence tomography; RMSE, root mean square error.

References

1. Flaxman SR, Bourne RR, Resnikoff S, Ackland P, Braithwaite T, Cicinelli MV, et al. Global causes of blindness and distance vision impairment 1990–2020: a systematic review and meta-analysis. Lancet Global Health. (2017) 5:e1221–34. doi: 10.1016/S2214-109X(17)30393-5

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Lundström M, Dickman M, Henry Y, Manning S, Rosen P, Tassignon MJ, et al. Femtosecond laser–assisted cataract surgeries reported to the European Registry of Quality Outcomes for Cataract and Refractive Surgery: Baseline characteristics, surgical procedure, and outcomes. J Cataract Refract Surg. (2017) 43:1549–56. doi: 10.1016/j.jcrs.2017.09.029

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Lundström M, Behndig A, Kugelberg M, Montan P, Stenevi U, Thorburn W. Decreasing rate of capsule complications in cataract surgery: eight-year study of incidence, risk factors, and data validity by the Swedish National Cataract Register. J Cataract Refract Surg. (2011) 37:1762–7. doi: 10.1016/j.jcrs.2011.05.022

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Lundström M, Barry P, Henry Y, Rosen P, Stenevi U. Visual outcome of cataract surgery; study from the European Registry of Quality Outcomes for Cataract and Refractive Surgery. J Cataract Refract Surg. (2013) 39:673–9. doi: 10.1016/j.jcrs.2012.11.026

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Lundström M, Dickman M, Henry Y, Manning S, Rosen P, Tassignon MJ, et al. Changing practice patterns in European cataract surgery as reflected in the European Registry of Quality Outcomes for Cataract and Refractive Surgery 2008 to 2017. J Cataract Refract Surg. (2021) 47:373–8. doi: 10.1097/j.jcrs.0000000000000457

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Galan A, Tavolato M, Babighian S. Grading the surgical difficulty of cataract phacoemulsification. Ophthal Surg Lasers Imaging Retina. (2009) 40:361–5. doi: 10.3928/15428877-20096030-02

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Lapid-Gortzak R. Gauging the difficulty of phacoemulsification: new grading systems. Expert Rev Ophthalmol. (2009) 4:455–6. doi: 10.1586/eop.09.41

CrossRef Full Text | Google Scholar

8. Stifter E, Sacu S, Benesch T, Weghaupt H. Impairment of visual acuity and reading performance and the relationship with cataract type and density. Invest Ophthalmol Vis Sci. (2005) 46:2071–5. doi: 10.1167/iovs.04-0890

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Karbassi M, Khu PM, Singer DM, Chylack L Jr. Evaluation of lens opacities classification system III applied at the slitlamp. Optomet Vis Sci. (1993) 70:923–8. doi: 10.1097/00006324-199311000-00009

PubMed Abstract | CrossRef Full Text | Google Scholar

10. McGwin G Jr, Scilley K, Brown J, Owsley C. Impact of cataract surgery on self-reported visual difficulties: comparison with a no-surgery reference group. J Cataract Refract Surg. (2003) 29:941–8. doi: 10.1016/S0886-3350(02)01846-1

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Dooley IJ, O'Brien PD. Subjective difficulty of each stage of phacoemulsification cataract surgery performed by basic surgical trainees. J Cataract Refract Surg. (2006) 32:604–8. doi: 10.1016/j.jcrs.2006.01.045

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Chu CJ, Johnston RL, Buscombe C, Sallam AB, Mohamed Q, Yang YC, et al. Risk factors and incidence of macular edema after cataract surgery: a database study of 81984 eyes. Ophthalmology. (2016) 123:316–23. doi: 10.1016/j.ophtha.2015.10.001

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Yoeruek E, Deuter C, Gieselmann S, Saygili O, Spitzer MS, Tatar O, et al. Long-term visual acuity and its predictors after cataract surgery in patients with uveitis. Eur J Ophthalmol. (2010) 20:694–701. doi: 10.1177/112067211002000409

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Ostri C, Lund-Andersen H, Sander B, La Cour M. Phacoemulsification cataract surgery in a large cohort of diabetes patients: visual acuity outcomes and prognostic factors. J Cataract Refract Surg. (2011) 37:2006–12. doi: 10.1016/j.jcrs.2011.05.030

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Li J-PO, Liu H, Ting DS, Jeon S, Chan RP, Kim JE, et al. Digital technology, tele-medicine and artificial intelligence in ophthalmology: a global perspective. Prog Retin Eye Res. (2021) 82:100900. doi: 10.1016/j.preteyeres.2020.100900

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Prevedello LM, Halabi SS, Shih G, Wu CC, Kohli MD, Chokshi FH, et al. Challenges related to artificial intelligence research in medical imaging and the importance of image analysis competitions. Radiol Artif Intelligence. (2019) 1:e180031. doi: 10.1148/ryai.2019180031

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Xu X, Zhang L, Li J, Guan Y, Zhang L. A hybrid global-local representation CNN model for automatic cataract grading. IEEE J Biomed Health Inform. (2019) 24:556–67. doi: 10.1109/JBHI.2019.2914690

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Zhang H, Niu K, Xiong Y, Yang W, He Z, Song H. Automatic cataract grading methods based on deep learning. Comput Methods Programs Biomed. (2019) 182:104978. doi: 10.1016/j.cmpb.2019.07.006

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Xiong L, Li H, Xu L. An approach to evaluate blurriness in retinal images with vitreous opacity for cataract diagnosis. J Healthc Eng. (2017) 2017:1–16. doi: 10.1155/2017/5645498

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Yang J-J, Li J, Shen R, Zeng Y, He J, Bi J, et al. Exploiting ensemble learning for automatic cataract detection and grading. Comput Methods Programs Biomed. (2016) 124:45–57. doi: 10.1016/j.cmpb.2015.10.007

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Guo L, Yang J-J, Peng L, Li J, Liang Q. A computer-aided healthcare system for cataract classification and grading based on fundus image analysis. Comput Industry. (2015) 69:72–80. doi: 10.1016/j.compind.2014.09.005

CrossRef Full Text | Google Scholar

22. Zhang X, Xiao Z, Higashita R, Chen W, Yuan J, Fang J, et al. A novel deep learning method for nuclear cataract classification based on anterior segment optical coherence tomography images. in IEEE. (2020). p. 662–8. doi: 10.1109/SMC42975.2020.9283218

CrossRef Full Text | Google Scholar

23. Szegedy C, Ioffe S, Vanhoucke V, Alemi AA. Inception-v4, inception-resnet and the impact of residual connections on learning. In: Thirty-First AAAI Conference on Artificial Intelligence (2017).

Google Scholar

24. He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. (2016). p. 770–8.

PubMed Abstract | Google Scholar

25. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale image recognition. arXiv preprint arXiv:14091556. (2014). Available online at: https://arxiv.org/abs/1409.1556

Google Scholar

26. Chylack LT, Wolfe JK, Singer DM, Leske MC, Bullimore MA, Bailey IL, et al. The lens opacities classification system III. Arch Ophthalmol. (1993) 111:831–6. doi: 10.1001/archopht.1993.01090060119035

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Mangione CM, Orav EJ, Lawrence MG, Phillips RS, Seddon JM, Goldman L. Prediction of visual function after cataract surgery: a prospectively validated model. Arch Ophthalmol. (1995) 113:1305–11. doi: 10.1001/archopht.1995.01100100093037

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Ahuja AS. The impact of artificial intelligence in medicine on the future role of the physician. PeerJ. (2019) 7:e7702. doi: 10.7717/peerj.7702

PubMed Abstract | CrossRef Full Text | Google Scholar

29. Ting DS, Lee AY, Wong TY. An ophthalmologist's guide to deciphering studies in artificial intelligence. Ophthalmology. (2019) 126:1475–9. doi: 10.1016/j.ophtha.2019.09.014

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Szolovits P. Artificial Intelligence in Medicine. New York, NY: Routledge (2019).

Google Scholar

31. Wolffsohn JS, Bhogal G, Shah S. Effect of uncorrected astigmatism on vision. J Cataract Refract Surg. (2011) 37:454–60. doi: 10.1016/j.jcrs.2010.09.022

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Hashimoto S, Yasuda M, Fujiwara K, Ueda E, Hata J, Hirakawa Y, et al. Association between axial length and myopic maculopathy: the Hisayama Study. Ophthalmol Retina. (2019) 3:867–73. doi: 10.1016/j.oret.2019.04.023

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Fujiwara A, Shiragami C, Manabe S, Izumibata S, Murata A, Morizane Y, et al. Normal values of retinal sensitivity determined by macular integrity assessment. Invest Ophthalmol Vis Sci. (2014) 55:5875–75. Available online at: https://iovs.arvojournals.org/article.aspx?articleid=2271519

PubMed Abstract | Google Scholar

34. Westall CA, Dhaliwal HS, Panton CM, Sigesmund D, Levin AV, Nischal KK, et al. Values of electroretinogram responses according to axial length. Documenta Ophthalmol. (2001) 102:115–30. doi: 10.1023/A:1017535207481

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Kannojia SP, Jaiswal G. Effects of varying resolution on performance of CNN based image classification: an experimental study. Int J Comput Sci Eng. (2018) 6:451–6. doi: 10.26438/ijcse/v6i9.451456

CrossRef Full Text | Google Scholar

36. Shorten C, Khoshgoftaar TM. A survey on image data augmentation for deep learning. J Big Data. (2019) 6:1–48. doi: 10.1186/s40537-019-0197-0

CrossRef Full Text | Google Scholar

37. Brad H. Feldman SH. Cataract. American Academy of Ophthalmology (2021). Available online at: https://eyewiki.aao.org/Cataract (accessed January 23, 2022).

38. Wei L, He W, Wang J, Zhang K, Du Y, Qi J, et al. An optical coherence tomography-based deep learning algorithm for visual acuity prediction of highly myopic eyes after cataract surgery. Front Cell Dev Biol. (2021) 9:1195. doi: 10.3389/fcell.2021.652848

PubMed Abstract | CrossRef Full Text | Google Scholar

39. Kallianos K, Mongan J, Antani S, Henry T, Taylor A, Abuya J, et al. How far have we come? Artificial intelligence for chest radiograph interpretation. Clin Radiol. (2019) 74:338–45. doi: 10.1016/j.crad.2018.12.015

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Alsharqi M, Woodward W, Mumith J, Markham D, Upton R, Leeson P. Artificial intelligence and echocardiography. Echo Res Pract. (2018) 5:R115–25. doi: 10.1530/ERP-18-0056

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Mendelson EB. Artificial intelligence in breast imaging: potentials and limitations. Am J Roentgenol. (2019) 212:293–9. doi: 10.2214/AJR.18.20532

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: artificial intelligence, cataract, convolutional neural network, optical coherence tomography, visual acuity

Citation: Ahn H, Jun I, Seo KY, Kim EK and Kim T-i (2022) Artificial Intelligence for the Estimation of Visual Acuity Using Multi-Source Anterior Segment Optical Coherence Tomographic Images in Senile Cataract. Front. Med. 9:871382. doi: 10.3389/fmed.2022.871382

Received: 08 February 2022; Accepted: 04 April 2022;
Published: 17 May 2022.

Edited by:

Michele Lanza, University of Campania Luigi Vanvitelli, Italy

Reviewed by:

Rajiv Raman, Sankara Nethralaya, India
Mukharram M. Bikbov, Ufa Eye Research Institute, Russia
Emmanuel Bui Quoc, Assistance Publique Hopitaux de Paris, France
Shiri Soudry, Rambam Health Care Campus, Israel
Georgios A. Kounis, GNEMS, Greece
Yingting Zhu, Sun Yat-sen University, China

Copyright © 2022 Ahn, Jun, Seo, Kim and Kim. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Tae-im Kim, tikim@yuhs.ac

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.