- 1Department of Ophthalmology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences, Beijing, China
- 2Key Laboratory of Ocular Fundus Diseases, Peking Union Medical College, Chinese Academy of Medical Sciences, Beijing, China
- 3Eight-Year Medical Doctor Program, Peking Union Medical College, Chinese Academy of Medical Sciences, Beijing, China
- 4Department of Ophthalmology, Beijing Ming Vision Clinic, Beijing, China
- 5Department of Ophthalmology, Beijing Fenglian Jiayue Lege Clinic, Beijing, China
Purpose: This study aims to evaluate the diagnostic performance of a machine learning model (ML model) to train junior ophthalmologists in detecting preclinical keratoconus (PKC).
Methods: A total of 1,334 corneal topography images (The Pentacam HR system) from 413 keratoconus eyes, 32 PKC eyes and 222 normal eyes were collected. Five junior ophthalmologists were trained and annotated the images with or without the suggestions proposed by the ML model. The diagnostic performance of PKC was evaluated among three groups: junior ophthalmologist group (control group), ML model group and ML model-training junior ophthalmologist group (test group).
Results: The accuracy of the ML model between the eyes of patients with KC and NEs in all three clinics (99% accuracy, area under the receiver operating characteristic (ROC) curve AUC of 1.00, 99% sensitivity, 99% specificity) was higher than that for Belin-Ambrósio enhanced ectasia display total deviation (BAD-D) (86% accuracy, AUC of 0.97, 97% sensitivity, 69% specificity). The accuracy of the ML model between eyes with PKC and NEs in all three clinics (98% accuracy, AUC of 0.96, 98% sensitivity, 98% specificity) was higher than that of BAD-D (69% accuracy, AUC of 0.73, 67% sensitivity, 69% specificity). The diagnostic accuracy of PKC was 47.5% (95%CI, 0.5–71.6%), 100% (95%CI, 100–100%) and 94.4% (95%CI, 14.7–94.7%) in the control group, ML model group and test group. With the assistance of the proposed ML model, the diagnostic accuracy of junior ophthalmologists improved with statistical significance (p < 0.05). According to the questionnaire of all the junior ophthalmologists, the average score was 4 (total 5) regarding to the comprehensiveness that the AI model has been in their keratoconus diagnosis learning; the average score was 4.4 (total 5) regarding to the convenience that the AI model has been in their keratoconus diagnosis learning.
Conclusion: The proposed ML model provided a novel approach for the detection of PKC with high diagnostic accuracy and assisted to improve the performance of junior ophthalmologists, resulting especially in reducing the risk of missed diagnoses.
Introduction
Keratoconus (KC) is a noninflammatory condition. This resulted in irregular astigmatism and reduced vision (1). A recent epidemiological study demonstrated that KC is not a rare disease, with a prevalence of 1.2% (2). However, early detection of KC is challenging. In the earliest stages, preclinical keratoconus (PKC) does not show classical keratometric or biomicroscopic features. Misdiagnosis of PKC leads to an increased risk of iatrogenic ectasia after refractive surgery (3), which is the most severe and irreversible complication after corneal refractive surgery (4). It has been estimated that 50% of patients with PKC progress to KC within 16 years (5). In addition, with the availability of corneal cross-linking, early detection can also contribute to the delay or cessation of KC progression (6).
However, the detection of early stage of keratoconus is a challenging task, especially for the junior ophthalmologists. We develop a machine learning system using corneal topography images obtained from three clinics to detect PKC and use it to train the junior ophthalmologists, helping them improve their accuracy of PKC detection.
Methods
Study design and population
The study followed the tenets of the Declaration of Helsinki. The study protocol was approved by the Review Board and Human Ethics Committee of Peking Union Medical College Hospital, Chinese Academy of Medical Sciences. The Pentacam HR system (Oculus, Wetzlar, Germany) is a Scheimpflug priciple-based tomography system with a rotating camera that can record 25,000 corneal points. Five images obtained by the Pentacam HR system were analyzed using for analyses by an artificial intelligence (AI) model: anterior and posterior corneal curvatures, anterior and posterior elevations, and entire pachymetry distribution.
Both eyes of all patients were evaluated using the Pentacam HR system. Patients were classified into KC, PKC, or NE groups. KC eyes were diagnosed using the following criteria established by the Collaborative Longitudinal Evaluation of Keratoconus (CLEK) study (7): (1) at least one of the following slit-lamp signs: focal stromal thinning, Vogt’s strias, Fleischer’s ring >2-mm arc, or corneal scarring consistent with KC; (2) irregular cornea with focal steeping determined via distorted keratometry test, and distortion of the retinoscopic or ophthalmoscopic red reflex and (3) no history of contact lens wear, ocular surgery or extensive scarring.
According to the Global consensus on keratoconus and ectatic diseases (8), true unilateral keratoconus does not exist. Therefore, PKC eyes were the unaffected eyes of the keratoconus patients whose the other eye were diagnosed with KC. An explicit definition of subclinical keratoconus (SKC) is lacking. The most common definition of SKC used to be patients with a tomographic pattern of localized steepening in the posterior or anterior corneal surface or paracentral corneal thinning, but no clinical signs of KC. However, PKC refers to unaffected eye of keratoconus patient with no sign of keratoconus by tomography. PKC eyes were diagnosed using the following criteria established by the Keratoconus Severity Score (KSS) study (9): (1) No corneal scarring consistent with KC; (2) No slit-lamp signs for keratoconus; (3) Inferior or superior steepening no more than 3D steeper than average corneal power (ACP); (4) ACP no more than 48D.
The eyes of normal individuals undergoing routine ophthalmological examination met the following screening criteria (10): (1) normal clinical evaluations (defined as no clinical signs or suggestive topographic patterns for suspicious KC, KC, or pellucid marginal degeneration), (2) no family history of KC, (3) no history of ocular surgery or trauma. and (4) stopped contact lens wear for ≥3 weeks for rigid gas permeable, and ≥ 1 week for soft contact lenses.
We included 272 NEs, 413 KC eyes and 72 PKC eyes from the Ming Vision Clinic (Clinic 1). We included 100 NEs, 87 KC eyes and 15 PKC eyes from the Fenglian JiaYue LeGe Clinic (Clinic 2). We included 72 NEs, 51 KC eyes and 14 PKC eyes from the Peking Union Medical College Hospital (Clinic 3). The patient data obtained from the clinics were used for the training model where input into a private model and not a public model.
Datasets and pre-processing
Patients were divided into three datasets. The training set consisted of three groups. The NE group comprised 35 eyes from Clinic 1, 25 eyes from clinic 2, and 4 eyes from clinic 3. The KC group comprised 25 eyes from Clinic 2. The PKC group comprised 30 eyes from Clinic 1, five eyes from Clinic 2, and four eyes from clinic 3. The validation set comprised of three groups. The NE group comprised 15 eyes from Clinic 1, 15 from Clinic 2, and 4 from Clinic 3. The KC group comprised 15 eyes from Clinic 2. The PKC group comprised 10 eyes from Clinic 1, 5 eyes from Clinic 2, and 4 eyes from Clinic 3. The independent test group consisted of three groups. The NE group comprised 222 eyes from Clinic 1, 60 from Clinic 2, and 64 from Clinic 3. The KC group comprised 413 eyes from Clinic 1, 47 eyes from Clinic 2, and 51 eyes from Clinic 3. The PKC group comprised 32 eyes from Clinic 1, 5 eyes from Clinic 2, and 6 eyes from Clinic 3.
Pre-processing of input images
Five standard topographical refractive maps (maps showing central corneal thickness, anterior (posterior) surface elevation, and anterior (posterior) surface curvature) were selected as the original data for our study. These maps were obtained from the Pentacam HR system and exported with a diameter of 8 mm. Each original image was then cropped to 540 × 540 × 3 pixels. We filtered out textual information and retained only color-gradient data from the exported images. The cropped images were then “stitched together” into a single row with dimensions of 2,700 × 540 × 3 pixels, following the order of maps of central corneal thickness, anterior (posterior) surface elevation, and anterior (posterior) surface curvature. This concatenation of five medical images allowed for joint learning, enabling the model to analyze related (but distinct) images simultaneously in one pass (11). This approach facilitated the identification of common features, alignments, correlations, and invariance across datasets.
To optimize training efficiency and minimize graphics processing unit (GPU) memory demands, we resized the input images to dimensions of 1,120 × 224 × 3 pixels. This resizing was performed to reduce the computational cost per iteration, to enable larger batch sizes, and to prevent out-of-memory errors. However, excessive downsizing can lead to performance degradation by eliminating discriminative details. Therefore, the chosen input dimensions were carefully selected to strike a balance between training speed and model accuracy.
In contrast to a conventional binary classification task that distinguishes between KC and NEs, we combined KC and PKC into one category to determine whether the model could identify the hidden distinctive features in PKC. To explain the possible impacts of interclass mixture and class imbalance in a multicenter study, we projected and analyzed corneal topographical images onto the feature space using t-distributed stochastic neighbor embedding (t-SNE), which helped design a strategy for the selection of training datasets that improved the detection accuracy of PKC as a form of KC against NE. We trained and optimized two models using the efficientnet-b0 architecture and the Train, Adapt, Optimize (TAO) Toolkit (NVIDIA, Santa Clara, CA, United States). The first model (referred to as “Model 1”) was trained using data from two centers. The second model (“Model 2”) used Model 1 as a pre-trained model and incorporated a small amount of additional data during training to enhance classification accuracy.
In the present study, t-SNE was employed to analyze clinical data collected from multiple centers (12). The objective of this study was to project the spatial distributions of high-dimensional features of topographical images onto a two-dimensional scatter plot. The ground truths of the three class labels (KC, PKC, and NEs) were determined by two expert annotator. The results showed that the KC group (including the PKC group) was distinguishable from the NE group. Notably, the scattered KC points within the normal cluster were patients with PKC who shared visual similarity with NEs. Furthermore, the distributions of the corresponding classes between each center were scattered as separate clusters, indicating a domain shift (i.e., a mismatch caused by differences in the picture quality between images obtained from different tasks) (Figure 1). In our study, this mismatch arose from variations in export settings across multiple clinical centers. Consequently, the model training process could be confounded by the interclass mixture and the domain shift resulting from the acquisition of multicenter data, thereby posing challenges to the classification task.
ML model
Our task involved distinguishing between two classes (PKC and KC) combined as one from the NE group using a small balanced dataset. Hence, we chose the TAO Toolkit (NVIDIA) (13) to train and optimize our model for the classification of KC/NEs. The TAO Toolkit is a low-code solution that allows non-AI professionals to expedite their processes for the training and optimization of a model. For this purpose, we selected efficientnet-b0 from TAO (14). This model employs a compound coefficient to scale network dimensions (e.g., depth, width, and resolution) uniformly. This balanced scaling approach results in improved performance compared with independent dimension scaling because it enables consistent growth of a network while maintaining a reasonable complexity of the model (15). Efficient net architectures achieve competitive accuracy with fewer parameters, owing to their sample efficiency resulting from balanced scaling (14). Therefore, the efficientnet-b0 model was well suited for our challenging binary classification task, which involved distinguishing between NEs and KC eyes, with limited subtle and readily confused cases.
Training sets
During the training phase, we conducted a two-step process to train the two models based on the efficientnet-b0 model. Model 1 was a binary classification model trained on data from Clinics 1 and 2. The dataset used for training consisted of 40 PKC samples from Clinic 1, 40 KC samples, 10 PKC samples from Clinic 2, 50 NE samples from Clinic 1, and 40 NE samples from Clinic 2. To ensure representative samples, we employed a data selection strategy that curated balanced subsets of the KC, PKC, and NE subgroups from the collected datasets. The aim was to optimize the robustness of the efficientnet-b0 model. Model 2 was a binary classification model trained on the data from all three clinics. To enhance the performance of Model 2, we added 8 PKCs and 8 NEs from Clinic 3 to the training dataset for Model 1. Model 1 was used as a pre-training model, following the concept of transfer learning.
Test sets
We employed two distinct test sets to evaluate the performance of the proposed two-step training model. The initial test set, referred to as “dataset 1,” comprised data obtained from two medical centers. Clinical 1 included 222 NEs, 413 KCs, and 32 PKCs. Clinic 2 included 60 NEs, 47 KCs, and 5 PKCs. The second test set, Dataset 2, encompassed data from three centers. It comprises data from Clinic 1 and Clinic 2 that were already present in Dataset 1, as well as 64 NEs, 51 KCs, and 6 PKCs from Clinic 3.
Given the limited availability of labeled data, we employed transfer learning to minimize the need for additional data when extending the model to adapt to data from new clinics (16). By providing a small sample from a new clinic, the pre-trained model can adapt quickly and maintain robust performance if deployed on larger, unseen datasets. As we repeated this process for additional clinics, the model progressively improved its versatility and robustness, despite the scarcity of data.
The Belin-Ambrósio enhanced ectasia display total deviation
The Belin-Ambrósio enhanced ectasia display total deviation (BAD-D) score was used as a diagnostic index by the Pentacam HR system. According to the Pentacam HR system, normal eyes are distinguished from the KC or KC suspects when the total D < 1.6, the standard have been used in various studies related to keratoconus or corneal ectasia (17–20). Ambrósio et al (17) used the index (total D > 1.6) as the risk factors for post-LASIK ectasia development. Herber et al. (18), Wang et al. (19) and Koc et al. (20) used the index (total D < 1.6) to help distinguish normal eyes from keratoconic eyes. In our study, cases from test sets with the total D < 1.6 were labelled “normal” as the diagnosis results of BAD-D. Finally, we compared the classification results of test sets of Model 2 with the BAD-D by the area under the receiver operating characteristic (ROC) curve (AUC). The sensitivity and specificity were also calculated on the basis of the test results of the Model 2 and the BAD-D to compare the two diagnostic tools.
Reading protocol and study groups
Two expert ophthalmologists were invited to make standard annotations for the images. Five junior ophthalmologists were trained and read the images for assessment in the study. The junior ophthalmologists included in the research were ophthalmologists who have accepted training of ophthalmology <5 years. Five junior ophthalmologists were selected from the Clinic 1. Before the formal annotation, they were trained on label principles. The diagnostic criteria of the disease and their clinical characteristics were not included in the training to maintain the real diagnostic capacity of junior ophthalmologists in clinical practice. After the training phase, each of the junior ophthalmologists was assigned to annotate portions of the datasets independently as the control group; specifically, each ophthalmologist annotated images obtained from the same test set as the ML model in Clinic 1. Then, they annotated the same groups of images, attached with labels previously annotated by the ML model, forming the test group. At last, all the junior ophthalmologists were request to sign up a questionnaire to evaluate the assistance of the ML model.
Statistical analyses
Data distribution was evaluated using the Kolmogorov–Smirnov goodness-of-fit test. Comparisons between groups were performed using the nonparametric Kruskal–Wallis test, followed by the post-hoc Dunn’s test to compare each pair of groups. We also computed confusion matrices and accuracy to objectively compare the performance and quality of learning. DeLong’s test was performed to pairwise compare different AUCs, and the binomial exact test was used to calculate 95% confidence intervals. Statistical analyses were performed using SPSS version 20 (IBM, Armonk, NY, United States).
Results
Study population
The mean age was 27.5 ± 7.6 years in the KC group and 28.1 ± 6.9 years in the NE group (p < 0.05). The mean K1 was 46.2 ± 6.0 in the KC group and 42.9 ± 1.4 in the NE group (p < 0.05). The mean K2 was 49.4 ± 7.0 in the KC group and 44.3 ± 1.6 in the NE group (p < 0.05). The mean value for astigmatism was 3.1 ± 2.5 in the KC group and 1.4 ± 0.8 in the NE group (p < 0.05). The mean Pachy min value was 463.8 ± 53.9 in the KC group and 534.8 ± 29.7 in the NE group (p < 0.05). The mean I-S (inferior–superior) was 4.4 ± 3.9 in the KC group and 0.3 ± 0.7 in the NE group (p < 0.05). The parameters for the different sets are listed in Table 1. In cases of severe KC, the K value and Pachy min values were unusual because the cornea was extremely distorted (Supplementary Figures S1–S3).
ML model characteristics
The accuracy of the ML model between the eyes of patients with KC and NEs in all three clinics (99% accuracy, area under the receiver operating characteristic (ROC) curve AUC of 1.00, 99% sensitivity, 99% specificity). The accuracy of the ML model between eyes with PKC and NEs in all three clinics (98% accuracy, AUC of 0.96, 98% sensitivity, 98% specificity). The sensitivity and specificity of ML model is higher than that of the previous studies (21–30) shown in Table 2.
The BAD-D is a widely used reference for assessing the KC (31). It calculates a total deviation value (known as a global “D”) by analyzing specific parameters and comparing them with a normative database. Our dataset was classified into NE and KC groups based on the BAD-D criteria, with a cut-off threshold of 1.6 (26). The BAD-D achieved an accuracy of 0.86, a sensitivity of 0.97, and a specificity of 0.69. The AUC of the BAD-D was 0.97, which was lower than that of model 2 (Figure 2). In comparison, Model 2 demonstrated significantly improved classification metrics compared to BAD-D. It achieved a 13% higher accuracy, 2% higher sensitivity, and a substantial 30% improvement in specificity. The higher specificity of model 2 indicated a lower false-positive rate. These results highlight the superior discriminative capability of Model 2 relative to the BAD-D, particularly for challenging borderline cases.
To assess the effectiveness of our model in identifying cases of PKC accurately, we further divided the results into two classes (PKC and KC) and demonstrated the classification results of PKC and NE using Models 2 and BAD-D. Model 2 accurately categorized 42 of the 43 PKCs in Dataset 2. In contrast, BAD-D correctly identified only 29 of 43 PKC cases, indicating its limited ability to discriminate early-stage disease. In addition, Model 2 achieved high accuracy in classifying the NE cases. In comparison, the BAD-D misclassified 106 NE cases. The AUC of the BAD-D was 0.73, which was significantly lower than that of Model 2 (Figure 3). Model 2 demonstrated superior discriminative ability over the BAD-D in distinguishing challenging borderline PKC and NE cases.
Figure 3. ROC curve of the model 2 (Left). ROC curve of the BAD-D for the classification of PKC and NE (Right).
Post-hoc analysis
The accuracy of the ML model between the eyes of patients with KC and NEs in clinic 1 was 98.74 (97.86% sensitivity, 100% specificity). The accuracy of the ML model between the eyes of patients with PKC and NEs in clinic 1 was 96.85 (72.72% sensitivity, 100% specificity). The diagnostic accuracy for junior ophthalmologists to detect PKC was 47.5% (95%CI, 0.5–71.6%), 100% (95%CI, 100–100%) and 94.4% (95%CI, 14.7–94.7%) in the control group, ML model group and test group. With the ML model training, the diagnostic accuracy of junior ophthalmologists improved with statistical significance (p < 0.05) as shown in Figure 4. According to the questionnaire, the average score (total 5) of the questionnaire was: 1. How do you think AI models help you understand the complexity of image judgment in PKC? (3.8) 2. How do you think AI models help you understand the complexity of image judgment in severe keratoconus? (4) 3. How comprehensive do you think the AI model has been in learning your keratoconus diagnosis? (4) 4. How convenient do you think the AI model is for you to learn the diagnosis of keratoconus? (4.4) 5. How do you think AI models help you learn to judge keratoconus as a whole? (3.6).
Figure 4. The diagnostic accuracy of the five junior ophthalmologists (1–5) of PKC before and after the training of ML model.
Typical PKC cases
The Pentacam HR images of five typical patients with PKC are shown in Figure 5. These cases were difficult to detect as PKC, and they were misdiagnosed as NEs by BAD-D. However, they were correctly identified using ML model.
Discussion
Detection of PKC is a difficult clinical problem. Researchers have developed various models based on ML to aid the diagnosis of PKC, but the results have been unsatisfactory (21–27, 29).
In some studies, sensitivity and specificity were relatively high. However, this study had several limitations. In a study by Ambrósio et al. (28), although the test results were high, they used Pentacam combined with Corvis®, which is not commonly used worldwide, thereby limiting the application of the model. The study by Shi et al. (30), they had the same limitations. They used a prototype of ultra-high-resolution optical coherence tomography, which has not been used in clinical practice, thereby endowing their model with limited practical value.
The model was trained by the Efficientnet-b0 net, which architectures have been confirmed to achieve competitive accuracy with fewer parameters, owing to their sample efficiency resulting from balanced scaling. From the review of the previous study, we can see that the sensitivity of the machine learning model to detect PKC is relatively higher. Also, rather than using multiple diagnosis devices, the model only utilized pentacam HR system to screen PKC, making the model with more practicability. We aimed to make the model more applicable because the Pentacam HR is commonly used to detect PKC by refractive surgeons globally.
By combining KC and PKC cases from three clinics, we included 551 KC eyes and 101 PKC eyes. Hence, this is one of the largest ML studies on KC and PKC. Depending only on Pentacam HR images, the model of our study could reach 99% accuracy, 99% sensitivity, 99% specificity, and an AUC of 1.00 to distinguish KC and NEs, much higher than those for BAD-D (86% accuracy, 97% sensitivity, 69% specificity, AUC of 0.97). The advantage of our model is the more prominent distinction between PKC and NEs. In this regard, the model could reach 98% accuracy, 98% sensitivity, and 98% specificity, with an AUC of 0.96, which was remarkably higher than those achieved using the BAD-D (69% accuracy, 67% sensitivity, 69% specificity, AUC of 0.73), making it one of the best ML models to distinguish PKC from NEs.
In contrast to a conventional training strategy (32, 33), we used a pre-trained model optimized using the TAO Toolkit. Our model was trained on a small (yet evenly balanced) training set to achieve the highest accuracy for the classification of PKC and NEs. The model was trained with only 25 KCs, 64 NEs, and 39 PKCs, modified in the validation set of 15 KCs, 34 NEs, and 19 PKCs, but obtained 99% accuracy and an AUC of 1.00 in a much larger test set (511 KCs, 346 NEs, 43 PKCs). The traditional training pattern usually uses 80% of the cases to train the model and tests only 20% of the cases. The pre-trained model that we describe can test much larger datasets by only training very small data, making the model more widely applicable in clinical practice.
In the research, with the help of the proposed ML model, the diagnostic accuracy of junior ophthalmologists to detect PKC improved significantly, indicating that the model may be potential tool for diagnosing PKC in clinic. Compared with the traditional training model, the model can achieve high accuracy with minimum training data. In addition, the model relies only on Pentacam HR images, making it very valuable for practical applications.
The machine learning model can provide high-level detection of PKC, making it potential assistant of refractive surgeons to screen qualified candidates for corneal refractive surgery. The model is of favourable practicability, since it can provide screening results only by using single diagnosis device pentacam HR system. However, more independent data from different clinics should be test to confirm the stability of the machine learning model.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by the Review Board and Human Ethics Committee of Peking Union Medical College Hospital. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants' legal guardians/next of kin in accordance with the national legislation and institutional requirements.
Author contributions
YJ: Writing – original draft. HJ: Writing – original draft. JZ: Writing – review & editing. TC: Writing – review & editing. YL: Writing – review & editing. YZ: Writing – review & editing. YC: Writing – review & editing. FL: Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Acknowledgments
The TAO Toolkit 3.0 (NVIDIA, Santa Clara, CA, United States) was used to train our model for the classification of keratoconus and normal eyes.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2024.1458356/full#supplementary-material
References
1. Li, X, Yang, H, and Rabinowitz, Y. Longitudinal study of keratoconus progression. Exp Eye Res. (2007) 85:502–7. doi: 10.1016/j.exer.2007.06.016
2. Chan, E, and Chong, EW. Prevalence of keratoconus based on Scheimpflug imaging: the Raine study. Opthalmology. (2021) 128:515–21. doi: 10.1016/j.ophtha.2020.08.020
3. Ambrosio, R Jr, and Wilson, SE. Complications of laser in situ keratomileusis: etiology, prevention, and treatment. J Refract Surg. (2001) 17:350–79. doi: 10.3928/1081-597X-20010501-09
4. Klein, SR, Epstein, RJ, and Randleman, JB. Corneal ectasia after laser in situ keratomileusis in patients without apparent preoperative risk factors. Cornea. (2006) 25:388–403. doi: 10.1097/01.ico.0000222479.68242.77
5. Li, X, Rabinowitz, YS, and Rasheed, K. Longitudinal study of the normal eyes in unilateral keratoconus patients. Ophthalmology. (2004) 111:440–6. doi: 10.1016/j.ophtha.2003.06.020
6. Caporossi, A, Mazzotta, C, and Baiocchi, S. Long-term results of riboflavin ultraviolet a corneal collagen cross-linking for keratoconus in Italy: the Siena eye cross study. Am J Ophthalmol. (2010) 149:585–93. doi: 10.1016/j.ajo.2009.10.021
7. Zadnik, K, Barr, JT, and Edrington, TB. Baseline findings in the collaborative longitudinal evaluation of keratoconus (CLEK) study. Invest Ophthalmol Vis Sci. (1998) 39:2537–46.
8. Gomes, JAP, Tan, D, Rapuano, CJ, Belin, MW, Guell, JL, Malecaze, F, et al. Group of Panelists for the global Delphi panel of keratoconus and Ectatic diseases. Global consensus on keratoconus and ectatic diseases. Cornea. (2015) 34:359–69. doi: 10.1097/ICO.0000000000000408
9. McMahon, TT, Szczotka-Flynn, L, Barr, JT, Anderson, RJ, Slaughter, ME, Lass, JH, et al. A new method for grading the severity of keratoconus: the keratoconus severity score (KSS). Cornea. (2006) 25:794–800. doi: 10.1097/01.ico.0000226359.26678.d1
10. Zhe, X, Feng, R, and Jin, X. Evaluation of artificial intelligence models for the detection of asymmetric keratoconus eyes using Scheimpflug tomography. Clin Experiment Ophthalmol. (2022) 50:714–23. doi: 10.1111/ceo.14126
11. Torralba, A, Fergus, R, and Freeman, WT. 80 million tiny images: a large data set for nonparametric object and scene recognition. IEEE Trans Pattern Anal Mach Intell. (2008) 30:1958–70. doi: 10.1109/TPAMI.2008.128
13. NVIDIA . TAO Toolkit. (2022). Available at: https://docs.nvidia.com/tao/tao-toolkit/text/guide.html (Accessed June 1, 2022).
14. Tan, M, and Le, QV. Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. Los Angeles, CA: PMLR (2019) 6105–6114.
15. Noord, V, Nanne,, and Postma, E. Learning scale-variant and scale-invariant features for deep image classification. Pattern Recogn. (2017) 61:583–92. doi: 10.1016/j.patcog.2016.06.005
16. Pan, SJ, and Yang, Q. A survey on transfer learning. IEEE Trans Knowl Data Eng. (2009) 10:1345–59.
17. el-Naggar, MT, Elkitkat, RS, Ziada, HE-D, Esporcatte, LPG, and Ambrósio Jr, R. Assessment of preoperative risk factors for post-LASIK ectasia development. Clin Ophthalmol. (2023) 17:3705–15. doi: 10.2147/OPTH.S425479
18. Herber, R, Hasanli, A, Lenk, J, Vinciguerra, R, Terai, N, Pillunat, LE, et al. Evaluation of corneal biomechanical indices in distinguishing between Normal, very asymmetric, and bilateral Keratoconic eyes. J Refract Surg. (2022) 38:364–72. doi: 10.3928/1081597X-20220601-01
19. Wang, Y, Liyan, X, Wang, S, Yang, K, Yuwei, G, Fan, Q, et al. A hospital-based study on the prevalence of keratoconus in first-degree relatives of patients with keratoconus in Central China. J Ophthalmol. (2022) 2022:6609531. doi: 10.1155/2022/6609531
20. Koc, M, Tekin, K, Tekin, MI, Uzel, MM, Kosekahya, P, Ozulken, K, et al. An early finding of keratoconus: increase in corneal densitometry. Cornea. (2018) 37:580–6. doi: 10.1097/ICO.0000000000001537
21. Bühren, J, Schäffeler, T, and Kohnen, T. Validation of metrics for the detection of PKC in a new patient collective. J Cataract Refract Surg. (2014) 40:259–68. doi: 10.1016/j.jcrs.2013.07.044
22. Chan, C, Ang, M, and Saad, A. Validation of an objective scoring system for Forme Fruste keratoconus detection and post-LASIK ectasia risk assessment in Asian eyes. Cornea. (2015) 34:996–1004. doi: 10.1097/ICO.0000000000000529
23. Kovács, I, Miháltz, K, and Kránitz, K. Accuracy of machine learning classifiers using bilateral data from a Scheimpflug camera for identifying eyes with preclinical signs of keratoconus. J Cataract Refract Surg. (2016) 42:275–83. doi: 10.1016/j.jcrs.2015.09.020
24. Saad, A, and Gatinel, D. Combining Placido and Corneal Wavefront data for the detection of Forme Fruste keratoconus. J Refract Surg. (2016) 32:510–6. doi: 10.3928/1081597X-20160523-01
25. Hidalgo, IR, Rodriguez, P, and Rozema, JJ. Evaluation of a machine-learning classifier for keratoconus detection based on Scheimpflug tomography. Cornea. (2016) 35:827–32. doi: 10.1097/ICO.00000000000000834
26. Hidalgo, IR, Rozema, JJ, and Saad, A. Validation of an objective keratoconus detection system implemented in a Scheimpflug Tomographer and comparison with other methods. Cornea. (2017) 36:689–95. doi: 10.1097/ICO.00000000000001194
27. Zhe, X, Li, W, and Jiang, J. Characteristic of entire corneal topography and tomography for the detection of sub-clinical keratoconus with Zernike polynomials using Pentacam. Sci Rep. (2017) 7:16486. doi: 10.1038/s41598-017-16568-y
28. Jr, RA, Lopes, BT, and Faria-Correia, F. Integration of Scheimpflug-based corneal tomography and biomechanical assessments for enhancing ectasia detection. J Refract Surg. (2017) 33:434–43. doi: 10.3928/1081597X-20170426-02
29. Steinberg, J, Siebert, M, and Katz, T. Tomographic and biomechanical Scheimpflug imaging for keratoconus characterization: a validation of current indices. J Refract Surg. (2018) 34:840–7. doi: 10.3928/1081597X-20181012-01
30. Shi, C, Wang, M, and Zhu, T. Machine learning helps improve diagnostic ability of PKC using Scheimpflug and OCT imaging modalities. Eye Vis. (2020) 7:48. doi: 10.1186/s40662-020-00213-3
31. Belin, MW, and Khachikian, SS. Keratoconus/ectasia detection with the oculus pentacam: Belin/Ambrósio enhanced ectasia display. Highlights Ophthalmol. (2007) 21:5–12.
32. Gao, H-B, Pan, Z-G, and Shen, M-X. Kerato screen: early keratoconus classification with Zernike polynomial using deep learning. Cornea. (2022) 41:1158–65. doi: 10.1097/ICO.0000000000003038
Keywords: machine learning, preclinical keratoconus, training, education, keratoconus
Citation: Jiang Y, Jiang H, Zhang J, Chen T, Li Y, Zhou Y, Chen Y and Li F (2024) The development of a machine learning model to train junior ophthalmologists in diagnosing the pre-clinical keratoconus. Front. Med. 11:1458356. doi: 10.3389/fmed.2024.1458356
Edited by:
Yong Tao, Capital Medical University, ChinaReviewed by:
Jiaxu Hong, Fudan University, ChinaCopyright © 2024 Jiang, Jiang, Zhang, Chen, Li, Zhou, Chen and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Ying Li, liyingpumch@126.com; Yuehua Zhou, yh0220@yahoo.com; Youxin Chen, chenyouxinpumch@163.com
†These authors share first authorship