Skip to main content

ORIGINAL RESEARCH article

Front. Med., 06 October 2023
Sec. Dermatology
This article is part of the Research Topic Artificial Intelligence in Cutaneous Lesions: Where do we Stand and What is Next? View all 11 articles

Development and validation of an artificial intelligence-powered acne grading system incorporating lesion identification

\r\nJiaqi Li,,&#x;Jiaqi Li1,2,3†Dan Du,,&#x;Dan Du1,2,3†Jianwei Zhang,Jianwei Zhang3,4Wenjie Liu,Wenjie Liu3,4Junyou Wang,Junyou Wang3,4Xin Wei,Xin Wei3,4Li Xue,Li Xue1,2Xiaoxue Li,Xiaoxue Li1,2Ping Diao,Ping Diao1,2Lei Zhang,Lei Zhang3,4Xian Jiang,,*Xian Jiang1,2,3*
  • 1Department of Dermatology, West China Hospital, Sichuan University, Chengdu, China
  • 2Laboratory of Dermatology, Frontiers Science Center for Disease-related Molecular Network, West China Hospital, Clinical Institute of Inflammation and Immunology, Sichuan University, Chengdu, China
  • 3Med-X Center for Informatics, Sichuan University, Chengdu, China
  • 4College of Computer Science, Sichuan University, Chengdu, Sichuan, China

Background: The management of acne requires the consideration of its severity; however, a universally adopted evaluation system for clinical practice is lacking. Artificial intelligence (AI) evaluation systems hold the promise of enhancing the efficiency and reproducibility of assessments. Artificial intelligence (AI) evaluation systems offer the potential to enhance the efficiency and reproducibility of assessments in this domain. While the identification of skin lesions represents a crucial component of acne evaluation, existing AI systems often overlook lesion identification or fail to integrate it with severity assessment. This study aimed to develop an AI-powered acne grading system and compare its performance with physician image-based scoring.

Methods: A total of 1,501 acne patients were included in the study, and standardized pictures were obtained using the VISIA system. The initial evaluation involved 40 stratified sampled frontal photos assessed by seven dermatologists. Subsequently, the three doctors with the highest inter-rater agreement annotated the remaining 1,461 images, which served as the dataset for the development of the AI system. The dataset was randomly divided into two groups: 276 images were allocated for training the acne lesion identification platform, and 1,185 images were used to assess the severity of acne.

Results: The average precision of our model for skin lesion identification was 0.507 and the average recall was 0.775. The AI severity grading system achieved good agreement with the true label (linear weighted kappa = 0.652). After integrating the lesion identification results into the severity assessment with fixed weights and learnable weights, the kappa rose to 0.737 and 0.696, respectively, and the entire evaluation on a Linux workstation with a Tesla K40m GPU took less than 0.1s per picture.

Conclusion: This study developed a system that detects various types of acne lesions and correlates them well with acne severity grading, and the good accuracy and efficiency make this approach potentially an effective clinical decision support tool.

Introduction

Acne vulgaris is the eighth most prevalent disease affecting 9.4% of the global population (1). Although acne can occur at all ages, adolescents are the most prevalent group of acne sufferers, and eighty-five percent of adolescents are affected by acne (2). As a condition that alters appearance, acne affects patients’ physical and psychological well-being and causes a strong desire for treatment (3). The large patient population and the strong desire for treatment seriously burden healthcare resources (4, 5). Assessment of acne severity is essential for the patient’s stepwise therapy. There are more than 20 published scales for evaluating acne, but none is adopted universally for clinical practice (6).

Most scales can be classified as lesion-counting scales or text description scales. Lesion counting scales correspond to the severity by measuring different types of acne lesions, such as the Global Acne Grading System (7, 8). Counting acne lesions is supposed to be a more objective method. However, it shows a high degree of variability between raters due to ambiguity between different categories of skin lesions and interevaluator differences in the definition of skin lesions (9). In addition, a single counting process ignores the degree of inflammation, postinflammatory hyperpigmentation, scarring, and other features that affect the severity. In contrast to quantitative scales, qualitative scales distinguish between different levels of severity through textual descriptions. Although qualitative scales require more clinical experience from the evaluator, they simplify the tedious counting process to a certain extent and take care of other acne characteristics beyond the number of lesions. For example, Investigator Global Assessment classifies acne into five levels through text descriptions (clear, almost clear, mild, moderate, severe, and very severe) (10). On this basis, a recent study found that replacing the qualitative labels with the corresponding treatment intensity labels effectively reduced the high interrater variability, although these labels are more unstable since treatment options may change depending on regional perceptions and disciplinary developments (11).

Artificial intelligence (AI) for acne grading has been considered a promising research direction to increase the consistency and efficiency of assessment. Some AI systems focus on identifying and counting different types of lesions, but as with lesion-counting scales, they ignore considerable information beyond the countable lesions (12, 13). Other AI systems analyze the image as a whole but leave the evaluation free from clinical interpretability (14, 15). We believe that the quantity of different types of lesions is an inadequate but crucial component of acne severity assessment. Therefore we sought to develop a novel AI system that could integrate the identification and counting of skin lesions into the overall facial evaluation process, thereby improving the predictive accuracy.

Materials and methods

Database

This study was conducted at sichuan university from January, 2020 to June, 2022, and was approved by the west china hospital institutional review board to use the patients’ deidentified images and records. This study followed the declaration of Helsinki and standards for reporting of diagnostic accuracy (STARD) reporting guidelines and the checklist for evaluation of image-based artificial intelligence algorithm reports in dermatology (CLEAR Derm) (16). We collected records of 3,098 visits to our dermatology specialist clinics with a diagnosis of acne without other inflammatory skin disease diagnoses. Of the 3,098 visits recorded, 1,501 had corresponding standardized pictures obtained via the VISIA system, including frontal, left and right profile photos, and information from these visits was included in the current study. To select labeling experts for the database and to evaluate the adequacy of the standardized frontal photo, 40 patients with acne (10 mild, 20 severe, 10 severe) were selected based on clinical records. seven experienced dermatologists first rated the frontal photos of the 40 patients, and the three evaluators with the highest average linear weighted Cohen’s κ were selected to complete the severity marking of the 1,461 records. The median of their ratings was considered the true label. After disrupting the order of the 40 images, the 7 dermatologists again rated the combined photos (frontal and left and right side photos) of the 40 patients. To improve interrater agreement, in this study we used the Treatment Intensity label to distinguish between the severity of patients (11), and due to the low number of extremely severe cases, we combined Level 8 and Level 9 (Table 1).

TABLE 1
www.frontiersin.org

Table 1. Severity label and corresponding treatment intensity list.

Development of the skin lesion identification platform

For the acne detection module, we used a publicly available deep-learning method to detect acne lesions (17). We used a VISIA complexion analysis system to photograph 276 facial images as our samples, where each sample has a resolution from 3128 × 4171 to 3456 × 5184 pixels. All the samples were split 9:1 into training samples (n = 248) and test samples (n = 28). Six dermatologists participated in annotating all the samples. A total of 15,922 skin lesions with 10 lesion categories, i.e., open comedone, closed comedone, papule, pustule, nodule/cyst, atrophic scar, hypertrophic scar, melasma and nevus were generated. Next, the network is trained by an SGD optimizer with 15 epochs, where the learning rate, momentum, and weight decay were 0.002, 0.9, and 0.0001, respectively.

Development of acne grading systems

We used ResNet50 as the training network for the baseline results (18). This network contains four large blocks, each with 3, 4, 6, and 3 small blocks, and each small block consists of three convolutional layers. In addition, the network contains jump connections to alleviate the problem of gradient explosion and gradient disappearance during training, thereby allowing the model to extract deeper features. A total of 1,185 images were used for the grading experiments, of which, 945 were used for training and 240 for testing. For the training set, all images were first resized to 256 × 256 pixels and later randomly cropped to 224 × 224 pixels to meet the input size of the network. Furthermore, the images are randomly flipped horizontally (50% probability) and randomly rotated from −20° to +20° to expand the data to prevent training overfitting. The model was trained using cross entropy loss with a total of 200 epochs and a batch size of 32. The initial learning rate was 0.001, and it decayed to 0.0001 using a cosine annealing function. The optimizer was the Adam optimizer with a weight decay of 0.0001. The training was conducted on a Tesla K40m GPU. For the acne grading task, the number of acne lesions as well as the overall assessment are an important reference for acne grading. Therefore, we propose a method that combines dermatologists’ a priori knowledge with a CNN to automatically grade pictures. The acne counts of all samples were semiautomatically labeled by the trained detection model and manually validated by an experienced dermatologist. The rule divides each image into a grading interval instead of a single grade to guide the network to better predict the image grading.

We propose two methods to integrate the proposed rules into the network, i.e., fixed weights and learnable weights, and the two methods are shown in Figure 1. For the fixed-weights approach, the probability weight of the interval is fixed. If the interval does not contain the grading, the weight is 0; otherwise, it is 1. Each input image is fed into the CNN first to learn the image features. The image features are average-pooled and mapped to an 8-dimensional vector to correspond to the probability of each classification. Then, the two vectors are multiplied by the corresponding position elements to obtain the predicted probability of each classification. Since the proposed rule reduces the weight of the intervals that do not belong to the image classification, only the predicted probability of the interval to which the image belongs is obtained. The classification corresponding to the highest probability is selected as the predicted class.

FIGURE 1
www.frontiersin.org

Figure 1. Overview of the development and validation of our AI systems. After confirming the adequacy of the frontal photo information, the doctors with the highest agreement with other peers were selected as true label raters for the remaining 1,461 frontal photos. Of the 1,461 photos, 276 were used to develop a skin lesion identification platform and 1,185 were used to develop an acne severity rating system. Then, we sought to incorporate skin lesion identification results into the severity evaluation and validated the feasibility in test set and rater selection subset.

For the learnable weights approach, the network is given an initial value, after which the weights are fine-tuned through training. As shown in Figure 2, after training, the network outputs the graded probability values and the learned interval weights. The prediction probability of each classification is obtained by multiplying the classification probabilities with the corresponding interval weights. Again, the classification with the highest probability is the grading predicted by the model.

FIGURE 2
www.frontiersin.org

Figure 2. Procedure for integrating skin lesion identification with acne severity assessment based on AI. (A) Fixed weights approach. (B) Learnable weights approach.

Statistical analysis

To determine the sample size of rater selection, assuming the interrater correlation coefficients were approximately 0.8, at least 7 raters and 40 subjects were needed. No formal sample size was calculated for validation of AI systems. Cohen’s kappa with linear weights was used to evaluate the AI’s performance against the true label or the 7 dermatologists on the rater selection dataset. A kappa value of less than 0.6 was considered unacceptably low. The statistical analyses were performed using Prism software (GraphPad Prism 8.0) and R (version 4.2.1).

Results

The database was divided into three subsets, and the baseline characteristics are summarized in Table 1. Forty records were enrolled to select the true label rater. The mean age of the 23 female and 17 male patients was 24.8 years, ranging from 16 to 39 years. Of the 560 assessments (7 raters, 40 patients and 2 rounds), each grading of severity was represented by at least 2 subjects. The evaluations obtained through the frontal photos are in good agreement with those obtained through the three-sided photos, indicating that the frontal photos are sufficiently informative as samples for the AI evaluation (Table 2). For interrater agreement of frontal photo assessment, the pairwise Cohen’s kappa for each dermatologist ranked in descending order is shown in Supplementary Figure 1, and the three raters with the greatest average kappa value were selected to rate all the photos in the database. For consistency of the assessment of frontal photographs and 3-side photographs, the overall ICC for frontal photo assessment and 3-side photograph was 0.878 (0.814, 0.916), which suggests that a frontal photograph taken with VISIA alone can yield a similar amount of information for acne as three-sided photos.

TABLE 2
www.frontiersin.org

Table 2. Baseline characteristics.

For the development of the acne lesion identification platform, 276 frontal photos were labeled by five doctors and reviewed by a senior doctor. In total, 3,060 closed pimples, 2,192 open pimples, 3,861 papules, 884 pustules, 113 nodules or cysts, 5,410 atrophic scars and 302 hypertrophic scars were marked in 276 images (Figure 3). The 276 images were divided into a training set and a test set at a ratio of 9:1. The average precision of our model for skin lesion identification was 0.507, and the average recall was 0.775, which outperformed state-of-the-art one-stage and two-stage generic object detection methods. As previously anticipated, skin lesion counts are not sufficient for severity determination, and we were not able to build a decision tree model with good performance for acne severity evaluation, either based on the number of manually annotated lesions or the number of lesions identified by the algorithm (data not shown). However, different types of lesions have different distribution patterns on the face (Supplementary Figure 2). Inflammatory lesions (papules, pustules, nodules/cysts) are more evenly distributed, and non-inflammatory lesions and secondary lesions have unique distribution characteristics. Closed acne tends to be located on the forehead and midface, while open acne tends to cluster on the forehead. Atrophic scarring is concentrated on both cheeks, while hyperplastic scarring often occurs on the skin of the lower jaw.

FIGURE 3
www.frontiersin.org

Figure 3. An example of seven types of acne-related lesions identified by the Lesion Recognition System in a patient with moderate to less severe acne.

For the development and validation of the severity grading systems, totally 945 images were used for training and 240 for testing, and the kappa obtained by the AI system relative to the true label was 0.652 (Figure 4A). To further enhance the predictive power, we further constructed a fixed-weight model a learnable-weight model to integrate the lesion identification results of papule, pustule and nodule/cyst into the severity assessment based on lesion identification platform, which improved the kappa relative to the true label to 0.737 and 0.696, respectively (Figures 4B, C). The 40 images that were initially used to select database annotators were applied to the three models, and the mean pairwise kappa achieved by the three AI models ranked 7th, 2nd and 4th (Figure 5).

FIGURE 4
www.frontiersin.org

Figure 4. Confusion matrices for acne grading. (A) Original model. (B) Fixed weight model. (C) Learnable weight model.

FIGURE 5
www.frontiersin.org

Figure 5. Acne grading performance on rater selection subsets. (A) Original model. (B) Fixed weight model. (C) Learnable weight model. Linear weighted Cohen’s kappa for each pathologist ranked from lowest to the highest. Each kappa value is the average pairwise kappa for each of the dermatologists compared with the others. The AI is highlighted with an orange dot.

Discussion

In this study, we found that the artificial intelligence acne severity evaluation system we developed produced a reasonable evaluation of the frontal part of acne patients’ photos, and its evaluation results were in good agreement with the true labels. Furthermore, we innovatively incorporated the lesion identification results into the severity evaluation with fixed weights and learnable weights, which improved the performance of the model. The AI system, whether weighted or not, can grade acne within the performance range of experienced dermatologists.

Artificial intelligence has powerful learning capabilities that enable it to capture the nuances of lesion images, including size, color and texture, etc (19). The morphological manifestation of the lesion is an important basis for diagnosing and evaluating dermatologic diseases, making AI even more distinctive in dermatology (20). Currently, AI research in dermatology is focused on multiclassification tasks (21, 22) for disease diagnosis and binary classification (23, 24) for benign or malignant skin lesions, but the evaluation of the severity of a specific disease is also a research direction with great potential for application. The high prevalence and the lack of widely accepted evaluation criteria make acne a perfect fit for AI research. As the eighth most prevalent disease in the world, acne creates a medical need that cannot be met due to the current shortage and uneven distribution of dermatologists. AI can act as a decision aid for clinicians to improve the efficiency of evaluation, particularly in the identification and counting of acne lesions. In recent years, many advances have also been made in the evaluation of acne by AI. Sophie Seité made several optimizations to their model to improve the recognition of inflammatory and non-inflammatory acne lesions, and their model achieves a GEA score similar to that of the dermatologists (13). Quan Thanh Huynh applied different models to complete the identification of acne lesions and the evaluation of severity with good accuracy, but their study did not incorporate the results of lesion identification into the severity evaluation (12). To the best of our knowledge, no previous studies have integrated skin lesion identification with severity assessment and consequently improved the accuracy of severity assessment. According to the principles of AI, skin lesion identification may no longer be important for severity evaluation when the sample size is sufficiently large, however, for more limited sample sizes, lesion identification can emphasize important information in the evaluation of severity and make the results more interpretable by doctors.

One of the major strengths of our study is that we have a much more detailed classification of severity (eight scales) than what is used by other common scales. One study found that the interobserver agreement using a crude acne severity scale was quite low (25). In order to improve interrater agreement, we referenced the treatment intensity label used by the Elena Bernardis’s study to represent acne severity (11). The physicians in this study strongly endorsed the logic of this intensity label after discussion, although it differed slightly from the current Chinese Guidelines for the Management of Acne Vulgaris and medication habits of Chinese dermatologists. The use of treatment intensity for labeling, in addition to increasing interrater consistency, provides doctors with an indication of the patient’s treatment regimen. However, the doctors will need to take into account other information about the patient as well as the results prompted by the AI, because our model does not consider patient information outside of the image data, including but not limited to pregnancy and breastfeeding status, drug allergy history, financial situation, personal wishes etc. In addition we are more rigorous in testing of the models. Besides comparing the differences between the AI model predictions and the true labels, this study compared the AI predictions with the ratings of several experienced dermatologists. This step is important for grading systems that lack objective indicators such as acne severity.

Our study also suffered from a number of shortcomings. First, all of the patients we included were Chinese, and although there were different ethnic groups, all of the patients had skin types II to IV; thus, further validation of our model’s ability to identify lesions and evaluate severity in patients with other skin types is needed. Second, our samples were sourced from hospital specialist clinics, and due to the low willingness of mild patients to seek treatment and the small proportion of patients with extremely severe illnesses, our sample is not evenly distributed at different levels. Finally, to obtain more reliable results, we included only patients with a diagnosis of acne and no other facial inflammatory diseases; however, in the real world acne is not exclusive to diseases such as rosacea and seborrheic dermatitis, and the AI evaluation for this group of patients requires a broader sample resource.

Conclusion

This study developed a system that detects various types of acne lesions and correlates them well with acne severity grading, and the good accuracy and efficiency make this approach potentially a very effective clinical decision support tool. However, further research is needed to validate the effectiveness of this AI system in real-world clinical settings.

Data availability statement

The original contributions presented in this study are included in this article/Supplementary material, further inquiries can be directed to the corresponding author.

Ethics statement

This study was approved by the West China Hospital Institutional Review Board to use the patients’ de-identified images and records.

Author contributions

JL: Data curation, Formal analysis, Investigation, Methodology, Project administration, Resources, Validation, Visualization, Writing – original draft. DD: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Validation, Writing – original draft. JZ: Methodology, Project administration, Software, Writing – original draft. WL: Methodology, Project administration, Software, Writing – original draft. JW: Methodology, Project administration, Software, Writing – original draft. XW: Methodology, Project administration, Software, Writing – original draft. LX: Conceptualization, Investigation, Writing – original draft. XL: Investigation, Writing – original draft. PD: Investigation, Writing – original draft. LZ: Software, Supervision, Writing – review and editing. XJ: Funding acquisition, Resources, Supervision, Writing – review and editing.

Funding

This manuscript was supported by grants from the Med-X Center for Informatics Funding Project, Sichuan University (YGJC-003).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2023.1255704/full#supplementary-material

References

1. Tan J, Bhate KA. Global perspective on the epidemiology of acne. Br J Dermatol. (2015) 172(Suppl. 1):3–12. doi: 10.1111/bjd.13462

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Karimkhani C, Dellavalle R, Coffeng L, Flohr C, Hay R, Langan S, et al. Global skin disease morbidity and mortality: an update from the Global Burden of Disease Study 2013. JAMA Dermatol. (2017) 153:406–12. doi: 10.1001/jamadermatol.2016.5538

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Bickers D, Lim H, Margolis D, Weinstock M, Goodman C, Faulkner E, et al. The burden of skin diseases: 2004 a joint project of the american academy of dermatology association and the society for investigative dermatology. J Am Acad Dermatol. (2006) 55:490–500. doi: 10.1016/j.jaad.2006.05.048

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Layton A, Thiboutot D, Tan J. Reviewing the global burden of acne: how could we improve care to reduce the burden? Br J Dermatol. (2021) 184:219–25. doi: 10.1111/bjd.19477

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Chilicka K, Rusztowicz M, Rogowska A, Szyguła R, Nowicka D. Efficacy of oxybrasion and cosmetic acids on selected skin parameters in the treatment with acne vulgaris. Clin Cosmet Investig Dermatol. (2023) 16:1309–17. doi: 10.2147/ccid.S407976

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Eichenfield D, Sprague J, Eichenfield L. Management of acne vulgaris: a review. JAMA. (2021) 326:2055–67. doi: 10.1001/jama.2021.17633

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Doshi A, Zaheer A, Stiller MJA. Comparison of current acne grading systems and proposal of a novel system. Int J Dermatol. (1997) 36:416–8. doi: 10.1046/j.1365-4362.1997.00099.x

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Tan J, Jones E, Allen E, Pripotnev S, Raza A, Wolfe B. Evaluation of essential clinical components and features of current acne global grading scales. J Am Acad Dermatol. (2013) 69:754–61. doi: 10.1016/j.jaad.2013.07.029

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Lucky A, Barber B, Girman C, Williams J, Ratterman J, Waldstreicher JA. Multirater validation study to assess the reliability of acne lesion counting. J Am Acad Dermatol. (1996) 35:559–65. doi: 10.1016/s0190-9622(96)90680-5

PubMed Abstract | CrossRef Full Text | Google Scholar

10. US Department of Health and Human Services, Food and Drug Administration, Center for Drug Evaluation and Research [CDER]. Guidance for Industry: Acne Vulgaris: Developing Drugs for Treatment. Washington, DC: US Department of Health and Human Services (2005)

Google Scholar

11. Bernardis E, Shou H, Barbieri J, McMahon P, Perman M, Rola L, et al. Development and initial validation of a multidimensional acne global grading system integrating primary lesions and secondary changes. JAMA Dermatol. (2020) 156:296–302. doi: 10.1001/jamadermatol.2019.4668

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Huynh Q, Nguyen P, Le H, Ngo L, Trinh N, Tran M, et al. Automatic acne object detection and acne severity grading using smartphone images and artificial intelligence. Diagnostics. (2022) 12:1879. doi: 10.3390/diagnostics12081879

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Seite S, Khammari A, Benzaquen M, Moyal D, Dreno B. Development and accuracy of an artificial intelligence algorithm for acne grading from smartphone photographs. Exp Dermatol. (2019) 28:1252–7. doi: 10.1111/exd.14022

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Yang Y, Guo L, Wu Q, Zhang M, Zeng R, Ding H, et al. Construction and evaluation of a deep learning model for assessing acne vulgaris using clinical images. Dermatol Ther. (2021) 11:1239–48. doi: 10.1007/s13555-021-00541-9

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Lim Z, Akram F, Ngo C, Winarto A, Lee W, Liang K, et al. Automated grading of acne vulgaris by deep learning with convolutional neural networks. Skin Res Technol. (2020) 26:187–92. doi: 10.1111/srt.12794

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Daneshjou R, Barata C, Betz-Stablein B, Celebi M, Codella N, Combalia M, et al. Checklist for evaluation of image-based artificial intelligence reports in dermatology: clear derm consensus guidelines from the international skin imaging collaboration artificial intelligence working group. JAMA Dermatol. (2022) 158:90–6. doi: 10.1001/jamadermatol.2021.4915

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Zhang J, Zhang L, Wang J, Wei X, Li J, Jiang X, et al. Learning High-Quality Proposals for Acne Detection. (2022). Available online at: https://ui.adsabs.harvard.edu/abs/2022arXiv220703674Z (accessed July 01, 2022).

Google Scholar

18. He K, Zhang X, Ren S, Sun J editors. Deep residual learning for image recognition. Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). London (2016).

Google Scholar

19. Huang W, Tan K, Hu J, Zhang Z, Dong S. A review of fusion methods for omics and imaging data. IEEE ACM Trans Comput Biol Bioinform. (2022) 20:74–93. doi: 10.1109/tcbb.2022.3143900

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Bajaj S, Marchetti M, Navarrete-Dechent C, Dusza S, Kose K, Marghoob A. The role of color and morphologic characteristics in dermoscopic diagnosis. JAMA Dermatol. (2016) 152:676–82. doi: 10.1001/jamadermatol.2016.0270

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Liu Y, Jain A, Eng C, Way D, Lee K, Bui P, et al. A deep learning system for differential diagnosis of skin diseases. Nat Med. (2020) 26:900–8. doi: 10.1038/s41591-020-0842-3

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Esteva A, Kuprel B, Novoa R, Ko J, Swetter S, Blau H, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature. (2017) 542:115–8. doi: 10.1038/nature21056

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Haenssle H, Fink C, Toberer F, Winkler J, Stolz W, Deinlein T, et al. Man against machine reloaded: performance of a market-approved convolutional neural network in classifying a broad spectrum of skin lesions in comparison with 96 dermatologists working under less artificial conditions. Ann Oncol. (2020) 31:137–43. doi: 10.1016/j.annonc.2019.10.013

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Haenssle H, Fink C, Schneiderbauer R, Toberer F, Buhl T, Blum A, et al. Man against machine: diagnostic performance of a deep learning convolutional neural network for dermoscopic melanoma recognition in comparison to 58 dermatologists. Ann Oncol. (2018) 29:1836–42. doi: 10.1093/annonc/mdy166

PubMed Abstract | CrossRef Full Text | Google Scholar

25. Beylot C, Chivot M, Faure M, Pawin H, Poli F, Revuz J, et al. Inter-observer agreement on acne severity based on facial photographs. J Eur Acad Dermatol Venereol. (2010) 24:196–8. doi: 10.1111/j.1468-3083.2009.03278.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: dermatology, acne, artificial intelligence, acne lesions, grading system

Citation: Li J, Du D, Zhang J, Liu W, Wang J, Wei X, Xue L, Li X, Diao P, Zhang L and Jiang X (2023) Development and validation of an artificial intelligence-powered acne grading system incorporating lesion identification. Front. Med. 10:1255704. doi: 10.3389/fmed.2023.1255704

Received: 09 July 2023; Accepted: 12 September 2023;
Published: 06 October 2023.

Edited by:

Justin Ko, Stanford University, United States

Reviewed by:

Karolina Chilicka-Hebel, Opole University, Poland
Hongxiang Chen, Huazhong University of Science and Technology, China

Copyright © 2023 Li, Du, Zhang, Liu, Wang, Wei, Xue, Li, Diao, Zhang and Jiang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Xian Jiang, amlhbmd4aWFuQHNjdS5lZHUuY24=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.