Skip to main content

ORIGINAL RESEARCH article

Front. Big Data , 19 February 2025

Sec. Medicine and Public Health

Volume 8 - 2025 | https://doi.org/10.3389/fdata.2025.1503883

This article is part of the Research Topic Deep Transfer Learning in Public Health: Opportunities for Innovation and Improvement View all articles

Next-generation approach to skin disorder prediction employing hybrid deep transfer learning

\r\nYonis Gulzar
Yonis Gulzar1*Shivani AgarwalShivani Agarwal2Saira SoomroSaira Soomro3Meenakshi KandpalMeenakshi Kandpal4Sherzod Turaev
Sherzod Turaev5*Choo W. OnnChoo W. Onn6Shilpa SainiShilpa Saini7Abdenour BounsiarAbdenour Bounsiar8
  • 1Department of Management Information Systems, College of Business Administration, King Faisal University, Al-Ahsa, Saudi Arabia
  • 2Department of Information Technology, Ajay Kumar Garg Engineering College, Ghaziabad, India
  • 3Department of Distance Continuing and Computer Education, Faculty of Education, University of Sindh, Jamshoro, Pakistan
  • 4Computer Science and Engineering, Odisha University of Technology and Research, Bhubaneswar, India
  • 5Department of Computer Science and Software Engineering, College of Information Technology, United Arab Emirates University, Al Ain, United Arab Emirates
  • 6Faculty of Data Science and Information Technology, INTI International University, Persiaran Perdana BBN, Putra Nilai, Nilai Negri Sembilan, Malaysia
  • 7Department of Computer Science and Engineering, Chandigarh University, Mohali, India
  • 8Department of Computer Science, College of Computer Sciences and Information Technology, King Faisal University, Al Hofuf, Saudi Arabia

Introduction: Skin diseases significantly impact individuals' health and mental wellbeing. However, their classification remains challenging due to complex lesion characteristics, overlapping symptoms, and limited annotated datasets. Traditional convolutional neural networks (CNNs) often struggle with generalization, leading to suboptimal classification performance. To address these challenges, this study proposes a Hybrid Deep Transfer Learning Method (HDTLM) that integrates DenseNet121 and EfficientNetB0 for improved skin disease prediction.

Methods: The proposed hybrid model leverages DenseNet121's dense connectivity for capturing intricate patterns and EfficientNetB0's computational efficiency and scalability. A dataset comprising 19 skin conditions with 19,171 images was used for training and validation. The model was evaluated using multiple performance metrics, including accuracy, precision, recall, and F1-score. Additionally, a comparative analysis was conducted against state-of-the-art models such as DenseNet121, EfficientNetB0, VGG19, MobileNetV2, and AlexNet.

Results: The proposed HDTLM achieved a training accuracy of 98.18% and a validation accuracy of 97.57%. It consistently outperformed baseline models, achieving a precision of 0.95, recall of 0.96, F1-score of 0.95, and an overall accuracy of 98.18%. The results demonstrate the hybrid model's superior ability to generalize across diverse skin disease categories.

Discussion: The findings underscore the effectiveness of the HDTLM in enhancing skin disease classification, particularly in scenarios with significant domain shifts and limited labeled data. By integrating complementary strengths of DenseNet121 and EfficientNetB0, the proposed model provides a robust and scalable solution for automated dermatological diagnostics.

1 Introduction

Skin diseases affect millions of people globally, affecting both physical and mental health. These conditions range from minor infections to severe chronic illnesses and encompass infections, inflammatory responses, chronic conditions, and hereditary disorders. The symptoms—such as rashes, itching, discoloration, dryness, and texture changes—can vary widely in severity and are often influenced by genetics, immune responses, environmental factors, and lifestyle choices (Li et al., 2021; Meedeniya et al., 2024). This diversity not only makes conditions hard to distinguish visually but also means that early symptoms are often similar across different disorders, leading to frequent misdiagnosis. As a result, the accurate identification of skin diseases presents significant challenges, especially in clinical settings where visual assessments are the primary diagnostic method. Effective management generally involves medical intervention, lifestyle modification, and preventive care (Alshahrani et al., 2024). Given the large global population affected, skin diseases are a significant public health concern (Behara et al., 2024).

Early diagnosis is essential for both prevention and effective treatment, but the visual complexity of many skin conditions often makes accurate diagnosis challenging, which can delay treatment and increase risks. In some cases, late diagnosis may lead to severe outcomes, including skin cancer (Gulzar and Khan, 2022; Mehmood et al., 2023). The importance of precise and early identification for reducing morbidity and mortality rates underscores the need for ongoing dermatological research (Chan et al., 2020).

Skin disorders are caused by a complex interplay of genetic, immunological, and environmental factors. As the body's largest organ, the skin acts as a protective barrier; however, it is highly susceptible to environmental stressors such as ultraviolet (UV) radiation, chemicals, and pathogens (Inthiyaz et al., 2023). Chronic skin diseases such as psoriasis and vitiligo are frequently associated with genetic predispositions (Abdallah et al., 2023), whereas immune responses play a critical role, as autoimmune reactions can cause inflammation and weakened immunity may heighten infection risks (Bucsek et al., 2018). Recent research on the skin microbiome has revealed its essential role in barrier function and infection prevention, adding another dimension to the understanding of skin health (Harris-Tryon and Grice, 2022).

Traditional diagnostic methods often rely on subjective visual assessments, which can lead to inaccuracies, especially with the diverse presentations and overlapping symptoms of skin diseases. Artificial intelligence (AI) has been used in diverse domains, including agriculture (Gulzar, 2024; Amri et al., 2024; Gulzar et al., 2023b), finance (Gulzar et al., 2023a), healthcare (Gulzar and Khan, 2022; Mehmood et al., 2023), and environmental monitoring (Malik et al., 2023), to address complex challenges and improve accuracy in decision-making processes. In dermatology, AI has emerged as a powerful tool for enhancing diagnostic precision. AI-powered image recognition algorithms can detect skin abnormalities such as eczema, psoriasis, and melanoma, offering early detection that improves patient prognosis and accessibility, particularly through tools like digital dermatoscopes and smartphone apps (Sengupta, 2023; Ye and Chen, 2023). AI also enables personalized treatment plans by analyzing patient-specific factors like medical history, genetics, and lifestyle to predict treatment efficacy and minimize adverse effects, especially for chronic conditions (Khan et al., 2023a; Anand et al., 2023; Khan et al., 2023b). In telemedicine, AI-based applications facilitate remote consultations and triage, reducing wait times and prioritizing urgent cases (Majid et al., 2023a,b). Additionally, AI contributes to drug discovery by identifying potential treatments and optimizing regimens for conditions like dermatitis, psoriasis, and acne (Rokni et al., 2024), ultimately advancing evidence-based, individualized dermatological care (Meedeniya et al., 2024; Jain et al., 2024).

Recent advancements have shown that AI can further elevate the precision and effectiveness of skin disease management. By leveraging cutting-edge architectures, AI-based models can address the inherent challenges in classifying skin diseases, such as their visual complexity, overlapping symptoms, and variations in lesion patterns. Traditional Convolutional Neural Network (CNN) models often fall short in generalizing to new or diverse datasets, particularly when labeled data are limited. This study proposes a Hybrid Deep Transfer Learning Model (HDTLM) that combines DenseNet and EfficientNet architectures. DenseNet, known for its dense connections, captures fine-grained features, whereas EfficientNet fine-tunes these features, thereby improving computational efficiency without compromising performance. Furthermore, domain adversarial training was employed to ensure that the learned features remain relevant across different datasets, which improves the model's generalizability and robustness in real-world applications.

The proposed model was rigorously tested on various benchmark datasets, achieving a training accuracy of 98.18% and a validation accuracy of 86.68%, outperforming traditional transfer learning methods. This approach not only demonstrates the potential of AI to enhance diagnostic accuracy and reduce overfitting but also provides a promising solution for domains with limited labeled data and significant variability. By integrating AI into dermatological research, this study seeks to address key issues in skin disease classification and prediction, aiming for a scalable model that adapts to diverse clinical settings.

The contributions of this study include:

• Introducing a Hybrid Deep Transfer Learning Model (HDTLM) combining DenseNet and EfficientNet, optimized for skin disease classification.

• High training and validation accuracy surpassing traditional transfer learning methods in performance and robustness.

• Comprehensive data preprocessing techniques, including augmentation and normalization, to improve model generalizability with limited labeled data.

• Thorough benchmarking on multiple skin disease datasets to validate improved accuracy, precision, and reduced overfitting.

• Insights into overcoming challenges such as domain shifts, data imbalance, and hyperparameter tuning specific to skin disease classification tasks.

The remainder of this study is organized as follows. In Section Related work, the related works relevant to this study are reviewed and discussed. Materials and methods are detailed in Section Material and methods. Section Results and discussion describes the results and discusses the findings. This study has limitations, which are discussed in section Limitations and future work along with possible directions for future work. Section Conclusion presents the conclusion.

2 Related work

2.1 Challenges in skin disease diagnosis

Skin diseases pose significant challenges in healthcare because of their diverse appearances and the need for accurate, specific diagnoses. Traditional diagnostic methods rely heavily on dermatologists' expertise, which can be subjective, time-consuming, and resource-intensive. These challenges have driven the increasing adoption of deep learning techniques to automate skin disease classification, aiming to enhance diagnostic accuracy and efficiency. To address these challenges, numerous studies have explored the use of deep learning techniques for automating skin disease classification.

2.2 Application of deep learning in skin disease classification

Several studies have explored the application of deep learning in skin disease classification. For instance, De et al. (2024) automatically identified skin diseases using dermatoscopic imaging. The authors also mentioned that, traditionally, dermatologists manually examined pigmented skin lesions, which can be subjective and time-consuming. Agarwal and Godavarthi (2023) evaluated and compared various skin diseases in terms of cosmetics and common skin issues. The author's dataset included ~25,000 records for the eight most common skin conditions, and a convolutional neural network was used to achieve imaging performance comparable to or superior to that of humans. Naeem et al. (2024) proposed SNC_Net, which applies deep learning models that include features from dermoscopic images and handmade (HC) feature extraction approaches to improve classifier performance. Alshahrani et al. (2024) combined CNN models (DenseNet121, MobileNet, and VGG19) with Random Forest (Rf) and Feed Forward Neural Networks (FFNN) networks to obtain complex features from dermoscopy images, and the result was a hybrid system capable of early detection of various skin lesions. Behara et al. (2024) employed adaptive thresholding to extract regions of interest (ROI) and improved cancer detection accuracy through dynamic capabilities. Khan et al. (2021) employed the PAD-UFES-20 dataset, which includes six different types of imbalanced skin cancer types, to address the data imbalance using data augmentation. Wei et al. (2023) presented a convolutional neural network model for skin disease categorization using model fusion. Sharma et al. (2023) analyzed the HAM10000 dataset to assess the profitability of several Convolutional Neural Network (CNN) designs in concealing seven different types of skin lesions. In this study, we employed EfficientNets, which outperforms standard designs because of its lightweight design. Venugopal et al. (2023) implemented improved EfficientNet B4 and EfficientNet V2-M models to categorize malignant and benign skin lesions in dermoscopic images. Adegun and Viriri (2020) proposed a framework to segment and classify skin lesions to automatically detect skin cancer. The suggested structure of this problem is divided into two parts. The basic part of this network is used to analyze mixed and complex data problems. In this step, the algorithm acquires the surrounding details using decryption. Ravi (2022) proposed classifying and detecting learning-based mixed features grouped for skin cancer that consider attention costs. Shimu et al. (2022) proposed the implementation of transfer learning to six forms of skin diseases: peeling, acne, eczema, heat rash, melanoma, and cold sores. The skin conditions were classified using a Convolutional Neural Network. Sadik et al. (2023) focused on employing CNNs to become adept at skin disease recognition, similar to the renowned detective Sherlock Holmes. They wanted to know how well these architectures could perform the job. AlSuwaidan (2023) compared six popular CNNs (VGG16, EfficientNet, InceptionV3, MobileNet, NasNet, and ResNet50) to determine which one was the best at guessing the top three skin issues in the Middle East. They cleaned up the images a bit with some filtering and denoising to ensure the models had the best view. Jaisakthi et al. (2023) developed a Deep Convolutional Neural Network (DCNN)-based model to cartegorize skin cancer into melanoma and non-melanoma. Milantev et al. (2020) conducted the ISIC Skin Lesion Classification Challenge using dermoscopic images and patient information to study skin lesions. Tahir et al. (2023) developed DSCC_Net, which uses a CNN to classify skin cancer and evaluates it on three big datasets (ISIC 2020, HAM10000, and DermIS). Gupta et al. (2020) presented a system that employs transfer learning with pretrained models to improve the results. Furthermore, this study used a CNN to precisely detect and categorize skin cancer. Pham et al. (2020) suggested a mixed strategy to address class disparities in skin disease categorization. This process combined data-level balanced mini-batch logic with real-time image augmentation and algorithm-level generation of new loss functions. Abbasi et al. (2024) proposed a modified VGG16-based algorithm to distinguish real and AI-generated medical images. The model was trained and fine-tuned using hyperparameter tuning on a dataset of 10,000 synthetic skin lesion images produced by a GAN. It distinguished real images from AI-generated images with 99.82% accuracy. Gamage et al. (2023) trained an Xception-based model for melanoma classification using the HAM10000 dataset. Bayesian hyperparameter optimization was employed alongside Grad-CAM/Grad-CAM++ heatmaps for improved explainability, where the model reached an accuracy of 90.24% while providing critical input regions that influence model predictions.

2.3 Summary of related work

Table 1 summarizes the key contributions, findings, and limitations of recent studies. These studies highlight the significant progress made in applying deep learning to skin disease classification while also identifying areas for improvement, such as addressing overfitting, improving data diversity, and ensuring computational efficiency.

Table 1
www.frontiersin.org

Table 1. Contributions and limitations of recent skin disease classification studies.

3 Material and methods

As shown in Figure 1, the development of a skin disease prediction model using transfer learning with DenseNet121 and EfficientNetB0 required a structured, methodical approach. This approach encompassed multiple stages, including data collection, data preprocessing, data augmentation, model selection, model training, evaluation, and comprehensive analysis. Each step was meticulously executed to ensure robustness and reliability in predicting skin diseases across various categories. This section provides a detailed breakdown of the research methodology employed in constructing the model, emphasizing the integration of transfer learning to enhance model generalization and performance on limited domain-specific data.

Figure 1
www.frontiersin.org

Figure 1. Research framework.

3.1 Dataset and data augmentation

Following a thorough review of existing skin disease datasets, we selected a public resource that provides extensive coverage of 19 skin conditions (Skin Diseases Dataset, 2024),. including Vitiligo, Psoriasis, Acne, Actinic Carcinoma, Atopic Dermatitis, Cellulitis, Eczema, Drug Eruptions, Herpes HPV, Light Diseases, Lupus, Melanoma, Poison Ivy, Benign Tumors, Systemic Disease, Ringworm, Urticarial Hives, Vascular Tumors, Vasculitis, and Viral Infections. Each condition was represented by sufficient images for both training and testing, thereby enhancing the dataset's applicability to dermatological research. Figure 2 shows sample images from the dataset, demonstrating its diversity.

Figure 2
www.frontiersin.org

Figure 2. Sample images of the dataset.

We employed a systematic and targeted data augmentation strategy to enhance the robustness and generalizability of the dataset. This approach improved the diversity and representativeness of the dataset while addressing class imbalances in underrepresented categories.

Image augmentations included random 30-degree rotations, zooms in or out by 20%, and 10%-20% horizontal or vertical shifts. These augmentations represented the common real-world variation in image posture, scale, and location, facilitating a broadly varied dataset that accurately depicts realistic conditions (Ayoub et al., 2023). We also applied horizontal flipping to add more variance. These augmentation techniques were implemented by utilizing TensorFlow's ImageDataGenerator to implement real-time augmentation during the preparation of the dataset. This not only maintained the original image quality but also allowed for uniform expansion of the dataset.

Particular focus was given to classes where the number of original images was limited, specifically classes with 5, 16, 37, 26, 29, and 46 samples. For these classes, the augmentation process was repeated numerous times by using different combinations of transformations to accrete a much larger dataset. This allowed a more proportionate representation of each class, reducing the bias and creating a more robust diversity in the overall dataset. This ensured that the dataset was broadening its representativeness of clinical variability found in the real world by generating more samples for rare conditions, which made analyses performed later much more robust. Table 2 presents the number of images per class before and after augmentation and divides the augmented dataset into training, validation, and testing subsets based on a 70:15:15 ratio.

Table 2
www.frontiersin.org

Table 2. Number of images per class before and after data augmentation.

3.2 Data preprocessing

To standardize the dataset for consistent model input, several preprocessing steps were applied. Images were resized to a uniform 128 × 128 pixels, and color values were adjusted to ensure consistency across samples. Pixel values were normalized to a 0–1 scale, improving training stability and convergence speed. The labels were one-hot encoded to format them appropriately for classification tasks, thereby enabling the model to distinguish between different classes effectively (Ayoub et al., 2022).

For model evaluation, the dataset was divided into training, validation, and testing sets, with 70% of the data allocated to training and 15% each for validation and testing. This balanced split supports robust model training, validation, and testing, ensuring generalizability across unseen data.

3.3 Proposed model

The proposed model for skin disease prediction combines the strengths of DenseNet121 and EfficientNetB0 through a hybrid transfer learning approach. Each architecture was selected for its unique capabilities, resulting in a robust model that could handle the complexity and variability of skin disease images.

DenseNet121: Known for its dense connections, this architecture reuses features by directly connecting each layer to subsequent layers. The layer output l can be expressed mathematically as follows:

xi=Hi([x0, x1, xl-1])

where Hi represents a composite function of operations such as batch normalization, ReLU activation, and convolution, and x0, x1, xl−1 is the concatenated output of all previous layers. This dense connectivity ensures efficient feature reuse and mitigates the vanishing gradient problem, enabling the network to learn fine-grained features that are essential for distinguishing between similar skin conditions.

EfficientNetB0: This architecture is designed with computational efficiency in mind, employing compound scaling to balance depth, width, and resolution. The scaling approach is defined as:

depth :d=αφ,   width :w=aβφ,  resolution :r=γφ

where φ is a user-defined scaling coefficient, and α, β, and γ are constants determined via a grid search. This allows EfficientNetB0 to maintain a manageable size while capturing high-level features effectively.

By combining DenseNet121 and EfficientNetB0, the hybrid model leverages DenseNet121′s capability for detailed pattern recognition and EfficientNetB0′s efficiency, resulting in a scalable and accurate architecture suitable for real-time applications.

3.3.1 Integration strategy and additional layers

After extracting features from both DenseNet121 and EfficientNetB0, the outputs of their final layers are concatenated as follows:

Fconcat=[FDenseNet, FEficientNet]

This operation combines detailed and high-level features into a single comprehensive representation.

To refine this combined feature set, we added:

1. Dense Layer: A fully connected layer with 256 units:

y= σ (wx+b)

where WW is the weight matrix, y is the input feature vector, b is the bias term, and σ is the ReLU activation function. This layer captures the interactions between the combined features.

2. Dropout Layer: Regularization at a rate of 0.5:

yi={xip,0, if the neuron is kept

where p is the probability of retaining a neuron during training. This mitigates overfitting, especially for underrepresented classes.

3. Softmax Layer: Final output layer for classification:

P(y=kx)=ezkj=1Kezj

where zk is the logit of class k, and K is the total number of classes (19 in this case).

These layers ensure that the model learns non-linear decision boundaries, handles feature interactions effectively, and outputs probabilities for each class, thereby enabling confident predictions.

3.3.2 Training strategy and benefits of the hybrid model

The hybrid model was trained as follows:

• Optimizer: Adam

• Learning Rate: 0.001

• Batch Size: 3,232

To ensure effective learning, the initial layers of DenseNet121 and EfficientNetB0 were frozen during the first phase of training to retain the generic features learned from ImageNet. This can be expressed as:

Frozen layers : Wpretrained  and fine-tuned layers : Wtrainable 

Freezing these layers allows the model to efficiently adapt to the dermatological domain during fine-tuning.

The architecture of the proposed model is shown in Figure 3.

Figure 3
www.frontiersin.org

Figure 3. Architectural diagram of the proposed hybrid skin disease prediction model.

3.4 Experimental environment settings and performance evaluation metrics

The proposed model was developed using Python 3.8, OpenCV 4.7, and the Keras library 2.8 for model building and image processing. The development environment was run on Windows 10 Pro with the following hardware configuration: Intel i5 processor (2.9 GHz), Nvidia RTX 2060 GPU, and 16 GB RAM.

To evaluate the model's performance, we used standard evaluation metrics, including accuracy, precision, recall, and the F1 score. These metrics provide a comprehensive assessment of the model's predictive accuracy and robustness, which is particularly important for multi-class classification tasks in skin disease prediction.

4 Results and discussion

4.1 Training and validation performances

In this section, we evaluate the performance of the proposed hybrid model for skin disease prediction using training and validation accuracy and training and validation loss. These metrics provide insights into the model's learning process, generalizability, and robustness against unseen data.

The training and validation accuracy (see Figure 4) shows a consistent increase over the 100 epochs, with the training accuracy at the last epoch reaching ~98.18%. The high training accuracy indicates that the model was able to learn complex functions that perfectly predicted the level of skin disease class from the training dataset. The validation accuracy stabilized at ~97.57%, indicating that the model generalized well on new data. The similarity between training and validation accuracy suggests that the model has not overfitted because it learned specific features of the training data while being able to perform well on images it has never seen before.

Figure 4
www.frontiersin.org

Figure 4. Hybrid model: Training and validation accuracy.

This may indicate that a specific hybrid architecture and training approach fit the given dataset well. Overall, the increasing trend of training vs. validation accuracy indicates that the model is robust and suitable for deployment. Moreover, these results emphasize the usefulness of the dropout and selective layer freezing techniques to handle overfitting.

Trends on training and validation loss curves shown in Figure 5 validate the behaviors observed for model performance. The training loss is consistently decreasing and converging to low levels, indicating successful optimization of model parameters. The validation loss also decreased, but with a few fluctuations, suggesting that the models were learning well while not overfitting significantly. The validation loss approximates the training loss in the last epoch, which indicates stable model performance and good generalizability.

Figure 5
www.frontiersin.org

Figure 5. Hybrid model: training and validation losses.

These observations validate the robustness of the hybrid model architecture and the training strategies employed, notably the dropout and selective freezing of the layers. The small difference between the training loss and validation loss indicates that the model does not overfit to unseen data.

The hybrid architecture, which combines DenseNet121′s fine-grained feature extraction with EfficientNetB0′s efficient scaling, clearly contributed to the model's high performance. The fluctuations observed in the validation accuracy and loss curves, although minimal, can be attributed to the variability in the dataset. This variability is especially common in medical imaging, in which similar visual features may appear across different classes (e.g., rashes and discoloration in multiple skin conditions). The use of dropout and selective layer freezing helped mitigate these fluctuations, thereby supporting the model's ability to generalize without compromising its learning depth.

4.2 Test set performance and confusion matrix analysis

To further assess the model's performance, we evaluated its predictions on the test set, which provided a realistic measure of its ability to generalize to unseen data. The confusion matrix for the test set (see Figure 6) provides a detailed view of how well the model performed across each skin disease class, with an overall accuracy of 97.57%. The high test accuracy underscores the robustness of the model because it consistently achieved correct classifications across most classes.

Figure 6
www.frontiersin.org

Figure 6. Confusion matrix.

The confusion matrix shown in Figure 6 demonstrates strong predictive power for the majority of classes with minimal misclassifications. Each class generally had a misclassification rate between 2.2 and 2.5%, suggesting that the model was capable of distinguishing between different skin disease categories with high precision. The low misclassification rate also indicates that the model successfully learned the nuanced features needed to separate visually similar skin conditions, which is particularly challenging in dermatology.

Within the matrix, certain classes, such as Melanoma, Psoriasis, and Vitiligo, showed very high precision and recall, as indicated by the nearly full diagonal dominance, reflecting the model's effectiveness in identifying the unique features of these conditions. However, a few classes with overlapping visual symptoms, such as Eczema and Dermatitis, experienced minor misclassifications. These errors, which generally involve one to three images per class, may stem from shared visual characteristics such as texture or color patterns. Such subtle misclassifications suggest that further data augmentation or fine-tuning can enhance the model's differentiation ability in these overlapping categories.

4.3 Model performance comparison

To evaluate the performance of the hybrid model, we compared it to state-of-the-art (SOTA) models such as DenseNet121, EfficientNetB0, VGG19, MobileNetV2, and AlexNet. As shown in Table 3, this comparison demonstrates that the proposed hybrid train-validation approach exhibits better performance for all three metrics. The adaptive hybrid model performs better than other architectures by leveraging DenseNet121′s ability to learn minimal/complex features along with EfficientB0 runtime performance, yielding an architecture with good accuracy and runtime performance for medical images.

Table 3
www.frontiersin.org

Table 3. Comparison of model performance metrics.

The use of transfer learning and extra dense layers allows the model to perform well on a small dataset. As transfer learning uses pretrained weights, the model can utilize the features learned earlier. Not only does this speed up training, but the proposed method also significantly improves the network's ability to learn complex representations and thereby increases accuracy. By integrating these techniques—a hybrid architecture to handle high-dimensional input data, transfer learning to leverage pretrained models, and a robust model design with performance-driven features—we established the hybrid model as a top-performing skin disease classifier, outperforming conventional architectures.

Although the hybrid model takes 5 h 14 min 37 s to be trained, it is a fair trade-off, especially considering VGG19, which trained for 8 h 15 min 32 s but has drastically lower performance (85% accuracy vs. 98.18%). Furthermore, its training time is only slightly worse than the standalone DenseNet121 or EfficientNetB0, regardless of having implemented both architectures and obtaining a significant jump in performance. This combination of DenseNet121 and EfficientNetB0 enabled the hybrid model to attain an outstanding accuracy rate (98.18%) along with remarkable precision (0.95), recall (0.96), and F1 Score (0.95), thus marking a revolution in skin disease classification.

The hybrid model also took 22 s to process 2,889 images during the testing, which is only a slight delay compared to lightweight models like MobileNetV2 or AlexNet. However, the slight increase in inference time was compensated by the near-optimal classification performance. Faster models that trade off large amounts of computation for time with little accuracy include AlexNet and MobileNetV2, both of which report significantly reduced accuracy on data such as medical images (80 and 83%, respectively) and are unable to handle much more complex data that may be found in the medical field. As evident from the performance metrics of the hybrid model, it has a greater ability to generalize and therefore is more reliable option for applications in high-stakes environments like clinical diagnosis.

The detailed results in Table 3 demonstrate that the hybrid model achieves an optimal trade-off between computational complexity and performance, thereby rendering it the most competitive model among the compared models. This approach is recognized as the best and most practical approach for medical image diagnosis due to its ability to provide excellent results without restricting itself from achieving predictions on bigger data, thereby providing scalable results.

Furthermore, the proposed model was benchmarked against several existing models in the literature (Table 4). The comparison includes key performance metrics such as recall, precision, and accuracy, which are essential for evaluating the effectiveness of classification models. Notably, the proposed model's performance is on par with or outperformes the majority of the existing models.

Table 4
www.frontiersin.org

Table 4. Comparison of classification accuracy with recent state-of-the-art methods.

From the table, it can be observed that the hybrid method proposed in the last row achieves the highest recall (0.9600) and precision (0.9500), although its accuracy (0.9818) is slightly lower than that of Naeem et al.'s model (0.9691). The models trained on ImageNet, such as Calderón et al. (2021), Jain et al. (2021), and Naeem et al. (2022), demonstrate competitive results in accuracy and recall, with Naeem et al.'s model achieving the highest accuracy. Interestingly, the model by Saarela and Geogieva (2022), which does not mention a pre-training method, has the lowest performance in terms of accuracy (0.8000), highlighting the importance of pre-training for enhancing model performance. In comparison to these models, the proposed model's performance in terms of recall, precision, and accuracy positions it as a strong contender in the field, showcasing the effectiveness of the model's architecture and the choice of pre-training techniques. The higher recall and precision values in the proposed model suggest better classification performance, particularly in terms of accurately distinguishing between classes and minimizing false positives.

5 Limitations and future work

Notably, the hybrid model performs well; however, it has its own limitations that must be considered. First, despite being extensive, the dataset lacks variety for rare skin conditions that could influence the model's generalizability to less frequently seen classes. The model's robustness across different lighting conditions and backgrounds was not analyzed in depth, which could influence generalizability. The dependency on transfer learning also implies that some pretrained characteristics are not fully dermatology-specific, which risks not capturing disease-specific features. Finally, the computational requirements during training restrict the model's use in low-resource devices, thus potentially limiting its availability in remote clinical scenarios.

In the future study, the dataset will be extended by sampling more varied cases and using advanced data augmentation techniques. The robust model will be further validated on external datasets, especially in various clinical settings. Additional fine-tuning to reduce misclassifications in overlapping classes may further improve their clinical applicability. Implementing the model in clinical workflows (e.g., telemedicine platform or mobile diagnostic application) can help improve the accessibility of dermatological care, especially in resource-limited areas (Gamage et al., 2024). Explainability techniques (e.g., heatmaps or saliency maps) also help build clinician trust and lead to real-world adoption. By overcoming these limitations and investigating the regulatory conditions and ethical considerations for their clinical implementation similarly, the hybrid approach can improve patient outcomes in dermatology (Gamage et al., 2024).

6 Conclusion

The increasing prevalence of skin diseases poses significant challenges to dermatology, particularly in terms of achieving accurate and timely diagnosis. To address this issue, we propose a hybrid model that combines the strengths of DenseNet121 and EfficientNetB0 through transfer learning. The model was meticulously constructed using a comprehensive dataset encompassing 19 distinct skin conditions, with systematic data augmentation and preprocessing ensuring robust feature extraction and generalization capabilities. The training results demonstrated exceptional performance, with the hybrid model achieving an accuracy of 98.18%. This high training accuracy reflects the model's effectiveness in learning complex patterns, which are essential for distinguishing between various skin disease classes. The validation results further confirmed the model's generalization ability, stabilizing ~at 97.57%, indicating a successful balance between learning and generalization without overfitting. A comparative analysis against several state-of-the-art (SOTA) models, including DenseNet121, EfficientNetB0, VGG19, MobileNetV2, and AlexNet, highlighted the superiority of the hybrid approach. The results, encapsulated in a comprehensive performance comparison table, reveal that the hybrid model outperformed all the other architectures, particularly in terms of accuracy, precision, and recall.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding authors.

Author contributions

YG: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing. SA: Conceptualization, Data curation, Investigation, Methodology, Software, Supervision, Writing – original draft, Writing – review & editing. SSo: Methodology, Supervision, Conceptualization, Formal analysis, Validation, Writing – review & editing. MK: Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Writing – original draft. ST: Conceptualization, Formal analysis, Funding acquisition, Investigation, Project administration, Resources, Validation, Visualization, Writing – original draft, Writing – review & editing. CO: Writing – original draft, Investigation, Conceptualization, Data curation, Methodology. SSa: Conceptualization, Data curation, Investigation, Methodology, Writing – review & editing. AB: Methodology, Writing – review & editing, Conceptualization.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This research was supported by the United Arab Emirates University (UAEU) Strategic Research Grant G00003676 (Fund No.: 12R136) through the Big Data Analytics Center.

Acknowledgments

The authors thank United Arab Emirates University for supporting this work through the UAEU Strategic Research Grant G00003676 and SURE+ Grant G00004748. The author also thank to Deanship of Scientific Research, Vice Presidency for Graduate Studies and Scientific Research, King Faisal University, Saudi Arabia, under the Project KFU242235.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Generative AI statement

The author(s) declare that no Gen AI was used in the creation of this manuscript.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Abbasi, S. F., Bilal, M., Mukherjee, T., Churm, J., Pournik, O., Epiphaniou, G., et al. (2024). Deep learning-based synthetic skin lesion image classification. Stud. Health Technol. Inform. 316, 1145–1150. doi: 10.3233/SHTI240612

PubMed Abstract | Crossref Full Text | Google Scholar

Abdallah, H. Y., Faisal, S., Tawfik, N. Z., Soliman, N. H., Kishk, R. M., and Ellawindy, A. (2023). Expression signature of immune-related microRNAs in autoimmune skin disease: psoriasis and vitiligo insights. Mol. Diagn. Ther. 27, 405–423. doi: 10.1007/s40291-023-00646-1

PubMed Abstract | Crossref Full Text | Google Scholar

Adegun, A. A., and Viriri, S. (2020). FCN-based densenet framework for automated detection and classification of skin lesions in dermoscopy images. IEEE Access 8, 150377–150396. doi: 10.1109/ACCESS.2020.3016651

Crossref Full Text | Google Scholar

Agarwal, R., and Godavarthi, D. (2023). Skin disease classification using CNN algorithms. EAI Endorsed. Trans. Pervasive Health Technol. 9, 1–8. doi: 10.4108/eetpht.9.4039

Crossref Full Text | Google Scholar

Alam, T. M., Shaukat, K., Khan, W. A., Hameed, I. A., Almuqren, L. A., Raza, M. A., et al. (2022). An efficient deep learning-based skin cancer classifier for an imbalanced dataset. Diagnostics 12, 2115. doi: 10.3390/diagnostics12092115

PubMed Abstract | Crossref Full Text | Google Scholar

Alshahrani, M., Al-Jabbar, M., Senan, E. M., Ahmed, I. A., and Saif, J. A. M. (2024). Analysis of dermoscopy images of multi-class for early detection of skin lesions by hybrid systems based on integrating features of CNN models. PLoS ONE 19:e0298305. doi: 10.1371/journal.pone.0298305

PubMed Abstract | Crossref Full Text | Google Scholar

AlSuwaidan, L. (2023). Deep learning based classification of dermatological disorders. Biomed. Eng. Comput. Biol. 14:1–9. doi: 10.1177/11795972221138470

PubMed Abstract | Crossref Full Text | Google Scholar

Amri, E., Gulzar, Y., Yeafi, A., Jendoubi, S., Dhawi, F., and Mir, M. S. (2024). Advancing automatic plant classification system in Saudi Arabia: introducing a novel dataset and ensemble deep learning approach. Model Earth Syst. Environ. 10, 2693–2709. doi: 10.1007/s40808-023-01918-9

Crossref Full Text | Google Scholar

Anand, V., Gupta, S., Gupta, D., Gulzar, Y., Xin, Q., Juneja, S., et al. (2023). Weighted average ensemble deep learning model for stratification of brain tumor in MRI images. Diagnostics 13:1320. doi: 10.3390/diagnostics13071320

PubMed Abstract | Crossref Full Text | Google Scholar

Ayoub, S., Gulzar, Y., Reegu, F. A., and Turaev, S. (2022). Generating image captions using bahdanau attention mechanism and transfer learning. Symmetry 14:2681. doi: 10.3390/sym14122681

Crossref Full Text | Google Scholar

Ayoub, S., Gulzar, Y., Rustamov, J., Jabbari, A., Reegu, F. A., and Turaev, S. (2023). Adversarial approaches to tackle imbalanced data in machine learning. Sustainability 15:7097. doi: 10.3390/su15097097

Crossref Full Text | Google Scholar

Behara, K., Bhero, E., and Agee, J. T. (2024). Grid-based structural and dimensional skin cancer classification with self-featured optimized explainable deep convolutional neural networks. Int. J. Mol. Sci. 25:1546. doi: 10.3390/ijms25031546

PubMed Abstract | Crossref Full Text | Google Scholar

Bucsek, M. J., Giridharan, T., MacDonald, C. R., Hylander, B. L., and Repasky, E. A. (2018). An overview of the role of sympathetic regulation of immune responses in infectious disease and autoimmunity. Int. J. Hyperthermia. 34, 135–143. doi: 10.1080/02656736.2017.1411621

PubMed Abstract | Crossref Full Text | Google Scholar

Calderón, C., Sanchez, K., Castillo, S., and Arguello, H. (2021). BILSK: a bilinear convolutional neural network approach for skin lesion classification. Comput. Methods Programs Biomed. 1:100036. doi: 10.1016/j.cmpbup.2021.100036

Crossref Full Text | Google Scholar

Chan, S., Reddy, V., Myers, B., Thibodeaux, Q., Brownstone, N., and Liao, W. (2020). Machine learning in dermatology: current applications, opportunities, and limitations. Dermatol. Ther. 10, 365–386. doi: 10.1007/s13555-020-00372-0

PubMed Abstract | Crossref Full Text | Google Scholar

De, A., Mishra, N., and Chang, H. T. (2024). An approach to the dermatological classification of histopathological skin images using a hybridized CNN-DenseNet model peer. J. Comput. Sci. 10:e1884. doi: 10.7717/peerj-cs.1884

PubMed Abstract | Crossref Full Text | Google Scholar

Fraiwan, M., and Faouri, E. (2022). On the automatic detection and classification of skin cancer using deep transfer learning. Sensors 22:4963. doi: 10.3390/s22134963

PubMed Abstract | Crossref Full Text | Google Scholar

Gamage, L., Isuranga, U., De Silva, S., and Meedeniya, D. (2023). Melanoma skin cancer classification with explainability ICAR 2023−3rd International Conference on Advanced Research in Computing: Digital Transformation for Sustainable Development (Belihuloya: IEEE), 30−35.

Google Scholar

Gamage, L., Isuranga, U., Meedeniya, D., De Silva, S., and Yogarajah, P. (2024). Melanoma skin cancer identification with explainability utilizing mask guided technique. Electronics 13:680. doi: 10.3390/electronics13040680

Crossref Full Text | Google Scholar

Gulzar, Y. (2024). Enhancing soybean classification with modified inception model: a transfer learning approach. Emirates J. Food Agric. 36, 1–9. doi: 10.3897/ejfa.2024.122928

Crossref Full Text | Google Scholar

Gulzar, Y., Alwan, A. A., Abdullah, R. M., Abualkishik, A. Z., and Oumrani, M. (2023a). OCA ordered clustering-based algorithm for e-commerce recommendation system. Sustainability 15:2947. doi: 10.3390/su15042947

Crossref Full Text | Google Scholar

Gulzar, Y., and Khan, S. A. (2022). Skin lesion segmentation based on vision transformers and convolutional neural networks—A comparative study. Appl. Sci. 12:5990. doi: 10.3390/app12125990

Crossref Full Text | Google Scholar

Gulzar, Y., Ünal, Z., Aktas, H. A., and Mir, M. S. (2023b). Harnessing the power of transfer learning in sunflower disease detection: a comparative study. Agriculture 13:1479. doi: 10.3390/agriculture13081479

Crossref Full Text | Google Scholar

Gupta, H., Bhatia, H., Giri, D., Saxena, R., and Singh, R. (2020). Comparison and analysis of skin lesion on pretrained architectures. Int. Res. J. Eng. Technol. 7:2704–7. Available at: https://www.researchgate.net/profile/Harsh-Gupta-39/publication/343125357_Comparison_and_Analysis_of_Skin_Lesion_on_Pretrained_Architectures/links/5f17d3af92851cd5fa3bf475/Comparison-and-Analysis-of-Skin-Lesion-on-Pretrained-Architectures.pdf

Google Scholar

Harris-Tryon, T. A., and Grice, E. A. (2022). Microbiota and maintenance of skin barrier function. Science 376, 940–945. doi: 10.1126/science.abo0693

PubMed Abstract | Crossref Full Text | Google Scholar

Inthiyaz, S., Altahan, B. R., Ahammad, S. H., Rajesh, V., Kalangi, R. R., Smirani, L. K., et al. (2023). Skin disease detection using deep learning. Adv. Eng. Softw. 175:103361. doi: 10.1016/j.advengsoft.2022.103361

Crossref Full Text | Google Scholar

Jain, P., Zameer, F., Khan, K., Alva, V., Huchegowda, R., Akki, A. J., et al. (2024). Artificial intelligence in diagnosis and monitoring of atopic dermatitis: from pixels to predictions. Artif. Intell. Health 1, 48–65. doi: 10.36922/aih.2775

Crossref Full Text | Google Scholar

Jain, S., Singhania, U., Tripathy, B., Nasr, E. A., Aboudaif, M. K., and Kamrani, A. K. (2021). Deep learning-based transfer learning for classification of skin cancer. Sensors 21:8142. doi: 10.3390/s21238142

PubMed Abstract | Crossref Full Text | Google Scholar

Jaisakthi, S. M., and Mirunalini, P. (2023). Classification of skin cancer from dermoscopic images using deep neural network architectures. Multimed. Tools Appl. 82, 15763–15778. doi: 10.1007/s11042-022-13847-3

PubMed Abstract | Crossref Full Text | Google Scholar

Khan, F., Ayoub, S., Gulzar, Y., Majid, M., Reegu, F. A., Mir, M. S., et al. (2023a). MRI-based effective ensemble frameworks for predicting human brain tumor. J. Imaging 9:163. doi: 10.3390/jimaging9080163

PubMed Abstract | Crossref Full Text | Google Scholar

Khan, F., Gulzar, Y., Ayoub, S., Majid, M., Mir, M. S., and Soomro, A. B. (2023b). Least square-support vector machine based brain tumor classification system with multi model texture features. Front. Appl. Math. Stat. 9:1324054. doi: 10.3389/fams.2023.1324054

Crossref Full Text | Google Scholar

Khan, I. U., Aslam, N., Anwar, T., Aljameel, S. S., Ullah, M., Khan, R., et al. (2021). Remote diagnosis and triaging model for skin cancer using efficientnet and extreme gradient boosting. Complexity 2021:5591614. doi: 10.1155/2021/5591614

Crossref Full Text | Google Scholar

Li, H., Pan, Y., Zhao, J., and Zhang, L. (2021). Skin disease diagnosis with deep learning: a review. Neurocomputing 464, 364–393. doi: 10.1016/j.neucom.2021.08.096

Crossref Full Text | Google Scholar

Majid, M., Gulzar, Y., Ayoub, S., Khan, F., Reegu, F. A., Mir, M. S., et al. (2023a). Enhanced transfer learning strategies for effective kidney tumor classification with CT imaging. Int. J. Adv. Comput. Sci. Appl. 14:2023. doi: 10.14569/IJACSA.2023.0140847

Crossref Full Text | Google Scholar

Majid, M., Gulzar, Y., Ayoub, S., Khan, F., Reegu, F. A., Mir, M. S., et al. (2023b). Using ensemble learning and advanced data mining techniques to improve the diagnosis of chronic kidney disease. Int. J. Adv. Comput. Sci. Appl. 14, 470–480. doi: 10.14569/IJACSA.2023.0141050

Crossref Full Text | Google Scholar

Malik, I., Ahmed, M., Gulzar, Y., Baba, S. H., Mir, M. S., Soomro, A. B., et al. (2023). Estimation of the extent of the vulnerability of agriculture to climate change using analytical and deep-learning methods: a case study in Jammu Kashmir and Ladakh. Sustainability 15:11465. doi: 10.3390/su151411465

Crossref Full Text | Google Scholar

Meedeniya, D., De Silva, S., Gamage, L., and Isuranga, U. (2024). Skin cancer identification utilizing deep learning: a survey. IET Image Process. 18, 3731–3749. doi: 10.1049/ipr2.13219

Crossref Full Text | Google Scholar

Mehmood, A., Gulzar, Y., Ilyas, Q. M., Jabbari, A., Ahmad, M., and Iqbal, S. (2023). SBXception: a shallower and broader xception architecture for efficient classification of skin lesions. Cancers. 15:3604. doi: 10.3390/cancers15143604

PubMed Abstract | Crossref Full Text | Google Scholar

Milantev, S., Olyunin, V., Milanteva, N., Bykov, I., and Bessmertny, I. (2020). “Skin lesion analysis using ensemble of cnn with dermoscopic images and metadata,” in Proceedings of the Proceedings of the 12th Majorov International Conference on Software Engineering and Computer Systems (Saint Petersburg: CEUR Workshop Proceedings).

Google Scholar

Naeem, A., Anees, T., Fiza, M., Naqvi, R. A., and Lee, S. W. (2022). SCDNet: a deep learning-based framework for the multiclassification of skin cancer using dermoscopy images. Sensors 22:5652. doi: 10.3390/s22155652

PubMed Abstract | Crossref Full Text | Google Scholar

Naeem, A., Anees, T., Khalil, M., Zahra, K., Naqvi, R. A., and Lee, S. W. (2024). SNC_Net: skin cancer detection by integrating handcrafted and deep learning-based features using dermoscopy images. Mathematics 12:1030. doi: 10.3390/math12071030

Crossref Full Text | Google Scholar

Pham, T. C., Doucet, A., Luong, C. M., Tran, C. T., and Hoang, V. D. (2020). Improving skin-disease classification based on customized loss function combined with balanced mini-batch logic and real-time image augmentation. IEEE Access 8, 150725–150737. doi: 10.1109/ACCESS.2020.3016653

Crossref Full Text | Google Scholar

Ravi, V. (2022). Attention cost-sensitive deep learning-based approach for skin cancer detection and classification. Cancers 14:5872. doi: 10.3390/cancers14235872

PubMed Abstract | Crossref Full Text | Google Scholar

Rokni, G. R., Gholizadeh, N., Babaei, M., Das, K., and Datta, S. (2024). Artificial intelligence in inflammatory skin disorders. Dermatol. Rev. 5:e243. doi: 10.1002/der2.243

Crossref Full Text | Google Scholar

Saarela, M., and Geogieva, L. (2022). Robustness, stability, and fidelity of explanations for a deep skin cancer classification model. Appl. Sci. 12:9545. doi: 10.3390/app12199545

Crossref Full Text | Google Scholar

Sadik, R., Majumder, A., Biswas, A. A., Ahammad, B., and Rahman, M. M. (2023). An in-depth analysis of convolutional neural network architectures with transfer learning for skin disease diagnosis. Healthc. Anal. 3:100143. doi: 10.1016/j.health.2023.100143

PubMed Abstract | Crossref Full Text | Google Scholar

Sengupta, D. (2023). Artificial intelligence in diagnostic dermatology: challenges and the way forward. Indian Dermatol Online J. 14, 782–787. doi: 10.4103/idoj.idoj_462_23

PubMed Abstract | Crossref Full Text | Google Scholar

Sharma, N., Mangla, M., Iqbal, M. M., and Mohanty, S. N. (2023). Deep learning framework for identification of skin lesions. EAI Endorsed. Trans. Pervasive Health Technol. 9, 1–16. doi: 10.4108/eetpht.9.3900

Crossref Full Text | Google Scholar

Shimu, S., Debnath, L. C., Sany, M. M. H., Keya, M., Khushbu, S. A., Noori, S. R. H., et al. (2022). Medical imaging a transfer learning process with multimodal CNN: dermis-disorder. Lect. Notes Netw. Syst. 514, 556–573. doi: 10.1007/978-3-031-12413-6_44

Crossref Full Text | Google Scholar

Skin Diseases Dataset (2024). Available at: https://www.kaggle.com/datasets/haroonalam16/20-skin-diseases-dataset/ (accessed August 17, 2024).

Google Scholar

Tahir, M., Naeem, A., Malik, H., Tanveer, J., Naqvi, R. A., and Lee, S. W. (2023). DSCC_Net: multi-classification deep learning models for diagnosing of skin cancer using dermoscopic images. Cancers 15:2179. doi: 10.3390/cancers15072179

PubMed Abstract | Crossref Full Text | Google Scholar

Venugopal, V., Raj, N. I., Nath, M. K., and Stephen, N. A. (2023). Deep neural network using modified efficientnet for skin cancer detection in dermoscopic images. Decis. Anal. J. 8:100278. doi: 10.1016/j.dajour.2023.100278

Crossref Full Text | Google Scholar

Wei, M., Wu, Q., Ji, H., Wang, J., Lyu, T., Liu, J., et al. (2023). Skin disease classification model based on densenet and convnext fusion. Electronics 12:438. doi: 10.3390/electronics12020438

Crossref Full Text | Google Scholar

Ye, S., and Chen, M. (2023). The emerging role of artificial intelligence in diagnosis and clinical analysis of dermatology. Dermatologica Sinica 41, 145–152. doi: 10.4103/ds.DS-D-23-00025

Crossref Full Text | Google Scholar

Keywords: skin disorder prediction, deep learning, transfer learning, DenseNet121, EfficientNetB0, computer vision, image classification

Citation: Gulzar Y, Agarwal S, Soomro S, Kandpal M, Turaev S, Onn CW, Saini S and Bounsiar A (2025) Next-generation approach to skin disorder prediction employing hybrid deep transfer learning. Front. Big Data 8:1503883. doi: 10.3389/fdata.2025.1503883

Received: 29 September 2024; Accepted: 03 February 2025;
Published: 19 February 2025.

Edited by:

Manjit Kaur, SR University, India

Reviewed by:

Dulani Meedeniya, University of Moratuwa, Sri Lanka
Saadullah Farooq Abbasi, University of Birmingham, United Kingdom
Badamasi Imam Ya'U, Abubakar Tafawa Balewa University, Nigeria
Uzair Bhatti, Hainan University, China

Copyright © 2025 Gulzar, Agarwal, Soomro, Kandpal, Turaev, Onn, Saini and Bounsiar. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yonis Gulzar, eWd1bHphckBrZnUuZWR1LnNh; Sherzod Turaev, c2hlcnpvZEB1YWV1LmFjLmFl

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

94% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more