Skip to main content

ORIGINAL RESEARCH article

Front. Med., 07 March 2024
Sec. Pathology
This article is part of the Research Topic Artificial Intelligence-Assisted Medical Imaging Solutions for Integrating Pathology and Radiology Automated Systems View all 9 articles

Enhancing accessibility for improved diagnosis with modified EfficientNetV2-S and cyclic learning rate strategy in women with disabilities and breast cancer

Moteeb Al MoteriMoteeb Al Moteri1T. R. MaheshT. R. Mahesh2Arastu ThakurArastu Thakur2V. Vinoth KumarV. Vinoth Kumar3Surbhi Bhatia Khan,
Surbhi Bhatia Khan4,5*Mohammed AlojailMohammed Alojail1
  • 1Department of Management Information Systems, College of Business Administration, King Saud University, Riyadh, Saudi Arabia
  • 2Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bangalore, India
  • 3School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, India
  • 4Department of Data Science, School of Science Engineering and Environment, University of Salford, Manchester, United Kingdom
  • 5Department of Electrical and Computer Engineering, Lebanese American University, Byblos, Lebanon

Breast cancer, a prevalent cancer among women worldwide, necessitates precise and prompt detection for successful treatment. While conventional histopathological examination is the benchmark, it is a lengthy process and prone to variations among different observers. Employing machine learning to automate the diagnosis of breast cancer presents a viable option, striving to improve both precision and speed. Previous studies have primarily focused on applying various machine learning and deep learning models for the classification of breast cancer images. These methodologies leverage convolutional neural networks (CNNs) and other advanced algorithms to differentiate between benign and malignant tumors from histopathological images. Current models, despite their potential, encounter obstacles related to generalizability, computational performance, and managing datasets with imbalances. Additionally, a significant number of these models do not possess the requisite transparency and interpretability, which are vital for medical diagnostic purposes. To address these limitations, our study introduces an advanced machine learning model based on EfficientNetV2. This model incorporates state-of-the-art techniques in image processing and neural network architecture, aiming to improve accuracy, efficiency, and robustness in classification. We employed the EfficientNetV2 model, fine-tuned for the specific task of breast cancer image classification. Our model underwent rigorous training and validation using the BreakHis dataset, which includes diverse histopathological images. Advanced data preprocessing, augmentation techniques, and a cyclical learning rate strategy were implemented to enhance model performance. The introduced model exhibited remarkable efficacy, attaining an accuracy rate of 99.68%, balanced precision and recall as indicated by a significant F1 score, and a considerable Cohen’s Kappa value. These indicators highlight the model’s proficiency in correctly categorizing histopathological images, surpassing current techniques in reliability and effectiveness. The research emphasizes improved accessibility, catering to individuals with disabilities and the elderly. By enhancing visual representation and interpretability, the proposed approach aims to make strides in inclusive medical image interpretation, ensuring equitable access to diagnostic information.

1 Introduction

Breast cancer remains a widespread and serious health issue worldwide, posing significant challenges for healthcare professionals and researchers. The intricate nature of this disease demands innovative approaches to diagnosis and treatment, with histopathological image analysis emerging as a crucial tool in this endeavor. The accurate interpretation and classification of these images are paramount for delivering timely and effective medical care to breast cancer patients. Against this backdrop, our research endeavors to elevate the accuracy and reliability of breast cancer histopathological image classification, aiming to make substantial contributions to the advancement of diagnostic and therapeutic practices.

To understand the significance of histopathological image analysis in breast cancer, it is essential to delve into the historical context of breast cancer diagnosis. Over the years, there has been a paradigm shift from traditional diagnostic methods to more sophisticated and precise techniques. Initially, clinical evaluation and palpation were the primary means of detecting breast abnormalities. However, the limitations of these methods in terms of sensitivity and specificity led to the exploration of imaging modalities such as mammography.

The introduction of mammography significantly transformed breast cancer screening, allowing for the observation of internal structures within the breast. While effective in identifying tumors at early stages, mammography has limitations, such as the occurrence of false positives and patient discomfort. With technological advancements, other imaging methods like ultrasound and magnetic resonance imaging (MRI) have been integrated into diagnostic processes, providing supplementary data to enhance diagnostic precision. Histopathological examination of breast tissue samples remains the gold standard for confirming the presence of cancer and determining its characteristics. When a suspicious lesion is detected through imaging, a biopsy is often performed, and the excised tissue undergoes histopathological analysis. This process involves the examination of cellular and tissue morphology under a microscope, allowing pathologists to identify cancerous cells, assess tumor grade, and characterize the extent of invasion.

In recent years, the integration of digital pathology and computer-aided diagnostic tools has transformed histopathological image analysis. Digital pathology involves the digitization of glass slides to create high-resolution digital images that can be analyzed using computational methods. This shift from traditional microscopy to digital platforms has paved the way for more efficient and objective assessments, reducing the dependence on subjective human interpretations. Even with progress in digital pathology, various obstacles remain in the precise evaluation of breast cancer histopathological images. The intricate nature of tissue formations, differences in staining methods, and the extensive amount of data produced present considerable challenges. Furthermore, the subjective aspect of conventional pathology assessments can lead to variability among observers, affecting the uniformity and dependability of diagnostic conclusions.

The need for rapid yet accurate results in clinical environments calls for expedited image analysis techniques without sacrificing precision. Consequently, there’s an increasing demand for automated and standardized methods in histopathological image classification. Machine learning (ML) and artificial intelligence (AI) are emerging as vital tools in this realm, offering swift, uniform, and unbiased analysis of images. ML and AI have become prominent in breast cancer diagnosis, especially in the analysis of histopathological images. These technologies can enhance the abilities of pathologists, refine diagnostic accuracy, and streamline the analysis workflow. ML algorithms, when trained on extensive datasets of labeled histopathological images, can identify patterns and features related to various types and stages of breast cancer. Once proficient, these algorithms can categorize new images, potentially surpassing traditional methods in consistency and performance. AI systems are also adept at recognizing subtle patterns and variations that might be difficult for human experts to notice.

The challenge in breast cancer diagnosis is the identification of its various histological subtypes, each requiring accurate classification for effective personalized treatment plans. Traditionally, pathologists manually review these images, a process that can be lengthy and subjective. Implementing machine learning and deep learning approaches to automate this task promises quicker, more precise diagnoses, potentially transforming the field of histopathology and improving patient outcomes.

As advancements in breast cancer diagnosis continue to evolve, it is imperative to address accessibility concerns in medical image interpretation, particularly for individuals with disabilities and the elderly. This paper presents a novel framework that modifies the EfficientNetV2-S architecture and integrates a cyclic learning rate strategy to optimize breast cancer histopathological image classification. By tailoring our approach to enhance interpretability, we seek to bridge gaps in accessibility, fostering inclusivity in the understanding and utilization of diagnostic information among diverse demographic groups, including those with disabilities and the elderly.

1.1 Objectives of the proposed study

The objective of this study is to develop a precise and efficient machine learning model for the automated classification of breast cancer histopathological images. Specifically, the goals are as follows:

• Implementing an advanced machine learning model based on EfficientNetV2 architecture to achieve high accuracy in classifying benign and malignant tumors.

• Addressing challenges related to generalizability, computational performance, and dataset imbalances encountered in existing models.

• Enhancing model transparency and interpretability to facilitate its applicability in medical diagnostic settings.

• Conducting rigorous training and validation using the BreakHis dataset to evaluate the performance of the proposed model.

• Assessing the model’s efficacy through metrics such as accuracy, precision, recall, F1 score, and Cohen’s Kappa value to demonstrate its superiority over existing methodologies.

1.2 Organization of the document

The structure of the remainder of this paper is as follows:

• Literature Review: In the subsequent section, we explore the existing body of research on breast cancer histopathological image classification, emphasizing the advantages and drawbacks of current methods.

• Methodology: Here, we describe our proposed deep learning model, elaborating on its architecture, the preprocessing of data, and the training process of the model.

• Experimental Results: This part details the outcomes of our experiments, demonstrating the enhanced performance of our model through both quantitative assessments and comparisons with existing methods.

• Discussion: In this segment, we analyze the implications of our results, focusing on the clinical relevance of precise classification of breast cancer histopathological images.

• Conclusion: The concluding section summarizes our research contributions and outlines the potential influence of our work on the diagnosis and treatment of breast cancer.

2 Literature review

The precise categorization of breast cancer histopathological images holds critical significance in clinical settings, as it directly impacts the diagnosis and treatment planning. There has been notable advancement over time in devising computational methods for the analysis of breast cancer images. This literature review delves into the development of these methodologies, underscoring both their strengths and their limitations.

2.1 Traditional approaches

Traditionally, the diagnosis of breast cancer relied on manual examination of histopathological images by experienced pathologists. While pathologists possess extensive expertise, this approach is time-consuming, subject to interobserver variability, and can be prone to errors, especially in distinguishing between subtle histological subtypes (1). The introduction of traditional machine learning approaches marked the beginning of automated image analysis. Characteristics like texture, shape, and color were extracted from histopathological images for the purpose of classification. Techniques such as Support Vector Machines (SVMs), Random Forests, and k-Nearest Neighbors (k-NN) were commonly employed for classification tasks. Although these methods enhanced efficiency, they faced difficulties in dealing with the complexity and variety present in breast cancer images, making accurate classification a persistent challenge (2).

2.2 Deep learning revolution

The rise of deep learning, especially Convolutional Neural Networks (CNNs), significantly transformed the realm of breast cancer histopathological image analysis. Deep learning models showed exceptional proficiency in tasks related to image classification, and their application in breast cancer research was a notable development.

• AlexNet and VGG: Early CNN architectures like AlexNet and VGG were adapted for breast cancer image classification. These models, pre-trained on large image datasets, showed promise in achieving higher accuracy compared to traditional methods. However, they struggled with handling diverse histological subtypes and required substantial computational resources (3).

• Transfer Learning: The adoption of transfer learning, involving the fine-tuning of pre-trained CNN models for breast cancer classification, became increasingly prevalent. This approach enabled the utilization of features learned from general image datasets, such as ImageNet, and their adaptation for specialized tasks. Models such as Inception and ResNet were broadly embraced for this purpose (4).

2.3 Challenges in deep learning

While deep learning models demonstrated significant improvements in breast cancer image classification, several challenges persisted:

• Diverse Histological Subtypes: Breast cancer encompasses various histological subtypes, each with distinct characteristics. Existing models struggled to accurately differentiate between these subtypes, potentially leading to misclassification (5).

• Imbalanced Datasets: Imbalanced datasets, where certain classes are underrepresented, posed challenges in model training. Models tended to favor the majority class, impacting overall accuracy (6).

• Overfitting: A prevalent challenge in deep learning, overfitting arises when models excel on training data but underperform on new, unseen data. Implementing strategies to counteract overfitting is crucial for the effectiveness of these models (7).

• Generalization: Ensuring that models could generalize to diverse image types, including different magnifying factors, was crucial for real-world applications (8).

2.4 Recent advances

• Data Augmentation: To tackle the issues of overfitting and generalization, data augmentation methods were applied. These techniques artificially enlarged the dataset through the application of random transformations to existing images, including rotations, flips, and adjustments in brightness. Data augmentation contributed to enhancing the robustness of the models.

• Attention Mechanisms: Attention mechanisms were introduced to enhance the discriminative capabilities of models. Attention modules allowed models to focus on relevant image regions, improving the accuracy of subtype classification (9).

• Ensemble Learning: Ensemble learning techniques, where multiple models were combined to make predictions, further improved accuracy. Ensemble methods reduced the impact of model bias and variance, leading to more reliable results (9).

In Table 1 different models with respect to their state of art technologies used is explored and a comprehensive survey is conducted.

Table 1
www.frontiersin.org

Table 1. Different existing studies.

Recent studies have showcased significant progress in utilizing machine learning and deep learning for various aspects of breast cancer diagnosis and treatment. Researchers have developed novel models for early detection and precise localization of tumors using convolutional neural networks (CNNs) on databases like BCDRD01. Transfer learning with architectures like InceptionResNetV2 has been effective in predicting breast cancer risk based on gene expression and imaging data. Another model, BCCNN, demonstrates high accuracy in classifying breast cancer types using MRI images, enhancing diagnostic capabilities. Additionally, the creation of synthetic datasets like SNOW has proven beneficial for training computer-vision models in pathology. One prevalent trend is the widespread adoption of deep learning techniques, particularly convolutional neural networks (CNNs), for automated analysis and classification of histopathological images. These deep learning models, often built upon architectures like EfficientNet, demonstrate remarkable performance in accurately identifying and distinguishing between benign and malignant tissue structures. Additionally, many studies emphasize the importance of advanced preprocessing and augmentation techniques to enhance the quality and diversity of training data, thereby improving model generalization and robustness. Furthermore, the integration of interpretability methods, such as attention mechanisms or feature visualization techniques, enhances the transparency and clinical utility of automated diagnostic systems. Overall, the collective efforts showcased in these studies underscore the significant advancements in breast cancer histopathological image analysis, paving the way for more accurate, efficient, and scalable diagnostic solutions in clinical practice.

3 Methodology

In our research, we adopted a systematic approach to categorize histopathological images of breast cancer, leveraging cutting-edge machine learning techniques. Our process utilized the extensive BreakHis dataset and included stages of data preprocessing, choosing the right model, training it, and then evaluating its performance. We selected the latest neural network frameworks, focusing on different versions of EfficientNetV2. To guarantee a sturdy and precise classification of benign and malignant breast cancer tissues, we employed comprehensive data augmentation and meticulous training methodologies. This section details the entire procedure and methods employed, ranging from the description of the dataset to the application of machine learning models and the metrics used for evaluation. Figure 1 illustrates a summarized architecture of our model which contains multiple pre processing steps along with the model architecture.

Figure 1
www.frontiersin.org

Figure 1. Brief model architecture.

3.1 Dataset description

The BreakHis dataset stands as a pivotal resource in medical imaging for breast cancer research. Comprising 9,109 microscopic images from 82 patients, it offers a profound look at breast tumor tissues. These images, captured at various magnifications (40×, 100×, 200×, and 400×), reveal intricate histopathological details. The dataset includes 2,480 benign and 5,429 malignant samples, each being a 700×460 pixel, 3-channel RGB image with 8-bit depth per channel, stored in PNG format.

Structured into two principal categories, benign and malignant tumors, the dataset further subdivides into specific tumor types. For benign tumors, it includes adenosis (A), fibroadenoma (F), phyllodes tumor (PT), and tubular adenoma (TA). The malignant tumors comprise ductal carcinoma (DC), lobular carcinoma (LC), mucinous carcinoma (MC), and papillary carcinoma (PC). Each image file is descriptively named, detailing biopsy method, tumor class, type, patient ID, and magnification factor, making it a valuable dataset for automated classification and analysis. Figure 2 visually represents this data distribution while Figures 3, 4 show some Benign and Malignant samples images from the dataset.

Figure 2
www.frontiersin.org

Figure 2. Dataset count.

Figure 3
www.frontiersin.org

Figure 3. Benign sample images.

Figure 4
www.frontiersin.org

Figure 4. Malignant sample images.

The BreakHis dataset, developed in partnership with the P&D Laboratory – Pathological Anatomy and Cytopathology in Paraná, Brazil, holds significant value for its ability to facilitate benchmarking and comparative research in the field of automated cancer diagnosis and histopathological image analysis. This collaboration has resulted in a dataset that is not only extensive but also deeply informative, making it an essential tool for advancements in medical imaging and machine learning applications in breast cancer research.

3.2 Preprocessing

The preprocessing of the BreakHis dataset was a vital phase in preparing the images for efficient machine learning analysis. This procedure incorporated various essential steps to normalize and improve the image data, thereby optimizing the performance of the model. These steps included techniques such as resizing images to a uniform dimension, adjusting brightness and contrast for better clarity, and potentially applying normalization or standardization to align the pixel value distributions. This careful preprocessing ensured that the images were suitably formatted and enhanced for the subsequent stages of machine learning and deep learning analysis.

• Resizing Images: Given the variation in image sizes within the dataset, all images were resized to a uniform dimension. The chosen size was dictated by the requirements of the EfficientNetV2 model variants used in the study. Specifically, images were resized to dimensions that matched the input size of the selected EfficientNetV2 model. For instance, by using EfficientNetV2-S, images were resized to 384×384 pixels.

• Normalization: To ensure that the model received input data within a consistent range, normalization of the images was performed. This step involved scaling the pixel values in each image from their original 8-bit range (0–255) to a floating-point range between 0 and 1. This normalization helps in speeding up the training process and improving model convergence.

• Color Consistency: As the images were in RGB format, no color conversion was necessary. However, we ensured that the color consistency was maintained across all images to prevent any color-related biases during model training.

• Data Augmentation: In order to enhance the variety within the training dataset and mitigate the risk of overfitting, we implemented data augmentation techniques. These techniques encompassed random horizontal flipping of images, rotations, adjustments in brightness and contrast, and random resized cropping. This augmentation process was dynamically integrated during the training phase, effectively expanding the dataset by creating varied versions of each image. Such augmentation not only diversified the training data but also helped the model in learning to generalize better from the augmented images, crucial for robust machine learning models.

• Dataset Splitting: The BreakHis dataset was systematically partitioned into training, validation, and test sets. To maintain a representative distribution of both benign and malignant cases across these sets, a stratified sampling method was utilized. This approach ensured that each set mirrored the overall composition of the dataset in terms of the proportion of benign to malignant cases. Additionally, to counteract any class imbalance, particularly evident in the training set, an upsampling technique was applied. This involved increasing the number of images in the underrepresented class (either benign or malignant) to match the count of the other class. Such a measure was crucial to prevent any bias in the model training process, thereby enhancing the accuracy and reliability of the model in distinguishing between benign and malignant breast cancer tissues.

• Batch Processing: Images were grouped into batches for efficient processing during the training phase. The batch size was determined based on the model’s requirements and the computational resources available.

• Image Parsing and Tensor Conversion: Each image file was read and decoded into a tensor format suitable for input into the TensorFlow-based neural network. This conversion process was essential for the compatibility of image data with the machine learning framework used.

In Figure 5 some images from the original Break-His dataset is shown.

Figure 5
www.frontiersin.org

Figure 5. Images from the Break-His dataset.

These preprocessing steps were vital in preparing the BreakHis dataset for the machine learning pipeline, ensuring that the input data was in an optimal format for model training and contributed to the high accuracy and f1 score achieved in the study.

3.3 EfficientNetV2 and its variants

The decision to utilize EfficientNetV2 models and their variants in this study was influenced by multiple compelling reasons, closely aligned with our goals of achieving high accuracy and efficiency in the classification of breast cancer histopathological images. EfficientNetV2, an advanced iteration of the original EfficientNet, offers substantial enhancements which render it especially appropriate for this task:

• State-of-the-Art Performance: The EfficientNetV2 models are renowned for their exceptional performance in image classification tasks. They have demonstrated impressive results on benchmark datasets, underscoring their potential for application in medical image analysis where accuracy is of utmost importance. This superior performance is attributed to their advanced architecture, which efficiently balances depth, width, and resolution, leading to highly accurate models. Their proven effectiveness in various image classification challenges makes them a promising option for the intricate task of classifying histopathological images in breast cancer research, where precision in diagnosis is critical.

• Scalability and Versatility: The EfficientNetV2 architecture comes in various sizes (small, medium, and large), providing a scalable solution that can be tailored to the computational resources available and the complexity of the task. This flexibility allows for choosing a model variant that balances between computational efficiency and classification performance.

• Improved Training Speed and Efficiency: EfficientNetV2 models are designed for faster training and higher efficiency compared to their predecessors. This is achieved through optimizations such as Fused-MBConv layers and a more efficient scaling method. For a dataset as large and complex as BreakHis, these improvements in training speed are crucial for timely model development and iteration.

• Advanced Regularization Techniques: These models incorporate the latest advancements in regularization and data augmentation strategies, such as progressive learning and model regularization methods. This is particularly beneficial in a medical imaging context, where the model needs to generalize well from limited and varied training data, reducing the risk of overfitting.

• Optimized for Diverse Image Resolutions: EfficientNetV2 is designed to adeptly manage a range of image resolutions. This trait is particularly important in the analysis of histopathological images, where recognizing details at multiple scales is essential for precise tumor categorization. Its capability to consistently deliver strong performance at various magnifications, such as 40×, 100×, 200×, and 400× as demonstrated in the BreakHis dataset, without requiring substantial changes to the architecture, stands out as a notable benefit.

• Robustness to Variations in Input Data: Given the inherent variations in medical images due to differences in staining techniques, tissue preparation, and imaging conditions, the robustness of EfficientNetV2 models to such variations makes them an apt choice for this application.

• Proven Effectiveness in Medical Imaging: EfficientNet architectures have shown remarkable success in a variety of medical imaging applications, showcasing their proficiency in processing intricate image data and identifying pertinent features for precise disease identification.

• Compatibility with Transfer Learning: The availability of pre-trained EfficientNetV2 models on standard image datasets like ImageNet allows for leveraging transfer learning. This approach is particularly useful in medical imaging tasks, where annotated data can be scarce. Pre-trained models can be fine-tuned on the specific medical dataset, significantly improving the learning process.

In the neural network structure developed for categorizing breast cancer histopathological images, every layer is carefully crafted to assist in the precise distinction between benign and malignant tumors. The architecture begins with the EfficientNetV2 as the base model, acting as a sophisticated feature extractor. This model outputs a 1,280-dimensional feature vector for each image and contains over 20 million trainable parameters, reflecting its capability to learn complex patterns in the image data. Following this, a dense layer compresses these features from 1,280 to 512 dimensions, allowing for a more compact yet informative representation of the image characteristics.

To enhance the training process and prevent overfitting, a batch normalization layer follows, normalizing the output of the dense layer by scaling and shifting. This is crucial for stabilizing learning and maintaining consistency across batches of data. A dropout layer is then employed, which randomly zeroes a fraction of the input units, further aiding in generalization and preventing the model from becoming overly reliant on specific features. The convolution process is depicted in Eq. 1, succeeded by the cyclic learning rate shown in Eq. 2, and the binary cross-entropy loss function is detailed in Eq. 3.

I K i j = m n I m n K i m , j n     (1)

This equation illustrates the convolution process, where “I” denotes the input image and “K”: represents the kernel or filter. Convolution is a core operation in CNN architectures, including EfficientNet.

C L R = L R min + 1 2 L R max L R min × 1 + cos epoch total epoch × π     (2)
B C E Loss = 1 N i = 1 N y i log p i + 1 y i log 1 p i     (3)

In Table 2 a complete description of the layers is given and in Table 3 the detailed algorithm is given.

Table 2
www.frontiersin.org

Table 2. Description of model architecture.

Table 3
www.frontiersin.org

Table 3. Algorithm of breast cancer image classification.

The network’s subsequent layers include another dense layer, which further compresses the feature vector to 128 dimensions, and a second batch normalization layer, continuing the process of feature refinement and stabilization. The final layer, a classifier, is a dense layer with a single neuron, responsible for making the final binary decision—classifying the tumor as benign or malignant. This layer’s output represents the culmination of a complex process of feature extraction, compression, and refinement, leveraging the intricate patterns learned from the histopathological images to arrive at a precise classification.

Choosing EfficientNetV2 models for the classification of breast cancer histopathological images is supported by their exceptional combination of efficiency, scalability, and cutting-edge performance. Their ability to handle diverse image resolutions and robustness to input variations aligns well with the challenges presented by the BreakHis dataset, ultimately contributing to the high accuracy and f1 scores achieved in this study.

3.4 Training process

The development of the breast cancer histopathological image classification model was rigorously planned to maximize performance. We commenced with data augmentation, a critical step in enhancing the model’s generalization capabilities. This process included applying transformations such as horizontal flipping, rotation, random brightness and contrast adjustments, and random resized cropping directly during training. This on-the-fly augmentation introduced a beneficial variability to the dataset, aiding the model in learning to identify tumors under different conditions.

Central to our training strategy was the implementation of a Cyclical Learning Rate (CLR). This technique involved dynamically varying the learning rate between a lower and an upper bound throughout the training epochs. Starting from a relatively high initial rate, it was adjusted to reach a peak and then cycled back, allowing the model to effectively navigate the loss landscape and avoid getting trapped in local minima.

The model training was conducted in batches, a decision driven by the need for efficient memory utilization and the advantages of parallel processing. Batch sizes were carefully chosen to balance the computational load and the model’s complexity. Additionally, batch processing complemented the effectiveness of batch normalization layers by normalizing activations across each batch.

In terms of model compilation, we chose the Binary Crossentropy loss function, suitable for our binary classification task, and paired it with a Stochastic Gradient Descent (SGD) optimizer. This optimizer was integrated with the CLR scheduler to facilitate dynamic learning rate adjustments. Using the SGD optimizer ensured steady and effective convergence during training.

The fitting of the model was a crucial phase, where we trained it using the prepared dataset, regularly evaluating its performance on a separate validation set. This phase also included the use of callbacks like ModelCheckpoint to save the best-performing model iteration, ensuring that we retained the model version with the highest validation accuracy. In Figure 6 an outline of the whole training process is given.

Figure 6
www.frontiersin.org

Figure 6. Outline of the process.

During the model’s training phase for classifying breast cancer histopathological images, vital metrics like accuracy, precision, recall, and the F1 score were meticulously tracked after each training epoch. These indicators were crucial in evaluating the model’s effectiveness and informed decisions on modifying its architecture, learning rates, or augmentation methods in further training cycles. This thorough, iterative process was key to crafting an efficient and robust model capable of accurately and reliably classifying breast cancer images. The training approach’s cornerstone was a blend of dynamic learning rate modification, strategic data augmentation, and careful batch processing. These components collectively ensured that the model not only achieved high performance metrics but also remained efficient in terms of training duration and resource consumption, establishing it as an effective instrument for medical image analysis in breast cancer detection.

3.5 Model evaluation metrics

For the assessment of our model dedicated to classifying breast cancer histopathological images, we utilized a variety of metrics to thoroughly evaluate its performance. These measures are essential for comprehensively gauging the model’s efficacy and dependability. The key metrics used were:

• Accuracy: Accuracy, a straightforward performance metric, is the ratio of correctly predicted cases to the total number of observations (20). It provides a fundamental insight into the frequency of correct predictions made by the model, regardless of class distribution. In binary classification tasks like ours, accuracy represents the fraction of accurate outcomes (encompassing both true positives and true negatives) in the total cases evaluated. This metric is determined using Eq. 4, where TP stands for True Positive, TN for True Negative, FP for False Positive, and FN for False Negative.

Accuracy = T P + T N T P + T N + F P + F N     (4)

• Precision: Precision, also referred to as the positive predictive value, measures the ratio of correctly predicted positive instances to the total predicted positives (21). It addresses the query: Among all the tumors predicted as malignant by the model, how many were truly malignant? High precision corresponds to a low false positive rate. In the field of medical diagnostics, precision is of paramount importance since false positives can result in unwarranted anxiety and medical interventions. It is computed using Eq. 5.

Precision = T P T P + F P     (5)

• Recall (Sensitivity): Recall, often referred to as sensitivity, quantifies the ratio of correctly predicted positive instances to all instances within the actual class. It answers the question: Out of all the malignant tumors present, how many did the model accurately detect? In the medical context, achieving a high recall rate is crucial because missing a malignant tumor (false negative) can carry significant consequences for the patient. Recall is computed using Eq. 6.

Recall = T P T P + F N     (6)

• F1 Score: The F1 Score serves as the weighted average of Precision and Recall. Consequently, this score considers both false positives and false negatives, offering a balanced assessment of a classifier’s performance (22). It is a valuable metric to demonstrate that a classifier attains high values for both recall and precision. Moreover, the F1 Score is particularly useful when dealing with imbalanced class distributions since it harmonizes the significance of precision and recall. This metric is calculated using Eq. 7.

F 1 Score = 2 × Precision × Recall Precision + Recall     (7)

• Cohen’s Kappa Coefficient: Cohen’s Kappa is a statistical measure that is used to evaluate the accuracy of a classification model while considering the agreement that occurs by chance. It’s particularly useful in scenarios where the class distribution is imbalanced. It is calculated using Eq. 8.

k = p o p e 1 p e     (8)

• where po is the relative observed agreement, and pe is the hypothetical probability of chance agreement.

These metrics provided a multi-faceted view of the model’s performance. Accuracy gave a quick snapshot of overall effectiveness, while precision and recall offered deeper insights into its ability to correctly identify malignant tumors and minimize false diagnoses. The F1 score played a crucial role as a unified metric that integrated precision and recall, delivering a well-balanced measure of the model’s diagnostic accuracy. This was particularly significant given the uneven distribution of benign and malignant cases within the dataset. The training and validation loss, as well as accuracy, of this model is visualized in Figure 7.

Figure 7
www.frontiersin.org

Figure 7. Training and validation accuracy and loss.

The amalgamation of these metrics provided a comprehensive insight into the model’s capabilities, assuring its reliability and suitability for clinical use in the context of breast cancer diagnosis.

4 Experimental results

The outcomes of our research, centered on the classification of histopathological images of breast cancer utilizing the BreakHis dataset, showcased the effectiveness and resilience of the selected machine learning model, EfficientNetV2. The evaluation metrics employed to appraise the model’s performance offered a thorough perspective on its capabilities.

• High Accuracy: The model attained an exceptionally high accuracy score of 99.68%. This signifies that the model adeptly categorized the majority of the images as either benign or malignant. Such a remarkable level of accuracy holds great importance in medical diagnostics, where the consequences of misclassification can be substantial.

• Precision and Recall: Both precision and recall metrics further underscored the model’s impressive performance. High precision signified that the model exhibited a low rate of false positives, a critical aspect in medical diagnostics to prevent unwarranted treatment or anxiety. Meanwhile, high recall indicated that the model effectively detected the majority of malignant cases, ensuring timely and appropriate medical intervention when needed.

• F1 Score: The F1 score, serving as a balance between precision and recall, also stood out with a notably high value of 99.77%. This illustrates that the model effectively struck a strong equilibrium between accurately identifying malignant cases (recall) and minimizing false positive diagnoses (precision).

• Cohen’s Kappa: The model achieved a Cohen’s Kappa score of 99.26% indicating high level of binary classification of this model. This high level of agreement, as measured by Cohen’s Kappa, complements the model’s high accuracy, reinforcing its reliability in classifying breast cancer images. Along with high accuracy, precision, recall, and F1 score, the Cohen’s Kappa score further validated the model’s effectiveness. Its ability to account for random chance in the agreement makes it a particularly valuable metric in the context of medical image classification, where the cost of misdiagnosis can be significant.

• Confusion Matrix and Classification Report: The results were additionally reinforced by a comprehensive classification report and a visualized confusion matrix. The classification report furnished insights into the model’s performance across various classes, while the confusion matrix provided a graphical representation of the model’s predictions compared to the actual labels. The classification report is presented in Table 4, followed by visual representations of the classification report in Figure 8 and the confusion matrix in Figure 9.

• Visual Representation of Results: The study also included visual aids such as graphs and charts to illustrate the model’s performance across various metrics. These visualizations helped in better understanding the distribution and nature of the model’s predictions. For the same ROC-AUC Curve along with Precision-Recall Curve is plotted which is been shown in Figures 10, 11.

• Comparison with Baseline Models: The results were compared with baseline models to highlight the improvements and effectiveness of the chosen architecture. This comparison provided a context for the model’s performance, showcasing its superiority in terms of accuracy and reliability over more traditional or less sophisticated models. In Table 5 a brief comparison with existing models is shown.

• Error Analysis: An analysis of the misclassified images was conducted to understand the circumstances under which the model failed. This analysis helped in identifying any patterns in errors, such as specific types of tumors that were consistently misclassified, or issues related to image quality or resolution. For the same calibration curve was plotted which is visualized in Figure 12.

• Robustness to Variations: The model’s robustness was evident in its consistent performance across different magnifications and subtypes of tumors within the dataset. This aspect was crucial, considering the diverse nature of histopathological images in terms of magnification, staining, and tumor types.

• Insights and Observations: The research yielded valuable insights into the model’s decision-making process. For instance, it delved into the regions of the images that garnered the model’s attention during the prediction process, providing insight into which features or patterns held the most significance in the classification procedure.

Table 4
www.frontiersin.org

Table 4. Classification report.

Figure 8
www.frontiersin.org

Figure 8. Classification report.

Figure 9
www.frontiersin.org

Figure 9. Confusion matrix.

Figure 10
www.frontiersin.org

Figure 10. ROC-AUC curve.

Figure 11
www.frontiersin.org

Figure 11. Precision-recall curve.

Table 5
www.frontiersin.org

Table 5. Comparison with existing studies.

Figure 12
www.frontiersin.org

Figure 12. Calibration curve.

Overall, the results were highly encouraging, indicating that the model not only achieved exceptional accuracy but also balanced precision and recall effectively. This balance is critical in the medical field, where both over-diagnosis and under-diagnosis carry significant consequences (33, 34). The high F1 score further confirmed the model’s ability to perform consistently well in both respects. These findings highlight the potential of employing advanced machine learning models such as EfficientNetV2 in the realm of medical image analysis, especially for critical tasks like cancer diagnosis.

5 Discussion

The outcomes of our study, which centered on the classification of breast cancer histopathological images utilizing the EfficientNetV2 model, offer numerous significant insights into the efficacy and potential applications of machine learning in the field of medical diagnostics.

• High Accuracy and Its Implications: The exceptionally high accuracy (99.68%) achieved by the model is indicative of its robustness and reliability in classifying histopathological images. In the context of medical imaging, such accuracy is paramount, as it directly impacts diagnostic decisions (35). The exceptional level of accuracy achieved by the model implies that it can serve as a valuable tool to aid pathologists, potentially enhancing the efficiency and throughput of diagnostic procedures in clinical settings (36).

• Precision, Recall, and F1 Score: The equilibrium between precision and recall, exemplified by the high F1 score, holds paramount importance in medical diagnostics. Elevated precision reduces the occurrence of false positives, a crucial aspect in avoiding unnecessary medical interventions (37). Simultaneously, a high recall rate ensures that the model accurately identifies the majority of malignant cases, a critical factor in patient care since missed diagnoses can lead to severe consequences (38). The results affirm that the model can effectively sustain this balance, establishing it as a dependable tool in the diagnostic process.

• Analysis of Misclassifications: The error analysis, where misclassified images were examined, provides valuable insights into the model’s limitations. Understanding the characteristics of these misclassified cases can guide future improvements in the model (39). For example, if certain tumor subtypes are consistently misclassified, additional training data or model adjustments targeting these subtypes might be necessary.

• Model Robustness: The consistent performance of the model across various magnifications and tumor types underscores its robustness. This aspect is particularly important given the diverse nature of medical images in real-world settings, where variability in factors like staining techniques and image quality is common (40).

• Clinical Integration and Future Work: While the results are promising, the integration of such models into clinical workflows requires careful consideration. Factors such as model interpretability, integration with existing health IT systems, and adherence to regulatory standards are critical for successful implementation (41). Future endeavors could center on multi-class classification to differentiate between various types of breast cancer, thereby expanding the model’s applicability in clinical diagnostics (42). Furthermore, assessing the model’s performance on external datasets and in real-world clinical environments would be crucial to validate its applicability and robustness further.

This study highlights the substantial potential of advanced machine learning models in the domain of medical image analysis, especially in critical applications like cancer diagnosis. The model’s impressive levels of accuracy, precision, recall, and F1 score not only validate its effectiveness but also underscore the capacity of such technologies to complement and elevate current diagnostic approaches in healthcare (43).

6 Conclusion

In this study, we delved into the utilization of the EfficientNetV2 model for the classification of breast cancer histopathological images and obtained promising outcomes. The model demonstrated a remarkable accuracy of 99.68%, a high F1 score, and an excellent balance between precision and recall. These findings suggest that the model has significant potential as an aid in diagnosing breast cancer, potentially enhancing the efficiency and accuracy of pathologists’ work. Nonetheless, it’s important to acknowledge the limitations of this study. The model’s training and evaluation were conducted solely on the BreakHis dataset, which, although extensive, may not encompass the entire spectrum of clinical scenarios. The absence of external validation raises concerns about the model’s ability to generalize to other datasets. Additionally, the inherent “black box” nature of deep learning models, including the one used in this study, presents challenges in terms of interpretability, a critical aspect in the context of medical diagnosis. Also, the inherent data imbalance in the dataset might have influenced the model’s learning and performance.

Looking forward, there are several avenues for further research. Testing the model on external datasets from diverse clinical environments is crucial for assessing its real-world applicability. Enhancing the model’s interpretability would be invaluable in gaining trust and acceptance in clinical settings. Expanding the model to include multi-class classification could provide more detailed diagnostic information, and exploring advanced techniques to handle data imbalance could improve its ability to recognize less common, yet clinically significant cases. Finally, conducting clinical trials would be a critical step in evaluating the model’s practical utility and integration into medical workflows. In conclusion, this study underscores the immense potential of advanced machine learning in the field of medical imaging, especially for crucial tasks such as cancer diagnosis. The high-performance metrics achieved by the model highlight its potential as a reliable diagnostic tool, yet further research and validation are necessary to overcome its limitations and fully harness its capabilities in clinical settings.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

MAM: Conceptualization, Formal analysis¸ Project administration, Writing – review & editing. TM: Conceptualization, Data curation, Software, Writing – original draft. AT: Conceptualization, Data curation, Software, Writing – review & editing. VV: Data curation, Formal analysis, Validation, Visualization, Writing – review & editing. SK: Conceptualization, Methodology, Supervision, Writing – review & editing. MAlo: Formal analysis, Project administration, Visualization¸ Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. The authors extended their appreciation to the King Salman Center for Disability Research for funding this work through Research Group (no. KSRG-2023-394).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. el-Tanani, M, al Khatib, AO, al-Najjar, BO, Shakya, AK, el-Tanani, Y, Lee, YF, et al. Cellular and molecular basis of therapeutic approaches to breast cancer. Cell Signal. (2023) 101:110492. doi: 10.1016/j.cellsig.2022.110492

PubMed Abstract | Crossref Full Text | Google Scholar

2. Atban, F, Ekinci, E, and Garip, Z. Traditional machine learning algorithms for breast cancer image classification with optimized deep features. Biomed Signal Process Control. (2023) 81:104534. doi: 10.1016/j.bspc.2022.104534

Crossref Full Text | Google Scholar

3. Chang, QQ, Chen, CY, Chen, Z, and Chang, S. LncRNA PVT1 promotes proliferation and invasion through enhancing Smad3 expression by sponging miR-140-5p in cervical cancer. Radiol Oncol. (2019) 53:443–52. doi: 10.2478/raon-2019-0048

Crossref Full Text | Google Scholar

4. Srikantamurthy, MM, Rallabandi, VPS, Dudekula, DB, Natarajan, S, and Park, J. Classification of benign and malignant subtypes of breast cancer histopathology imaging using hybrid CNN-LSTM based transfer learning. BMC Med Imaging. (2023) 23:19. doi: 10.1186/s12880-023-00964-0

PubMed Abstract | Crossref Full Text | Google Scholar

5. Αoun-Bacha, Z, Bitar, N, Abi Saleh, W, Assi, H, Bahous, J, Boukhalil, P, et al. Diagnosis and management of patients with stage III non-small cell lung cancer: a joint statement by the Lebanese Society of Medical Oncology and the Lebanese pulmonary society. Oncol Lett. (2023) 25:1–10. doi: 10.3892/ol.2023.13699

Crossref Full Text | Google Scholar

6. Nassar, HB. Classification for imbalanced breast cancer dataset using resampling methods. IJCSNS. (2023) 23:89. doi: 10.22937/IJCSNS.2023.23.1.12

Crossref Full Text | Google Scholar

7. Xu, C, Coen-Pirani, P, and Jiang, X. Empirical study of overfitting in deep learning for predicting breast Cancer metastasis. Cancers. (2023) 15:1969. doi: 10.3390/cancers15071969

PubMed Abstract | Crossref Full Text | Google Scholar

8. Huang, H, Wu, N, Liang, Y, Peng, X, and Shu, J. SLNL: a novel method for gene selection and phenotype classification. Int J Intell Syst. (2022) 37:6283–304. doi: 10.1002/int.22844

Crossref Full Text | Google Scholar

9. Zheng, Y, Li, C, Zhou, X, Chen, H, Xu, H, Li, Y, et al. Application of transfer learning and ensemble learning in image-level classification for breast histopathology. Intell Med. (2023) 3:115–28. doi: 10.1016/j.imed.2022.05.004

Crossref Full Text | Google Scholar

10. Jafarzadeh Ghoushchi, S, Ranjbarzadeh, R, Najafabadi, SA, Osgooei, E, and Tirkolaee, EB. An extended approach to the diagnosis of tumour location in breast cancer using deep learning. J Ambient Intell Humaniz Comput. (2023) 14:8487–97. doi: 10.1007/s12652-021-03613-y

Crossref Full Text | Google Scholar

11. Humayun, M, Khalil, MI, Almuayqil, SN, and Jhanjhi, NZ. Framework for detecting breast cancer risk presence using deep learning. Electronics. (2023) 12:403. doi: 10.3390/electronics12020403

Crossref Full Text | Google Scholar

12. Abunasser, BS, al-Hiealy, MR, Zaqout, I, and Abu-Naser, S. Convolution neural network for breast Cancer detection and classification using deep learning. Asian Pac J Cancer Prev. (2023) 24:531–44. doi: 10.31557/APJCP.2023.24.2.531

PubMed Abstract | Crossref Full Text | Google Scholar

13. Ding, K, Zhou, M, Wang, H, Gevaert, O, Metaxas, D, and Zhang, S. A large-scale synthetic pathological dataset for deep learning-enabled segmentation of breast cancer. Sci Data. (2023) 10:231. doi: 10.1038/s41597-023-02125-y

PubMed Abstract | Crossref Full Text | Google Scholar

14. Dammu, H, Ren, T, and Duong, TQ. Deep learning prediction of pathological complete response, residual cancer burden, and progression-free survival in breast cancer patients. PLoS One. (2023) 18:e0280148. doi: 10.1371/journal.pone.0280148

PubMed Abstract | Crossref Full Text | Google Scholar

15. Hekal, AA, Moustafa, HE-D, and Elnakib, A. Ensemble deep learning system for early breast cancer detection. Evol Intel. (2023) 16:1045–54. doi: 10.1007/s12065-022-00719-w

Crossref Full Text | Google Scholar

16. Tekin, E, Yazıcı, Ç, Kusetogullari, H, Tokat, F, Yavariabdi, A, Iheme, LO, et al. Tubule-U-net: a novel dataset and deep learning-based tubule segmentation framework in whole slide images of breast cancer. Sci Rep. (2023) 13:128. doi: 10.1038/s41598-022-27331-3

PubMed Abstract | Crossref Full Text | Google Scholar

17. Obayya, M, Maashi, MS, Nemri, N, Mohsen, H, Motwakel, A, Osman, AE, et al. Hyperparameter optimizer with deep learning-based decision-support systems for histopathological breast cancer diagnosis. Cancers. (2023) 15:885. doi: 10.3390/cancers15030885

PubMed Abstract | Crossref Full Text | Google Scholar

18. Raza, A, Ullah, N, Khan, JA, Assam, M, Guzzo, A, and Aljuaid, H. DeepBreastCancerNet: a novel deep learning model for breast cancer detection using ultrasound images. Appl Sci. (2023) 13:2082. doi: 10.3390/app13042082

Crossref Full Text | Google Scholar

19. Jabeen, K, Khan, MA, Balili, J, Alhaisoni, M, Almujally, NA, Alrashidi, H, et al. BC2NetRF: breast cancer classification from mammogram images using enhanced deep learning features and equilibrium-Jaya controlled regula falsi-based features selection. Diagnostics. (2023) 13:1238. doi: 10.3390/diagnostics13071238

PubMed Abstract | Crossref Full Text | Google Scholar

20. Chakravarthy, S, Nagarajan, B, Kumar, VV, Mahesh, TR, Sivakami, R, and Annand, JR. Breast tumor classification with enhanced transfer learning features and selection using chaotic map-based optimization. Int J Comput Intell Syst. (2024) 17:18. doi: 10.1007/s44196-024-00409-8

Crossref Full Text | Google Scholar

21. Albalawi, E, Thakur, A, Ramakrishna, MT, Bhatia Khan, S, SankaraNarayanan, S, Almarri, B, et al. Oral squamous cell carcinoma detection using EfficientNet on histopathological images. Front Med. (2024) 10:1349336. doi: 10.3389/fmed.2023.1349336

PubMed Abstract | Crossref Full Text | Google Scholar

22. R, MT, Thakur, A, Gupta, M, Sinha, DK, Mishra, KK, Venkatesan, VK, et al. Transformative breast cancer diagnosis using CNNs with optimized ReduceLROnPlateau and early stopping enhancements. Int J Comput Intell Syst. (2024) 17:14. doi: 10.1007/s44196-023-00397-1

Crossref Full Text | Google Scholar

23. Jha, PDS, Khalf, MF, and Karthick, M. Convolutional neural networks for breast cancer detection using regions of interest from infrared images. Tamjeed J Healthc Eng Sci Technol. (2023) 1:44–53. doi: 10.59785/tjhest.v1i2.25

Crossref Full Text | Google Scholar

24. Faris, A, and Badamasi, M. Feasibility of breast Cancer detection through a convolutional neural network in mammographs. Tamjeed J Healthc Eng Sci Technol. (2023) 1:36–43. doi: 10.59785/tjhest.v1i2.24

Crossref Full Text | Google Scholar

25. Rani, S., Memoria, M., Rani, M., and Kumar, R., "Breast cancer detection using mammographic images over convolutional neural network," 2023 13th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India, (2023), pp. 73–78.

Google Scholar

26. Gami, B, Chauhan, K, and Panchal, BY. Breast cancer detection using deep learning In: N Marriwala, C Tripathi, S Jain, and D Kumar, editors. Mobile radio communications and 5G networks, Lecture Notes in Networks and Systems, vol. 588. Singapore: Springer (2023)

Google Scholar

27. Balasubramaniam, S, Velmurugan, Y, Jaganathan, D, and Dhanasekaran, S. A modified LeNet CNN for breast Cancer diagnosis in ultrasound images. Diagnostics. (2023) 13:2746. doi: 10.3390/diagnostics13172746

PubMed Abstract | Crossref Full Text | Google Scholar

28. Aidossov, N, Zarikas, V, Mashekova, A, Zhao, Y, Ng, EYK, Midlenko, A, et al. Evaluation of integrated CNN, transfer learning, and BN with thermography for breast Cancer detection. Appl Sci. (2023) 13:600. doi: 10.3390/app13010600

Crossref Full Text | Google Scholar

29. Nadkarni, S., and Noronha, K., "Breast cancer detection with multiparametric magnetic resonance imaging using deep learning," 2023 international conference on smart systems for applications in electrical sciences (ICSSES), Tumakuru, India (2023), pp. 1–5.

Google Scholar

30. Sasirekha, N., Karuppaiah, J., Shekhar, H., and Naga Saranya, N. ‘Breast cancer detection using histopathology image with Mini-batch stochastic gradient descent and convolutional neural network’. (2023), pp. 4651–4667.

Google Scholar

31. Ouf, M., Abdul-Hamid, Y., and Mohammed, A. (2023). An enhanced deep learning approach for breast cancer detection in histopathology images. In: Hassanien, A.E., et al., eds. The 3rd International Conference on Artificial Intelligence and Computer Vision (AICV2023), March 5–7, 2023. AICV 2023, Lecture Notes on Data Engineering and Communications Technologies, vol. 164. Springer, Cham

Google Scholar

32. Rashid, M. H. Or, Shahriyar, S. M., Shamrat, F. M. J. M., Mahbub, T., Tasnim, Z., and Ahmed, M. Z., (2023) "A convolutional neural network based classification approach for breast Cancer detection,"2023 7th International Conference on Trends in Electronics and Informatics (ICOEI), Tirunelveli, India, pp. 761–768

Google Scholar

33. Agbley, BLY, Li, JP, Haq, AU, Bankas, EK, Mawuli, CB, Ahmad, S, et al. Federated fusion of magnified histopathological images for breast tumor classification in the internet of medical things. IEEE J Biomed Health Inform. (2023):1–12. doi: 10.1109/JBHI.2023.3256974

Crossref Full Text | Google Scholar

34. Haq, U, Li, JP, Khan, I, Agbley, BLY, Ahmad, S, Uddin, MI, et al. DEBCM: deep learning-based enhanced breast invasive ductal carcinoma classification model in IoMT healthcare systems. IEEE J Biomed Health Inform. (2024):1–12. doi: 10.1109/JBHI.2022.3228577

Crossref Full Text | Google Scholar

35. Zheng, W, Lu, S, Yang, Y, Yin, Z, and Yin, L. Lightweight transformer image feature extraction network. PeerJ Comput. Sci. (2024) 10:e1755. doi: 10.7717/peerj-cs.1755

Crossref Full Text | Google Scholar

36. Lu, S, Yang, J, Yang, B, Li, X, Yin, Z, Yin, L, et al. Surgical instrument posture estimation and tracking based on LSTM. ICT Express. (2024). doi: 10.1016/j.icte.2024.01.002

Crossref Full Text | Google Scholar

37. Jiang, Z, Yang, L, Jin, L, Yi, L, Bing, P, Zhou, J, et al. Identification of novel cuproptosis-related lncRNA signatures to predict the prognosis and immune microenvironment of breast cancer patients. Front Oncol. (2022) 12. doi: 10.3389/fonc.2022.988680

PubMed Abstract | Crossref Full Text | Google Scholar

38. Gerges, F., Shih, F., and Azar, D. (2023). Automated diagnosis of acne and rosacea using convolution neural networks. In Proceedings of the 2021 4th International Conference on Artificial Intelligence and Pattern Recognition September 20–21, 2023, Xiamen, China

Google Scholar

39. Moubayed, A., Ahmed, T., Haque, A., and Shami, A. (2020) Machine learning towards enabling spectrum-as-a-service dynamic sharing. In 2020 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE) (pp. 1–6). IEEE.

Google Scholar

40. Ibrahim, JN, and Audi, L. Anxiety symptoms among Lebanese health-care students: prevalence, risk factors, and relationship with vitamin D status. J. Health Sci. (2021) 11:29–36. doi: 10.17532/jhsci.2021.1191

Crossref Full Text | Google Scholar

41. Khouloud, S, El Akoum, F, and Tekli, J. Unsupervised knowledge representation of panoramic dental X-ray images using SVG image-and-object clustering. Multimed. Syst. (2023) 29:2293–322. doi: 10.1007/s00530-023-01099-6

Crossref Full Text | Google Scholar

42. Hassan, HF, Koaik, L, Khoury, AE, Atoui, A, El Obeid, T, and Karam, L. Dietary exposure and risk assessment of mycotoxins in thyme and thyme-based products marketed in Lebanon. Toxins. (2022) 14:331. doi: 10.3390/toxins14050331

PubMed Abstract | Crossref Full Text | Google Scholar

43. Slim, A., Al Yusuf, H., Abbas, N., Abdallah, C. T., Heileman, G. L., and Slim, A. (2021) A Markov decision processes modeling for curricular analytics. In 2021 20th IEEE International Conference on Machine Learning and Applications (ICMLA) (pp. 415–421). IEEE.

Google Scholar

Keywords: machine learning, EfficientNetV2, histopathological image classification, deep learning, image processing, BreakHis dataset, individuals with disabilities, medical image interpretation

Citation: Al Moteri M, Mahesh TR, Thakur A, Vinoth Kumar V, Khan SB and Alojail M (2024) Enhancing accessibility for improved diagnosis with modified EfficientNetV2-S and cyclic learning rate strategy in women with disabilities and breast cancer. Front. Med. 11:1373244. doi: 10.3389/fmed.2024.1373244

Received: 19 January 2024; Accepted: 27 February 2024;
Published: 07 March 2024.

Edited by:

Vinayakumar Ravi, Prince Mohammad Bin Fahd University, Saudi Arabia

Reviewed by:

Gemmachis Teshite Dalu, Haramaya University, Ethiopia
Santhosh Jayagopalan, British Applied College, United Arab Emirates
Prabhishek Singh, Bennett University, India

Copyright © 2024 Al Moteri, Mahesh, Thakur, Vinoth Kumar, Khan and Alojail. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Surbhi Bhatia Khan, surbhibhatia1988@yahoo.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.