Skip to main content

METHODS article

Front. Oncol., 19 July 2024
Sec. Cancer Imaging and Image-directed Interventions

Swin-GA-RF: genetic algorithm-based Swin Transformer and random forest for enhancing cervical cancer classification

Manal Abdullah AlohaliManal Abdullah Alohali1Nora El-RashidyNora El-Rashidy2Saad Alaklabi*Saad Alaklabi3*Hela ElmannaiHela Elmannai4Saleh AlharbiSaleh Alharbi3Hager Saleh,,*Hager Saleh5,6,7*
  • 1Department of Information Systems, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
  • 2Machine Learning and Information Retrieval Department, Faculty of Artificial Intelligence, Kafrelsheiksh University, Kafrelsheiksh, Egypt
  • 3Department of Computer Science, College of Science and Humanities in Dawadmi, Shaqra University, Shaqra, Saudi Arabia
  • 4Department of Information Technology, College of Computer and Information Sciences, Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia
  • 5Faculty of Computers and Artificial Intelligence, South Valley University, Hurghada, Egypt
  • 6Data Science Institute, Galway University, Galway, Ireland
  • 7Atlantic Technological University, Letterkenny, Ireland

Cervical cancer is a prevalent and concerning disease affecting women, with increasing incidence and mortality rates. Early detection plays a crucial role in improving outcomes. Recent advancements in computer vision, particularly the Swin transformer, have shown promising performance in image classification tasks, rivaling or surpassing traditional convolutional neural networks (CNNs). The Swin transformer adopts a hierarchical and efficient approach using shifted windows, enabling the capture of both local and global contextual information in images. In this paper, we propose a novel approach called Swin-GA-RF to enhance the classification performance of cervical cells in Pap smear images. Swin-GA-RF combines the strengths of the Swin transformer, genetic algorithm (GA) feature selection, and the replacement of the softmax layer with a random forest classifier. Our methodology involves extracting feature representations from the Swin transformer, utilizing GA to identify the optimal feature set, and employing random forest as the classification model. Additionally, data augmentation techniques are applied to augment the diversity and quantity of the SIPaKMeD1 cervical cancer image dataset. We compare the performance of the Swin-GA-RF Transformer with pre-trained CNN models using two classes and five classes of cervical cancer classification, employing both Adam and SGD optimizers. The experimental results demonstrate that Swin-GA-RF outperforms other Swin transformers and pre-trained CNN models. When utilizing the Adam optimizer, Swin-GA-RF achieves the highest performance in both binary and five-class classification tasks. Specifically, for binary classification, it achieves an accuracy, precision, recall, and F1-score of 99.012, 99.015, 99.012, and 99.011, respectively. In the five-class classification, it achieves an accuracy, precision, recall, and F1-score of 98.808, 98.812, 98.808, and 98.808, respectively. These results underscore the effectiveness of the Swin-GA-RF approach in cervical cancer classification, demonstrating its potential as a valuable tool for early diagnosis and screening programs.

1 Introduction

Cervical cancer poses a significant health burden for women worldwide, with a rising incidence and mortality rate in recent years Beckmann et al. (1). As the fourth most prevalent cancer according to the World Health Organization (WHO), it accounted for 570,000 new cases and over 311,000 deaths in 2018 alone. These alarming statistics can be attributed to various factors, including limited awareness, inadequate early screening options, and a shortage of skilled healthcare professionals Beckmann et al. (1).

Early diagnosis plays a crucial role in improving cervical cancer outcomes and saving lives. Currently, several screening techniques such as visual inspection, Pap tests, histopathology tests, and human papillomavirus (HPV) testing are employed for cervical cancer detection Chitra and Kumar (2). However, manual diagnosis is prone to misdiagnosis due to inter and intra-observer variability. For instance, Pap screening tests require extensive microscopic examinations, which are not only costly but also timeconsuming. Consequently, there is an urgent need to develop an advanced and reliable model capable of providing accurate decision-making in cervical cancer diagnosis Lellée and Küppers (3).

Medical imaging techniques, including magnetic resonance imaging (MRI), computed tomography (CT) scans, and ultrasound, offer detailed insights into infected tissues, and tumor characteristics, and guide treatment decisions such as radiation therapy and chemotherapy Sarhangi et al. (4). The integration of artificial intelligence (AI) with medical imaging has emerged as a promising approach to enhance the accuracy of cervical cancer diagnosis systems Tripathi et al. (5).

Deep learning, a subset of AI, has revolutionized various domains, including computer vision, natural language processing (NLP) Saleh et al. (6), time series analysis Saleh et al. (7), and continuous health monitoring Zhu et al. (8). There is research that applied CNN and pre-trained CNN models to classify cervical cancer. For example, In AlMubarak et al. (9), the authors proposed a hybrid method based on hybrid images and DL models. In Plissiti et al. (10), the authors conducted several experiments using Support Vector Machines, Multi-layer Perceptron (MLP), and CNN. In Alsubai et al. (11), the authors proposed a deep CNN using four convolutional layers. These studies just applied CNN models to classify cervical cancer. These models have limitations that hinder their performance in medical image analysis. These models often focus on local patterns, limiting their ability to grasp the global context of the images Tripathi et al. (5). To overcome these limitations, transfer learning has emerged as a promising approach. Transfer learning leverages pre-trained deep learning models, such as convolutional neural networks (CNNs) Tripathi et al. (5), to transfer learned representations from one task to another, thereby improving classification accuracy. In recent years, researchers have introduced revolutionary approaches like Swin Transformers Cao et al. (12), which utilize shifted windows and self-attention mechanisms to process images hierarchically and efficiently. These transformers excel in capturing both local and global contextual information, exhibiting superior scalability and flexibility compared to traditional CNN models Cao et al. (12). Other research was replace fully connected layers of simple CNN by extreme gradient boosting (XGBoost) to improve classification accuracy of X-ray Images Zivkovic et al. (13).

The main questions of our work are as follows:

● How can we develop an advanced and reliable model that improves the accuracy of cervical cancer diagnosis and overcomes the limitations of existing screening techniques?

● How can we extract local and global features?

● How can we reduce the complicity of feature representation?

To answer these questions, we propose a novel approach, Swin-GA-RF, to enhance the classification performance of cervical cells in Pap smear images. The Swin-GA-RF model combines the strengths of the Swin Transformer, a genetic algorithm (GA) for feature selection, and a random forest classifier. By leveraging the Swin Transformer, we can effectively capture both local and global contextual information from medical images. The GA is then employed to reduce feature complexity and select the most informative features extracted by the Swin Transformer. Finally, the softmax classifier is replaced with a random forest to further enhance the overall classification performance.

The primary contributions of this paper are as follows:

● Introducing the Swin-GA-RF approach, which combines the Swin Transformer, GA-based feature selection, and a random forest classifier for the classification of cervical cells in Pap smear images.

● Leveraging the Swin Transformer to capture both local and global contextual information from medical images, thereby improving the accuracy of classification.

● Employing genetic algorithms to reduce feature complexity and select the most significant features extracted by the Swin Transformer.

● Enhancing the classification performance by replacing the softmax classifier with a random forest.

● Conducting comprehensive comparisons between the proposed Swin-GA-RF model and other models, including Swin Transformer and pre-trained CNN models, using two optimizer methods (SDG and Adam) in both two-class and five-class classification scenarios.

● Demonstrating the superior performance of the proposed model compared to state-of-the-art methods using publicly available cervical cancer datasets.

The remainder of this paper is organized as follows: Section 2 provides a comprehensive review of the relevant literature in the field. In Section 3, we present the methodologies employed, including details about the dataset used and a thorough explanation of the proposed Swin-GA-RF model. The experimental results are discussed in Section 4, followed by the conclusions drawn in Section 6.

2 Background literature

The purpose of this section is to review relevant studies on classifying cervical cancer and gaps in the literature.

There are studies have been AI models for classifying cervical cancer.

In AlMubarak et al. (9), the authors proposed a hybrid method based on hybrid images and DL models to classify cervical cancer. The hybrid method improves upon the DL and imaging approaches alone. In Plissiti et al. (10), the authors conducted several experiments using Support Vector Machines, Multi-layer Perceptron (MLP), and CNN to classify cervical cancer. CNN performs better than MLP classifiers. In Alsubai et al. (11), the authors proposed a CNN model using the SIPaKMeD dataset to classify five classes of cervical cells. They used segmented Pap smear images to create augmented images of cervical cells, which were then processed by a deep CNN using four convolutional layers. In Li et al. (14), the authors proposed hybrid models based on CNN and Feature Pyramid Networks (FPN). In addition to the Region Proposal Network (RPN), a global contextual aware module is introduced to improve spatial correlation between the background and the foreground. The results showed that the DGCA-RCNN model achieved the highest performance. In Wu et al. (15), the authors applied CNN with 3012 original images and 108432 augmented images. The results indicated that CNN performed better on original images than on augmented ones. In Manna et al. (16), the authors proposed an ensemble-based model based on fuzzy rank-based fusion using InceptionV3, Xception, and DenseNet 169 to classify two and five classes of cervical cancer. Their proposed model recorded the highest performance. In Pramanik et al. (17), the authors proposed ensemble learning, which integrates the outputs of InceptionV3, MobileNet V2, and Inception ResNet V2. Additionally, they employed fuzzy distance-based measures and applied fuzzy distance measures using the product rule to calculate the final predictions. In Ghoneim et al. (18), the authors employed the proposed CNN-ELM, which is based on convolutional neural networks (CNN) and extreme learning machines (ELM). CNN was utilized for feature extraction, while ELM was employed for classifying normal and abnormal cervical cancer cases. CNN-ELM demonstrated superior performance compared to Multi-Layer Perceptron (MLP) and autoencoder (AE)-based classifiers. In Chen et al. (19), the authors proposed a hybrid loss function with label smoothing to improve the distinguishing power of CNN. The results showed that their proposed model achieved satisfactory accuracy. In Tripathi et al. (5), the authors applied different pre-trained models: InceptionResNetV2, VGG19, DenseNet201, and Xception, to classify cervical images using the SIPaKMeD dataset. In Yaman and Tuncer (20), the authors applied the pre-trained CNN models on the SIPAKMED pap-smear image dataset to the assessment of forthcoming classification techniques. The results showed that ResNet-152 recorded the highest accuracy.

The related work primarily applied DL models and pre-trained CNN models for cervical cancer classification, while others utilized vision transformer models for the same purpose that recorded the lowest performance. In this paper, we propose a novel approach based on Swin Transformer and GA to improve the performance of cervical cell classification. ViT Transformer is employed to capture both local and global contextual information from images, while GA is used to select optimal features from the representations extracted by Swin. Furthermore, the softmax classifier is replaced with a SVM classifier to further enhance performance.

3 Materials and method

The primary steps involved in classifying cervical cells are depicted in Figure 1. The main objective of this study is to propose a novel approach (Swin-GA-RF) aimed at enhancing classification performance by leveraging a combination of techniques, including the Swin Transformer, feature selection, and Random Forest, to classify cervical cells in Pap smear images.

Figure 1
www.frontiersin.org

Figure 1 The main steps of classifying cervical cancer detection.

3.1 Dataset description

A dataset named SIPaKMeD1, provided by Plissiti et al. dat (2018), is used to evaluate the proposed model (21). It is a balanced dataset and it includes five classes of Pap smear images which are superficial-intermediate (Superf), parabasal (Parab), koilocytes (Koilo), dyskeratotic (Dysker), and metaplastic (Metapl). The dataset includes 3644 images. The number of images in each class is shown in Table 1.

Table 1
www.frontiersin.org

Table 1 The number of images in each class.

3.2 Image augmentation

Image augmentation involves making changes to an image in terms of color and position. Positional manipulation is achieved by altering the position of pixels, while color manipulation involves changing the pixel values. Data augmentation aims to enhance the visual characteristics of an image. It includes techniques such as flipping, cropping, and resizing, Bao et al. (22). These techniques contribute to improving the overall generalization performance of the model by exposing it to a wide variety of images during the training process. Example of images for each class as shown in Figure 2.

Figure 2
www.frontiersin.org

Figure 2 Example of images for SIPaKMeD1.

3.3 CNN model

CNN consists of input, an output, and multiple hidden layers.

● The Convolutional layer is the first layer utilized to gather different features, handling the majority of the processing Rai and Rivas (23). Convolution occurs between the input and a window filter by sliding the filter over the input and calculating the dot product over the filter and the input regarding the filter’s size Kuo (24).

● The Pooling layer’s primary goal is to lower the size of the convolved feature map to lower computational expenditures and the number of calculations performed Bailer et al. (25). It summarizes the properties of a convolution layer Nasr-Esfahani et al. (26) The feature map is used to obtain the largest element in max pooling.

● In fully connected layers, each neuron’s input is the weighted sum of all the outputs from the neurons in the layer above Basha et al. (27).

3.4 Pre-trained CNN models

DenseNet121, VGG16, ResNet18, and AlexNet are utilized for the classification of cervical cancer.

● DenseNet is a special architecture of the CNN that was developed to address the vanishing gradient problem and improve feature extraction within the network Jaiswal et al. (28).

● ResNet is a residual network that was developed to address the challenges associated with degradation in deep neural networks. ResNet tackles this problem by introducing residual connections, which enable the network to learn residual mappings instead of relying solely on the underlying mappings Deng et al. (29).

● VGG (Visual Geometry Group) is an improved iteration of convolutional neural networks renowned for its effectiveness. The architecture of VGG comprises multiple convolutional layers followed by fully connected layers.

● AlexNet is a deep neural network architecture composed of eight layers, featuring five convolutional layers followed by three fully connected layers. The success of this model can be attributed to several key factors Alzubaidi et al. (30).

● MobileNet is CNN architecture for mobile and embedded devices with minimal computing resources.

These models are tuned for accuracy, minimal computational complexity, and memory footprint Howard et al. (31). Depth-wise separable and point-wise convolutions minimize computational cost in MobileNet models. They are commonly employed in mobile real-time image processing applications Howard et al. (31).

3.5 Swin Transformer

Swin Transformer represents a particular type of deep learning model proposed by academics from the Chinese University of Hong Kong and other establishments Liu et al. (32). Swin Transformer’s core principle is to separate the input image into non-overlapping patches and analyze them hierarchically, resulting in a two-stage architecture Liu et al. (33). In the first stage, the model applies self-attention to each local patch to capture local dependencies. In the second stage, the model applies self-attention across patches to capture global dependencies. To handle images with a high resolution efficiently, the Swin Transformer employs a shifted window method Zhang et al. (34). Instead of moving the fixedsized window across the image, the model shifts its position recursively. This enables the model to efficiently capture information at various spatial resolutions Wang et al. (35). The Swin Transformer hierarchical processing mechanism performed well on a variety of computer vision tasks, including picture categorization and object detection, while using fewer computation resources than competing methods Wang et al. (36).

Swin Transformer pairing local self-attention within shifted windows in hierarchical processing to model both local and global relationships in high-resolution images. Input images are first separated into nonoverlapping patch tokens. The patches are then projected linearly into a lower-dimensional embedding space. Swin transforms the patches recursively into shifted windows, allowing the model to gather data at various spatial resolutions Ma et al. (37). The Swin Transformer employs a stack of Transformer encoder blocks. Each encoder block is made up of two major components: a shifting window self-attention module that captures local dependencies between patches inside a window, and a feed-forward neural network that allows the model to record local information within a patch Xiao et al. (38).

The Swin Transformer performs hierarchical processing by stacking several Transformer encoder blocks. The output of one block is used as the input for the next, allowing the model to capture global dependencies across patches Zhou et al. (39). Finally, following hierarchical processing, a classification head is attached to the Swin Transformer to predict class labels for image classification tasks. The classification head is often made up of one or more fully connected layers, which are then activated using softmax Taslimi et al. (40). In general, Swin Transformer has performed well on a variety of image recognition benchmarks, demonstrating its ability to handle enormous amounts of imagery effectively while reaching cutting-edge accuracy and computational efficiency.

Figure 3 shows the general architecture of the swin transformer. As shown in Figure 3, each swim transformer consists of LayerNorm(LN) with multi head attention with 2 MLP layers. Window-based multi-head attention (W MSA) and the shifted Window-based multi-head attention (SW MSA). The swim transformer block could be formulated as follows:

Figure 3
www.frontiersin.org

Figure 3 General architecture of swim transformer (32).

Zl=WMSA (LN(Zl1))+Zl11
Zl=MLP (LN(Zl))+Zl2
Z+1=SWMSA (LN(Zl))+Zl3
Zl+1=MLP (LN(Zl+1))+Zl+14

3.6 Genetic algorithm

The utilization of Genetic Algorithms (GAs) for feature optimization has been extensively explored across various domains, demonstrating their effectiveness in selecting informative features that yield valuable insights Wang (41). In the context of feature optimization, GAs treat each feature as a gene within the population, forming chromosomes that represent potential solutions comprising specific subsets of features. Evaluating the quality of these solutions is achieved through the fitness function, which assesses their performance based on predefined criteria Reeves (42) El-Rashidy et al. (43).

The GA process initiates by randomly selecting an initial solution from the feature space. Through iterative steps of selection, crossover, and mutation, the process continues, refining the solutions over generations. Solutions with higher fitness values, indicating better performance, have an increased likelihood of being selected for subsequent iterations. GAs have proven to be particularly effective in handling high-dimensional data and supporting multi-objective feature optimization Reeves (42). The strength of GAs lies in their ability to treat each feature as a gene, allowing for comprehensive exploration of the feature space. By iteratively evaluating and evolving solutions through selection, crossover, and mutation, GAs guide the search toward promising subsets of features that improve the classification performance. This capability is especially valuable when dealing with high-dimensional data, enabling the identification of relevant and influential features that contribute to superior classification accuracy Sivanandam et al. (44).

3.7 The proposed model

In our pursuit of building a classifier that achieves exceptional performance, we have devised a groundbreaking model that integrates the Swin transformer, genetic algorithm, and RF. This fusion of techniques allows us to leverage the strengths of each component to create a robust and accurate classifier. The process of constructing this high-performance classifier follows a meticulously designed sequence, encompassing the following steps as shown in Figure 4.

● The images are initially divided into small non-overlapping patches, where each patch represents an isolated token.

● All patches are then flattened into 2D tokens, which serve as input for the transformer.

● The Swin transformer incorporates positional information by adding it to the spatial information.

● Window-based positions are utilized to transform the shifted window into the position embedding scheme. This approach enables the Swin transformer to capture both local and global information.

● The Swin transformer processes the aggregated features using stage-wise processing, which includes multiple stages with transformer layers. This strategy enhances processing performance by reducing token resolution in earlier stages and subsequently improving it in subsequent layers.

● Each stage in the transformer consists of multiple layers, with each layer incorporating self-attention and feed-forward layers. The self-attention layers enable the model to capture contextual relationships and relevant patches.

● After processing each window, all tokens are fused to reintegrate the information, facilitating the capture of both local and global information.

● Feature representations are extracted from the layer preceding the softmax layer. Genetic Algorithms (GA) are then applied to reduce feature complexity and select the most significant features. Finally, in the last layer, SoftMax is replaced with a random forest to enhance results.

Figure 4
www.frontiersin.org

Figure 4 The proposed model.

3.8 Evaluation models

The models are evaluated using evaluation metrics, as shown in Equations 14. Where True positive (TP)— is the result of the correctly classified positive class. True negative (TN)—the results of the correctly classified negative class. False-positive (FP)—the results of the incorrectly classified positive class. False-negative (FN)—the results of the incorrectly classified negative class.

Accuracy=TP+TNTP+FP+TN+FN5
Recall=TPTP+FN6
Precision=TPTP+FP7
F1score=2precisionrecallprecision+recall8

4 Experiments results

This section presents the results of comparing the proposed model (Swin-GA-RF), the Swin Transformer, CNN, and pre-trained CNN models for classifying cervical cells in two classification scenarios: binary and five classes.

4.1 Experimental setup

The models were implemented using Python, PyTorch, and the Monai library. The sklearn-genetic library is used to implement genetic feature selection methods Calzolari (45) The image dataset was divided into 70% for training, 10% for validation, and 20% for testing. The number of images in each set is shown in Table 2. We conducted experiments using both the Adam and SGD optimizers. The cross-entropy loss function was employed for models. The hyperparameters of Swin, CNN models, and the pre-trained CNN models were CrossEntropy as loss function, epochs=70, and batch size=30. The hyperparameters of RF are Max depth=10, min samples split=2, bootstrap=True, and criterion=gini. A summary of the GA parameters for selecting the best features can be found in Table 3.

Table 2
www.frontiersin.org

Table 2 The number of images in set.

Table 3
www.frontiersin.org

Table 3 Parameters of GA.

4.2 The results of applying models with five classes

4.2.1 The results of Adam optimizer

Table 4 presents the results of each class obtained from different models, including CNN, DenseNet121, VGG16, ResNet18, AlexNet, Swin Transformer, and Swin-GA-RF using Adam to classify five classes of Cervical Cancer with Adam optimizer. We can see that, CNN records the highest precision, recall, and F1-score at 94.44, 96.84, and 95.62, respectively for the Parab class. CNN records the lowest results for the Koilo and Metapl classes. DenseNet121 consistently demonstrates high precision, recall, and F1-score for Parab, achieving a high recall at 99.37. It particularly excels in the Dysker and Parab classes but struggles with the Koilo class, indicating challenges in accurate classification. VGG16 records the highest precision, recall, and F1-score at 98.70, 96.20, and 97.44, respectively for the Parab class, it faces difficulties in accurately classifying the Koilo class, exhibiting relatively lower precision and recall rates. ResNet18 showcases notable improvements in classification performance, achieving an F1-score of 98.73 for Parab. It demonstrates considerable enhancement in the Dysker, Superf, and Parab classes. However, it encounters challenges in accurately classifying the Koilo class, where its recall rate is comparatively lower. AlexNet records the highest precision, recall, and F1-score at 94.44, 96.84, and 95.62, respectively for the Parab class. The Swin Transformer emerges as a significant improvement over CNN and the pre-trained CNN models, achieving an F1-score of 99.37 for the Parab class. It exhibits significant enhancements in all classes, with notable improvements in the Parab class, where it achieves a precision of 93.90, recall of 93.33, and F1-score of 93.62 with the Koilo class. Finally, the proposed Swin-GA-RF model outperforms all other models, recording the highest overall performance. It records the highest precision of 100, recall of 100, and F1-score of 100 for the Superf class, The Swin-GA-RF model showcases perfect precision, recall, and F1-score, highlighting its ability to accurately classify this type of cervical cancer. These improvements can be attributed to the utilization of the GA for feature selection and the Random Forest classifier, which reduces feature complexity and enhances the overall performance of the model.

Table 4
www.frontiersin.org

Table 4 The results of each class for the models to classify five classes of cervical cancer with Adam optimizer.

Figure 5 presents the average accuracy, precision, recall, and F1-score of CNN, pre-trained CNN models, Swin Transformer, and Swin-GA-RF models for five classes using Adam optimizer. We can see that the Swin transformer achieves the highest results compared to pre-trained models. Furthermore, the improved version of the Swin transformer, referred to as Swin-GA-RF, records the highest performance with the highest accuracy, F1-score, and improved results at 1.5. This enhancement is attributed to the reduction in feature complexity by the Genetic Algorithm (GA) and Random Forest (RF). Both DenseNet121 and VGG16 exhibit identical results, with accuracy above 92 and recall above 92.308. The third-highest performance is recorded by ResNet18. To compare these results, we conducted a detailed comparison between the Swin transformer and Swin-GA-RF models in each class, as illustrated in Figure 6. Swin-GA-RF improved TP by 8% compared to the Swin transformer for the Koilo class.

Figure 5
www.frontiersin.org

Figure 5 The average of results for models with Adam optimizer for five classes.

Figure 6
www.frontiersin.org

Figure 6 Confusion matrix of Swin and Swin-GA-RF with Adam optimizer for five classes.

4.2.2 The results of SGD optimizer

Table 5 presents the results of each class obtained from different models, including CNN, DenseNet121, VGG16, ResNet18, AlexNet, Swin Transformer, and Swin-GA-RF using SDG to classify five classes of Cervical Cancer with Adam optimizer. CNN records the highest precision, recall, and F1-score at 91.00, 91.09, and 92.60, respectively for class Parab. DenseNet121 consistently demonstrates high precision, recall, and F1-score across various classes, achieving a precision of 97.37 for the Parab class. VGG16 records the highest precision, recall, and F1-score at 91.00, 91.09, and 92.60, respectively for class Parab. ResNet18 showcases notable improvements in classification performance, achieving an average F1-score of 92.57. It demonstrates considerable enhancement in the Dysker and Parab classes, with a precision of 98.70, recall of 96.20, and an F1-score exceeding 97. However, it encounters challenges in accurately classifying the Koilo class, where its recall rate is comparatively lower. VGG16 records the highest precision, recall, and F1-score for the Parab and Superf classes, and the lowest results for Koilo and Metapl classes. ResNet18 showcases notable improvements in classification performance, achieving an F1-score of 98.73 for Parab. AlexNet records the highest precision, recall, and F1-score at 94.44, 96.84, and 95.62, respectively for the Parab class. The Swin Transformer emerges as a significant improvement over CNN and the pre-trained CNN models, achieving an F1-score of 99.37 for the Parab class. It exhibits significant enhancements in all classes, with notable improvements in the Parab class, where it achieves a precision of 93.90, recall of 93.33, and F1-score of 93.62 with the Koilo class. Finally, the proposed Swin-GA-RF model outperforms all other models, recording the highest overall performance. It achieves notable improvements in the Dysker, Koilo, and Metapl classes, demonstrating enhanced precision, recall, and F1-score compared to the Swin Transformer. The Swin-GA-RF model showcases perfect precision, recall, and F1-score in the Parab class, highlighting its ability to accurately classify this type of cervical cancer. These improvements can be attributed to the utilization of the Genetic Algorithm (GA) for feature selection and the Random Forest classifier, which reduces feature complexity and enhances the overall performance of the model.

Table 5
www.frontiersin.org

Table 5 The results of each class for the models to classify five classes of cervical cancer with SDG optimizer.

Figure 7 presents the average accuracy, precision, recall, and F1-score of CNN, pre-trained CNN models, Swin Transformer, and Swin-GA-RF models for five classes using Adam optimizer. The Swin Transformer achieves the highest results compared to the pre-trained models. Furthermore, the improved version of the Swin Transformer, referred to as Swin-GA-RF, demonstrates the highest performance with the highest accuracy, F1-score, and improved results at 1.5. The third-highest performance is recorded by ResNet18, with accuracy, precision, recall, and F1-score at 95.813, 95.887, 95.813, and 95.785, respectively. To compare these results, we conducted a detailed comparison between the Swin Transformer and Swin-GARF models in each class, as illustrated in Figure 8. Swin-GA-RF improved TP by 3% compared to the Swin Transformer for the Metapl class. Additionally, it records 100 TP for Parab and Superf classes.

Figure 7
www.frontiersin.org

Figure 7 The average of results for models with SDG optimizer for five classes.

Figure 8
www.frontiersin.org

Figure 8 Confusion matrix of Swin and Swin-GA-RF with SDG optimizer for five classes.

4.3 The results of applying models with two classes

4.3.1 The results of Adam optimizer

Table 6 presents the results of each class obtained from different models, including CNN, DenseNet121, VGG16, ResNet18, AlexNet, Swin Transformer, and Swin-GA-RF using Adam to classify two classes of Cervical Cancer with Adam optimizer. We can see that, CNN records the highest recall, and F1-score at 94.44, 96.84, and 95.62, respectively for the normal cells class. CNN records the lowest recall for the abnormal cells class. DenseNet121 records the highest precision, and F1-score at 96.79, 96.84, and 95.17, respectively for normal cells class. VGG16 records the lowest recall at 88.72 for abnormal cells class. ResNet18 records the highest results compared to CNN DenseNet121, VGG16, and AlexNet. The Swin Transformer emerges as a significant improvement over CNN and the pre-trained CNN models, the proposed Swin-GA-RF model outperforms all other models, recording the highest overall performance and exhibits significant enhancements in all classes. It records the highest precision, recall, and F1-score above 99 for each class.

Table 6
www.frontiersin.org

Table 6 The results of each class for the models to classify two classes of cervical cancer with Adam optimizer.

Figure 9 presents the average of accuracy, precision, recall, and F1-score pre-trained CNN models, Swin Transformer, and Swin-GA-RF models using Adam to classify two classes: abnormal cells and normal cells. The Swin Transformer achieves the highest results compared to the pre-trained models. Furthermore, the improved version of the Swin Transformer, referred to as Swin-GA-RF, records the highest performance with the highest accuracy, F1-score, and improved results at 2.5. Specifically, it achieves 99.012 accuracy, 99.555 precision, 99.61 recall, and an F1-score of 99.58. The third-highest performance is recorded by ResNet18. To further compare these results, we conducted a detailed comparison between the Swin Transformer and Swin-GA-RF models in each class, as illustrated in Figure 10. Swin-GA-RF improved TP by 4% compared to the Swin Transformer for abnormal cells and normal cells.

Figure 9
www.frontiersin.org

Figure 9 The average of results for models with Adam optimizer for two classes.

Figure 10
www.frontiersin.org

Figure 10 Confusion matrix of Swin and Swin-GA-RF with Adam optimizer for two classes.

4.3.2 The results of SDG optimizer

Table 7 presents the results of each class obtained from different models, including CNN, DenseNet121, VGG16, ResNet18, AlexNet, Swin Transformer, and Swin-GA-RF using SDG to classify two classes of Cervical Cancer with Adam optimizer. We can see that, CNN records the highest recall, and F1-score at 92.08, 92.87, and 92.47, respectively for the normal cells class. CNN records the lowest recall for abnormal cells class. DenseNet121 records the highest precision, and F1-score at 93.88, 94.66, and 94.96, respectively for normal cells class. VGG16 records the lowest recall at 90.86 for the abnormal cells class. ResNet18 records the highest results compared to CNN DenseNet121, VGG16, and AlexNet. The Swin Transformer emerges as a significant improvement over CNN and the pre-trained CNN models. It exhibits significant enhancements in all classes, the proposed Swin-GA-RF model outperforms all other models, recording the highest overall performance. The Swin-GA-RF model showcases perfect precision, recall, and F1-score, highlighting its ability to accurately classify this type of cervical cancer.

Table 7
www.frontiersin.org

Table 7 The results of each class for the models to classify two classes of cervical cancer with SDG optimizer.

Figure 11 presents the average of accuracy, precision, recall, and F1-score pre-trained CNN models, Swin Transformer, and Swin-GA-RF models using SDG to classify two classes: abnormal cells and normal cells. We can observe that the Swin Transformer achieves the highest results compared to the pre-trained models. Furthermore, the improved version of the Swin Transformer, referred to as Swin-GA-RF, achieves the highest performance with the highest accuracy, F1-score, and improved results at 1.5. It records an accuracy of 98.06, a precision of 98.240, a recall of 98.400, and an F1-score of 98.06. Both DenseNet121 and AlexNet exhibit identical results, with accuracy at 93.605, and 93.88, respectively. The third-highest performance is recorded by ResNet18. To compare these results, we conducted a detailed comparison between the Swin Transformer and Swin-GA-RF models in each class, as illustrated in Figure 12. Swin-GA-RF improved TP by 3% compared to the Swin Transformer for abnormal cells and normal cells.

Figure 11
www.frontiersin.org

Figure 11 The average of results for models with SDG optimizer for two classes.

Figure 12
www.frontiersin.org

Figure 12 Confusion matrix of Swin and Swin-GA-RF with SDG optimizer for two classes.

4.4 Discussion

4.4.1 The best models

We conducted various experiments to compare the proposed model with different models, including Swin Transformer and pre-trained CNN models, using two optimization methods: SGD and Adam. Additionally, we classified cervical cells in Pap smear images into two classes (abnormal cells and normal cells) and five classes (Dysker, Koilo, Metapl, Parab, and Superf). From Figure 13 and Figure 14, we can observe that the proposed model (Swin-GA-RF) records the highest performance compared to other models. Also, the results using Adam Optimizer have better performance compared to SDG optimizer. Figure 14 shows a comparison of the best models between two optimizers for five classes. We can see that Swin-GA-RF achieves the highest accuracy at 98.808 with the Adam optimizer, while DenseNet121 has the lowest accuracy at 90.764 with the SGD optimizer. The Swin transformer records the second-highest accuracy at 97.044 for the Adam optimizer.

Figure 13
www.frontiersin.org

Figure 13 Best models for binary classes.

Figure 14
www.frontiersin.org

Figure 14 Best models for five classes.

4.4.2 Statistical analysis

To ensure a rigorous and comprehensive statistical analysis, we employed well-established methodologies in our study. We utilized The Friedman test, Derrac et al. (46), which was chosen due to its suitability for comparing multiple related samples. This non-parametric test allowed us to thoroughly evaluate the presence of significant variations among the models, enabling us to determine whether there were noteworthy differences based on the observed data. The test shows if there is a significant difference between the observed data. Following the Friedman test, we conducted the Nemenyi test as a post-hoc analysis to extract the ranks of each model. This additional step provided a deeper understanding of the differences between the models by facilitating multiple comparisons Brown and Mues (47).

Specifically, the Nemenyi test allowed us to identify specific pairs of models that exhibited statistically significant differences in their performance, providing valuable insights into their relative strengths and weaknesses. The Nemenyi test represents the difference in terms of critical difference. To enhance the interpretability of the results, we created a critical difference diagram (CD diagram) as depicted in Figure 15. The CD diagram, a widely utilized visualization tool in multiple comparison analyses, played a crucial role in highlighting the significant differences among the models. By displaying the average ranks of the models, it provided a clear and concise representation of their performance disparities. Notably, the CD diagram helped identify which pairs of models showed statistically significant differences, further aiding in the comprehensive understanding of their relative performances. The CD diagram offers a concise way to interpret the results of multiple comparison analyses.

Figure 15
www.frontiersin.org

Figure 15 Critical difference diagram.

4.4.3 Comparison with existing state-of-the-art methods

In this study, we proposed Swin-GA-RF to classify cervical cancer. Table 8 provided a comparison to the existing studies and the proposed model for classifying cervical cancer. As can be seen, studies are based on different CNN architectures, ViT transformers and do not use Swin transformer and optimization feature extraction methods. We can see that the Swin-GA-RF achieved the highest accuracy in each class case. The authors in AlMubarak et al. (9) Plissiti et al. (10) Alsubai et al. (11) Manna et al. (16) Pramanik et al. (17) Chen et al. (19) to classify five classes of cervical cancer. In AlMubarak et al. (9), the authors applied a hybrid method and recorded an accuracy of 80.72. In Plissiti et al. (10), SVM was utilized, resulting in an accuracy of 94.44. In Alsubai et al. (11), CNN achieved an accuracy of 91.13.

Table 8
www.frontiersin.org

Table 8 Comparison with literature studies.

Some authors used advanced methods of DL and ensemble learning. For example, ensemble learning models applied in Manna et al. (16) Pramanik et al. (17) Win et al. (48). In Manna et al. (16), an ensemble-based model achieved an accuracy of 98.55. In Win et al. (48), a bagging ensemble method achieved an accuracy of 98.27. In Win et al. (48), a bagging ensemble method attained an accuracy of 94.09. In Manna et al. (16) Pramanik et al. (17), ensemble-based models recorded accuracies of 97.55 and 96.96, respectively. Other applied ViT transformer Pacal and Kılıcarslan (50) recorded 91.93 accuracy. Hybrid models such as ViT-CNN ensemble based on VIT and CNN recorded 97.65 accuracy Maurya et al. (52). SE-ResNet152 is transformer learning that was optimized by DHO and recorded 97.68 accuracy Ravindran et al. (51). In Li et al. (53), VTCNet mobed that combined CNN-SPPF and ViT and recorded 98.02 accuracy. In Deo et al. (49), CerviFormer achieved an accuracy of 93.70. In Chen et al. (19), CNN with loss recorded an accuracy of 96.18. In Ghoneim et al. (18), CNN-ELM achieved an accuracy of 91.20. We can see that the Swin-GA-RF achieved the highest accuracy in each class case.

5 The practical implications and potential applications of the proposed methodology

The proposed methodology in our research study has several practical implications and potential applications in the field of cervical cancer diagnosis and beyond. Some of these implications and applications include:

Improved Cervical Cancer Diagnosis: The primary practical implication of our methodology is the potential for significantly improving the accuracy of cervical cancer diagnosis. By leveraging the advanced features of the Swin Transformer, genetic algorithm-based feature selection, and a random forest classifier, our model can effectively classify cervical cells in Pap smear images.

Enhanced Screening Efficiency: Our methodology has the potential to enhance the efficiency of cervical cancer screening programs. This targeted approach can optimize the utilization of healthcare resources and reduce the burden on healthcare systems.

Support for Healthcare Professionals: The proposed methodology can serve as a valuable tool for healthcare professionals involved in cervical cancer diagnosis. By providing a reliable and automated classification system. This support can lead to improved workflow management and better allocation of resources.

Generalizability to Other Medical Imaging Tasks: The combination of the Swin Transformer, genetic algorithm-based feature selection, and a random forest classifier has the potential to be extended to other medical imaging domains. The methodology can be applied to various image-based diagnostic tasks, such as the classification of different types of cancer cells or the detection of other diseases. thereby benefiting a wider range of diagnostic processes.

Finally, it is important to note that while our study focuses on cervical cancer diagnosis, the practical implications and potential applications of our methodology extend beyond this specific domain. The combination of advanced deep learning techniques, feature selection algorithms, and classification models can be adapted and applied to other medical imaging tasks.

6 Conclusion

Women’s cervical cancer is frequently fatal, so early detection is crucial to reducing the number of cases. This paper proposes a novel approach to enhancing classification performance by leveraging a combination of techniques, including Swin Transformer, GA feature selection, and replacing the softmax layer with Random Forest. To propose Swin-GA-RF, the Swin Transformer is employed to capture both local and global contextual information from images. Then, we extracted feature representation from the Swin transformer, then A genetic algorithm-based feature selection was used to determine the best feature set from the extracted features, then the SoftMax classifier was replaced by random forest to enhance accuracy. we applied different image augmentation including flipping, cropping, and resizing to enhance the quality of the images. Through extensive experiments, we compared the performance of the proposed Swin-GA-RF model with other models, including the Swin Transformer, CNN, and pre-trained CNN models, using two optimizer methods, SGD and Adam. The results demonstrate that Swin-GA-RF, particularly when utilizing the Adam optimizer, achieved the highest performance in both binary and five-class classification tasks. For binary classification, it achieved an accuracy, precision, recall, and F1-score of 99.012, 99.015, 99.012, and 99.011, respectively. In the five-class classification, it achieved an accuracy, precision, recall, and F1-score of 98.808, 98.812, 98.808, and 98.808, respectively.

While the proposed Swin-GA-RF approach shows promising results, it is essential to acknowledge its limitations. Firstly, The detection and classification were performed at the cell level, which may not fully represent the complexity of cancerous lesions at the tissue or organ level. Additionally, the dataset used in the study is large and contains various classes that are closely related, posing challenges in accurately distinguishing between them. The overlapping features among classes can lead to misclassifications or lower accuracy rates, particularly in distinguishing closely related types of cervical cancer. Future work in this area could focus on expanding the dataset to include a broader range of cervical cell abnormalities, exploring additional optimization techniques for the Swin-GA-RF model, and conducting further comparative analyses with other state-of-the-art approaches. Additionally, efforts can be made to address the practical limitations by improving dataset collection, and data preprocessing techniques, and optimizing the model for efficient deployment in real-world scenarios.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material. Further inquiries can be directed to the corresponding author.

Author contributions

MAA: Writing – original draft, Writing – review & editing, Methodology, Validation. NE-R: Investigation, Methodology, Writing – original draft, Writing – review & editing. SAla: Writing – original draft, Writing – review & editing, Software. HE: Writing – original draft, Writing – review & editing, Investigation. SAlh: Writing – original draft, Writing – review & editing, Methodology. HS: Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Resources, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This study was supported by Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2024R330), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia.

Acknowledgments

The authors would like to thank Princess Nourah bint Abdulrahman University Researchers Supporting Project number (PNURSP2024R330), Princess Nourah bint Abdulrahman University, Riyadh, Saudi Arabia. The authors would like to thank the deanship of scientific research at Shaqra University for supporting this work.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Beckmann MW, Stübs FA, Koch MC, Mallmann P, Dannecker C, Dietl A, et al. Diagnosis, therapy and follow-up of cervical cancer. guideline of the dggg, dkg and dkh (s3-level, awmf registry no. 032/033ol, may 2021)–part 1 with recommendations on epidemiology, screening, diagnostics and therapy. Geburtshilfe und Frauenheilkunde. (2022) 82:139–80. doi: 10.1055/a-1671-2158

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Chitra B, Kumar S. Recent advancement in cervical cancer diagnosis for automated screening: a detailed review. J Ambient Intell Humanized Computing. (2022), 1–19. doi: 10.1007/s12652-021-02899-2

CrossRef Full Text | Google Scholar

3. Lellé RJ, Küppers V. (2023). Cervical cancer prevention, diagnosis, and management within a low-resource environment, in: Colposcopy: Comprehensive Textbook and Atlas, . pp. 225–86. Springer.

Google Scholar

4. Sarhangi HA, Beigifard D, Farmani E, Bolhasani H. Deep learning techniques for cervical cancer diagnosis based on pathology and colposcopy images. arXiv preprint arXiv:2310.16662. (2023).

Google Scholar

5. Tripathi A, Arora A, Bhan A. (2021). Classification of cervical cancer using deep learning algorithm, in: 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), . pp. 1210–8. IEEE.

Google Scholar

6. Saleh H, Mostafa S, Gabralla LA, O. Aseeri A, El-Sappagh S. Enhanced arabic sentiment analysis using a novel stacking ensemble of hybrid and deep learning models. Appl Sci. (2022) 12:8967.

Google Scholar

7. Saleh H, Younis EM, Sahal R, Ali AA. Predicting systolic blood pressure in real-time using streaming data and deep learning. Mobile Networks Appl. (2021) 26:326–35.

Google Scholar

8. Zhu P-C, Wan J-J, Shao W, Meng X-C, Chen B. Colorectal image analysis for polyp diagnosis. Front Comput Neurosci. (2024) 18:1356447.

PubMed Abstract | Google Scholar

9. AlMubarak HA, Stanley J, Guo P, Long R, Antani S, Thoma G, et al. A hybrid deep learning and handcrafted feature approach for cervical cancer digital histology image classification. Int J Healthcare Inf Syst Inf (IJHISI). (2019) 14:66–87.

Google Scholar

10. Plissiti ME, Dimitrakopoulos P, Sfikas G, Nikou C, Krikoni O, Charchanti A. (2018). Sipakmed: A new dataset for feature and image based classification of normal and pathological cervical cells in pap smear images, in: 2018 25th IEEE International Conference on Image Processing (ICIP), . pp. 3144–8. IEEE.

Google Scholar

11. Alsubai S, Alqahtani A, Sha M, Almadhor A, Abbas S, Mughal H, et al. Privacy preserved cervical cancer detection using convolutional neural networks applied to pap smear images. Comput Math Methods Med. (2023) 2023. doi: 10.1155/2023/9676206

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Cao H, Wang Y, Chen J, Jiang D, Zhang X, Tian Q, et al. (2022). Swin-unet: Unet-like pure transformer for medical image segmentation, in: European conference on computer vision, . pp. 205–18. Springer.

Google Scholar

13. Zivkovic M, Bacanin N, Antonijevic M, Nikolic B, Kvascev G, Marjanovic M, et al. Hybrid cnn and xgboost model tuned by modified arithmetic optimization algorithm for covid-19 early diagnostics from x-ray images. Electronics. (2022) 11:3798. doi: 10.3390/electronics11223798

CrossRef Full Text | Google Scholar

14. Li X, Xu Z, Shen X, Zhou Y, Xiao B, Li T-Q. Detection of cervical cancer cells in whole slide images using deformable and global context aware faster rcnn-fpn. Curr Oncol. (2021) 28:3585–601.

PubMed Abstract | Google Scholar

15. Wu M, Yan C, Liu H, Liu Q, Yin Y. Automatic classification of cervical cancer from cytological images by using convolutional neural network. Bioscience Rep. (2018) 38:BSR20181769. doi: 10.1042/BSR20181769

CrossRef Full Text | Google Scholar

16. Manna A, Kundu R, Kaplun D, Sinitca A, Sarkar R. A fuzzy rank-based ensemble of cnn models for classification of cervical cytology. Sci Rep. (2021) 11:14538.

PubMed Abstract | Google Scholar

17. Pramanik R, Biswas M, Sen S, de Souza Júnior LA, Papa JP, Sarkar R. A fuzzy distance-based ensemble of deep models for cervical cancer detection. Comput Methods Programs Biomedicine. (2022) 219:106776.

Google Scholar

18. Ghoneim A, Muhammad G, Hossain MS. Cervical cancer classification using convolutional neural networks and extreme learning machines. Future Generation Comput Syst. (2020) 102:643–9.

Google Scholar

19. Chen W, Shen W, Gao L, Li X. Hybrid loss-constrained lightweight convolutional neural networks for cervical cell classification. Sensors. (20223272) 22. doi: 10.3390/s22093272

CrossRef Full Text | Google Scholar

20. Yaman O, Tuncer T. Exemplar pyramid deep feature extraction based cervical cancer image classification model using pap-smear images. Biomed Signal Process Control. (2022) 73:103428.

Google Scholar

21. Dataset. Cervical cancer largest dataset (sipakmed) (2018). Available online at: https://www.kaggle.com/datasets/prahladmehandiratta/cervical-cancer-largest-dataset-sipakmed (Accessed 2024).

Google Scholar

22. Bao H, Dong L, Piao S, Wei F. Beit: Bert pre-training of image transformers. arXiv preprint arXiv:2106.08254. (2021).

Google Scholar

23. Rai M, Rivas P. 2020 international conference on computational science and computational intelligence (CSCI). IEEE (2020). p. 1560–7.

Google Scholar

24. Kuo C-CJ. Understanding convolutional neural networks with a mathematical model. J Visual Communication Image Representation. (2016) 41:406–13.

Google Scholar

25. Bailer C, Habtegebrial T, varanasi K, Stricker D. Fast feature extraction with cnns with pooling layers. arXiv preprint arXiv:1805.03096. (2018).

Google Scholar

26. Nasr-Esfahani E, Rafiei S, Jafari MH, Karimi N, Wrobel JS, Samavi S, et al. Dense pooling layers in fully convolutional network for skin lesion segmentation. Computerized Med Imaging Graphics. (2019) 78:101658.

Google Scholar

27. Basha SS, Dubey SR, Pulabaigari V, Mukherjee S. Impact of fully connected layers on performance of convolutional neural networks for image classification. Neurocomputing. (2020) 378:112–9.

Google Scholar

28. Jaiswal A, Gianchandani N, Singh D, Kumar V, Kaur M. Classification of the covid-19 infected patients using densenet201 based deep transfer learning. J Biomolecular Structure Dynamics. (2021) 39:5682–9.

Google Scholar

29. Deng X, Shao H, Shi L, Wang X, Xie T. A classification-detection approach of covid-19 based on chest x-ray and ct by using keras pre-trained deep learning models. CMES-Computer Modeling Eng Sci. (2020) 125. doi: 10.32604/cmes.2020.011920

CrossRef Full Text | Google Scholar

30. Alzubaidi L, Zhang J, Humaidi AJ, Al-Dujaili A, Duan Y, Al-Shamma O, et al. Review of deep learning: Concepts, cnn architectures, challenges, applications, future directions. J big Data. (2021) 8:1–74. doi: 10.1186/s40537-021-00444-8

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Howard AG, Zhu M, Chen B, Kalenichenko D, Wang W, Weyand T, et al. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861. (2017).

Google Scholar

32. Liu Z, Lin Y, Cao Y, Hu H, Wei Y, Zhang Z, et al. (2021). Swin transformer: Hierarchical vision transformer using shifted windows, in: Proceedings of the IEEE/CVF international conference on computer vision, . pp. 10012–22.

Google Scholar

33. Liu J, Liu S, Xu S, Zhou C. Two-stage underwater object detection network using swin transformer. IEEE Access. (2022) 10:117235–47.

Google Scholar

34. Zhang B, Gu S, Zhang B, Bao J, Chen D, Wen F, et al. (2022). Styleswin: Transformer-based gan for high-resolution image generation, in: Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, . pp. 11304–14.

Google Scholar

35. Wang S, Gao Z, Liu D. Swin-gan: generative adversarial network based on shifted windows transformer architecture for image generation. Visual Comput. (2023) 39:6085–95. doi: 10.1007/s00371-022-02714-9

CrossRef Full Text | Google Scholar

36. Wang C, Xu H, Zhang X, Wang L, Zheng Z, Liu H. (2022). Convolutional embedding makes hierarchical vision transformer stronger, in: European Conference on Computer Vision, . pp. 739–56. Springer.

Google Scholar

37. Ma Z, Wu X, Chu A, Huang L, Wei Z. Swinfg: A fine-grained recognition scheme based on swin transformer. Expert Syst Appl. (2023) 123021.

Google Scholar

38. Xiao X, Guo W, Chen R, Hui Y, Wang J, Zhao H. A swin transformer-based encoding booster integrated in u-shaped network for building extraction. Remote Sens. (2022) 14:2611.

Google Scholar

39. Zhou X, Zhou L, Gong S, Zhong S, Yan W, Huang Y. Swin transformer embedding dual-stream for semantic segmentation of remote sensing imagery. IEEE J Selected Topics Appl Earth Observations Remote Sens. (2023). doi: 10.1109/JSTARS.2023.3326967

CrossRef Full Text | Google Scholar

40. Taslimi S, Taslimi S, Fathi N, Salehi M, Rohban MH. Swinchex: Multi-label classification on chest x-ray images with transformers. arXiv preprint arXiv:2206.04246. (2022).

Google Scholar

41. Wang R. An improved nondominated sorting genetic algorithm for multiobjective problem. Math Problems Eng. (2016) 2016. doi: 10.1155/2016/1519542

CrossRef Full Text | Google Scholar

42. Reeves CR. Genetic algorithms. Handb metaheuristics. (2010), 109–39. doi: 10.1007/978-1-4419-1665-5_5

CrossRef Full Text | Google Scholar

43. El-Rashidy N, Abuhmed T, Alarabi L, El-Bakry HM, Abdelrazek S, Ali F, et al. Sepsis prediction in intensive care unit based on genetic feature optimization and stacked deep ensemble learning. Neural Computing Appl. (2022), 1–30. doi: 10.1007/s00521-021-06631-1

CrossRef Full Text | Google Scholar

44. Sivanandam S, Deepa S, Sivanandam S, Deepa S. Genetic algorithm optimization problems. Introduction to Genet algorithms. (2008) 165–209. doi: 10.1007/978-3-540-73190-0_7

CrossRef Full Text | Google Scholar

45. Calzolari M. Datasetmanuel-calzolari/sklearn-genetic: sklearn-genetic 0.2. (2019).

Google Scholar

46. Derrac J, Garcia S, Molina D, Herrera F. A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evolutionary Comput. (2011) 1:3–18.

Google Scholar

47. Brown I, Mues C. An experimental comparison of classification algorithms for imbalanced credit scoring data sets. Expert Syst Appl. (2012) 39:3446–53.

Google Scholar

48. Win KP, Kitjaidure Y, Hamamoto K, Myo Aung T. Computer-assisted screening for cervical cancer using digital image processing of pap smear images. Appl Sci. (2020) 10:1800.

Google Scholar

49. Deo BS, Pal M, Panigarhi PK, Pradhan A. Cerviformer: A pap-smear based cervical cancer classification method using cross attention and latent transformer. arXiv preprint arXiv:2303.10222. (2023). doi: 10.1002/ima.23043

CrossRef Full Text | Google Scholar

50. Pacal I, Kılıcarslan S. Deep learning-based approaches for robust classification of cervical cancer. Neural Computing Appl. (2023) 35:18813–28.

Google Scholar

51. Ravindran K, Rajkumar S, Muthuvel K. An investigation on cervical cancer with image processing and hybrid classification. Int J Performability Eng. (2021) 17:918.

Google Scholar

52. Maurya R, Pandey NN, Dutt MK. Visioncervix: Papanicolaou cervical smears classification using novel cnn-vision ensemble approach. Biomed Signal Process Control. (2023) 79:104156.

Google Scholar

53. Li M, Que N, Zhang J, Du P, Dai Y. Vtcnet: A feature fusion dl model based on cnn and vit for the classification of cervical cells. Available at SSRN 4748846. (2024). doi: 10.2139/ssrn.4748846

CrossRef Full Text | Google Scholar

Keywords: image processing, image classification, image cancer classification, Swin Transformer, CNN models, genetic algorithm, random forest

Citation: Alohali MA, El-Rashidy N, Alaklabi S, Elmannai H, Alharbi S and Saleh H (2024) Swin-GA-RF: genetic algorithm-based Swin Transformer and random forest for enhancing cervical cancer classification. Front. Oncol. 14:1392301. doi: 10.3389/fonc.2024.1392301

Received: 27 February 2024; Accepted: 19 June 2024;
Published: 19 July 2024.

Edited by:

Miodrag Zivkovic, Singidunum University, Serbia

Reviewed by:

Zafer Cömert, Samsun University, Türkiye
Bosko Nikolic, University of Belgrade, Serbia

Copyright © 2024 Alohali, El-Rashidy, Alaklabi, Elmannai, Alharbi and Saleh. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hager Saleh, hager.saleh.fci@gmail.com; Saad Alaklabi, salaklabi@su.edu.sa

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.