- 1Department of Computer Science and Engineering, C. V. Raman Global University, Bhubaneswar, Odisha, India
- 2Department of Computer Science and Engineering, Siksha ‘O’ Anusandhan (Deemed to be University), Bhubaneswar, Odisha, India
- 3Department of Information Technology, Vardhaman College of Engineering (Autonomous), Hyderabad, Telangana, India
- 4Key Laboratory of Biomedical Engineering of Hainan Province, School of Biomedical Engineering, Hainan University, Sanya, China
- 5Centre for Intelligent Healthcare, Coventry University, Coventry, United Kingdom
- 6Department of Mechatronics Engineering, Parul Institute of Technology, Parul University, Vadodara, Gujarat, India
Background: Cancer rates are rising rapidly, causing global mortality. According to the World Health Organization (WHO), 9.9 million people died from cancer in 2020. Machine learning (ML) helps identify cancer early, reducing deaths. An ML-based cancer diagnostic model can use the patient’s genetic information, such as microarray data. Microarray data are high dimensional, which can degrade the performance of the ML-based models. For this, feature selection becomes essential.
Methods: Swarm Optimization Algorithm (SSA), Improved Maximum Relevance and Minimum Redundancy (IMRMR), and Boruta form the basis of this work’s ML-based model BIMSSA. The BIMSSA model implements a pipelined feature selection method to effectively handle high-dimensional microarray data. Initially, Boruta and IMRMR were applied to extract relevant gene expression aspects. Then, SSA was implemented to optimize feature size. To optimize feature space, five separate machine learning classifiers, Support Vector Machine (SVM), Random Forest (RF), Extreme Learning Machine (ELM), AdaBoost, and XGBoost, were applied as the base learners. Then, majority voting was used to build an ensemble of the top three algorithms. The ensemble ML-based model BIMSSA was evaluated using microarray data from four different cancer types: Adult acute lymphoblastic leukemia and Acute myelogenous leukemia (ALL-AML), Lymphoma, Mixed-lineage leukemia (MLL), and Small round blue cell tumors (SRBCT).
Results: In terms of accuracy, the proposed BIMSSA (Boruta + IMRMR + SSA) achieved 96.7% for ALL-AML, 96.2% for Lymphoma, 95.1% for MLL, and 97.1% for the SRBCT cancer datasets, according to the empirical evaluations.
Conclusion: The results show that the proposed approach can accurately predict different forms of cancer, which is useful for both physicians and researchers.
1 Introduction
Cancer has been one of the main causes of mortality across the globe for many decades, making it a significant public health issue on a global scale. Cancer is an umbrella term for illnesses caused by the unchecked proliferation and metastasis of aberrant cells. It significantly affects public health and affects individuals of all ages and walks of life. Cancer mortality rates change both annually and geographically. WHO and the International Agency for Research on Cancer (IARC) frequently release statistics on cancer fatalities. It was predicted that there were around 9.9 million deaths from cancer globally in 2020 (Panigrahi et al., 2023). Deaths from cancer are not distributed in the same way all across the world. Developed nations often have higher incidence and survival rates because of better healthcare, earlier diagnosis, and more effective treatments. However, cancer may be a significant problem in locations with inadequate healthcare infrastructure (Venkatesan et al., 2022). Screening for cancer at an early stage is another way to boost treatment success. Many forms of cancer now have better survival rates because of advances in cancer research and therapy (Khalsan et al., 2022).
Infants with ALL rearranged to MLL have even worse prognoses, with survival rates of 40%–50%; the overall death rate for ALL is 20%–25% in children and up to 50% in adults. The death rate for lymphomas varies greatly depending on subtype. For example, aggressive non-Hodgkin lymphomas, such as diffuse large B-cell lymphoma, have a mortality rate of 30%–40%, whereas Hodgkin lymphomas have a rate below 10%. In children, the survival rate for leukemias with MLL rearrangements is around 50%, but in adults, it is lower, especially in cases of acute myeloid leukemia (AML) (Lewis et al., 2020). The 5-year survival rate for SRBCTs drops to 15%–30% in metastatic instances from over 70% in localized illness (de Leval and Jaffe, 2020).
Through enhancing early detection and diagnosis, treatment, and patient outcomes, machine learning has the potential to contribute significantly to reducing cancer mortality rates (Bolón-Canedo et al., 2014). Medical imaging like mammograms and CT scans may be analyzed using machine learning algorithms for the early detection of cancer. Earlier diagnosis is associated with better treatment results (Ma et al., 2020). It may ensure that patients get the best possible treatment promptly by lowering the percentage of erroneous diagnostic results. Machine learning may predict treatment outcomes by analyzing a patient’s medical record, genetic data, and tumor features (Alghunaim and Al-Baity, 2019). By aiding in medication research and development, machine learning may help produce more effective and tailored medicines for cancer.
Several forms of medical data have been analyzed and interpreted using machine learning (ML) for cancer detection. Medical imaging data and genomic data account for the vast majority of ML-based cancer research. Mammograms, X-rays, CT scans, and MRI scans are all examples of medical imaging data (Almugren and Alshamlan, 2019). DNA sequencing data and gene expression profiling or microarray data are further examples of “genomic data.” Microarray data, or genetic information, is crucial for cancer diagnosis because it reveals important information about the disease itself and its genetic disorders. This knowledge facilitates personalized therapy, boosts treatment efficacy, and facilitates educated decision-making in cancer care and prevention (López-García et al., 2020). When processing microarray data, ML encounters several challenges. The most difficult aspects of the microarray data for ML to handle are the high dimensionality, small sample size, and class imbalance. The high dimensionality of microarray data results from the fact that thousands of characteristics (genes or probes) are often assessed for each sample or patient. Owing to the “curse of dimensionality,” conventional ML methods may become ineffective in required resources (both computational and cognitive) (Shukla et al., 2020a). There are often fewer samples available than there are characteristics to analyze. Creating reliable and generalizable models might be difficult when working with a small sample size.
The best possible solution is to overcome the above-stated challenges, employing strategies such as dimensionality reduction, feature selection, and optimization algorithms to effectively select genes for cancer diagnosis (Shukla et al., 2019). Deploying a single feature selection algorithm may reduce the number of features to some extent. However, if the feature space is larger, employing a single method to select the appropriate number of features may not be sufficient. Hence, the current work aims to deal with the high dimensionality issues of microarray data with the help of the pipelined featured selection algorithms followed by a nature-inspired optimization algorithm to reduce the feature space to the extent upon which the machine learning models can be used to develop a more effective ML model. The reported research includes the Boruta and IMRMR feature selection algorithm in a pipelined manner. The motivation behind using the Boruta Feature selection algorithm is its ability to identify all relevant genes while considering the complex feature interaction with the target feature using Random Forest (RF). In addition, it also deals with the challenges, including the overfitting and interaction of the features with the target variable. Improved Maximum Relevance and Minimum Redundancy (IMRMR) IMRMR improves traditional mRMR by selecting feature subsets highly related to the target class and mainly uncorrelated. This makes features more informative and efficient. This balance is a significant benefit of the approach in complicated datasets where inter-feature relationships may not be evident and impair each feature’s prediction ability. The Salp Swarm Optimization Algorithm (SSA) effectively lowers the dimensionality of datasets and enhances model performance by identifying the most influential features that contribute to the model’s predictive capability. Its innate simplicity and resilience make it highly suitable for managing intricate, high-dimensional data where conventional feature selection approaches may encounter difficulties. It utilizes the swarming characteristics of the salp in an ocean, which balances the exploration and exploitation with efficacy, which helps the SSA to converge on a globally optimal solution while avoiding the local optima. In the current work, BIMSSA, the Boruta is initially applied to eliminate the irrelevant features from the dataset, thus reducing the feature space. It creates the shadow features by shuffling the original features randomly. Using the base classifier RF, the importance score is calculated for original and shadowed features. The highest score of the shadowed feature set is termed the threshold value. The original features with a score greater than the threshold value are considered relevant. Then, the IMRMR is used to select the most relevant features from the feature selection by calculating the IMRMR score. Finally, to the selected features, the SSA is applied to select an optimal number through multiple iterations of features by having a stopping criterion, such as maximum iteration for current work. The algorithms mentioned above are dedicated to selecting the features from the entire feature space. Thus, these algorithms work in a vertical approach.
1.1 Objective
This study aims to create an ensemble ML-based model using the Boruta IMRMR technique to select features and the SSA algorithm to optimize those features. The research’s key contribution is summed up as follows.
• To develop an ensemble ML model for efficient Cancer diagnosis.
• To compare the performance of various feature selection approaches, specifically Boruta and IMRMR, in reducing the dimensionality of the microarray data.
• To access the impact of SSA in optimizing the selected features in the context of the Cancer gene expression data.
• To analyze the performance of hybrid models with feature selection methods along with four conventional ML classifiers.
• To analyze the efficacy of the ensemble learning model over the hybrid model.
• Finally, four different cancer microarray data are considered to evaluate the proposed model.
1.2 Literature survey
For the current work, 130 reports are initially identified. Several records are excluded from the study with different steps. Twenty duplicate records are removed before the screening process. The remaining 110 numbers of records are considered for the screening phase. In this phase, ten irrelevant records are excluded. From the screening phase, 100 records are processed for the retrieval phase, out of which 7 records are excluded as those records could not be retrieved. Hence, the remaining 93 full-text records are considered. From the considered full texts, 17 records are excluded as sufficient data for analysis are unavailable. In addition, 20 records are excluded as those dealing with diseases other than cancer, and 7 records are found to be irrelevant to the study. Finally, the remaining 49 numbers of articles are considered for the current work. From the considered records, 20 articles are used in the literature survey part, and the remaining 29 numbers of records are considered in the rest of the manuscript. The Figure in Appendix I shows the current work’s Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA).
Ibrahim et al. (2017) introduced a novel hybrid approach, the Salp Swarm Algorithm in feature selection (SSA-FS), on the real datasets obtained from Iraqi hospitals for breast, bladder, and colon cancers. Hegazy et al. (2018a) introduced a novel method, the chaotic salp swarm algorithm (CSSA), to enhance the SSA on 28 datasets. Hegazy et al. (2018b) introduced a novel method, the improved salp swarm algorithm (ISSA), which is consolidated with the KNN classifier for feature selection on 23 UCI-ML datasets and claimed to achieve enhanced accuracies on Breast Cancer, Lung Cancer, and BeastEW disease datasets respectively. Using a neighborhood entropy-based uncertainty measures model, Sun et al. (Sun et al., 2019) successfully applied machine learning (ML) methods, including k-nearest neighbor (KNN), C4.5, and Support Vector Machine (SVM), to the classification of the colon, diffuse large B-cell lymphoma (DLBCL), leukemia, lung, and small round blue cell tumor (SRBCT). Ghoniem (2020) introduced a novel bio-inspired liver cancer diagnosis model considering a deep learning (DL) approach, i.e., Convolutional Neural Network (CNN) along with SegNet and UNet, and the optimization technique, i.e., Artificial Bee Colony optimization (ABC) on Radiopaedia and LiTS datasets. Shukla et al. (2020b) introduced an adaptive inertia weight teaching-learning model considering machine learning approaches, i.e., Support Vector Machine (SVM), Extreme Learning Machine (ELM), and Naïve Bayes (NB) on Breast Cancer, Colon Cancer, DLBCL, Leukaemia, SRBCT, Lung Cancer.
Meenachi and Ramakrishnan (2020) introduced differential evolution and global optimal feature selection for cancer data classification model considering the machine learning (ML) approach, i.e., Decision Tree (DT) and optimization technique, i.e., Ant colony optimization (ACO) on five datasets, i.e., DLBCL, Breast Cancer, Leukemia, SRBCT, Gisette datasets. Nouri-Moghaddam et al. (2021) introduced a new hybrid solution based on a multi-filter and adaptive chaotic multi-objective forest optimization algorithm (AC-MOFOA) considering Forest optimization algorithm (FOA), Extreme learning machine (ELM), multi-objective optimization (MOO), and five filter methods, i.e., IG, mRMR, RelifF, CFS, and Fisher-score, on nine datasets, i.e., SRBCT, Tumors_9, Leukaemia3, colon_prostate, Lung, GCM, Breast, Rsctc_5, Rsctc_6.
Yan et al. (2021) introduced a novel feature selection model considering a machine learning (ML) approach, i.e., K-nearest neighbor (KNN) and a wrapper feature selection algorithm FS_SSA based on Salp swarm, on five datasets, i.e., ALL-AML-4, Colon Cancer, Lymphoma, MLL, SRBCT datasets. Sarala et al. (Arun Prabha et al., 2021) introduced a decision-based Salp Swarm Optimization (DT-SWO) algorithm considering machine learning (ML) approaches, i.e., Decision Tree (DT), Support Vector Machine (SVM), Naïve Bayes (NB), Kernel Support Vector Machine (KSVM), and optimization technique, i.e., Salp Swarm Optimization (SWO) on four datasets, i.e., DLBCL, Leukemia, Lung Cancer and colon datasets. Alomari et al. (2021) introduced a hybrid filter-wrapper approach considering robust Minimum Redundancy Maximum Relevancy (rMRMR) as a filter approach, Modified Gray Wolf Optimization (MGWO) as a wrapper approach, and ML approaches including Random Forest (RF), Elastic Networks (EN) and Decision Tree (DT) on nine datasets. Balakrishnan et al. (2021) introduced an improved salp swarm algorithm (iSSA) based on the levy flight for feature selection model and Support Vector Machine (SVM) classifier on six datasets, i.e., Oral Squamous Cell Carcinoma (OSCC), Ovarian cancer, Breast Cancer, CNS, Colon Cancer, Leukemia datasets. Hameed et al. (2021) included the binary particle swarm optimization (BPSO), the genetic algorithm (GA), and the cuckoo search algorithm (CS) for selecting the features. ML approaches, including SVM, NB, KNN, and RF, are applied to twelve datasets.
Gene selection for microarray data categorization was developed by Rostami et al. (2022) using a multi-objective graph theoretic-based approach model that considers the idea of community detection with node centrality (CDNC). On six datasets, Aziz (2022) presented a metaheuristics model that was inspired by nature. The model utilized ML approaches such as Support Vector Machine (SVM), Naïve Bayes (NB), Artificial Neural Network (ANN), cuckoo search (CS), genetic algorithm (GA), and artificial bee colony (ABC). Alromema et al. (2023) used logistic regression, Support Vector Machine, K-Nearest Neighbours, Neural Networks, Naive Bayes, Decision Tree, and eXtreme Gradient Boosting to gene expression datasets. On 17 microarray expression datasets, including CNS, Colon, Leukemia_3C, Leukemia_4C, Leukaemia, Hungtington Disease, DLBCL, Lymphoma66 ö4026_3c, Lymphoma, Prostate, SRBCT, Lung Cancer, Breast Cancer, Sarcoma, Mycloma, and Ovarian, Ke et al. (2022) suggested a population initialization method based on ranking criteria (PIRC) using NB, C4.5, genetic algorithm (GA), and ant colony optimization (ACO). In their study on the Wisconsin Breast Cancer Dataset (WBCD), Rustagi et al. (2024) presented a method for breast cancer detection that relies on Salp Swarm and Grey Wolf Optimisation. They took SVM and KNN classifiers into account. Ünalan et al. (2024) highlighted ensemble learning methodologies’ efficacy in classifying breast cancer. It was revealed that performance improved over standalone classifiers. Classifiers AdaBoost, GBM, and RGF gave an impressive accuracy of 99.5%. However, ensembles of this kind surpassed the respective individual algorithms LGBM for accuracy and gave an F1 score of 99.2% alongside an accuracy of 98.9%. Incorporating stratified shuffle split and k-fold cross-validation raises the question of the strict evaluation technique in obtaining credible and clinically relevant classification outputs. Table 1 shows the analytical study of the literature mentioned above. Batool and Byun (2024) showed that ensemble learning applied to the task of breast cancer classification was doing well by enhancing the general predictive accuracy through a set of multiple models. Several studies on the WBCD dataset showed that the ensemble method, for example, voting classifier involving ETC, LightGBM, RC, and LDA models, performed better than individual models. The proposed model succeeded in surpassing all known state-of-the-art classifiers utilized in the detection and diagnosis cases of breast cancer with an average accuracy of 97.6% and an F1 value of 98.1%. Mahesh et al. (2022) proposes an early prediction of breast cancer through a blended ensemble learning approach with SVM, KNN, DT, RF, and LR as base classifiers. The model’s performance is checked using a breast cancer dataset, which has yielded considerable improvements in accuracy at 98.14%. Accurate, recall, precision, and F1-score metrics validate the ensemble model’s effectiveness over individual classifiers.
Table 1 shows the analytical study of the considered literature in which it has been observed that the reported literature follows a two-stage feature selection process. In addition, the adopted approaches follow a hybrid approach for cancer classification. Applying the two-stage feature selection does not impact the undertaken dataset more, i.e., ALL-AML, Lymphoma, MLL, and SRBCT cancer dataset. Thus, the current research aims to apply a pipelined feature selection method consisting of three different feature selection approaches, starting from Boruta and followed by IMRMR and SSA. However, to make it more effective, Table 2 has been taken from Table 1, which is dedicated to the literature that deals with the dataset considered for the current study, such as ALL-AML, Lymphoma, MLL, and SRBCT.
1.3 Paper structure
The remaining parts of the paper are organized as follows. Section 2 depicts the approach used in developing the suggested model and describes the utilized dataset. The core ideas behind the proposed paradigm are discussed in Section 3. The empirical evaluation of the suggested model is presented in Section 4. In Section 5, we do a critical analysis of the proposed model. The conclusion is presented in Section 6.
2 Methodology of BIMSSA
This section details the methods used to create the reported model and the dataset it was built from. The BIMSSA model utilizes a pipelined feature selection technique to efficiently address the dimensionality problem in the microarray data. Initially, Boruta and IMRMR are used with the objective of extracting pertinent gene expression features. In the next step, the SSA method is used to optimize the size of the feature set acquired by Boruta and IMRMR. The Support Vector Machine (SVM), Extreme Learning Machine (ELM), Random Forest (RF), AdaBoost, and XGBoost are the foundation learners that are used in the construction of an ensemble model.
2.1 Dataset description
Four publicly available cancer gene expression data, including ALL-AML (D1), Lymphoma (D2), MLL (D3), and SRBCT (D4), are considered for developing the current work (Zhu et al., 2007). Among the above datasets, D1, D2, and D3 have three classes, and D4 has four classes. ALL-AML dataset contains three classes labeled B-Cell, T-Cell, and AML. The Lymphoma cancer dataset includes DLBCL, FL, and CLL classes. Considering the MLL cancer dataset, it contains three classes, including ALL, AML, and MLL. Similarly, the class names for the SRBCT cancer dataset are BL, EWS, NB, and RMS. Table 3 shows the proposed model’s dataset description and the class-wise distribution in each dataset.
2.2 Boruta feature selection
Boruta is a wrapper-based feature selection that aims to improve the efficiency and interpretability of machine learning models. Boruta finds and chooses important features from a dataset by combining a random forest classifier with a significance testing method. Shadow features, essentially randomized replicas of the original features, are first generated by the algorithm. After that, the original and shadow feature sets are pooled and used to train a random forest model. Based on how each feature affects the model’s performance, Boruta ranks them and compares them to their corresponding shadow characteristics to determine their relative relevance. Initially, dataset (D) contains the total number of features.
Where Shuffle is a function that shuffles the values of the feature
To the integrated feature set,
After obtaining the I for each feature, the threshold value (Ith) is calculated as the maximum score among the shadow features. This can be represented by Equation 4. Equation 5 shows the criteria based on which the Boruta feature selection algorithms identify the important features present in the dataset, thus discarding the others to modify the feature set (F)
Once Boruta determines if a feature is relevant, the algorithm iteratively confirms or rejects it until all features have been categorized. A refined collection of characteristics expected to improve model generalization and prediction accuracy is the final output, a subset of the relevant features. Boruta provides a strong method for feature selection in the ML process, making it ideal for dealing with high-dimensional datasets (Maurya et al., 2023).
2.3 Improved maximum relevance and minimum redundancy (IMRMR)
Maximum relevance and minimum redundancy (MRMR) is a filter-based feature selection technique used to select a subset of features from a more extensive set of features. The objective is to select the most useful features while minimizing redundancy. This method is often used in machine learning and data analysis for better model performance, reduced overfitting, and more interpretable models. The IMRMR is the modified version of the MRMR feature section algorithm. Like MRMR, it also selects relevant features with low redundancy scores. MRMR adopts the Mutual Information (MI) to select those features that depend on the target feature. In addition to MI, IMRMR aids Pearson correlation to focus on the linear relation between the features. The following shows the workings of MRMR and IMRMR to determine the difference between these two for selecting features (Ding and Peng, 2005; Yan and Jia, 2019; Zhao et al., 2019).
2.3.1 Relevance calculation
• Calculating each feature’s relevance score requires assessing its correlation with the target variable.
• It is generally accepted that characteristics with high relevance scores should be weighed more for classification.
• The Relevance score (
F is the total feature set, |F| is the total number of features,
2.3.2 Redundancy calculation
• Perform a pairwise redundancy analysis between the characteristics. The word “redundancy” refers to the degree to which two characteristics are comparable regarding the information they contain.
• Metrics such as mutual information, correlation, and distance-based metrics are common redundancy measurements.
• It is best to avoid using features in the final decision that are extremely repetitive with one another since they give comparable information. The redundancy score (
The MRMR score of the considered feature
By determining the significance and duplication of each feature, the mRMR technique quantifies the contribution of features. It does not take into account the combined effect of several characteristics. Mutual information measures are the only basis for relevance and redundancy. To choose the best feature subset, IMRMR employs two metrics—the Pearson correlation coefficient and mutual information—to assess the subsets’ relevance and redundancy with weight factors ranging [0.1,1] with step size 0.1.
Calculate the relevance (
The modified IMRMR(i) can be calculated using Equation 14.
2.4 Salp swarm optimization algorithm (SSA)
After applying the IMRMR feature selection, the relevant features remain in the dataset. However, applying the IMRMR does not ensure that the number of selected features has a role in the diagnosis process. Some features will be relevant, but removing them does not impact the diagnosis model. The Salp Swarm Optimization Algorithm (SSA), a wrapper-based feature selection algorithm, is implemented to ensure that the optimal set of relevant features is in the selected feature set, through which the processing time of the developed model can be decreased.
The SSA has been implemented to solve optimization challenges. It mimics the group dynamics of salps, sea animals similar to jellyfish. When applied to an optimization issue, SSA seeks to identify the best possible outcome (Mirjalili et al., 2017; Castelli et al., 2022). There are more than 1.2 million known marine creature species. Most of these species have similar habits and traits, including ways of communication, speed of movement, and foraging strategies. The salp’s habitats are notoriously tough to reach, yet scientists think this behavior aids the animals in movement and feeding. Inspired by the coordinated movement of salps (gelatinous sea organisms), the SSA uses natural selection to find optimal solutions. Collective behavior in these species serves as a paradigm for the optimization difficulties SSA seeks to solve (Ibrahim et al., 2018; Thawkar, 2021; Sayed et al., 2018).
2.4.1 Mathematical model
A member of the family Salpidae, salps are found in the ocean. Its cylindrical form and end apertures evoke images of jellyfish, which pump water through their gelatinous bodies using internal feeding filters to propel themselves and eat. Some of the aquatic creatures exhibit similar behaviors, such as swarming. In the case of fish, this group is known as a school, whereas in the case of salps, it is referred to as a salp chain.
The SSA starts with the swarm X of n numbers of salps. Equation 15 shows the two-dimensional matrix of the scalp position.
The working of the SSA is started by determining the fitness function (F ()) population size, maximum iteration, initial step size (
Where
The fitness of the salp is updated by comparing the fitness of the new position with the previous position. This can be determined by using Equation 20. Personal best is based on the fitness function obtained. The position of salp can be obtained by using Equation 21.
The exploitation components of salp aim to improve the current obtained solution to an optimized one. This process ensures that the obtained solution is the best. Exploitation is the process of leader-follower dynamics where the entire salp group is divided into two groups. The first salp is the leader salp, and the others are called the follower salp. The position of the leader salp and follower salp is updated using Equations 22 and 23, respectively. Let x, y be the leader and follower salp belonging to the salp community X.
Where lb and ub are the lower and upper bounds of the exploring dimension, x(t) is the current position of the leader salp at current time t, and c1 and c2 are the random numbers between the interval [0,1].
2.4.2 Working of SSA
The selected features from MRMR are taken as input for the SSA algorithm. The workings of the SSA are described below. Figure 1 shows how the SSA algorithm works.
Step 1: The parameter initialization is the primary of the algorithm. Various parameters have been initiated, including maximum iteration (Max_It), number of salps (num_salps), etc. The parameter initialization is shown in Table 4.
Step 2: To get the best solution, the proper fitness function needs to be calculated. For the current work, average accuracy (Equation 24) determines the fitness function of SSA. For determining the across-validation, the fitness function
Step 3: The best solution is identified based on the objective function.
Step 4: Update the salp position as per Equation 15.
Step 5: Local and global exploration is performed to find the local and global solution. This algorithm aims to strike a balance between global exploration, which searches the whole solution space, and local exploration, which exploits the neighborhood of specific solutions.
Step 6: Update the best solution.
Step 7: Repeat Step-3 to Step-6 for each feature until the number of iterations is less than Max_It.
2.5 Voting ensemble classifier
The current work employs ML classifiers such as SVM (Alfian et al., 2022), RF (Alfian et al., 2022), ELM (Ding et al., 2013), AdaBoost (Asselman et al., 2021), and XGBoost (Asselman et al., 2021) to make the initial prediction. Then, the majority voting classifier is used as an ensemble learning approach to make the final prediction. It is a simple and powerful strategy for integrating the results of numerous models into a single prediction. When many models provide varying results and a group choice must be made based on those results, majority voting is an effective tool (Naji et al., 2021; Pati et al., 2023).
3 Workflow of BIMSSA
Initially, the dataset is considered for normalization to remove the noisy data. The MRMR is then applied to the normalized data to choose the best feature. As a last step, the SSA is used as an optimizer. Algorithm 1 shows the working description of BIMSSA. The functionality of the suggested model is shown in Figure 2. Initially, the dataset is considered for normalization to remove the noisy data. The MRMR is then applied to the normalized data to choose the best feature. As a last step, the SSA is used as an optimizer. Figure 2 and Algorithm 1 show the suggested model’s functionality. The suggested method’s operation is described below.
3.1 Step-1: dataset considered for preprocessing
The dataset is first subjected to a preprocessing stage to ensure data quality and consistency. This involves several key steps.
• Data Cleaning: Remove or correct any noisy data, missing values, or inconsistencies within the dataset.
• Normalization: Apply normalization techniques to scale the features to a standard range, typically [0, 1] or [-1, 1]. This helps in improving the performance and convergence of machine learning algorithms.
• Encoding: If applicable, convert categorical variables into numerical values using label encoding.
• Balancing the Dataset: If the dataset is imbalanced, techniques like oversampling, undersampling, or synthetic data generation (e.g., SMOTE) are applied to ensure that the model does not become biased towards the majority class.
3.2 Step 2: dataset splitting
The preprocessed dataset is then split into training and testing sets. Two different splitting ratios are considered: 80–20. This helps in evaluating the robustness of the model.
3.3 Step 3: Feature selection
To enhance model performance, feature selection algorithms are applied.
• Boruta Algorithm: An all-relevant feature selection method that identifies relevant features by comparing original attributes with shadow attributes.
• Improved Minimum Redundancy Maximum Relevance (MRMR): Select the best features with maximum relevance with the target variable and minimal redundancy among themselves.
3.4 Step 4: optimization using salp swarm algorithm (SSA)
The SSA is used to optimize the feature set selected in the previous step.
i. Initiate Population
ii. Parameter Initialization
iii. Fitness Function Calculation
iv. Optimization Process Starts: Iteratively update the positions of salps to find the optimal feature set.
v. Check the Maximum Iteration
vi. If Not Exceeding, Return to iii
vii. If Exceeds, Obtain the Best Feature Set
3.5 Step 5: model training
Based on the training data and the optimized feature set, train the following five models.
• Support Vector Machine (SVM)
• Random Forest (RF)
• Extreme Learning Machine (ELM)
• AdaBoost
• XGBoost
3.6 Step 6: Classifier Selection
Select the top three classifiers from the trained models based on their highest accuracy during training.
3.7 Step 7: Model Evaluation
Test the ensemble classifier using the testing dataset to obtain evaluative parameters such as accuracy, precision, recall, F1-score, and ROC-AUC.
Algorithm 1.Working of Proposed BIMSSA model.
Input: Dataset D← {D1, D2, D3, D4}, Feature set F← {f1, f2, ……fn}, Max Iteration, num_salps, Lb, Ub, Γ1, Γ2, BL: {SVM (BL1), RF (BL2), ELM (BL3), AdaBoost (BL4), XGBoost (BL5),
Output: Performance Measures
For k = 1 → 4
Normalize Dk
End for
Split the dataset with an aspect ratio of 85:15 for Train: Test
Invoke Boruta () to D and return the resulting feature subset is
Invoke IMRMR () to
for k = 1 → 4
for i = 1 → n
Find
for j = i+1 → n
Find
End for
Find
End for
End For
Invoke SSA () to
for k = 1 → 4
for i = 1 → n
Find Fitness function F (
Calculate
Calculate
Calculate
Find
Find
Update position salp
End for
Update
End for
Apply Majority Voting ensemble technique with the three best BL
Measure the performance of the proposed model
for k = 1 → 4
for i = 1 → 5
Apply BLi to
Calculate the performance measures
Find best three BLi
End for
End for
4 Empirical analysis
The suggested model uses Python 3.11 on Ubuntu 20.04 with 32 GB of RAM, an Intel Core i7 CPU, and a 1 TB SSD. The Boruta and IMRMR feature selection techniques extract the important and relevant features from the dataset. After identifying the relevant features, the SSA optimization technique selects a set of relevant features. The implemented Boruta feature selection algorithm selects 2067, 1,043, 3,767, and 643 features for ALL-AML, Lymphoma, MLL, and SRBCT datasets, respectively. While applying the Boruta + IMRMR hybrid feature selection model, it selects 413, 197, 656, and 118 numbers of features for ALL-AML, Lymphoma, MLL, and SRBCT datasets, respectively. Finally, SSA is applied to the feature subset selected by IMRMR, resulting in 57, 29, 194, and 23 numbers of features selected by the hybrid model Boruta + IMRMR + SSA for the final features classification purpose. The empirical analysis has been done in 3 different approaches. Approach 1 shows the performance of all considered classifiers with the IMRMR feature selection algorithm. Approach 2 shows the performance of all considered classifiers with IMRMR and SSA optimizer. Approach 3 shows the performance of the proposed ensemble classifier with the three best classifiers from Approach 2 based on accuracy. The performance in all of the above-mentioned approaches is based on ten different parameters, including Accuracy (ACC), Precision (PRE), Fβ-score (F1-Score (F1-S) and F2 Score (F 2)), Specificity (SPE), Misclassification Rate (MCR), False Negative Rate (FNR), False Positive Rate (FPR), and MCC. Additionally, we employed confidence interval (CI) statistical analysis of the obtained results (Asselman et al., 2021; Naji et al., 2021). The relative importance of precision and recall can be adjusted using the beta parameter of the F-beta score. Precision and recall are of equal value when β = 1 (F1 score). When β is higher than 1 (F 2 score), the model is more sensitive to positive cases since recall is given more weight. As beta decreases (to values of F0.5), the model becomes more concerned with delivering accurate positive predictions. In the proposed model, two values of β have been considered, such as F-1 and F2. The above-said parameters can be calculated by using Equations 26–35. The training set contains 61 features, and the test set contains 11 samples for the ALL-AML dataset. The training and testing set contains 52 and 10 samples for the Lymphoma dataset. The training set contains 61 features, and the test set contains 11 samples for the MLL dataset. Considering the SRBCT dataset, the train set includes 70 and 13 samples in the test set. Equations 26–35 show the calculation of all evaluative parameters.
4.1 Analysis of hybrid models with Boruta, IMRMR and SSA
The performance analysis of the mentioned hybrid model, such as Boruta + IMRMR + SSA + SVM, Boruta + IMRMR + SSA + RF, Boruta + IMRMR + SSA + ELM, Boruta + IMRMR + SSA + AdaBoost, Boruta + IMRMR + SSA + XGBoost for different cancer gene expression datasets is summarized as follows. Figures 3–6 show the performance of the hybrid mentioned above models in contrast to different datasets.
• For the ALL-AML dataset, the Boruta (B)+ IMRMR (IM) +XGBoost with SSA shows the highest accuracy at 0.917. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.929, 0.929, 0.929, and 0.929, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.071, 0.100, 0.829, and 0.083, respectively.
• For the Lymphoma dataset, B + IM + SSA + AdaBoost shows the best performance with an accuracy of 0.919. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.938, 0.957, 0.947, 0.953, and 0.800, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.043, 0.200, 0.776, and 0.081, respectively.
• Similarly, for the MLL dataset, B + IM + AdaBoost with SSA shows the highest performance with an accuracy of 0.917 compared to other hybrid models, as resented. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.939, 0.939, 0.939, 0.939, and 0.870, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.061, 0.130, 0.808, and 0.083, respectively.
• With the SRBCT gene expression dataset, the B + IM + SSA + XGBoost performs best with an accuracy of 0.916. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.941, 0.923, 0.932, 0.927, and 0.903, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.077, 0.097, 0.821, and 0.084, respectively.
Figure 3. Out-of-sample performance measure of the different classifiers with MRMR and SSA for the ALL-AML dataset.
Figure 4. Out-of-sample performance measure of the different classifiers with MRMR and SSA for the Lymphoma dataset.
Figure 5. Out-of-sample performance measure of different classifiers with MRMR and SSA for the MLL dataset.
Figure 6. Out-of-sample performance measure of the different classifiers with MRMR and SSA for the SRBCT dataset.
The analysis demonstrates that the B + IM + SSA + AdaBoost model, resulting in an accuracy of 0.919, uses the highest obtained Lymphoma dataset. In order to enhance the accuracy even further, it is possible to include an ensemble machine learning approach. The present study uses the majority voting approach to combine predictions from various models and make a final choice based on the majority of votes. This strategy may improve the accuracy by using the advantages of many models while reducing the limitations of each model. By using majority voting, the ultimate model may amalgamate these advantages, resulting in enhanced overall accuracy and more resilient categorization outcomes. The hybrid models mentioned demonstrate remarkable accuracy across several cancer gene expression datasets, reaching a maximum accuracy of 0.919. A majority voting ensemble classifier may help overcome the problems with the provided hybrid models, namely, their high computational cost and the possibility of overfitting caused by combining several algorithms. By integrating several models’ strengths, this method streamlines decision-making, leads to more accurate forecasts with less hyperparameter tweaking, and reduces the danger of overfitting. To further enhance accuracy, the current work integrates the prediction of the best three classifiers using the majority voting technique. This improves overall precision, minimizes mistakes, and results in more dependable and transferable results in cancer categorization assignments.
4.2 Analysis of the proposed model
The performance analysis of the proposed BIMSSA model is shown in Table 5. Table 5 quantifies the ACC, PRE, REC, F-1S, F-2, and SPE measures with Clopper–Pearson confidence interval (CI) (Puza and O’neill, 2006).
• For the ALL-AML dataset, the proposed BIMSSA model shows the accuracy as 0.967 with a CI of 0.887–0.99. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.967, 0.974, 0.974, 0.974, 0.974, and 0.955, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.026, 0.046, 0.929, and 0.033, respectively.
• For the Lymphoma dataset, the proposed BIMSSA shows an accuracy level of 0.962% with a CI of 0.862–0.995. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.972, 0.972, 0.972, 0.972, and 0.938respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.028, 0.063, 0.910, and 0.039, respectively.
• Similarly, for the MLL dataset, the proposed BIMSSA shows the accuracy level as 0.951 with a CI of 0.887–0.99. The other parameters, such as PRE, REC, F1-S, F 2, and SPE, are 0.949, 0.974, 0.961, 0.969, and 0.913, respectively. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.026, 0.087, 0.895, and 0.049, respectively.
• With SRBCT, the BIMSSA shows a 0.971 accuracy level with a CI of 0.901–0.991. The other parameters, such as PRE, REC, F1-S, and F 2, are 0.978. Specificity (SPE) is 0.958. The other parameters, such as FNR, FPR, MCC, and MCR, are 0.022, 0.042, 0.937, and 029, respectively.
• The model maintains a good level of performance throughout all four datasets (D1, D2, D3, and D4) in regards to accuracy, recall, precision, F1-Score, F2-Score, and specificity. The Clopper-Pearson method’s confidence intervals support the measures’ dependability; narrow intervals show little fluctuation and strong faith in the model’s forecasts. As shown by its low false positive rate (FPR) of 0.042, high Matthews correlation coefficient (MCC) of 0.937, and low Misclassification Rate (MCR) of 0.029, dataset D4 has the best overall performance, especially in accuracy and specificity. These numbers show how well D4 can detect real positives and genuine negatives, which allows it to make strong and trustworthy predictions. D1 additionally demonstrates excellent outcomes; it is a balanced model with a little advantage in predictive accuracy, with a FNR of 0.026, FPR of 0.046, MCC of 0.929, and MCR of 0.033. The great predictive power for D2 is shown in its FNR of 0.028, FPR of 0.063, MCC of 0.910, and MCR of 0.039, which are somewhat lower than D1 and D4, but it still maintains a high level of efficiency. Having a greater FPR of 0.087 affects D3’s MCC of 0.895 and MCR of 0.049, which in turn causes a little drop in specificity. Regardless, the model’s great predictive power and balance between recall and accuracy make it quite useful across all datasets.
• The developed BIMSSA model’s ROC curve is shown in Figure 7 for ALL-AML, Lymphoma, MLL, and SRBCT datasets. The suggested model’s AUC for the ALL-AML dataset is 0.971. The suggested model’s AUC for the Lymphoma dataset is 0.961. The suggested model has AUC values of 0.950 and 0.985 for the MLL and SRBCT datasets.
• Figure 8 shows the training and test time of the developed BIMSSA compared to cancer datasets.
Table 5. Performance Analysis of Proposed BIMSSA Model with Clopper–Pearson confidence interval (CI).
The proposed BIMSSA does include the Boruta, IMRMR, and SSA as the feature selection algorithm. In addition, the BIMSSA includes SVM, RF, ELM, AdaBoost, and XGBoost as the base classifiers, including majority voting as the ensemble classifier. The model is evaluated over 4 different high dimensional datasets for evaluation. The computational complexity of the proposed BIMSSA model becomes O (T⋅F⋅log(F)) + O(P⋅F⋅I) + O(N2⋅F), with N << F. Where, N is the number of sample present in the dataset, F is the number of features, T is the number of trees in the RF classifier, I is the number of iterations, and P is the number of population for SSA.
Although the proposed model is efficient, it still retains several limitations that must be mitigated. First is the increased computational complexity based on the iterative structure of the Boruta algorithm, the added overhead of SSA, and the computation-intensive nature of classifiers used, such as SVM and methods based on trees. Then, hybrid approaches for models increase the complexity of their implementation and call for careful coordination between feature selection and classification phases for optimal performance. The pipeline’s repetitive nature might make it seem to extend training time, making it clumsy for situations requiring deployment and results in a very fast manner.
5 Critical analysis
In this section the proposed BIMSSA is compared with some existing literature based on four cancer datasets including ALL-AML, Lymphoma, MLL, and SRBCT. The detailed comparison of BIMSSA with the existing works (Sun et al., 2019; Meenachi and Ramakrishnan, 2020; Nouri-Moghaddam et al., 2021; Yan et al., 2021; Alomari et al., 2021; Rostami et al., 2022; Ke et al., 2022), emphasizing the performance differences across these datasets is given below.
5.1 Lymphoma dataset
• For Lymphoma dataset the proposed BIMSSA shows an accuracy of 96.2%, BIMSSA again shows strong performance in classifying Lymphoma, indicating the model’s versatility.
• The work in Yan et al. (2021) reports an accuracy of 88.57% for Lymphoma, which is significantly lower than that of BIMSSA. The proposed model outperforms Yan et al. (2021) by ∼8.61%. This difference may highlight the superiority of the BIMSSA model in handling Lymphoma data, possibly due to better feature selection, model training, and data preprocessing techniques.
• The work in Ke et al. (2022) reports an accuracy of 93.21% for Lymphoma, which is still lower than that of BIMSSA by ∼3.21%. This further confirms BIMSSA’s effectiveness in this context.
5.2 MLL dataset
• BIMSSA achieves an accuracy of 95.1%, showcasing its strong performance in this dataset as well.
• The work in Yan et al. (2021) obtains an accuracy of 86.19%, which is lower than that of BIMSSA by a significant margin of ∼10.33%. This difference further emphasizes BIMSSA’s advantage in handling complex datasets.
5.3 SRBCT dataset
• BIMSSA with an accuracy of 97.1%, performs exceptionally well on the SRBCT dataset, indicating its robustness and effectiveness across different cancer types.
• The work in Sun et al. (2019) reports a high accuracy of 93.6% for SRBCT, which is noteworthy but still lower than BIMSSA’s accuracy by ∼3.73%.
• In Meenachi and Ramakrishnan (2020) the accuracy is 81.58% which is lower than that of the BIMSSA with a significant margin of ∼19.02%, suggesting that the feature selection, classification and preprocessing techniques used in Meenachi and Ramakrishnan (2020) struggles with this dataset.
• The work in Nouri-Moghaddam et al. (2021) reports an accuracy of 90.72% for SRBCT, which, while remarkable, is still outperformed by BIMSSA by ∼6.92%.
• The work in Yan et al. (2021) reported accuracy here is 76.74%, the lowest among the all compared literature, and BIMSSA outperforms it by ∼26.53%. It is indicating that the model in Yan et al. (2021) may not be as effective for SRBCT.
• In Rostami et al. (2022) the accuracy of 82.82% is reported, which is again significantly low by ∼17.24% when compared to BIMSSA.
5.4 ALL-AML dataset
• BIMSSA achieves an accuracy of 96.7%, which demonstrates strong performance and reliability in classifying this cancer type.
• The work in Alomari et al. (2021) reports an exceptionally high accuracy of 99.86%, which outperforms the proposed BIMSSA model by ∼3.27%.
BIMSSA proves to be a very efficient and dependable model for categorizing cancer, surpassing the accuracy of other models from the existing literature in all tested datasets except ALL-AML. Though the accuracy is low as compared to Alomari et al. (2021) in case of ALL-AML dataset, still it shows an accuracy of 96.7% that can be considered as a noteworthy performance. The model’s constant and superior performance in classifying many forms of cancer, together with its excellent generalization, sets it out as an exceptional model in the field of cancer classification. Therefore, BIMSSA showcases the flexibility and applicability needed for various cancer datasets. Figure 9 shows the comparison of BIMSSA with other existing models in contrast to diverse datasets.
6 Conclusion
The primary focus of this study is developing an ensemble machine learning-based model for cancer detection. Gene expression data, also known as microarray data, leaves its unique imprint when used in a cancer detection model. However, there are challenges unique to dealing with microarray data, such as a limited sample size, which diminishes the model’s performance. To deal with this issue, the proposed BIMSSA considers a pipeline feature selection approach with Boruta, IMRMR, and SSA feature selection algorithm to select relevant features. The selected features, such as SVM, RF, ELM, AdaBoost, and XGBoost, are applied to the optimized feature set as the base classifiers. Based on the performance, three classifiers, ELM, AdaBoost, and XGBoost with Boruta, IMRMR, and SSA feature selection, are considered for developing the ensemble model with a majority voting classifier. After selecting three classifiers, we use majority voting to create an ensemble ML-based cancer diagnostic model called BIMSSA. Empirical results from this study using the developed BIMSSA reveal an accuracy of 0.967, 0.962, 0.951, and 0.971 for ALL-AML, Lymphoma, MLL, and SRBCT datasets. The suggested model’s AUC for the ALL-AML dataset is 0.973. The proposed model’s AUC for the Lymphoma dataset is 0.969. The suggested model achieves an AUC of 0.951 for the MLL dataset and 0.979 for the SRBCT dataset. The proposed model does not perform well despite the existing literature regarding the ALL-MLL dataset. However, there are some limitations to the current work. In addition, the current research does not aim to consider the concept of class imbalance.
As a future scope of this manuscript, the feature selection process will be reversed, starting with SSA to explore different combinations of features and identify the promising subset, followed by IMRMR to refine the feature subset further and prioritize the most informative features. Finally, Boruta feature selection is applied to validate the selected feature subset and provide additional insights into the importance of the feature. Testing the model’s performance on more kinds of cancer gene expression datasets is planned for the future of this study. In addition, future plans for this study include using deep learning approaches to increase the model’s capability for working with picture datasets.
Data availability statement
Publicly available datasets were analyzed in this study. This data can be found here: https://csse.szu.edu.cn/staff/zhuzx/Datasets.html.
Author contributions
PP: Conceptualization, Writing–original draft, Writing–review and editing. SB: Investigation, Writing–original draft, Writing–review and editing. AmP: Data curation, Investigation, Writing–original draft, Writing–review and editing. AbP: Conceptualization, Investigation, Writing–original draft, Writing–review and editing. BS: Software, Writing–original draft, Writing–review and editing. ZG: Conceptualization, Data curation, Investigation, Methodology, Visualization, Writing–original draft, Writing–review and editing. HL: Methodology, Validation, Writing–original draft, Writing–review and editing. PJ: Conceptualization, Software, Writing–original draft, Writing–review and editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded by the Innovational Fund for Scientific and Technological Personnel of Hainan Province (Grant No. KJRC 2023L01), and the South China Sea Rising Star Project of Hainan Province.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers at the time of submission. This had no impact on the peer review process and the final decision.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alfian, G., Syafrudin, M., Fahrurrozi, I., Fitriyani, N. L., Atmaji, F. T. D., Widodoet al, T., et al. (2022). Predicting breast cancer from risk factors using SVM and extra-trees-based feature selection method. Computers 11 (9), 136. doi:10.3390/computers11090136
Alghunaim, S., and Al-Baity, H. H. (2019). On the scalability of machine-learning algorithms for breast cancer prediction in big data context. IEEE Access 7, 91535–91546. doi:10.1109/ACCESS.2019.2927080
Almugren, N., and Alshamlan, H. (2019). A survey on hybrid feature selection methods in microarray gene expression data for cancer classification. IEEE Access 7, 78533–78548. doi:10.1109/ACCESS.2019.2922987
Alomari, O. A., Makhadmeh, S. N., Al-Betar, M. A., Alyasseri, Z. A. A., Doush, I. A., Abasi, A. K., et al. (2021). Gene selection for microarray data classification based on Gray Wolf Optimizer enhanced with TRIZ-inspired operators. Knowledge-Based Syst. 223, 107034. doi:10.1016/j.knosys.2021.107034
Alromema, N., Syed, A. H., and Khan, T. (2023). A hybrid machine learning approach to screen optimal predictors for the classification of primary breast tumors from gene expression microarray data. Diagnostics 13 (4), 708. doi:10.3390/diagnostics13040708
Arun Prabha, K., Mahesh, C., and Raja, S. P. (2021). Decision tree based salp swarm optimization for multi medical data classification with feature reduction technique. Braz. Archives Biol. Technol. 64 (Jan). doi:10.1590/1678-4324-2021210240
Asselman, A., Khaldi, M., and Aammou, S. (2021). Enhancing the prediction of student performance based on the machine learning XGBoost algorithm. Interact. Learn. Environ. 31, 3360–3379. doi:10.1080/10494820.2021.1928235
Aziz, R. M. (2022). Nature-inspired metaheuristics model for gene selection and classification of biomedical microarray data. Med. and Biol. Eng. and Comput. 60 (6), 1627–1646. doi:10.1007/s11517-022-02555-7
Balakrishnan, K., Dhanalakshmi, R., and Khaire, U. M. (2021). Improved salp swarm algorithm based on the levy flight for feature selection. J. Supercomput. 77 (11), 12399–12419. doi:10.1007/s11227-021-03773-w
Batool, A., and Byun, Y.-C. (2024). Toward improving breast cancer classification using an adaptive voting ensemble learning algorithm. IEEE Access 12, 12869–12882. doi:10.1109/access.2024.3356602
Bolón-Canedo, V., Sánchez-Maroño, N., Alonso-Betanzos, A., Benítez, J. M., and Herrera, F. (2014). A review of microarray datasets and applied feature selection methods. Inf. Sci. 282, 111–135. doi:10.1016/j.ins.2014.05.042
Castelli, M., Manzoni, L., Mariot, L., Nobile, M. S., and Tangherloni, A. (2022). Salp swarm optimization: a critical review. Expert Syst. Appl. 189, 116029. doi:10.1016/j.eswa.2021.116029
de Leval, L., and Jaffe, E. S. (2020). Lymphoma classification. Cancer J. 26 (3), 176–185. doi:10.1097/PPO.0000000000000451
Ding, C., and Peng, H. (2005). Minimum Redundancy feature selection from microarray gene expression data. J. Bioinforma. Comput. Biol. 03 (02), 185–205. doi:10.1142/s0219720005001004
Ding, S., Xu, X., and Nie, R. (2013). Extreme learning machine and its applications. Neural Comput. Appl. 25 (3–4), 549–556. doi:10.1007/s00521-013-1522-8
Ghoniem, R. M. (2020). A novel bio-inspired deep learning approach for liver cancer diagnosis. Information 11 (2), 80. doi:10.3390/info11020080
Hameed, S. S., Hassan, W. H., Latiff, L. A., and Muhammadsharif, F. F. (2021). A comparative study of nature-inspired metaheuristic algorithms using a three-phase hybrid approach for gene selection and classification in high-dimensional cancer datasets. Soft Comput. 25 (13), 8683–8701. doi:10.1007/s00500-021-05726-0
Hegazy, Ah. E., Makhlouf, M. A., and El-Tawel, Gh. S. (2018a). Feature selection using chaotic salp swarm algorithm for data classification. Arabian J. Sci. Eng. 44 (4), 3801–3816. doi:10.1007/s13369-018-3680-6
Hegazy, Ah. E., Makhlouf, M. A., and El-Tawel, Gh. S. (2018b). Improved salp swarm algorithm for feature selection. J. King Saud Univ. - Comput. Inf. Sci. 32, 335–344. doi:10.1016/j.jksuci.2018.06.003
Ibrahim, H. T., Mazher, W. J., Ucan, O. N., and Bayat, O. (2017). Feature selection using salp swarm algorithm for real biomedical datasets. Int. J. Comput. Netw. Inf. Secur., 1738–7906.
Ibrahim, R. A., Ewees, A. A., Oliva, D., Abd Elaziz, M., and Lu, S. (2018). Improved salp swarm algorithm based on particle swarm optimization for feature selection. J. Ambient Intell. Humaniz. Comput. 10 (8), 3155–3169. doi:10.1007/s12652-018-1031-9
Ke, L., Li, M., Wang, L., Deng, S., Ye, J., and Yu, X. (2022). Improved swarm-optimization-based filter-wrapper gene selection from microarray data for gene expression tumor classification. Pattern Analysis Appl. 26 (2), 455–472. doi:10.1007/s10044-022-01117-9
Khalsan, M., Machado, L. R., Al-Shamery, E. S., Ajit, S., Anthony, K., Mu, M., et al. (2022). A survey of machine learning approaches applied to gene expression analysis for cancer prediction. IEEE Access 10, 27522–27534. doi:10.1109/ACCESS.2022.3146312
Lewis, W. D., Lilly, S., and Jones, K. L. (2020). Lymphoma: diagnosis and treatment. Am. Fam. Physician 101 (1), 34–41.
López-García, G., Jerez, J. M., Franco, L., and Veredas, F. J. (2020). Transfer learning with convolutional neural networks for cancer survival prediction using gene-expression data. PLOS ONE 15 (3), e0230536. doi:10.1371/journal.pone.0230536
Ma, B., Meng, F., Yan, G., Yan, H., Chai, B., and Song, F. (2020). Diagnostic classification of cancers using extreme gradient boosting algorithm and multi-omics data. Comput. Biol. Med. 121, 103761. doi:10.1016/j.compbiomed.2020.103761
Mahesh, T. R., Vinoth Kumar, V., Vivek, V., Karthick Raghunath, K. M., and Sindhu Madhuri, G. (2022). Early predictive model for breast cancer classification using blended ensemble learning. Int. J. Syst. Assur. Eng. Manag. 15, 188–197. doi:10.1007/s13198-022-01696-0
Maurya, N. S., Kushwah, S., Kushwaha, S., Chawade, A., and Mani, A. (2023). Prognostic model development for classification of colorectal adenocarcinoma by using machine learning model based on feature selection technique boruta. Sci. Rep. 13 (1), 6413. doi:10.1038/s41598-023-33327-4
Meenachi, L., and Ramakrishnan, S. (2020). Differential evolution and ACO based global optimal feature selection with fuzzy rough set for cancer data classification. Soft Comput. 24 (24), 18463–18475. doi:10.1007/s00500-020-05070-9
Mirjalili, S., Gandomi, A. H., Mirjalili, S. Z., Saremi, S., Faris, H., and Mirjalili, S. M. (2017). Salp Swarm Algorithm: a bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 114, 163–191. doi:10.1016/j.advengsoft.2017.07.002
Naji, M. A., Filali, S. E., Bouhlal, M., Benlahmar, E. H., Abdelouhahid, R. A., and Debauche, O. (2021). Breast cancer prediction and diagnosis through a new approach based on majority voting ensemble classifier. Procedia Comput. Sci. 191, 481–486. doi:10.1016/j.procs.2021.07.061
Nouri-Moghaddam, B., Ghazanfari, M., and Fathian, M. (2021). A novel bio-inspired hybrid multi-filter wrapper gene selection method with ensemble classifier for microarray data. Neural Comput. Appl. 35, 11531–11561. doi:10.1007/s00521-021-06459-9
Panigrahi, A., Pati, A., Sahu, B., Das, M. N., Nayak, D. S. K., Sahoo, G., et al. (2023). En-MinWhale: an ensemble approach based on MRMR and whale optimization for cancer diagnosis. IEEE Access 11, 113526–113542. doi:10.1109/ACCESS.2023.3318261
Pati, A., Panigrahi, A., Nayak, D. S. K., Sahoo, G., and Singh, D. (2023). Predicting pediatric appendicitis using ensemble learning techniques. Procedia Comput. Sci. 218, 1166–1175. doi:10.1016/j.procs.2023.01.095
Puza, B., and O’neill, T. (2006). Generalised Clopper–Pearson confidence intervals for the binomial proportion. J. Stat. Comput. Simul. 76 (6), 489–508. doi:10.1080/10629360500107527
Rostami, M., Forouzandeh, S., Berahmand, K., Soltani, M., Shahsavari, M., and Oussalah, M. (2022). Gene selection for microarray data classification via multi-objective graph theoretic-based method. Artif. Intell. Med. 123, 102228. doi:10.1016/j.artmed.2021.102228
Rustagi, K., Bhatnagar, P., Mathur, R., Singh, I., and K. G, S. (2024). Hybrid salp swarm and grey wolf optimizer algorithm based ensemble approach for breast cancer diagnosis. Multimedia tools Appl. 83, 70117–70141. doi:10.1007/s11042-023-18015-9
Sayed, G. I., Khoriba, G., and Haggag, M. H. (2018). A novel chaotic salp swarm algorithm for global optimization and feature selection. Appl. Intell. 48 (10), 3462–3481. doi:10.1007/s10489-018-1158-6
Shukla, A. K., Singh, P., and Vardhan, M. (2020a). Gene selection for cancer types classification using novel hybrid metaheuristics approach. Swarm Evol. Comput. 54, 100661. doi:10.1016/j.swevo.2020.100661
Shukla, A. K., Singh, P., and Vardhan, M. (2020b). An adaptive inertia weight teaching-learning-based optimization algorithm and its applications. Appl. Math. Model. 77, 309–326. doi:10.1016/j.apm.2019.07.046
Shukla, A. K., Singh, P., and Vardhan, M. (2019). A new hybrid wrapper TLBO and SA with SVM approach for gene expression data. Inf. Sci. 503, 238–254. doi:10.1016/j.ins.2019.06.063
Sun, L., Zhang, X., Qian, Y., Xu, J., and Zhang, S. (2019). Feature selection using neighborhood entropy-based uncertainty measures for gene expression data classification. Inf. Sci. (N. Y). 502, 18–41. doi:10.1016/j.ins.2019.05.072
Thawkar, S. (2021). A hybrid model using teaching–learning-based optimization and Salp swarm algorithm for feature selection and classification in digital mammography. J. Ambient Intell. Humaniz. Comput. 12 (9), 8793–8808. doi:10.1007/s12652-020-02662-z
Ünalan, S., Günay, O., Akkurt, I., Gunoglu, K., and Tekin, H. O. (2024). A comparative study on breast cancer classification with stratified shuffle split and K-fold cross validation via ensembled machine learning. J. Radiat. Res. Appl. Sci. 17 (4), 101080. doi:10.1016/j.jrras.2024.101080
Venkatesan, C., Balamurugan, D., Thamaraimanalan, T., and Ramkumar, M. (2022). “Efficient machine learning technique for tumor classification based on gene expression data,” in 2022 8th international conference on advanced computing and communication systems (ICACCS). Coimbatore, India, 1982–1986. doi:10.1109/ICACCS54159.2022.9785294
Yan, C., Suo, Z., Guan, X., and Luo, H. (2021). “A novel feature selection method based on salp swarm algorithm,” in 2021 IEEE international conference on information communication and software engineering (ICICSE). Chengdu, China, 126–130. doi:10.1109/ICICSE52190.2021.9404129
Yan, X., and Jia, M. (2019). Intelligent fault diagnosis of rotating machinery using improved multiscale dispersion entropy and mRMR feature selection. Knowledge-Based Syst. 163, 450–471. doi:10.1016/j.knosys.2018.09.004
Zhao, Z., Anand, R., and Wang, M. (2019). “Maximum relevance and minimum redundancy feature selection methods for a marketing machine learning platform,” in 2019 IEEE international conference on data science and advanced analytics (DSAA). Washington, DC, USA, 442–452. doi:10.1109/DSAA.2019.00059
Zhu, Z., Ong, Y.-S., and Dash, M. (2007). Markov blanket-embedded genetic algorithm for gene selection. Pattern Recognit. 40 (11), 3236–3248. doi:10.1016/j.patcog.2007.02.007
Appendix-I
Keywords: cancer prediction, microarray data, feature selection, swarm intelligence, ensemble learning
Citation: Panda P, Bisoy SK, Panigrahi A, Pati A, Sahu B, Guo Z, Liu H and Jain P (2025) BIMSSA: enhancing cancer prediction with salp swarm optimization and ensemble machine learning approaches. Front. Genet. 15:1491602. doi: 10.3389/fgene.2024.1491602
Received: 06 October 2024; Accepted: 11 December 2024;
Published: 06 January 2025.
Edited by:
Lei Chen, Shanghai Maritime University, ChinaReviewed by:
Puspanjali Mohapatra, International Institute of Information Technology, IndiaAbdelkader Benyettou, Centre Universitaire de Relizane, Algeria
Copyright © 2025 Panda, Bisoy, Panigrahi, Pati, Sahu, Guo, Liu and Jain. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Zheshan Guo, Z3Vvemhlc2hhbkBoYWluYW51LmVkdS5jbg==; Prince Jain, cHJpbmNlZWNlNDhAZ21haWwuY29t
†ORCID ID: Haipeng Liu, https://orcid.org/0000-0002-4212-2503; Prince Jain, https://orcid.org/0000-0002-7950-7263